From 2e3f2ea20bcca76cb37c95363aa597d6757aa543 Mon Sep 17 00:00:00 2001 From: David Oppenheimer Date: Sun, 5 Jul 2015 21:31:54 -0700 Subject: [PATCH] gofmt --- .../algorithm/priorities/priorities.go | 6 +-- .../algorithm/priorities/priorities_test.go | 39 +++++++++---------- 2 files changed, 22 insertions(+), 23 deletions(-) diff --git a/plugin/pkg/scheduler/algorithm/priorities/priorities.go b/plugin/pkg/scheduler/algorithm/priorities/priorities.go index 1dffbbcc8a..4a4e333131 100644 --- a/plugin/pkg/scheduler/algorithm/priorities/priorities.go +++ b/plugin/pkg/scheduler/algorithm/priorities/priorities.go @@ -20,8 +20,8 @@ import ( "math" "github.com/GoogleCloudPlatform/kubernetes/pkg/api" - "github.com/GoogleCloudPlatform/kubernetes/pkg/labels" "github.com/GoogleCloudPlatform/kubernetes/pkg/api/resource" + "github.com/GoogleCloudPlatform/kubernetes/pkg/labels" "github.com/GoogleCloudPlatform/kubernetes/plugin/pkg/scheduler/algorithm" "github.com/GoogleCloudPlatform/kubernetes/plugin/pkg/scheduler/algorithm/predicates" "github.com/golang/glog" @@ -47,8 +47,8 @@ func calculateScore(requested int64, capacity int64, node string) int { // pods will not all be scheduled to the machine with the smallest in-use limit, // and that when scheduling regular pods, such pods will not see zero-limit pods as // consuming no resources whatsoever. -const defaultMilliCpuLimit int64 = 100 // 0.1 core -const defaultMemoryLimit int64 = 60 * 1024 * 1024 // 60 MB +const defaultMilliCpuLimit int64 = 100 // 0.1 core +const defaultMemoryLimit int64 = 60 * 1024 * 1024 // 60 MB // TODO: Consider setting default as a fixed fraction of machine capacity (take "capacity api.ResourceList" // as an additional argument here) rather than using constants diff --git a/plugin/pkg/scheduler/algorithm/priorities/priorities_test.go b/plugin/pkg/scheduler/algorithm/priorities/priorities_test.go index 4662786fed..b6af3446f2 100644 --- a/plugin/pkg/scheduler/algorithm/priorities/priorities_test.go +++ b/plugin/pkg/scheduler/algorithm/priorities/priorities_test.go @@ -73,9 +73,9 @@ func TestZeroLimit(t *testing.T) { Resources: api.ResourceRequirements{ Limits: api.ResourceList{ "cpu": resource.MustParse( - strconv.FormatInt(defaultMilliCpuLimit * 3, 10) + "m"), + strconv.FormatInt(defaultMilliCpuLimit*3, 10) + "m"), "memory": resource.MustParse( - strconv.FormatInt(defaultMemoryLimit * 3, 10)), + strconv.FormatInt(defaultMemoryLimit*3, 10)), }, }, }, @@ -86,30 +86,30 @@ func TestZeroLimit(t *testing.T) { large2 := large large2.NodeName = "machine2" tests := []struct { - pod *api.Pod - pods []*api.Pod - nodes []api.Node - test string + pod *api.Pod + pods []*api.Pod + nodes []api.Node + test string }{ // The point of these tests is to show you get the same priority for a zero-limit pod // as for a pod with the defaults limits, both when the zero-limit pod is already on the machine // and when the zero-limit pod is the one being scheduled. { - pod: &api.Pod{Spec: noResources}, + pod: &api.Pod{Spec: noResources}, // match current f1-micro on GCE - nodes: []api.Node{makeMinion("machine1", 1000, defaultMemoryLimit * 10), makeMinion("machine2", 1000, defaultMemoryLimit * 10)}, - test: "test priority of zero-limit pod with machine with zero-limit pod", - pods: []*api.Pod { + nodes: []api.Node{makeMinion("machine1", 1000, defaultMemoryLimit*10), makeMinion("machine2", 1000, defaultMemoryLimit*10)}, + test: "test priority of zero-limit pod with machine with zero-limit pod", + pods: []*api.Pod{ {Spec: large1}, {Spec: noResources1}, {Spec: large2}, {Spec: small2}, }, }, { - pod: &api.Pod{Spec: small}, + pod: &api.Pod{Spec: small}, // match current f1-micro on GCE - nodes: []api.Node{makeMinion("machine1", 1000, defaultMemoryLimit * 10), makeMinion("machine2", 1000, defaultMemoryLimit * 10)}, - test: "test priority of nonzero-limit pod with machine with zero-limit pod", - pods: []*api.Pod { + nodes: []api.Node{makeMinion("machine1", 1000, defaultMemoryLimit*10), makeMinion("machine2", 1000, defaultMemoryLimit*10)}, + test: "test priority of nonzero-limit pod with machine with zero-limit pod", + pods: []*api.Pod{ {Spec: large1}, {Spec: noResources1}, {Spec: large2}, {Spec: small2}, }, @@ -130,14 +130,13 @@ func TestZeroLimit(t *testing.T) { t.Errorf("unexpected error: %v", err) } for _, hp := range list { - if (hp.Score != expectedPriority) { + if hp.Score != expectedPriority { t.Errorf("%s: expected 25 for all priorities, got list %#v", list) } } } } - func TestLeastRequested(t *testing.T) { labels1 := map[string]string{ "foo": "bar", @@ -162,7 +161,7 @@ func TestLeastRequested(t *testing.T) { { Resources: api.ResourceRequirements{ Limits: api.ResourceList{ - "cpu": resource.MustParse("1000m"), + "cpu": resource.MustParse("1000m"), "memory": resource.MustParse("0"), }, }, @@ -170,7 +169,7 @@ func TestLeastRequested(t *testing.T) { { Resources: api.ResourceRequirements{ Limits: api.ResourceList{ - "cpu": resource.MustParse("2000m"), + "cpu": resource.MustParse("2000m"), "memory": resource.MustParse("0"), }, }, @@ -494,7 +493,7 @@ func TestBalancedResourceAllocation(t *testing.T) { { Resources: api.ResourceRequirements{ Limits: api.ResourceList{ - "cpu": resource.MustParse("1000m"), + "cpu": resource.MustParse("1000m"), "memory": resource.MustParse("0"), }, }, @@ -502,7 +501,7 @@ func TestBalancedResourceAllocation(t *testing.T) { { Resources: api.ResourceRequirements{ Limits: api.ResourceList{ - "cpu": resource.MustParse("2000m"), + "cpu": resource.MustParse("2000m"), "memory": resource.MustParse("0"), }, },