From a9ec31209eeea4a0e710aa53dbd5cb2c23834391 Mon Sep 17 00:00:00 2001 From: Wojciech Tyczynski Date: Wed, 7 Dec 2016 15:40:26 +0100 Subject: [PATCH] GetOptions - fix tests --- test/e2e/addon_update.go | 7 +- test/e2e/autoscaling_utils.go | 7 +- test/e2e/batch_v1_jobs.go | 7 +- test/e2e/cluster_logging_es.go | 3 +- test/e2e/cluster_logging_utils.go | 3 +- test/e2e/cluster_size_autoscaling.go | 3 +- test/e2e/common/container_probe.go | 11 ++- test/e2e/common/downwardapi_volume.go | 3 +- test/e2e/common/pods.go | 7 +- test/e2e/common/sysctl.go | 5 +- test/e2e/common/volumes.go | 2 +- test/e2e/cronjob.go | 9 +- test/e2e/daemon_set.go | 4 +- test/e2e/deployment.go | 38 ++++---- test/e2e/dns.go | 4 +- test/e2e/dns_autoscaling.go | 3 +- test/e2e/events.go | 3 +- test/e2e/examples.go | 3 +- test/e2e/federated-ingress.go | 13 +-- test/e2e/federated-namespace.go | 7 +- test/e2e/federated-secret.go | 13 +-- test/e2e/federated-service.go | 3 +- test/e2e/federation-apiserver.go | 3 +- test/e2e/federation-daemonset.go | 13 +-- test/e2e/federation-deployment.go | 10 +- test/e2e/federation-replicaset.go | 10 +- test/e2e/federation-util-14.go | 5 +- test/e2e/federation-util.go | 13 +-- test/e2e/framework/exec_util.go | 5 +- test/e2e/framework/framework.go | 3 +- test/e2e/framework/networking_utils.go | 8 +- test/e2e/framework/pods.go | 5 +- test/e2e/framework/util.go | 92 ++++++++++--------- test/e2e/garbage_collector.go | 4 +- test/e2e/ingress_utils.go | 9 +- test/e2e/job.go | 7 +- test/e2e/kibana_logging.go | 3 +- test/e2e/kubectl.go | 26 +++--- test/e2e/kubelet.go | 3 +- test/e2e/limit_range.go | 7 +- test/e2e/mesos.go | 2 +- test/e2e/namespace.go | 9 +- test/e2e/network_partition.go | 11 ++- test/e2e/node_problem_detector.go | 3 +- test/e2e/nodeoutofdisk.go | 7 +- test/e2e/pd.go | 2 +- test/e2e/persistent_volumes.go | 24 ++--- test/e2e/petset.go | 10 +- test/e2e/pods.go | 3 +- test/e2e/pre_stop.go | 3 +- test/e2e/rc.go | 7 +- test/e2e/reboot.go | 3 +- test/e2e/replica_set.go | 6 +- test/e2e/resize_nodes.go | 2 +- test/e2e/resource_quota.go | 3 +- test/e2e/restart.go | 5 +- test/e2e/scheduler_predicates.go | 21 +++-- test/e2e/security_context.go | 3 +- test/e2e/service.go | 35 +++---- test/e2e/service_accounts.go | 19 ++-- test/e2e/ubernetes_lite.go | 2 +- test/e2e/volume_provisioning.go | 4 +- test/e2e/volumes.go | 2 +- test/e2e_node/apparmor_test.go | 3 +- test/e2e_node/benchmark_util.go | 2 +- test/e2e_node/container.go | 9 +- test/e2e_node/disk_eviction_test.go | 7 +- test/e2e_node/garbage_collector_test.go | 3 +- test/e2e_node/image_id_test.go | 3 +- test/e2e_node/kubelet_test.go | 6 +- test/e2e_node/log_path_test.go | 3 +- test/e2e_node/memory_eviction_test.go | 7 +- test/e2e_node/mirror_pod_test.go | 15 +-- test/e2e_node/util.go | 3 +- test/integration/client/client_test.go | 12 +-- test/integration/evictions/evictions_test.go | 4 +- test/integration/framework/master_utils.go | 3 +- .../garbage_collector_test.go | 4 +- test/integration/master/master_test.go | 7 +- .../integration/replicaset/replicaset_test.go | 8 +- .../replicationcontroller_test.go | 8 +- test/integration/scheduler/extender_test.go | 2 +- test/integration/scheduler/scheduler_test.go | 4 +- .../serviceaccount/service_account_test.go | 11 ++- test/integration/utils.go | 3 +- .../volume/persistent_volumes_test.go | 32 +++---- test/utils/density_utils.go | 5 +- test/utils/runners.go | 2 +- 88 files changed, 398 insertions(+), 338 deletions(-) diff --git a/test/e2e/addon_update.go b/test/e2e/addon_update.go index 0ebe73af98..6f04e893d7 100644 --- a/test/e2e/addon_update.go +++ b/test/e2e/addon_update.go @@ -26,6 +26,7 @@ import ( "golang.org/x/crypto/ssh" "k8s.io/kubernetes/pkg/api/v1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" clientset "k8s.io/kubernetes/pkg/client/clientset_generated/release_1_5" "k8s.io/kubernetes/test/e2e/framework" @@ -329,11 +330,11 @@ var _ = framework.KubeDescribe("Addon update", func() { waitForReplicationControllerInAddonTest(f.ClientSet, addonNsName, "addon-test-v2", false) By("verify invalid API addons weren't created") - _, err = f.ClientSet.Core().ReplicationControllers(addonNsName).Get("invalid-addon-test-v1") + _, err = f.ClientSet.Core().ReplicationControllers(addonNsName).Get("invalid-addon-test-v1", metav1.GetOptions{}) Expect(err).To(HaveOccurred()) - _, err = f.ClientSet.Core().Services(addonNsName).Get("ivalid-addon-test") + _, err = f.ClientSet.Core().Services(addonNsName).Get("ivalid-addon-test", metav1.GetOptions{}) Expect(err).To(HaveOccurred()) - _, err = f.ClientSet.Core().Services(defaultNsName).Get("ivalid-addon-test-v2") + _, err = f.ClientSet.Core().Services(defaultNsName).Get("ivalid-addon-test-v2", metav1.GetOptions{}) Expect(err).To(HaveOccurred()) // invalid addons will be deleted by the deferred function diff --git a/test/e2e/autoscaling_utils.go b/test/e2e/autoscaling_utils.go index 41ea143ec2..1a30408235 100644 --- a/test/e2e/autoscaling_utils.go +++ b/test/e2e/autoscaling_utils.go @@ -22,6 +22,7 @@ import ( "time" "k8s.io/kubernetes/pkg/api/v1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" "k8s.io/kubernetes/pkg/client/clientset_generated/internalclientset" clientset "k8s.io/kubernetes/pkg/client/clientset_generated/release_1_5" "k8s.io/kubernetes/pkg/util/intstr" @@ -247,21 +248,21 @@ func (rc *ResourceConsumer) sendConsumeCustomMetric(delta int) { func (rc *ResourceConsumer) GetReplicas() int { switch rc.kind { case kindRC: - replicationController, err := rc.framework.ClientSet.Core().ReplicationControllers(rc.framework.Namespace.Name).Get(rc.name) + replicationController, err := rc.framework.ClientSet.Core().ReplicationControllers(rc.framework.Namespace.Name).Get(rc.name, metav1.GetOptions{}) framework.ExpectNoError(err) if replicationController == nil { framework.Failf(rcIsNil) } return int(replicationController.Status.Replicas) case kindDeployment: - deployment, err := rc.framework.ClientSet.Extensions().Deployments(rc.framework.Namespace.Name).Get(rc.name) + deployment, err := rc.framework.ClientSet.Extensions().Deployments(rc.framework.Namespace.Name).Get(rc.name, metav1.GetOptions{}) framework.ExpectNoError(err) if deployment == nil { framework.Failf(deploymentIsNil) } return int(deployment.Status.Replicas) case kindReplicaSet: - rs, err := rc.framework.ClientSet.Extensions().ReplicaSets(rc.framework.Namespace.Name).Get(rc.name) + rs, err := rc.framework.ClientSet.Extensions().ReplicaSets(rc.framework.Namespace.Name).Get(rc.name, metav1.GetOptions{}) framework.ExpectNoError(err) if rs == nil { framework.Failf(rsIsNil) diff --git a/test/e2e/batch_v1_jobs.go b/test/e2e/batch_v1_jobs.go index 82b9cde6cb..a350894d45 100644 --- a/test/e2e/batch_v1_jobs.go +++ b/test/e2e/batch_v1_jobs.go @@ -28,6 +28,7 @@ import ( "k8s.io/kubernetes/pkg/api/v1" batchinternal "k8s.io/kubernetes/pkg/apis/batch" batch "k8s.io/kubernetes/pkg/apis/batch/v1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" clientset "k8s.io/kubernetes/pkg/client/clientset_generated/release_1_5" "k8s.io/kubernetes/pkg/kubectl" "k8s.io/kubernetes/pkg/labels" @@ -279,7 +280,7 @@ func newTestV1Job(behavior, name string, rPol v1.RestartPolicy, parallelism, com } func getV1Job(c clientset.Interface, ns, name string) (*batch.Job, error) { - return c.Batch().Jobs(ns).Get(name) + return c.Batch().Jobs(ns).Get(name, metav1.GetOptions{}) } func createV1Job(c clientset.Interface, ns string, job *batch.Job) (*batch.Job, error) { @@ -316,7 +317,7 @@ func waitForAllPodsRunningV1(c clientset.Interface, ns, jobName string, parallel // Wait for job to reach completions. func waitForV1JobFinish(c clientset.Interface, ns, jobName string, completions int32) error { return wait.Poll(framework.Poll, v1JobTimeout, func() (bool, error) { - curr, err := c.Batch().Jobs(ns).Get(jobName) + curr, err := c.Batch().Jobs(ns).Get(jobName, metav1.GetOptions{}) if err != nil { return false, err } @@ -327,7 +328,7 @@ func waitForV1JobFinish(c clientset.Interface, ns, jobName string, completions i // Wait for job fail. func waitForV1JobFail(c clientset.Interface, ns, jobName string, timeout time.Duration) error { return wait.Poll(framework.Poll, timeout, func() (bool, error) { - curr, err := c.Batch().Jobs(ns).Get(jobName) + curr, err := c.Batch().Jobs(ns).Get(jobName, metav1.GetOptions{}) if err != nil { return false, err } diff --git a/test/e2e/cluster_logging_es.go b/test/e2e/cluster_logging_es.go index e37656e64f..3e76ba5342 100644 --- a/test/e2e/cluster_logging_es.go +++ b/test/e2e/cluster_logging_es.go @@ -25,6 +25,7 @@ import ( "k8s.io/kubernetes/pkg/api" "k8s.io/kubernetes/pkg/api/v1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" "k8s.io/kubernetes/pkg/labels" "k8s.io/kubernetes/test/e2e/framework" @@ -92,7 +93,7 @@ func checkElasticsearchReadiness(f *framework.Framework) error { // being run as the first e2e test just after the e2e cluster has been created. var err error for start := time.Now(); time.Since(start) < graceTime; time.Sleep(5 * time.Second) { - if _, err = s.Get("elasticsearch-logging"); err == nil { + if _, err = s.Get("elasticsearch-logging", metav1.GetOptions{}); err == nil { break } framework.Logf("Attempt to check for the existence of the Elasticsearch service failed after %v", time.Since(start)) diff --git a/test/e2e/cluster_logging_utils.go b/test/e2e/cluster_logging_utils.go index 950b789bf4..b28b46c820 100644 --- a/test/e2e/cluster_logging_utils.go +++ b/test/e2e/cluster_logging_utils.go @@ -23,6 +23,7 @@ import ( "k8s.io/kubernetes/pkg/api" "k8s.io/kubernetes/pkg/api/v1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" "k8s.io/kubernetes/pkg/labels" "k8s.io/kubernetes/test/e2e/framework" ) @@ -62,7 +63,7 @@ func createSynthLogger(f *framework.Framework, linesCount int) { } func reportLogsFromFluentdPod(f *framework.Framework) error { - synthLoggerPod, err := f.PodClient().Get(synthLoggerPodName) + synthLoggerPod, err := f.PodClient().Get(synthLoggerPodName, metav1.GetOptions{}) if err != nil { return fmt.Errorf("Failed to get synth logger pod due to %v", err) } diff --git a/test/e2e/cluster_size_autoscaling.go b/test/e2e/cluster_size_autoscaling.go index 68dd136d03..8c9801255d 100644 --- a/test/e2e/cluster_size_autoscaling.go +++ b/test/e2e/cluster_size_autoscaling.go @@ -27,6 +27,7 @@ import ( "time" "k8s.io/kubernetes/pkg/api/v1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" clientset "k8s.io/kubernetes/pkg/client/clientset_generated/release_1_5" "k8s.io/kubernetes/pkg/fields" "k8s.io/kubernetes/pkg/util/sets" @@ -240,7 +241,7 @@ var _ = framework.KubeDescribe("Cluster size autoscaling [Slow]", func() { By(fmt.Sprintf("New nodes: %v\n", newNodesSet)) registeredNodes := sets.NewString() for nodeName := range newNodesSet { - node, err := f.ClientSet.Core().Nodes().Get(nodeName) + node, err := f.ClientSet.Core().Nodes().Get(nodeName, metav1.GetOptions{}) if err == nil && node != nil { registeredNodes.Insert(nodeName) } else { diff --git a/test/e2e/common/container_probe.go b/test/e2e/common/container_probe.go index 4a693f6bde..2d0dbe35e6 100644 --- a/test/e2e/common/container_probe.go +++ b/test/e2e/common/container_probe.go @@ -21,6 +21,7 @@ import ( "time" "k8s.io/kubernetes/pkg/api/v1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" "k8s.io/kubernetes/pkg/util/intstr" "k8s.io/kubernetes/pkg/util/uuid" "k8s.io/kubernetes/test/e2e/framework" @@ -50,7 +51,7 @@ var _ = framework.KubeDescribe("Probing container", func() { p := podClient.Create(makePodSpec(probe.withInitialDelay().build(), nil)) f.WaitForPodReady(p.Name) - p, err := podClient.Get(p.Name) + p, err := podClient.Get(p.Name, metav1.GetOptions{}) framework.ExpectNoError(err) isReady, err := testutils.PodRunningReady(p) framework.ExpectNoError(err) @@ -76,14 +77,14 @@ var _ = framework.KubeDescribe("Probing container", func() { It("with readiness probe that fails should never be ready and never restart [Conformance]", func() { p := podClient.Create(makePodSpec(probe.withFailing().build(), nil)) Consistently(func() (bool, error) { - p, err := podClient.Get(p.Name) + p, err := podClient.Get(p.Name, metav1.GetOptions{}) if err != nil { return false, err } return v1.IsPodReady(p), nil }, 1*time.Minute, 1*time.Second).ShouldNot(BeTrue(), "pod should not be ready") - p, err := podClient.Get(p.Name) + p, err := podClient.Get(p.Name, metav1.GetOptions{}) framework.ExpectNoError(err) isReady, err := testutils.PodRunningReady(p) @@ -366,7 +367,7 @@ func runLivenessTest(f *framework.Framework, pod *v1.Pod, expectNumRestarts int, // Check the pod's current state and verify that restartCount is present. By("checking the pod's current state and verifying that restartCount is present") - pod, err := podClient.Get(pod.Name) + pod, err := podClient.Get(pod.Name, metav1.GetOptions{}) framework.ExpectNoError(err, fmt.Sprintf("getting pod %s in namespace %s", pod.Name, ns)) initialRestartCount := v1.GetExistingContainerStatus(pod.Status.ContainerStatuses, containerName).RestartCount framework.Logf("Initial restart count of pod %s is %d", pod.Name, initialRestartCount) @@ -376,7 +377,7 @@ func runLivenessTest(f *framework.Framework, pod *v1.Pod, expectNumRestarts int, lastRestartCount := initialRestartCount observedRestarts := int32(0) for start := time.Now(); time.Now().Before(deadline); time.Sleep(2 * time.Second) { - pod, err = podClient.Get(pod.Name) + pod, err = podClient.Get(pod.Name, metav1.GetOptions{}) framework.ExpectNoError(err, fmt.Sprintf("getting pod %s", pod.Name)) restartCount := v1.GetExistingContainerStatus(pod.Status.ContainerStatuses, containerName).RestartCount if restartCount != lastRestartCount { diff --git a/test/e2e/common/downwardapi_volume.go b/test/e2e/common/downwardapi_volume.go index f6c2f57138..8e82aebb29 100644 --- a/test/e2e/common/downwardapi_volume.go +++ b/test/e2e/common/downwardapi_volume.go @@ -22,6 +22,7 @@ import ( "k8s.io/kubernetes/pkg/api/resource" "k8s.io/kubernetes/pkg/api/v1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" "k8s.io/kubernetes/pkg/util/uuid" "k8s.io/kubernetes/test/e2e/framework" @@ -133,7 +134,7 @@ var _ = framework.KubeDescribe("Downward API volume", func() { By("Creating the pod") podClient.CreateSync(pod) - pod, err := podClient.Get(pod.Name) + pod, err := podClient.Get(pod.Name, metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred(), "Failed to get pod %q", pod.Name) Eventually(func() (string, error) { diff --git a/test/e2e/common/pods.go b/test/e2e/common/pods.go index bc49a7af81..c663c2bbb1 100644 --- a/test/e2e/common/pods.go +++ b/test/e2e/common/pods.go @@ -27,6 +27,7 @@ import ( "golang.org/x/net/websocket" "k8s.io/kubernetes/pkg/api/v1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" "k8s.io/kubernetes/pkg/kubelet" "k8s.io/kubernetes/pkg/labels" "k8s.io/kubernetes/pkg/util/intstr" @@ -54,7 +55,7 @@ func testHostIP(podClient *framework.PodClient, pod *v1.Pod) { hostIPTimeout := 2 * time.Minute t := time.Now() for { - p, err := podClient.Get(pod.Name) + p, err := podClient.Get(pod.Name, metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred(), "Failed to get pod %q", pod.Name) if p.Status.HostIP != "" { framework.Logf("Pod %s has hostIP: %s", p.Name, p.Status.HostIP) @@ -100,7 +101,7 @@ func getRestartDelay(podClient *framework.PodClient, podName string, containerNa beginTime := time.Now() for time.Since(beginTime) < (2 * maxBackOffTolerance) { // may just miss the 1st MaxContainerBackOff delay time.Sleep(time.Second) - pod, err := podClient.Get(podName) + pod, err := podClient.Get(podName, metav1.GetOptions{}) framework.ExpectNoError(err, fmt.Sprintf("getting pod %s", podName)) status, ok := v1.GetContainerStatus(pod.Status.ContainerStatuses, containerName) if !ok { @@ -201,7 +202,7 @@ var _ = framework.KubeDescribe("Pods", func() { // may be carried out immediately rather than gracefully. framework.ExpectNoError(f.WaitForPodRunning(pod.Name)) // save the running pod - pod, err = podClient.Get(pod.Name) + pod, err = podClient.Get(pod.Name, metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred(), "failed to GET scheduled pod") framework.Logf("running pod: %#v", pod) diff --git a/test/e2e/common/sysctl.go b/test/e2e/common/sysctl.go index 2dbf5228e7..1d149ddcc4 100644 --- a/test/e2e/common/sysctl.go +++ b/test/e2e/common/sysctl.go @@ -20,6 +20,7 @@ import ( "fmt" "k8s.io/kubernetes/pkg/api/v1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" "k8s.io/kubernetes/pkg/kubelet/events" "k8s.io/kubernetes/pkg/kubelet/sysctl" "k8s.io/kubernetes/pkg/util/uuid" @@ -107,7 +108,7 @@ var _ = framework.KubeDescribe("Sysctls", func() { By("Waiting for pod completion") err = f.WaitForPodNoLongerRunning(pod.Name) Expect(err).NotTo(HaveOccurred()) - pod, err = podClient.Get(pod.Name) + pod, err = podClient.Get(pod.Name, metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred()) By("Checking that the pod succeeded") @@ -148,7 +149,7 @@ var _ = framework.KubeDescribe("Sysctls", func() { By("Waiting for pod completion") err = f.WaitForPodNoLongerRunning(pod.Name) Expect(err).NotTo(HaveOccurred()) - pod, err = podClient.Get(pod.Name) + pod, err = podClient.Get(pod.Name, metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred()) By("Checking that the pod succeeded") diff --git a/test/e2e/common/volumes.go b/test/e2e/common/volumes.go index 94156b696e..2da8ff4f2b 100644 --- a/test/e2e/common/volumes.go +++ b/test/e2e/common/volumes.go @@ -148,7 +148,7 @@ func startVolumeServer(f *framework.Framework, config VolumeTestConfig) *v1.Pod serverPod = podClient.CreateSync(serverPod) By("locating the server pod") - pod, err := podClient.Get(serverPod.Name) + pod, err := podClient.Get(serverPod.Name, metav1.GetOptions{}) framework.ExpectNoError(err, "Cannot locate the server pod %v: %v", serverPod.Name, err) By("sleeping a bit to give the server time to start") diff --git a/test/e2e/cronjob.go b/test/e2e/cronjob.go index 3200a968b9..2b23c02378 100644 --- a/test/e2e/cronjob.go +++ b/test/e2e/cronjob.go @@ -26,6 +26,7 @@ import ( "k8s.io/kubernetes/pkg/api/v1" batchv1 "k8s.io/kubernetes/pkg/apis/batch/v1" batch "k8s.io/kubernetes/pkg/apis/batch/v2alpha1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" clientset "k8s.io/kubernetes/pkg/client/clientset_generated/release_1_5" "k8s.io/kubernetes/pkg/controller/job" "k8s.io/kubernetes/pkg/runtime/schema" @@ -235,7 +236,7 @@ func createCronJob(c clientset.Interface, ns string, cronJob *batch.CronJob) (*b } func getCronJob(c clientset.Interface, ns, name string) (*batch.CronJob, error) { - return c.BatchV2alpha1().CronJobs(ns).Get(name) + return c.BatchV2alpha1().CronJobs(ns).Get(name, metav1.GetOptions{}) } func deleteCronJob(c clientset.Interface, ns, name string) error { @@ -245,7 +246,7 @@ func deleteCronJob(c clientset.Interface, ns, name string) error { // Wait for at least given amount of active jobs. func waitForActiveJobs(c clientset.Interface, ns, cronJobName string, active int) error { return wait.Poll(framework.Poll, cronJobTimeout, func() (bool, error) { - curr, err := c.BatchV2alpha1().CronJobs(ns).Get(cronJobName) + curr, err := c.BatchV2alpha1().CronJobs(ns).Get(cronJobName, metav1.GetOptions{}) if err != nil { return false, err } @@ -256,7 +257,7 @@ func waitForActiveJobs(c clientset.Interface, ns, cronJobName string, active int // Wait for no jobs to appear. func waitForNoJobs(c clientset.Interface, ns, jobName string) error { return wait.Poll(framework.Poll, cronJobTimeout, func() (bool, error) { - curr, err := c.BatchV2alpha1().CronJobs(ns).Get(jobName) + curr, err := c.BatchV2alpha1().CronJobs(ns).Get(jobName, metav1.GetOptions{}) if err != nil { return false, err } @@ -312,7 +313,7 @@ func waitForAnyFinishedJob(c clientset.Interface, ns string) error { // checkNoUnexpectedEvents checks unexpected events didn't happen. // Currently only "UnexpectedJob" is checked. func checkNoUnexpectedEvents(c clientset.Interface, ns, cronJobName string) error { - sj, err := c.BatchV2alpha1().CronJobs(ns).Get(cronJobName) + sj, err := c.BatchV2alpha1().CronJobs(ns).Get(cronJobName, metav1.GetOptions{}) if err != nil { return fmt.Errorf("error in getting cronjob %s/%s: %v", ns, cronJobName, err) } diff --git a/test/e2e/daemon_set.go b/test/e2e/daemon_set.go index b5cda9b357..dcfd3485df 100644 --- a/test/e2e/daemon_set.go +++ b/test/e2e/daemon_set.go @@ -303,7 +303,7 @@ func setDaemonSetNodeLabels(c clientset.Interface, nodeName string, labels map[s var newNode *v1.Node var newLabels map[string]string err := wait.Poll(dsRetryPeriod, dsRetryTimeout, func() (bool, error) { - node, err := nodeClient.Get(nodeName) + node, err := nodeClient.Get(nodeName, metav1.GetOptions{}) if err != nil { return false, err } @@ -385,7 +385,7 @@ func checkRunningOnNoNodes(f *framework.Framework, selector map[string]string) f } func checkDaemonStatus(f *framework.Framework, dsName string) error { - ds, err := f.ClientSet.Extensions().DaemonSets(f.Namespace.Name).Get(dsName) + ds, err := f.ClientSet.Extensions().DaemonSets(f.Namespace.Name).Get(dsName, metav1.GetOptions{}) if err != nil { return fmt.Errorf("Could not get daemon set from v1.") } diff --git a/test/e2e/deployment.go b/test/e2e/deployment.go index f8901e0da7..ca002191fc 100644 --- a/test/e2e/deployment.go +++ b/test/e2e/deployment.go @@ -150,7 +150,7 @@ func newDeploymentRollback(name string, annotations map[string]string, revision // checkDeploymentRevision checks if the input deployment's and its new replica set's revision and images are as expected. func checkDeploymentRevision(c clientset.Interface, ns, deploymentName, revision, imageName, image string) (*extensions.Deployment, *extensions.ReplicaSet) { - deployment, err := c.Extensions().Deployments(ns).Get(deploymentName) + deployment, err := c.Extensions().Deployments(ns).Get(deploymentName, metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred()) // Check revision of the new replica set of this deployment newRS, err := deploymentutil.GetNewReplicaSet(deployment, c) @@ -180,7 +180,7 @@ func stopDeployment(c clientset.Interface, internalClient internalclientset.Inte } func stopDeploymentMaybeOverlap(c clientset.Interface, internalClient internalclientset.Interface, ns, deploymentName, overlapWith string) { - deployment, err := c.Extensions().Deployments(ns).Get(deploymentName) + deployment, err := c.Extensions().Deployments(ns).Get(deploymentName, metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred()) framework.Logf("Deleting deployment %s", deploymentName) @@ -191,7 +191,7 @@ func stopDeploymentMaybeOverlap(c clientset.Interface, internalClient internalcl Expect(err).NotTo(HaveOccurred()) framework.Logf("Ensuring deployment %s was deleted", deploymentName) - _, err = c.Extensions().Deployments(ns).Get(deployment.Name) + _, err = c.Extensions().Deployments(ns).Get(deployment.Name, metav1.GetOptions{}) Expect(err).To(HaveOccurred()) Expect(errors.IsNotFound(err)).To(BeTrue()) framework.Logf("Ensuring deployment %s's RSes were deleted", deploymentName) @@ -259,7 +259,7 @@ func testNewDeployment(f *framework.Framework) { err = framework.WaitForDeploymentStatus(c, deploy) Expect(err).NotTo(HaveOccurred()) - deployment, err := c.Extensions().Deployments(ns).Get(deploymentName) + deployment, err := c.Extensions().Deployments(ns).Get(deploymentName, metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred()) newRS, err := deploymentutil.GetNewReplicaSet(deployment, c) Expect(err).NotTo(HaveOccurred()) @@ -291,7 +291,7 @@ func testDeleteDeployment(f *framework.Framework) { err = framework.WaitForDeploymentStatus(c, deploy) Expect(err).NotTo(HaveOccurred()) - deployment, err := c.Extensions().Deployments(ns).Get(deploymentName) + deployment, err := c.Extensions().Deployments(ns).Get(deploymentName, metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred()) newRS, err := deploymentutil.GetNewReplicaSet(deployment, c) Expect(err).NotTo(HaveOccurred()) @@ -337,7 +337,7 @@ func testRollingUpdateDeployment(f *framework.Framework) { Expect(err).NotTo(HaveOccurred()) // There should be 1 old RS (nginx-controller, which is adopted) - deployment, err := c.Extensions().Deployments(ns).Get(deploymentName) + deployment, err := c.Extensions().Deployments(ns).Get(deploymentName, metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred()) _, allOldRSs, err := deploymentutil.GetOldReplicaSets(deployment, c) Expect(err).NotTo(HaveOccurred()) @@ -388,7 +388,7 @@ func testRollingUpdateDeploymentEvents(f *framework.Framework) { err = framework.WaitForDeploymentStatus(c, deploy) Expect(err).NotTo(HaveOccurred()) // Verify that the pods were scaled up and down as expected. We use events to verify that. - deployment, err := c.Extensions().Deployments(ns).Get(deploymentName) + deployment, err := c.Extensions().Deployments(ns).Get(deploymentName, metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred()) framework.WaitForEvents(c, ns, deployment, 2) events, err := c.Core().Events(ns).Search(deployment) @@ -441,7 +441,7 @@ func testRecreateDeployment(f *framework.Framework) { Expect(err).NotTo(HaveOccurred()) // Verify that the pods were scaled up and down as expected. We use events to verify that. - deployment, err := c.Extensions().Deployments(ns).Get(deploymentName) + deployment, err := c.Extensions().Deployments(ns).Get(deploymentName, metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred()) framework.WaitForEvents(c, ns, deployment, 2) events, err := c.Core().Events(ns).Search(deployment) @@ -572,7 +572,7 @@ func testRolloverDeployment(f *framework.Framework) { Expect(err).NotTo(HaveOccurred()) // Verify that the pods were scaled up and down as expected. - deployment, err := c.Extensions().Deployments(ns).Get(deploymentName) + deployment, err := c.Extensions().Deployments(ns).Get(deploymentName, metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred()) framework.Logf("Make sure deployment %q performs scaling operations", deploymentName) // Make sure the deployment starts to scale up and down replica sets by checking if its updated replicas >= 1 @@ -619,7 +619,7 @@ func testPausedDeployment(f *framework.Framework) { _, err := c.Extensions().Deployments(ns).Create(d) Expect(err).NotTo(HaveOccurred()) // Check that deployment is created fine. - deployment, err := c.Extensions().Deployments(ns).Get(deploymentName) + deployment, err := c.Extensions().Deployments(ns).Get(deploymentName, metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred()) // Verify that there is no latest state realized for the new deployment. @@ -846,7 +846,7 @@ func testRollbackDeploymentRSNoRevision(f *framework.Framework) { Expect(err).NotTo(HaveOccurred()) // Check that the replica set we created still doesn't contain revision information - rs, err = c.Extensions().ReplicaSets(ns).Get(rsName) + rs, err = c.Extensions().ReplicaSets(ns).Get(rsName, metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred()) Expect(rs.Annotations[deploymentutil.RevisionAnnotation]).Should(Equal("")) @@ -976,7 +976,7 @@ func testDeploymentLabelAdopted(f *framework.Framework) { Expect(err).NotTo(HaveOccurred()) // There should be no old RSs (overlapping RS) - deployment, err := c.Extensions().Deployments(ns).Get(deploymentName) + deployment, err := c.Extensions().Deployments(ns).Get(deploymentName, metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred()) oldRSs, allOldRSs, newRS, err := deploymentutil.GetAllReplicaSets(deployment, c) Expect(err).NotTo(HaveOccurred()) @@ -1011,7 +1011,7 @@ func testScalePausedDeployment(f *framework.Framework) { Expect(err).NotTo(HaveOccurred()) // Check that deployment is created fine. - deployment, err := c.Extensions().Deployments(ns).Get(deploymentName) + deployment, err := c.Extensions().Deployments(ns).Get(deploymentName, metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred()) err = framework.WaitForObservedDeployment(c, ns, deploymentName, deployment.Generation) @@ -1091,7 +1091,7 @@ func testScaledRolloutDeployment(f *framework.Framework) { err = framework.WaitForObservedDeployment(c, ns, deploymentName, deployment.Generation) Expect(err).NotTo(HaveOccurred()) - deployment, err = c.Extensions().Deployments(ns).Get(deploymentName) + deployment, err = c.Extensions().Deployments(ns).Get(deploymentName, metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred()) if deployment.Status.AvailableReplicas < deploymentutil.MinAvailable(deployment) { @@ -1102,7 +1102,7 @@ func testScaledRolloutDeployment(f *framework.Framework) { second, err := deploymentutil.GetNewReplicaSet(deployment, c) Expect(err).NotTo(HaveOccurred()) - first, err = c.Extensions().ReplicaSets(first.Namespace).Get(first.Name) + first, err = c.Extensions().ReplicaSets(first.Namespace).Get(first.Name, metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred()) firstCond := replicaSetHasDesiredReplicas(c.Extensions(), first) @@ -1151,7 +1151,7 @@ func testScaledRolloutDeployment(f *framework.Framework) { err = framework.WaitForObservedDeployment(c, ns, deploymentName, deployment.Generation) Expect(err).NotTo(HaveOccurred()) - deployment, err = c.Extensions().Deployments(ns).Get(deploymentName) + deployment, err = c.Extensions().Deployments(ns).Get(deploymentName, metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred()) if deployment.Status.AvailableReplicas < deploymentutil.MinAvailable(deployment) { @@ -1159,7 +1159,7 @@ func testScaledRolloutDeployment(f *framework.Framework) { } By(fmt.Sprintf("Checking that the replica sets for %q are synced", deploymentName)) - oldRs, err := c.Extensions().ReplicaSets(rs.Namespace).Get(rs.Name) + oldRs, err := c.Extensions().ReplicaSets(rs.Namespace).Get(rs.Name, metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred()) newRs, err := deploymentutil.GetNewReplicaSet(deployment, c) @@ -1459,7 +1459,7 @@ func testIterativeDeployments(f *framework.Framework) { } // unpause the deployment if we end up pausing it - deployment, err = c.Extensions().Deployments(ns).Get(deployment.Name) + deployment, err = c.Extensions().Deployments(ns).Get(deployment.Name, metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred()) if deployment.Spec.Paused { deployment, err = framework.UpdateDeploymentWithRetries(c, ns, deployment.Name, func(update *extensions.Deployment) { @@ -1480,7 +1480,7 @@ func testIterativeDeployments(f *framework.Framework) { func replicaSetHasDesiredReplicas(rsClient extensionsclient.ReplicaSetsGetter, replicaSet *extensions.ReplicaSet) wait.ConditionFunc { desiredGeneration := replicaSet.Generation return func() (bool, error) { - rs, err := rsClient.ReplicaSets(replicaSet.Namespace).Get(replicaSet.Name) + rs, err := rsClient.ReplicaSets(replicaSet.Namespace).Get(replicaSet.Name, metav1.GetOptions{}) if err != nil { return false, err } diff --git a/test/e2e/dns.go b/test/e2e/dns.go index 50cf871707..3727f7ba6e 100644 --- a/test/e2e/dns.go +++ b/test/e2e/dns.go @@ -237,7 +237,7 @@ func validateDNSResults(f *framework.Framework, pod *v1.Pod, fileNames []string) framework.ExpectNoError(f.WaitForPodRunning(pod.Name)) By("retrieving the pod") - pod, err := podClient.Get(pod.Name) + pod, err := podClient.Get(pod.Name, metav1.GetOptions{}) if err != nil { framework.Failf("Failed to get pod %s: %v", pod.Name, err) } @@ -266,7 +266,7 @@ func validateTargetedProbeOutput(f *framework.Framework, pod *v1.Pod, fileNames framework.ExpectNoError(f.WaitForPodRunning(pod.Name)) By("retrieving the pod") - pod, err := podClient.Get(pod.Name) + pod, err := podClient.Get(pod.Name, metav1.GetOptions{}) if err != nil { framework.Failf("Failed to get pod %s: %v", pod.Name, err) } diff --git a/test/e2e/dns_autoscaling.go b/test/e2e/dns_autoscaling.go index 4f0832849d..cd7553e71c 100644 --- a/test/e2e/dns_autoscaling.go +++ b/test/e2e/dns_autoscaling.go @@ -26,6 +26,7 @@ import ( "k8s.io/kubernetes/pkg/api" "k8s.io/kubernetes/pkg/api/resource" "k8s.io/kubernetes/pkg/api/v1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" clientset "k8s.io/kubernetes/pkg/client/clientset_generated/release_1_5" "k8s.io/kubernetes/pkg/labels" "k8s.io/kubernetes/pkg/util/wait" @@ -230,7 +231,7 @@ func getScheduableCores(nodes []v1.Node) int64 { } func fetchDNSScalingConfigMap(c clientset.Interface) (*v1.ConfigMap, error) { - cm, err := c.Core().ConfigMaps(api.NamespaceSystem).Get(DNSAutoscalerLabelName) + cm, err := c.Core().ConfigMaps(api.NamespaceSystem).Get(DNSAutoscalerLabelName, metav1.GetOptions{}) if err != nil { return nil, err } diff --git a/test/e2e/events.go b/test/e2e/events.go index 1049bc1fd5..c174b32808 100644 --- a/test/e2e/events.go +++ b/test/e2e/events.go @@ -22,6 +22,7 @@ import ( "time" "k8s.io/kubernetes/pkg/api/v1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" "k8s.io/kubernetes/pkg/fields" "k8s.io/kubernetes/pkg/labels" "k8s.io/kubernetes/pkg/util/uuid" @@ -79,7 +80,7 @@ var _ = framework.KubeDescribe("Events", func() { Expect(len(pods.Items)).To(Equal(1)) By("retrieving the pod") - podWithUid, err := podClient.Get(pod.Name) + podWithUid, err := podClient.Get(pod.Name, metav1.GetOptions{}) if err != nil { framework.Failf("Failed to get pod: %v", err) } diff --git a/test/e2e/examples.go b/test/e2e/examples.go index 149ae66ba3..31d161cbf5 100644 --- a/test/e2e/examples.go +++ b/test/e2e/examples.go @@ -27,6 +27,7 @@ import ( "time" "k8s.io/kubernetes/pkg/api/v1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" clientset "k8s.io/kubernetes/pkg/client/clientset_generated/release_1_5" "k8s.io/kubernetes/pkg/labels" "k8s.io/kubernetes/pkg/util/wait" @@ -398,7 +399,7 @@ var _ = framework.KubeDescribe("[Feature:Example]", func() { err := framework.WaitForPodNameRunningInNamespace(c, podName, ns) Expect(err).NotTo(HaveOccurred()) for t := time.Now(); time.Since(t) < timeout; time.Sleep(framework.Poll) { - pod, err := c.Core().Pods(ns).Get(podName) + pod, err := c.Core().Pods(ns).Get(podName, metav1.GetOptions{}) framework.ExpectNoError(err, fmt.Sprintf("getting pod %s", podName)) stat := v1.GetExistingContainerStatus(pod.Status.ContainerStatuses, podName) framework.Logf("Pod: %s, restart count:%d", stat.Name, stat.RestartCount) diff --git a/test/e2e/federated-ingress.go b/test/e2e/federated-ingress.go index 7ee43d930d..b584ef72b0 100644 --- a/test/e2e/federated-ingress.go +++ b/test/e2e/federated-ingress.go @@ -29,6 +29,7 @@ import ( "k8s.io/kubernetes/pkg/api/errors" "k8s.io/kubernetes/pkg/api/v1" "k8s.io/kubernetes/pkg/apis/extensions/v1beta1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" kubeclientset "k8s.io/kubernetes/pkg/client/clientset_generated/release_1_5" "k8s.io/kubernetes/pkg/util/intstr" "k8s.io/kubernetes/pkg/util/wait" @@ -227,7 +228,7 @@ func verifyCascadingDeletionForIngress(clientset *fedclientset.Clientset, cluste // ingress should be present in underlying clusters unless orphanDependents is false. shouldExist := orphanDependents == nil || *orphanDependents == true for clusterName, clusterClientset := range clusters { - _, err := clusterClientset.Extensions().Ingresses(nsName).Get(ingressName) + _, err := clusterClientset.Extensions().Ingresses(nsName).Get(ingressName, metav1.GetOptions{}) if shouldExist && errors.IsNotFound(err) { errMessages = append(errMessages, fmt.Sprintf("unexpected NotFound error for ingress %s in cluster %s, expected ingress to exist", ingressName, clusterName)) } else if !shouldExist && !errors.IsNotFound(err) { @@ -247,7 +248,7 @@ func waitForIngressOrFail(clientset *kubeclientset.Clientset, namespace string, By(fmt.Sprintf("Fetching a federated ingress shard of ingress %q in namespace %q from cluster", ingress.Name, namespace)) var clusterIngress *v1beta1.Ingress err := wait.PollImmediate(framework.Poll, timeout, func() (bool, error) { - clusterIngress, err := clientset.Ingresses(namespace).Get(ingress.Name) + clusterIngress, err := clientset.Ingresses(namespace).Get(ingress.Name, metav1.GetOptions{}) if (!present) && errors.IsNotFound(err) { // We want it gone, and it's gone. By(fmt.Sprintf("Success: shard of federated ingress %q in namespace %q in cluster is absent", ingress.Name, namespace)) return true, nil // Success @@ -293,7 +294,7 @@ func waitForIngressShardsUpdatedOrFail(namespace string, ingress *v1beta1.Ingres func waitForIngressUpdateOrFail(clientset *kubeclientset.Clientset, namespace string, ingress *v1beta1.Ingress, timeout time.Duration) { By(fmt.Sprintf("Fetching a federated ingress shard of ingress %q in namespace %q from cluster", ingress.Name, namespace)) err := wait.PollImmediate(framework.Poll, timeout, func() (bool, error) { - clusterIngress, err := clientset.Ingresses(namespace).Get(ingress.Name) + clusterIngress, err := clientset.Ingresses(namespace).Get(ingress.Name, metav1.GetOptions{}) if err == nil { // We want it present, and the Get succeeded, so we're all good. if equivalentIngress(*clusterIngress, *ingress) { By(fmt.Sprintf("Success: shard of federated ingress %q in namespace %q in cluster is updated", ingress.Name, namespace)) @@ -326,7 +327,7 @@ func deleteIngressOrFail(clientset *fedclientset.Clientset, namespace string, in framework.ExpectNoError(err, "Error deleting ingress %q from namespace %q", ingressName, namespace) // Wait for the ingress to be deleted. err = wait.Poll(framework.Poll, wait.ForeverTestTimeout, func() (bool, error) { - _, err := clientset.Extensions().Ingresses(namespace).Get(ingressName) + _, err := clientset.Extensions().Ingresses(namespace).Get(ingressName, metav1.GetOptions{}) if err != nil && errors.IsNotFound(err) { return true, nil } @@ -456,7 +457,7 @@ func waitForFederatedIngressAddress(c *fedclientset.Clientset, ns, ingName strin // waitForFederatedIngressExists waits for the Ingress object exists. func waitForFederatedIngressExists(c *fedclientset.Clientset, ns, ingName string, timeout time.Duration) error { err := wait.PollImmediate(10*time.Second, timeout, func() (bool, error) { - _, err := c.Extensions().Ingresses(ns).Get(ingName) + _, err := c.Extensions().Ingresses(ns).Get(ingName, metav1.GetOptions{}) if err != nil { framework.Logf("Waiting for Ingress %v, error %v", ingName, err) return false, nil @@ -468,7 +469,7 @@ func waitForFederatedIngressExists(c *fedclientset.Clientset, ns, ingName string // getFederatedIngressAddress returns the ips/hostnames associated with the Ingress. func getFederatedIngressAddress(client *fedclientset.Clientset, ns, name string) ([]string, error) { - ing, err := client.Extensions().Ingresses(ns).Get(name) + ing, err := client.Extensions().Ingresses(ns).Get(name, metav1.GetOptions{}) if err != nil { return nil, err } diff --git a/test/e2e/federated-namespace.go b/test/e2e/federated-namespace.go index 4281838064..40acee5c0f 100644 --- a/test/e2e/federated-namespace.go +++ b/test/e2e/federated-namespace.go @@ -26,6 +26,7 @@ import ( "k8s.io/kubernetes/pkg/api/errors" "k8s.io/kubernetes/pkg/api/v1" api_v1 "k8s.io/kubernetes/pkg/api/v1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" "k8s.io/kubernetes/pkg/util/wait" "k8s.io/kubernetes/test/e2e/framework" @@ -132,7 +133,7 @@ var _ = framework.KubeDescribe("Federation namespace [Feature:Federation]", func f.FederationClientset_1_5.Core().Namespaces().Delete) By(fmt.Sprintf("Verify that event %s was deleted as well", event.Name)) - latestEvent, err := f.FederationClientset_1_5.Core().Events(nsName).Get(event.Name) + latestEvent, err := f.FederationClientset_1_5.Core().Events(nsName).Get(event.Name, metav1.GetOptions{}) if !errors.IsNotFound(err) { framework.Failf("Event %s should have been deleted. Found: %v", event.Name, latestEvent) } @@ -150,7 +151,7 @@ func verifyNsCascadingDeletion(nsClient clientset.NamespaceInterface, clusters m By(fmt.Sprintf("Waiting for namespace %s to be created in all underlying clusters", nsName)) err := wait.Poll(5*time.Second, 2*time.Minute, func() (bool, error) { for _, cluster := range clusters { - _, err := cluster.Core().Namespaces().Get(nsName) + _, err := cluster.Core().Namespaces().Get(nsName, metav1.GetOptions{}) if err != nil && !errors.IsNotFound(err) { return false, err } @@ -170,7 +171,7 @@ func verifyNsCascadingDeletion(nsClient clientset.NamespaceInterface, clusters m // namespace should be present in underlying clusters unless orphanDependents is false. shouldExist := orphanDependents == nil || *orphanDependents == true for clusterName, clusterClientset := range clusters { - _, err := clusterClientset.Core().Namespaces().Get(nsName) + _, err := clusterClientset.Core().Namespaces().Get(nsName, metav1.GetOptions{}) if shouldExist && errors.IsNotFound(err) { errMessages = append(errMessages, fmt.Sprintf("unexpected NotFound error for namespace %s in cluster %s, expected namespace to exist", nsName, clusterName)) } else if !shouldExist && !errors.IsNotFound(err) { diff --git a/test/e2e/federated-secret.go b/test/e2e/federated-secret.go index ae765bad0d..eb4d18e007 100644 --- a/test/e2e/federated-secret.go +++ b/test/e2e/federated-secret.go @@ -27,6 +27,7 @@ import ( "k8s.io/kubernetes/federation/pkg/federation-controller/util" "k8s.io/kubernetes/pkg/api/errors" "k8s.io/kubernetes/pkg/api/v1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" kubeclientset "k8s.io/kubernetes/pkg/client/clientset_generated/release_1_5" "k8s.io/kubernetes/pkg/util/wait" "k8s.io/kubernetes/test/e2e/framework" @@ -116,7 +117,7 @@ func verifyCascadingDeletionForSecret(clientset *fedclientset.Clientset, cluster By(fmt.Sprintf("Waiting for secret %s to be created in all underlying clusters", secretName)) err := wait.Poll(5*time.Second, 2*time.Minute, func() (bool, error) { for _, cluster := range clusters { - _, err := cluster.Core().Secrets(nsName).Get(secretName) + _, err := cluster.Core().Secrets(nsName).Get(secretName, metav1.GetOptions{}) if err != nil { if !errors.IsNotFound(err) { return false, err @@ -136,7 +137,7 @@ func verifyCascadingDeletionForSecret(clientset *fedclientset.Clientset, cluster // secret should be present in underlying clusters unless orphanDependents is false. shouldExist := orphanDependents == nil || *orphanDependents == true for clusterName, clusterClientset := range clusters { - _, err := clusterClientset.Core().Secrets(nsName).Get(secretName) + _, err := clusterClientset.Core().Secrets(nsName).Get(secretName, metav1.GetOptions{}) if shouldExist && errors.IsNotFound(err) { errMessages = append(errMessages, fmt.Sprintf("unexpected NotFound error for secret %s in cluster %s, expected secret to exist", secretName, clusterName)) } else if !shouldExist && !errors.IsNotFound(err) { @@ -173,7 +174,7 @@ func deleteSecretOrFail(clientset *fedclientset.Clientset, nsName string, secret // Wait for the secret to be deleted. err = wait.Poll(5*time.Second, wait.ForeverTestTimeout, func() (bool, error) { - _, err := clientset.Core().Secrets(nsName).Get(secretName) + _, err := clientset.Core().Secrets(nsName).Get(secretName, metav1.GetOptions{}) if err != nil && errors.IsNotFound(err) { return true, nil } @@ -191,7 +192,7 @@ func updateSecretOrFail(clientset *fedclientset.Clientset, nsName string, secret var newSecret *v1.Secret for retryCount := 0; retryCount < MaxRetries; retryCount++ { - secret, err := clientset.Core().Secrets(nsName).Get(secretName) + secret, err := clientset.Core().Secrets(nsName).Get(secretName, metav1.GetOptions{}) if err != nil { framework.Failf("failed to get secret %q: %v", secretName, err) } @@ -223,7 +224,7 @@ func waitForSecretOrFail(clientset *kubeclientset.Clientset, nsName string, secr By(fmt.Sprintf("Fetching a federated secret shard of secret %q in namespace %q from cluster", secret.Name, nsName)) var clusterSecret *v1.Secret err := wait.PollImmediate(framework.Poll, timeout, func() (bool, error) { - clusterSecret, err := clientset.Core().Secrets(nsName).Get(secret.Name) + clusterSecret, err := clientset.Core().Secrets(nsName).Get(secret.Name, metav1.GetOptions{}) if (!present) && errors.IsNotFound(err) { // We want it gone, and it's gone. By(fmt.Sprintf("Success: shard of federated secret %q in namespace %q in cluster is absent", secret.Name, nsName)) return true, nil // Success @@ -252,7 +253,7 @@ func waitForSecretShardsUpdatedOrFail(nsName string, secret *v1.Secret, clusters func waitForSecretUpdateOrFail(clientset *kubeclientset.Clientset, nsName string, secret *v1.Secret, timeout time.Duration) { By(fmt.Sprintf("Fetching a federated secret shard of secret %q in namespace %q from cluster", secret.Name, nsName)) err := wait.PollImmediate(framework.Poll, timeout, func() (bool, error) { - clusterSecret, err := clientset.Core().Secrets(nsName).Get(secret.Name) + clusterSecret, err := clientset.Core().Secrets(nsName).Get(secret.Name, metav1.GetOptions{}) if err == nil { // We want it present, and the Get succeeded, so we're all good. if util.SecretEquivalent(*clusterSecret, *secret) { By(fmt.Sprintf("Success: shard of federated secret %q in namespace %q in cluster is updated", secret.Name, nsName)) diff --git a/test/e2e/federated-service.go b/test/e2e/federated-service.go index f232dcd3bb..6ab8725117 100644 --- a/test/e2e/federated-service.go +++ b/test/e2e/federated-service.go @@ -25,6 +25,7 @@ import ( "k8s.io/kubernetes/pkg/api" "k8s.io/kubernetes/pkg/api/v1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" "k8s.io/kubernetes/test/e2e/framework" . "github.com/onsi/ginkgo" @@ -129,7 +130,7 @@ var _ = framework.KubeDescribe("[Feature:Federation]", func() { By(fmt.Sprintf("Deletion of service %q in namespace %q succeeded.", service.Name, nsName)) By(fmt.Sprintf("Verifying that services in underlying clusters are not deleted")) for clusterName, clusterClientset := range clusters { - _, err := clusterClientset.Core().Services(service.Namespace).Get(service.Name) + _, err := clusterClientset.Core().Services(service.Namespace).Get(service.Name, metav1.GetOptions{}) if err != nil { framework.Failf("Unexpected error in fetching service %s in cluster %s, %s", service.Name, clusterName, err) } diff --git a/test/e2e/federation-apiserver.go b/test/e2e/federation-apiserver.go index c57f69b291..1285154a3a 100644 --- a/test/e2e/federation-apiserver.go +++ b/test/e2e/federation-apiserver.go @@ -24,6 +24,7 @@ import ( federationapi "k8s.io/kubernetes/federation/apis/federation/v1beta1" "k8s.io/kubernetes/federation/client/clientset_generated/federation_release_1_5" "k8s.io/kubernetes/pkg/api/v1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" "k8s.io/kubernetes/pkg/util/wait" "k8s.io/kubernetes/test/e2e/framework" @@ -124,7 +125,7 @@ func newService(name, namespace string) *v1.Service { // Verify that the cluster is marked ready. func isReady(clusterName string, clientset *federation_release_1_5.Clientset) error { return wait.PollImmediate(time.Second, 5*time.Minute, func() (bool, error) { - c, err := clientset.Federation().Clusters().Get(clusterName) + c, err := clientset.Federation().Clusters().Get(clusterName, metav1.GetOptions{}) if err != nil { return false, err } diff --git a/test/e2e/federation-daemonset.go b/test/e2e/federation-daemonset.go index b5facdcaaf..389d9cb264 100644 --- a/test/e2e/federation-daemonset.go +++ b/test/e2e/federation-daemonset.go @@ -28,6 +28,7 @@ import ( "k8s.io/kubernetes/pkg/api/errors" "k8s.io/kubernetes/pkg/api/v1" "k8s.io/kubernetes/pkg/apis/extensions/v1beta1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" kubeclientset "k8s.io/kubernetes/pkg/client/clientset_generated/release_1_5" "k8s.io/kubernetes/pkg/util/wait" "k8s.io/kubernetes/test/e2e/framework" @@ -120,7 +121,7 @@ func verifyCascadingDeletionForDS(clientset *fedclientset.Clientset, clusters ma By(fmt.Sprintf("Waiting for daemonset %s to be created in all underlying clusters", daemonsetName)) err := wait.Poll(5*time.Second, 2*time.Minute, func() (bool, error) { for _, cluster := range clusters { - _, err := cluster.Extensions().DaemonSets(nsName).Get(daemonsetName) + _, err := cluster.Extensions().DaemonSets(nsName).Get(daemonsetName, metav1.GetOptions{}) if err != nil && errors.IsNotFound(err) { return false, nil } @@ -140,7 +141,7 @@ func verifyCascadingDeletionForDS(clientset *fedclientset.Clientset, clusters ma // daemon set should be present in underlying clusters unless orphanDependents is false. shouldExist := orphanDependents == nil || *orphanDependents == true for clusterName, clusterClientset := range clusters { - _, err := clusterClientset.Extensions().DaemonSets(nsName).Get(daemonsetName) + _, err := clusterClientset.Extensions().DaemonSets(nsName).Get(daemonsetName, metav1.GetOptions{}) if shouldExist && errors.IsNotFound(err) { errMessages = append(errMessages, fmt.Sprintf("unexpected NotFound error for daemonset %s in cluster %s, expected daemonset to exist", daemonsetName, clusterName)) } else if !shouldExist && !errors.IsNotFound(err) { @@ -193,7 +194,7 @@ func deleteDaemonSetOrFail(clientset *fedclientset.Clientset, nsName string, dae // Wait for the daemonset to be deleted. err = wait.Poll(5*time.Second, wait.ForeverTestTimeout, func() (bool, error) { - _, err := clientset.Extensions().DaemonSets(nsName).Get(daemonsetName) + _, err := clientset.Extensions().DaemonSets(nsName).Get(daemonsetName, metav1.GetOptions{}) if err != nil && errors.IsNotFound(err) { return true, nil } @@ -211,7 +212,7 @@ func updateDaemonSetOrFail(clientset *fedclientset.Clientset, namespace string) var newDaemonSet *v1beta1.DaemonSet for retryCount := 0; retryCount < FederatedDaemonSetMaxRetries; retryCount++ { - daemonset, err := clientset.Extensions().DaemonSets(namespace).Get(FederatedDaemonSetName) + daemonset, err := clientset.Extensions().DaemonSets(namespace).Get(FederatedDaemonSetName, metav1.GetOptions{}) if err != nil { framework.Failf("failed to get daemonset %q: %v", FederatedDaemonSetName, err) } @@ -241,7 +242,7 @@ func waitForDaemonSetOrFail(clientset *kubeclientset.Clientset, namespace string By(fmt.Sprintf("Fetching a federated daemonset shard of daemonset %q in namespace %q from cluster", daemonset.Name, namespace)) var clusterDaemonSet *v1beta1.DaemonSet err := wait.PollImmediate(framework.Poll, timeout, func() (bool, error) { - clusterDaemonSet, err := clientset.Extensions().DaemonSets(namespace).Get(daemonset.Name) + clusterDaemonSet, err := clientset.Extensions().DaemonSets(namespace).Get(daemonset.Name, metav1.GetOptions{}) if (!present) && errors.IsNotFound(err) { // We want it gone, and it's gone. By(fmt.Sprintf("Success: shard of federated daemonset %q in namespace %q in cluster is absent", daemonset.Name, namespace)) return true, nil // Success @@ -270,7 +271,7 @@ func waitForDaemonSetShardsUpdatedOrFail(namespace string, daemonset *v1beta1.Da func waitForDaemonSetUpdateOrFail(clientset *kubeclientset.Clientset, namespace string, daemonset *v1beta1.DaemonSet, timeout time.Duration) { By(fmt.Sprintf("Fetching a federated daemonset shard of daemonset %q in namespace %q from cluster", daemonset.Name, namespace)) err := wait.PollImmediate(framework.Poll, timeout, func() (bool, error) { - clusterDaemonSet, err := clientset.Extensions().DaemonSets(namespace).Get(daemonset.Name) + clusterDaemonSet, err := clientset.Extensions().DaemonSets(namespace).Get(daemonset.Name, metav1.GetOptions{}) if err == nil { // We want it present, and the Get succeeded, so we're all good. if util.ObjectMetaAndSpecEquivalent(clusterDaemonSet, daemonset) { By(fmt.Sprintf("Success: shard of federated daemonset %q in namespace %q in cluster is updated", daemonset.Name, namespace)) diff --git a/test/e2e/federation-deployment.go b/test/e2e/federation-deployment.go index de4b8379b0..f4bfb0de2f 100644 --- a/test/e2e/federation-deployment.go +++ b/test/e2e/federation-deployment.go @@ -154,7 +154,7 @@ func verifyCascadingDeletionForDeployment(clientset *fedclientset.Clientset, clu By(fmt.Sprintf("Waiting for deployment %s to be created in all underlying clusters", deploymentName)) err := wait.Poll(5*time.Second, 2*time.Minute, func() (bool, error) { for _, cluster := range clusters { - _, err := cluster.Extensions().Deployments(nsName).Get(deploymentName) + _, err := cluster.Extensions().Deployments(nsName).Get(deploymentName, metav1.GetOptions{}) if err != nil && errors.IsNotFound(err) { return false, nil } @@ -174,7 +174,7 @@ func verifyCascadingDeletionForDeployment(clientset *fedclientset.Clientset, clu // deployment should be present in underlying clusters unless orphanDependents is false. shouldExist := orphanDependents == nil || *orphanDependents == true for clusterName, clusterClientset := range clusters { - _, err := clusterClientset.Extensions().Deployments(nsName).Get(deploymentName) + _, err := clusterClientset.Extensions().Deployments(nsName).Get(deploymentName, metav1.GetOptions{}) if shouldExist && errors.IsNotFound(err) { errMessages = append(errMessages, fmt.Sprintf("unexpected NotFound error for deployment %s in cluster %s, expected deployment to exist", deploymentName, clusterName)) } else if !shouldExist && !errors.IsNotFound(err) { @@ -193,13 +193,13 @@ func waitForDeploymentOrFail(c *fedclientset.Clientset, namespace string, deploy func waitForDeployment(c *fedclientset.Clientset, namespace string, deploymentName string, clusters map[string]*cluster) error { err := wait.Poll(10*time.Second, FederatedDeploymentTimeout, func() (bool, error) { - fdep, err := c.Deployments(namespace).Get(deploymentName) + fdep, err := c.Deployments(namespace).Get(deploymentName, metav1.GetOptions{}) if err != nil { return false, err } specReplicas, statusReplicas := int32(0), int32(0) for _, cluster := range clusters { - dep, err := cluster.Deployments(namespace).Get(deploymentName) + dep, err := cluster.Deployments(namespace).Get(deploymentName, metav1.GetOptions{}) if err != nil && !errors.IsNotFound(err) { By(fmt.Sprintf("Failed getting deployment: %q/%q/%q, err: %v", cluster.name, namespace, deploymentName, err)) return false, err @@ -265,7 +265,7 @@ func deleteDeploymentOrFail(clientset *fedclientset.Clientset, nsName string, de // Wait for the deployment to be deleted. err = wait.Poll(5*time.Second, wait.ForeverTestTimeout, func() (bool, error) { - _, err := clientset.Extensions().Deployments(nsName).Get(deploymentName) + _, err := clientset.Extensions().Deployments(nsName).Get(deploymentName, metav1.GetOptions{}) if err != nil && errors.IsNotFound(err) { return true, nil } diff --git a/test/e2e/federation-replicaset.go b/test/e2e/federation-replicaset.go index 7e6b547c3e..ef637d67a6 100644 --- a/test/e2e/federation-replicaset.go +++ b/test/e2e/federation-replicaset.go @@ -156,7 +156,7 @@ func verifyCascadingDeletionForReplicaSet(clientset *fedclientset.Clientset, clu By(fmt.Sprintf("Waiting for replica sets %s to be created in all underlying clusters", replicaSetName)) err := wait.Poll(5*time.Second, 2*time.Minute, func() (bool, error) { for _, cluster := range clusters { - _, err := cluster.Extensions().ReplicaSets(nsName).Get(replicaSetName) + _, err := cluster.Extensions().ReplicaSets(nsName).Get(replicaSetName, metav1.GetOptions{}) if err != nil && errors.IsNotFound(err) { return false, nil } @@ -174,7 +174,7 @@ func verifyCascadingDeletionForReplicaSet(clientset *fedclientset.Clientset, clu By(fmt.Sprintf("Verifying replica sets %s in underlying clusters", replicaSetName)) errMessages := []string{} for clusterName, clusterClientset := range clusters { - _, err := clusterClientset.Extensions().ReplicaSets(nsName).Get(replicaSetName) + _, err := clusterClientset.Extensions().ReplicaSets(nsName).Get(replicaSetName, metav1.GetOptions{}) if (orphanDependents == nil || *orphanDependents == true) && errors.IsNotFound(err) { errMessages = append(errMessages, fmt.Sprintf("unexpected NotFound error for replica set %s in cluster %s, expected replica set to exist", replicaSetName, clusterName)) } else if (orphanDependents != nil && *orphanDependents == false) && (err == nil || !errors.IsNotFound(err)) { @@ -193,13 +193,13 @@ func waitForReplicaSetOrFail(c *fedclientset.Clientset, namespace string, replic func waitForReplicaSet(c *fedclientset.Clientset, namespace string, replicaSetName string, clusters map[string]*cluster) error { err := wait.Poll(10*time.Second, FederatedReplicaSetTimeout, func() (bool, error) { - frs, err := c.ReplicaSets(namespace).Get(replicaSetName) + frs, err := c.ReplicaSets(namespace).Get(replicaSetName, metav1.GetOptions{}) if err != nil { return false, err } specReplicas, statusReplicas := int32(0), int32(0) for _, cluster := range clusters { - rs, err := cluster.ReplicaSets(namespace).Get(replicaSetName) + rs, err := cluster.ReplicaSets(namespace).Get(replicaSetName, metav1.GetOptions{}) if err != nil && !errors.IsNotFound(err) { By(fmt.Sprintf("Failed getting replicaset: %q/%q/%q, err: %v", cluster.name, namespace, replicaSetName, err)) return false, err @@ -251,7 +251,7 @@ func deleteReplicaSetOrFail(clientset *fedclientset.Clientset, nsName string, re // Wait for the replicaSet to be deleted. err = wait.Poll(5*time.Second, wait.ForeverTestTimeout, func() (bool, error) { - _, err := clientset.Extensions().ReplicaSets(nsName).Get(replicaSetName) + _, err := clientset.Extensions().ReplicaSets(nsName).Get(replicaSetName, metav1.GetOptions{}) if err != nil && errors.IsNotFound(err) { return true, nil } diff --git a/test/e2e/federation-util-14.go b/test/e2e/federation-util-14.go index a64ab984ff..6f4c1c3976 100644 --- a/test/e2e/federation-util-14.go +++ b/test/e2e/federation-util-14.go @@ -22,13 +22,14 @@ import ( federationapi "k8s.io/kubernetes/federation/apis/federation/v1beta1" apiv1 "k8s.io/kubernetes/pkg/api/v1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" "k8s.io/kubernetes/pkg/util/wait" "k8s.io/kubernetes/test/e2e/framework" ) func createClusterObjectOrFail_14(f *framework.Framework, context *framework.E2EContext) { framework.Logf("Looking up cluster: %s", context.Name) - foundCluster, err := f.FederationClientset_1_5.Federation().Clusters().Get(context.Name) + foundCluster, err := f.FederationClientset_1_5.Federation().Clusters().Get(context.Name, metav1.GetOptions{}) if err == nil && foundCluster != nil { return } @@ -68,7 +69,7 @@ func buildClustersOrFail_14(f *framework.Framework) []*federationapi.Cluster { // Wait for all clusters to become ready for up to 5 min. if err := wait.PollImmediate(5*time.Second, 5*time.Minute, func() (bool, error) { for _, context := range contexts { - cluster, err := f.FederationClientset_1_5.Federation().Clusters().Get(context.Name) + cluster, err := f.FederationClientset_1_5.Federation().Clusters().Get(context.Name, metav1.GetOptions{}) if err != nil { return false, err } diff --git a/test/e2e/federation-util.go b/test/e2e/federation-util.go index f01694cf6f..2961c910c0 100644 --- a/test/e2e/federation-util.go +++ b/test/e2e/federation-util.go @@ -27,6 +27,7 @@ import ( "k8s.io/kubernetes/pkg/api" "k8s.io/kubernetes/pkg/api/errors" "k8s.io/kubernetes/pkg/api/v1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" kubeclientset "k8s.io/kubernetes/pkg/client/clientset_generated/release_1_5" "k8s.io/kubernetes/pkg/client/restclient" "k8s.io/kubernetes/pkg/client/unversioned/clientcmd" @@ -85,7 +86,7 @@ func createClusterObjectOrFail(f *framework.Framework, context *framework.E2ECon } func clusterIsReadyOrFail(f *framework.Framework, context *framework.E2EContext) { - c, err := f.FederationClientset_1_5.Federation().Clusters().Get(context.Name) + c, err := f.FederationClientset_1_5.Federation().Clusters().Get(context.Name, metav1.GetOptions{}) framework.ExpectNoError(err, fmt.Sprintf("get cluster: %+v", err)) if c.ObjectMeta.Name != context.Name { framework.Failf("cluster name does not match input context: actual=%+v, expected=%+v", c, context) @@ -140,7 +141,7 @@ func createNamespaceInClusters(clusters map[string]*cluster, f *framework.Framew for name, c := range clusters { // The e2e Framework created the required namespace in federation control plane, but we need to create it in all the others, if it doesn't yet exist. // TODO(nikhiljindal): remove this once we have the namespace controller working as expected. - if _, err := c.Clientset.Core().Namespaces().Get(nsName); errors.IsNotFound(err) { + if _, err := c.Clientset.Core().Namespaces().Get(nsName, metav1.GetOptions{}); errors.IsNotFound(err) { ns := &v1.Namespace{ ObjectMeta: v1.ObjectMeta{ Name: nsName, @@ -164,7 +165,7 @@ func unregisterClusters(clusters map[string]*cluster, f *framework.Framework) { nsName := f.FederationNamespace.Name for name, c := range clusters { if c.namespaceCreated { - if _, err := c.Clientset.Core().Namespaces().Get(nsName); !errors.IsNotFound(err) { + if _, err := c.Clientset.Core().Namespaces().Get(nsName, metav1.GetOptions{}); !errors.IsNotFound(err) { err := c.Clientset.Core().Namespaces().Delete(nsName, &v1.DeleteOptions{}) framework.ExpectNoError(err, "Couldn't delete the namespace %s in cluster %q: %v", nsName, name, err) } @@ -217,7 +218,7 @@ func waitForServiceOrFail(clientset *kubeclientset.Clientset, namespace string, By(fmt.Sprintf("Fetching a federated service shard of service %q in namespace %q from cluster", service.Name, namespace)) var clusterService *v1.Service err := wait.PollImmediate(framework.Poll, timeout, func() (bool, error) { - clusterService, err := clientset.Services(namespace).Get(service.Name) + clusterService, err := clientset.Services(namespace).Get(service.Name, metav1.GetOptions{}) if (!present) && errors.IsNotFound(err) { // We want it gone, and it's gone. By(fmt.Sprintf("Success: shard of federated service %q in namespace %q in cluster is absent", service.Name, namespace)) return true, nil // Success @@ -297,7 +298,7 @@ func cleanupServiceShardsAndProviderResources(namespace string, service *v1.Serv err := wait.PollImmediate(framework.Poll, FederatedServiceTimeout, func() (bool, error) { var err error - cSvc, err = c.Clientset.Services(namespace).Get(service.Name) + cSvc, err = c.Clientset.Services(namespace).Get(service.Name, metav1.GetOptions{}) if err != nil && !errors.IsNotFound(err) { // Get failed with an error, try again. framework.Logf("Failed to find service %q in namespace %q, in cluster %q: %v. Trying again in %s", service.Name, namespace, name, err, framework.Poll) @@ -385,7 +386,7 @@ func podExitCodeDetector(f *framework.Framework, name, namespace string, code in } return func() error { - pod, err := f.ClientSet.Core().Pods(namespace).Get(name) + pod, err := f.ClientSet.Core().Pods(namespace).Get(name, metav1.GetOptions{}) if err != nil { return logerr(err) } diff --git a/test/e2e/framework/exec_util.go b/test/e2e/framework/exec_util.go index f35de2e78c..ffe74fb7a6 100644 --- a/test/e2e/framework/exec_util.go +++ b/test/e2e/framework/exec_util.go @@ -24,6 +24,7 @@ import ( "k8s.io/kubernetes/pkg/api" "k8s.io/kubernetes/pkg/api/v1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" "k8s.io/kubernetes/pkg/client/restclient" "k8s.io/kubernetes/pkg/client/unversioned/remotecommand" remotecommandserver "k8s.io/kubernetes/pkg/kubelet/server/remotecommand" @@ -112,14 +113,14 @@ func (f *Framework) ExecShellInContainer(podName, containerName string, cmd stri } func (f *Framework) ExecCommandInPod(podName string, cmd ...string) string { - pod, err := f.PodClient().Get(podName) + pod, err := f.PodClient().Get(podName, metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred(), "failed to get pod") Expect(pod.Spec.Containers).NotTo(BeEmpty()) return f.ExecCommandInContainer(podName, pod.Spec.Containers[0].Name, cmd...) } func (f *Framework) ExecCommandInPodWithFullOutput(podName string, cmd ...string) (string, string, error) { - pod, err := f.PodClient().Get(podName) + pod, err := f.PodClient().Get(podName, metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred(), "failed to get pod") Expect(pod.Spec.Containers).NotTo(BeEmpty()) return f.ExecCommandInContainerWithFullOutput(podName, pod.Spec.Containers[0].Name, cmd...) diff --git a/test/e2e/framework/framework.go b/test/e2e/framework/framework.go index 1130261e96..0cbc63a090 100644 --- a/test/e2e/framework/framework.go +++ b/test/e2e/framework/framework.go @@ -34,6 +34,7 @@ import ( apierrs "k8s.io/kubernetes/pkg/api/errors" "k8s.io/kubernetes/pkg/api/v1" "k8s.io/kubernetes/pkg/apimachinery/registered" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" "k8s.io/kubernetes/pkg/client/clientset_generated/internalclientset" clientset "k8s.io/kubernetes/pkg/client/clientset_generated/release_1_5" "k8s.io/kubernetes/pkg/client/restclient" @@ -283,7 +284,7 @@ func (f *Framework) deleteFederationNs() { } // Verify that it got deleted. err := wait.PollImmediate(5*time.Second, timeout, func() (bool, error) { - if _, err := clientset.Core().Namespaces().Get(ns.Name); err != nil { + if _, err := clientset.Core().Namespaces().Get(ns.Name, metav1.GetOptions{}); err != nil { if apierrs.IsNotFound(err) { return true, nil } diff --git a/test/e2e/framework/networking_utils.go b/test/e2e/framework/networking_utils.go index 80e8d2f2fa..b7de5b9cad 100644 --- a/test/e2e/framework/networking_utils.go +++ b/test/e2e/framework/networking_utils.go @@ -423,12 +423,12 @@ func (config *NetworkingTestConfig) createTestPods() { ExpectNoError(config.f.WaitForPodRunning(hostTestContainerPod.Name)) var err error - config.TestContainerPod, err = config.getPodClient().Get(testContainerPod.Name) + config.TestContainerPod, err = config.getPodClient().Get(testContainerPod.Name, metav1.GetOptions{}) if err != nil { Failf("Failed to retrieve %s pod: %v", testContainerPod.Name, err) } - config.HostTestContainerPod, err = config.getPodClient().Get(hostTestContainerPod.Name) + config.HostTestContainerPod, err = config.getPodClient().Get(hostTestContainerPod.Name, metav1.GetOptions{}) if err != nil { Failf("Failed to retrieve %s pod: %v", hostTestContainerPod.Name, err) } @@ -441,7 +441,7 @@ func (config *NetworkingTestConfig) createService(serviceSpec *v1.Service) *v1.S err = WaitForService(config.f.ClientSet, config.Namespace, serviceSpec.Name, true, 5*time.Second, 45*time.Second) Expect(err).NotTo(HaveOccurred(), fmt.Sprintf("error while waiting for service:%s err: %v", serviceSpec.Name, err)) - createdService, err := config.getServiceClient().Get(serviceSpec.Name) + createdService, err := config.getServiceClient().Get(serviceSpec.Name, metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred(), fmt.Sprintf("Failed to create %s service: %v", serviceSpec.Name, err)) return createdService @@ -542,7 +542,7 @@ func (config *NetworkingTestConfig) createNetProxyPods(podName string, selector runningPods := make([]*v1.Pod, 0, len(nodes)) for _, p := range createdPods { ExpectNoError(config.f.WaitForPodReady(p.Name)) - rp, err := config.getPodClient().Get(p.Name) + rp, err := config.getPodClient().Get(p.Name, metav1.GetOptions{}) ExpectNoError(err) runningPods = append(runningPods, rp) } diff --git a/test/e2e/framework/pods.go b/test/e2e/framework/pods.go index 47ea4c0edc..efc3cbe97f 100644 --- a/test/e2e/framework/pods.go +++ b/test/e2e/framework/pods.go @@ -24,6 +24,7 @@ import ( "k8s.io/kubernetes/pkg/api/errors" "k8s.io/kubernetes/pkg/api/v1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" v1core "k8s.io/kubernetes/pkg/client/clientset_generated/release_1_5/typed/core/v1" "k8s.io/kubernetes/pkg/labels" "k8s.io/kubernetes/pkg/util/sets" @@ -66,7 +67,7 @@ func (c *PodClient) CreateSync(pod *v1.Pod) *v1.Pod { p := c.Create(pod) ExpectNoError(c.f.WaitForPodRunning(p.Name)) // Get the newest pod after it becomes running, some status may change after pod created, such as pod ip. - p, err := c.Get(p.Name) + p, err := c.Get(p.Name, metav1.GetOptions{}) ExpectNoError(err) return p } @@ -92,7 +93,7 @@ func (c *PodClient) CreateBatch(pods []*v1.Pod) []*v1.Pod { // pod object. func (c *PodClient) Update(name string, updateFn func(pod *v1.Pod)) { ExpectNoError(wait.Poll(time.Millisecond*500, time.Second*30, func() (bool, error) { - pod, err := c.PodInterface.Get(name) + pod, err := c.PodInterface.Get(name, metav1.GetOptions{}) if err != nil { return false, fmt.Errorf("failed to get pod %q: %v", name, err) } diff --git a/test/e2e/framework/util.go b/test/e2e/framework/util.go index 07156a86ce..6a85fc4ffd 100644 --- a/test/e2e/framework/util.go +++ b/test/e2e/framework/util.go @@ -367,7 +367,7 @@ func SkipUnlessFederated(c clientset.Interface) { federationNS = "federation" } - _, err := c.Core().Namespaces().Get(federationNS) + _, err := c.Core().Namespaces().Get(federationNS, metav1.GetOptions{}) if err != nil { if apierrs.IsNotFound(err) { Skipf("Could not find federation namespace %s: skipping federated test", federationNS) @@ -756,7 +756,7 @@ func waitForServiceAccountInNamespace(c clientset.Interface, ns, serviceAccountN func WaitForPodCondition(c clientset.Interface, ns, podName, desc string, timeout time.Duration, condition podCondition) error { Logf("Waiting up to %[1]v for pod %[2]s status to be %[3]s", timeout, podName, desc) for start := time.Now(); time.Since(start) < timeout; time.Sleep(Poll) { - pod, err := c.Core().Pods(ns).Get(podName) + pod, err := c.Core().Pods(ns).Get(podName, metav1.GetOptions{}) if err != nil { if apierrs.IsNotFound(err) { Logf("Pod %q in namespace %q disappeared. Error: %v", podName, ns, err) @@ -828,7 +828,7 @@ func WaitForFederationApiserverReady(c *federation_release_1_5.Clientset) error func WaitForPersistentVolumePhase(phase v1.PersistentVolumePhase, c clientset.Interface, pvName string, Poll, timeout time.Duration) error { Logf("Waiting up to %v for PersistentVolume %s to have phase %s", timeout, pvName, phase) for start := time.Now(); time.Since(start) < timeout; time.Sleep(Poll) { - pv, err := c.Core().PersistentVolumes().Get(pvName) + pv, err := c.Core().PersistentVolumes().Get(pvName, metav1.GetOptions{}) if err != nil { Logf("Get persistent volume %s in failed, ignoring for %v: %v", pvName, Poll, err) continue @@ -848,7 +848,7 @@ func WaitForPersistentVolumePhase(phase v1.PersistentVolumePhase, c clientset.In func WaitForPersistentVolumeDeleted(c clientset.Interface, pvName string, Poll, timeout time.Duration) error { Logf("Waiting up to %v for PersistentVolume %s to get deleted", timeout, pvName) for start := time.Now(); time.Since(start) < timeout; time.Sleep(Poll) { - pv, err := c.Core().PersistentVolumes().Get(pvName) + pv, err := c.Core().PersistentVolumes().Get(pvName, metav1.GetOptions{}) if err == nil { Logf("PersistentVolume %s found and phase=%s (%v)", pvName, pv.Status.Phase, time.Since(start)) continue @@ -868,7 +868,7 @@ func WaitForPersistentVolumeDeleted(c clientset.Interface, pvName string, Poll, func WaitForPersistentVolumeClaimPhase(phase v1.PersistentVolumeClaimPhase, c clientset.Interface, ns string, pvcName string, Poll, timeout time.Duration) error { Logf("Waiting up to %v for PersistentVolumeClaim %s to have phase %s", timeout, pvcName, phase) for start := time.Now(); time.Since(start) < timeout; time.Sleep(Poll) { - pvc, err := c.Core().PersistentVolumeClaims(ns).Get(pvcName) + pvc, err := c.Core().PersistentVolumeClaims(ns).Get(pvcName, metav1.GetOptions{}) if err != nil { Logf("Get persistent volume claim %s in failed, ignoring for %v: %v", pvcName, Poll, err) continue @@ -973,7 +973,7 @@ func deleteNS(c clientset.Interface, clientPool dynamic.ClientPool, namespace st // wait for namespace to delete or timeout. err := wait.PollImmediate(5*time.Second, timeout, func() (bool, error) { - if _, err := c.Core().Namespaces().Get(namespace); err != nil { + if _, err := c.Core().Namespaces().Get(namespace, metav1.GetOptions{}); err != nil { if apierrs.IsNotFound(err) { return true, nil } @@ -1046,7 +1046,7 @@ func logNamespaces(c clientset.Interface, namespace string) { // logNamespace logs detail about a namespace func logNamespace(c clientset.Interface, namespace string) { - ns, err := c.Core().Namespaces().Get(namespace) + ns, err := c.Core().Namespaces().Get(namespace, metav1.GetOptions{}) if err != nil { if apierrs.IsNotFound(err) { Logf("namespace: %v no longer exists", namespace) @@ -1452,7 +1452,7 @@ func WaitForRCPodToDisappear(c clientset.Interface, ns, rcName, podName string) // WaitForService waits until the service appears (exist == true), or disappears (exist == false) func WaitForService(c clientset.Interface, namespace, name string, exist bool, interval, timeout time.Duration) error { err := wait.PollImmediate(interval, timeout, func() (bool, error) { - _, err := c.Core().Services(namespace).Get(name) + _, err := c.Core().Services(namespace).Get(name, metav1.GetOptions{}) switch { case err == nil: if !exist { @@ -1507,7 +1507,7 @@ func countEndpointsNum(e *v1.Endpoints) int { // WaitForReplicationController waits until the RC appears (exist == true), or disappears (exist == false) func WaitForReplicationController(c clientset.Interface, namespace, name string, exist bool, interval, timeout time.Duration) error { err := wait.PollImmediate(interval, timeout, func() (bool, error) { - _, err := c.Core().ReplicationControllers(namespace).Get(name) + _, err := c.Core().ReplicationControllers(namespace).Get(name, metav1.GetOptions{}) if err != nil { Logf("Get ReplicationController %s in namespace %s failed (%v).", name, namespace, err) return !exist, nil @@ -1525,7 +1525,7 @@ func WaitForReplicationController(c clientset.Interface, namespace, name string, func WaitForEndpoint(c clientset.Interface, ns, name string) error { for t := time.Now(); time.Since(t) < EndpointRegisterTimeout; time.Sleep(Poll) { - endpoint, err := c.Core().Endpoints(ns).Get(name) + endpoint, err := c.Core().Endpoints(ns).Get(name, metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred()) if len(endpoint.Subsets) == 0 || len(endpoint.Subsets[0].Addresses) == 0 { Logf("Endpoint %s/%s is not ready yet", ns, name) @@ -2178,7 +2178,7 @@ func (f *Framework) MatchContainerOutput( } // Grab its logs. Get host first. - podStatus, err := podClient.Get(createdPod.Name) + podStatus, err := podClient.Get(createdPod.Name, metav1.GetOptions{}) if err != nil { return fmt.Errorf("failed to get pod status: %v", err) } @@ -2309,7 +2309,7 @@ func dumpAllNodeInfo(c clientset.Interface) { func DumpNodeDebugInfo(c clientset.Interface, nodeNames []string, logFunc func(fmt string, args ...interface{})) { for _, n := range nodeNames { logFunc("\nLogging node info for node %v", n) - node, err := c.Core().Nodes().Get(n) + node, err := c.Core().Nodes().Get(n, metav1.GetOptions{}) if err != nil { logFunc("Error getting node info %v", err) } @@ -2482,7 +2482,7 @@ func AddOrUpdateLabelOnNode(c clientset.Interface, nodeName string, labelKey, la func ExpectNodeHasLabel(c clientset.Interface, nodeName string, labelKey string, labelValue string) { By("verifying the node has the label " + labelKey + " " + labelValue) - node, err := c.Core().Nodes().Get(nodeName) + node, err := c.Core().Nodes().Get(nodeName, metav1.GetOptions{}) ExpectNoError(err) Expect(node.Labels[labelKey]).To(Equal(labelValue)) } @@ -2499,7 +2499,7 @@ func RemoveLabelOffNode(c clientset.Interface, nodeName string, labelKey string) func AddOrUpdateTaintOnNode(c clientset.Interface, nodeName string, taint v1.Taint) { for attempt := 0; attempt < UpdateRetries; attempt++ { - node, err := c.Core().Nodes().Get(nodeName) + node, err := c.Core().Nodes().Get(nodeName, metav1.GetOptions{}) ExpectNoError(err) nodeTaints, err := v1.GetTaintsFromNodeAnnotations(node.Annotations) @@ -2553,7 +2553,7 @@ func taintExists(taints []v1.Taint, taintToFind v1.Taint) bool { func ExpectNodeHasTaint(c clientset.Interface, nodeName string, taint v1.Taint) { By("verifying the node has the taint " + taint.ToString()) - node, err := c.Core().Nodes().Get(nodeName) + node, err := c.Core().Nodes().Get(nodeName, metav1.GetOptions{}) ExpectNoError(err) nodeTaints, err := v1.GetTaintsFromNodeAnnotations(node.Annotations) @@ -2586,7 +2586,7 @@ func deleteTaint(oldTaints []v1.Taint, taintToDelete v1.Taint) ([]v1.Taint, erro func RemoveTaintOffNode(c clientset.Interface, nodeName string, taint v1.Taint) { By("removing the taint " + taint.ToString() + " off the node " + nodeName) for attempt := 0; attempt < UpdateRetries; attempt++ { - node, err := c.Core().Nodes().Get(nodeName) + node, err := c.Core().Nodes().Get(nodeName, metav1.GetOptions{}) ExpectNoError(err) nodeTaints, err := v1.GetTaintsFromNodeAnnotations(node.Annotations) @@ -2622,7 +2622,7 @@ func RemoveTaintOffNode(c clientset.Interface, nodeName string, taint v1.Taint) time.Sleep(100 * time.Millisecond) } - nodeUpdated, err := c.Core().Nodes().Get(nodeName) + nodeUpdated, err := c.Core().Nodes().Get(nodeName, metav1.GetOptions{}) ExpectNoError(err) By("verifying the node doesn't have the taint " + taint.ToString()) taintsGot, err := v1.GetTaintsFromNodeAnnotations(nodeUpdated.Annotations) @@ -2774,11 +2774,11 @@ func WaitForPodsWithLabelRunningReady(c clientset.Interface, ns string, label la func getRuntimeObjectForKind(c clientset.Interface, kind schema.GroupKind, ns, name string) (runtime.Object, error) { switch kind { case api.Kind("ReplicationController"): - return c.Core().ReplicationControllers(ns).Get(name) + return c.Core().ReplicationControllers(ns).Get(name, metav1.GetOptions{}) case extensionsinternal.Kind("ReplicaSet"): - return c.Extensions().ReplicaSets(ns).Get(name) + return c.Extensions().ReplicaSets(ns).Get(name, metav1.GetOptions{}) case extensionsinternal.Kind("Deployment"): - return c.Extensions().Deployments(ns).Get(name) + return c.Extensions().Deployments(ns).Get(name, metav1.GetOptions{}) default: return nil, fmt.Errorf("Unsupported kind when getting runtime object: %v", kind) } @@ -3020,7 +3020,7 @@ func waitForPodsGone(ps *testutils.PodStore, interval, timeout time.Duration) er // Delete a ReplicaSet and all pods it spawned func DeleteReplicaSet(clientset clientset.Interface, internalClientset internalclientset.Interface, ns, name string) error { By(fmt.Sprintf("deleting ReplicaSet %s in namespace %s", name, ns)) - rc, err := clientset.Extensions().ReplicaSets(ns).Get(name) + rc, err := clientset.Extensions().ReplicaSets(ns).Get(name, metav1.GetOptions{}) if err != nil { if apierrs.IsNotFound(err) { Logf("ReplicaSet %s was already deleted: %v", name, err) @@ -3068,7 +3068,7 @@ func waitForReplicaSetPodsGone(c clientset.Interface, rs *extensions.ReplicaSet) // WaitForReadyReplicaSet waits until the replica set has all of its replicas ready. func WaitForReadyReplicaSet(c clientset.Interface, ns, name string) error { - rs, err := c.Extensions().ReplicaSets(ns).Get(name) + rs, err := c.Extensions().ReplicaSets(ns).Get(name, metav1.GetOptions{}) if err != nil { return err } @@ -3134,7 +3134,7 @@ func WaitForDeploymentStatusValid(c clientset.Interface, d *extensions.Deploymen err := wait.Poll(Poll, 5*time.Minute, func() (bool, error) { var err error - deployment, err = c.Extensions().Deployments(d.Namespace).Get(d.Name) + deployment, err = c.Extensions().Deployments(d.Namespace).Get(d.Name, metav1.GetOptions{}) if err != nil { return false, err } @@ -3206,7 +3206,7 @@ func WaitForDeploymentStatus(c clientset.Interface, d *extensions.Deployment) er err := wait.Poll(Poll, 5*time.Minute, func() (bool, error) { var err error - deployment, err = c.Extensions().Deployments(d.Namespace).Get(d.Name) + deployment, err = c.Extensions().Deployments(d.Namespace).Get(d.Name, metav1.GetOptions{}) if err != nil { return false, err } @@ -3260,7 +3260,7 @@ func WaitForDeploymentStatus(c clientset.Interface, d *extensions.Deployment) er // WaitForDeploymentUpdatedReplicasLTE waits for given deployment to be observed by the controller and has at least a number of updatedReplicas func WaitForDeploymentUpdatedReplicasLTE(c clientset.Interface, ns, deploymentName string, minUpdatedReplicas int, desiredGeneration int64) error { err := wait.Poll(Poll, 5*time.Minute, func() (bool, error) { - deployment, err := c.Extensions().Deployments(ns).Get(deploymentName) + deployment, err := c.Extensions().Deployments(ns).Get(deploymentName, metav1.GetOptions{}) if err != nil { return false, err } @@ -3279,7 +3279,7 @@ func WaitForDeploymentUpdatedReplicasLTE(c clientset.Interface, ns, deploymentNa // Note that rollback should be cleared shortly, so we only wait for 1 minute here to fail early. func WaitForDeploymentRollbackCleared(c clientset.Interface, ns, deploymentName string) error { err := wait.Poll(Poll, 1*time.Minute, func() (bool, error) { - deployment, err := c.Extensions().Deployments(ns).Get(deploymentName) + deployment, err := c.Extensions().Deployments(ns).Get(deploymentName, metav1.GetOptions{}) if err != nil { return false, err } @@ -3303,7 +3303,7 @@ func WaitForDeploymentRevisionAndImage(c clientset.Interface, ns, deploymentName var reason string err := wait.Poll(Poll, 1*time.Minute, func() (bool, error) { var err error - deployment, err = c.Extensions().Deployments(ns).Get(deploymentName) + deployment, err = c.Extensions().Deployments(ns).Get(deploymentName, metav1.GetOptions{}) if err != nil { return false, err } @@ -3360,7 +3360,7 @@ func WaitForDeploymentRevisionAndImage(c clientset.Interface, ns, deploymentName func WaitForOverlappingAnnotationMatch(c clientset.Interface, ns, deploymentName, expected string) error { return wait.Poll(Poll, 1*time.Minute, func() (bool, error) { - deployment, err := c.Extensions().Deployments(ns).Get(deploymentName) + deployment, err := c.Extensions().Deployments(ns).Get(deploymentName, metav1.GetOptions{}) if err != nil { return false, err } @@ -3373,7 +3373,7 @@ func WaitForOverlappingAnnotationMatch(c clientset.Interface, ns, deploymentName // CheckNewRSAnnotations check if the new RS's annotation is as expected func CheckNewRSAnnotations(c clientset.Interface, ns, deploymentName string, expectedAnnotations map[string]string) error { - deployment, err := c.Extensions().Deployments(ns).Get(deploymentName) + deployment, err := c.Extensions().Deployments(ns).Get(deploymentName, metav1.GetOptions{}) if err != nil { return err } @@ -3410,7 +3410,7 @@ func WaitForPodsReady(c clientset.Interface, ns, name string, minReadySeconds in // Waits for the deployment to clean up old rcs. func WaitForDeploymentOldRSsNum(c clientset.Interface, ns, deploymentName string, desiredRSNum int) error { return wait.Poll(Poll, 5*time.Minute, func() (bool, error) { - deployment, err := c.Extensions().Deployments(ns).Get(deploymentName) + deployment, err := c.Extensions().Deployments(ns).Get(deploymentName, metav1.GetOptions{}) if err != nil { return false, err } @@ -3435,13 +3435,15 @@ func logReplicaSetsOfDeployment(deployment *extensions.Deployment, allOldRSs []* } func WaitForObservedDeployment(c clientset.Interface, ns, deploymentName string, desiredGeneration int64) error { - return deploymentutil.WaitForObservedDeployment(func() (*extensions.Deployment, error) { return c.Extensions().Deployments(ns).Get(deploymentName) }, desiredGeneration, Poll, 1*time.Minute) + return deploymentutil.WaitForObservedDeployment(func() (*extensions.Deployment, error) { + return c.Extensions().Deployments(ns).Get(deploymentName, metav1.GetOptions{}) + }, desiredGeneration, Poll, 1*time.Minute) } func WaitForDeploymentWithCondition(c clientset.Interface, ns, deploymentName, reason string, condType extensions.DeploymentConditionType) error { var conditions []extensions.DeploymentCondition pollErr := wait.PollImmediate(time.Second, 1*time.Minute, func() (bool, error) { - deployment, err := c.Extensions().Deployments(ns).Get(deploymentName) + deployment, err := c.Extensions().Deployments(ns).Get(deploymentName, metav1.GetOptions{}) if err != nil { return false, err } @@ -3516,7 +3518,7 @@ func UpdateDeploymentWithRetries(c clientset.Interface, namespace, name string, deployments := c.Extensions().Deployments(namespace) var updateErr error pollErr := wait.Poll(10*time.Millisecond, 1*time.Minute, func() (bool, error) { - if deployment, err = deployments.Get(name); err != nil { + if deployment, err = deployments.Get(name, metav1.GetOptions{}); err != nil { return false, err } // Apply the update, then attempt to push it to the apiserver. @@ -3541,7 +3543,7 @@ func UpdateReplicaSetWithRetries(c clientset.Interface, namespace, name string, var updateErr error pollErr := wait.PollImmediate(10*time.Millisecond, 1*time.Minute, func() (bool, error) { var err error - if rs, err = c.Extensions().ReplicaSets(namespace).Get(name); err != nil { + if rs, err = c.Extensions().ReplicaSets(namespace).Get(name, metav1.GetOptions{}); err != nil { return false, err } // Apply the update, then attempt to push it to the apiserver. @@ -3566,7 +3568,7 @@ func UpdateReplicationControllerWithRetries(c clientset.Interface, namespace, na var updateErr error pollErr := wait.PollImmediate(10*time.Millisecond, 1*time.Minute, func() (bool, error) { var err error - if rc, err = c.Core().ReplicationControllers(namespace).Get(name); err != nil { + if rc, err = c.Core().ReplicationControllers(namespace).Get(name, metav1.GetOptions{}); err != nil { return false, err } // Apply the update, then attempt to push it to the apiserver. @@ -3590,7 +3592,7 @@ func UpdateStatefulSetWithRetries(c clientset.Interface, namespace, name string, statefulSets := c.Apps().StatefulSets(namespace) var updateErr error pollErr := wait.Poll(10*time.Millisecond, 1*time.Minute, func() (bool, error) { - if statefulSet, err = statefulSets.Get(name); err != nil { + if statefulSet, err = statefulSets.Get(name, metav1.GetOptions{}); err != nil { return false, err } // Apply the update, then attempt to push it to the apiserver. @@ -3614,7 +3616,7 @@ func UpdateJobWithRetries(c clientset.Interface, namespace, name string, applyUp jobs := c.Batch().Jobs(namespace) var updateErr error pollErr := wait.PollImmediate(10*time.Millisecond, 1*time.Minute, func() (bool, error) { - if job, err = jobs.Get(name); err != nil { + if job, err = jobs.Get(name, metav1.GetOptions{}); err != nil { return false, err } // Apply the update, then attempt to push it to the apiserver. @@ -3932,7 +3934,7 @@ func IsNodeConditionUnset(node *v1.Node, conditionType v1.NodeConditionType) boo func WaitForNodeToBe(c clientset.Interface, name string, conditionType v1.NodeConditionType, wantTrue bool, timeout time.Duration) bool { Logf("Waiting up to %v for node %s condition %s to be %t", timeout, name, conditionType, wantTrue) for start := time.Now(); time.Since(start) < timeout; time.Sleep(Poll) { - node, err := c.Core().Nodes().Get(name) + node, err := c.Core().Nodes().Get(name, metav1.GetOptions{}) if err != nil { Logf("Couldn't get node %s", name) continue @@ -4259,7 +4261,7 @@ func WaitForMasters(masterPrefix string, c clientset.Interface, size int, timeou // address. Returns an error if the node the pod is on doesn't have an External // address. func GetHostExternalAddress(client clientset.Interface, p *v1.Pod) (externalAddress string, err error) { - node, err := client.Core().Nodes().Get(p.Spec.NodeName) + node, err := client.Core().Nodes().Get(p.Spec.NodeName, metav1.GetOptions{}) if err != nil { return "", err } @@ -4335,7 +4337,7 @@ func OpenWebSocketForURL(url *url.URL, config *restclient.Config, protocols []st // getIngressAddress returns the ips/hostnames associated with the Ingress. func getIngressAddress(client clientset.Interface, ns, name string) ([]string, error) { - ing, err := client.Extensions().Ingresses(ns).Get(name) + ing, err := client.Extensions().Ingresses(ns).Get(name, metav1.GetOptions{}) if err != nil { return nil, err } @@ -4406,7 +4408,7 @@ func LookForString(expectedString string, timeout time.Duration, fn func() strin // getSvcNodePort returns the node port for the given service:port. func getSvcNodePort(client clientset.Interface, ns, name string, svcPort int) (int, error) { - svc, err := client.Core().Services(ns).Get(name) + svc, err := client.Core().Services(ns).Get(name, metav1.GetOptions{}) if err != nil { return 0, err } @@ -4471,7 +4473,7 @@ func ScaleRCByLabels(clientset clientset.Interface, internalClientset internalcl if err := ScaleRC(clientset, internalClientset, ns, name, replicas, false); err != nil { return err } - rc, err := clientset.Core().ReplicationControllers(ns).Get(name) + rc, err := clientset.Core().ReplicationControllers(ns).Get(name, metav1.GetOptions{}) if err != nil { return err } @@ -4739,7 +4741,7 @@ func LaunchWebserverPod(f *Framework, podName, nodeName string) (ip string) { _, err := podClient.Create(pod) ExpectNoError(err) ExpectNoError(f.WaitForPodRunning(podName)) - createdPod, err := podClient.Get(podName) + createdPod, err := podClient.Get(podName, metav1.GetOptions{}) ExpectNoError(err) ip = fmt.Sprintf("%s:%d", createdPod.Status.PodIP, port) Logf("Target pod IP:port is %s", ip) @@ -4799,7 +4801,7 @@ func CoreDump(dir string) { func UpdatePodWithRetries(client clientset.Interface, ns, name string, update func(*v1.Pod)) (*v1.Pod, error) { for i := 0; i < 3; i++ { - pod, err := client.Core().Pods(ns).Get(name) + pod, err := client.Core().Pods(ns).Get(name, metav1.GetOptions{}) if err != nil { return nil, fmt.Errorf("Failed to get pod %q: %v", name, err) } @@ -5061,7 +5063,7 @@ func getMaster(c clientset.Interface) Address { master := Address{} // Populate the internal IP. - eps, err := c.Core().Endpoints(v1.NamespaceDefault).Get("kubernetes") + eps, err := c.Core().Endpoints(v1.NamespaceDefault).Get("kubernetes", metav1.GetOptions{}) if err != nil { Failf("Failed to get kubernetes endpoints: %v", err) } diff --git a/test/e2e/garbage_collector.go b/test/e2e/garbage_collector.go index 7eab78fe67..e1215aac2b 100644 --- a/test/e2e/garbage_collector.go +++ b/test/e2e/garbage_collector.go @@ -182,7 +182,7 @@ var _ = framework.KubeDescribe("Garbage collector", func() { } // wait for rc to create pods if err := wait.Poll(5*time.Second, 30*time.Second, func() (bool, error) { - rc, err := rcClient.Get(rc.Name) + rc, err := rcClient.Get(rc.Name, metav1.GetOptions{}) if err != nil { return false, fmt.Errorf("Failed to get rc: %v", err) } @@ -242,7 +242,7 @@ var _ = framework.KubeDescribe("Garbage collector", func() { } // wait for rc to create some pods if err := wait.Poll(5*time.Second, 30*time.Second, func() (bool, error) { - rc, err := rcClient.Get(rc.Name) + rc, err := rcClient.Get(rc.Name, metav1.GetOptions{}) if err != nil { return false, fmt.Errorf("Failed to get rc: %v", err) } diff --git a/test/e2e/ingress_utils.go b/test/e2e/ingress_utils.go index 807d32c3b8..96e1e11b25 100644 --- a/test/e2e/ingress_utils.go +++ b/test/e2e/ingress_utils.go @@ -44,6 +44,7 @@ import ( "google.golang.org/api/googleapi" apierrs "k8s.io/kubernetes/pkg/api/errors" extensions "k8s.io/kubernetes/pkg/apis/extensions/v1beta1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" clientset "k8s.io/kubernetes/pkg/client/clientset_generated/release_1_5" gcecloud "k8s.io/kubernetes/pkg/cloudprovider/providers/gce" "k8s.io/kubernetes/pkg/labels" @@ -307,7 +308,7 @@ func createSecret(kubeClient clientset.Interface, ing *extensions.Ingress) (host }, } var s *v1.Secret - if s, err = kubeClient.Core().Secrets(ing.Namespace).Get(tls.SecretName); err == nil { + if s, err = kubeClient.Core().Secrets(ing.Namespace).Get(tls.SecretName, metav1.GetOptions{}); err == nil { // TODO: Retry the update. We don't really expect anything to conflict though. framework.Logf("Updating secret %v in ns %v with hosts %v for ingress %v", secret.Name, secret.Namespace, host, ing.Name) s.Data = secret.Data @@ -767,7 +768,7 @@ func (j *testJig) update(update func(ing *extensions.Ingress)) { var err error ns, name := j.ing.Namespace, j.ing.Name for i := 0; i < 3; i++ { - j.ing, err = j.client.Extensions().Ingresses(ns).Get(name) + j.ing, err = j.client.Extensions().Ingresses(ns).Get(name, metav1.GetOptions{}) if err != nil { framework.Failf("failed to get ingress %q: %v", name, err) } @@ -878,7 +879,7 @@ func ingFromManifest(fileName string) *extensions.Ingress { func (cont *GCEIngressController) getL7AddonUID() (string, error) { framework.Logf("Retrieving UID from config map: %v/%v", api.NamespaceSystem, uidConfigMap) - cm, err := cont.c.Core().ConfigMaps(api.NamespaceSystem).Get(uidConfigMap) + cm, err := cont.c.Core().ConfigMaps(api.NamespaceSystem).Get(uidConfigMap, metav1.GetOptions{}) if err != nil { return "", err } @@ -932,7 +933,7 @@ func (cont *NginxIngressController) init() { framework.Logf("initializing nginx ingress controller") framework.RunKubectlOrDie("create", "-f", mkpath("rc.yaml"), fmt.Sprintf("--namespace=%v", cont.ns)) - rc, err := cont.c.Core().ReplicationControllers(cont.ns).Get("nginx-ingress-controller") + rc, err := cont.c.Core().ReplicationControllers(cont.ns).Get("nginx-ingress-controller", metav1.GetOptions{}) framework.ExpectNoError(err) cont.rc = rc diff --git a/test/e2e/job.go b/test/e2e/job.go index 6bde390f70..70807d0071 100644 --- a/test/e2e/job.go +++ b/test/e2e/job.go @@ -24,6 +24,7 @@ import ( "k8s.io/kubernetes/pkg/api/v1" batchinternal "k8s.io/kubernetes/pkg/apis/batch" batch "k8s.io/kubernetes/pkg/apis/batch/v1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" clientset "k8s.io/kubernetes/pkg/client/clientset_generated/release_1_5" "k8s.io/kubernetes/pkg/kubectl" "k8s.io/kubernetes/pkg/labels" @@ -276,7 +277,7 @@ func newTestJob(behavior, name string, rPol v1.RestartPolicy, parallelism, compl } func getJob(c clientset.Interface, ns, name string) (*batch.Job, error) { - return c.Batch().Jobs(ns).Get(name) + return c.Batch().Jobs(ns).Get(name, metav1.GetOptions{}) } func createJob(c clientset.Interface, ns string, job *batch.Job) (*batch.Job, error) { @@ -313,7 +314,7 @@ func waitForAllPodsRunning(c clientset.Interface, ns, jobName string, parallelis // Wait for job to reach completions. func waitForJobFinish(c clientset.Interface, ns, jobName string, completions int32) error { return wait.Poll(framework.Poll, jobTimeout, func() (bool, error) { - curr, err := c.Batch().Jobs(ns).Get(jobName) + curr, err := c.Batch().Jobs(ns).Get(jobName, metav1.GetOptions{}) if err != nil { return false, err } @@ -324,7 +325,7 @@ func waitForJobFinish(c clientset.Interface, ns, jobName string, completions int // Wait for job fail. func waitForJobFail(c clientset.Interface, ns, jobName string, timeout time.Duration) error { return wait.Poll(framework.Poll, timeout, func() (bool, error) { - curr, err := c.Batch().Jobs(ns).Get(jobName) + curr, err := c.Batch().Jobs(ns).Get(jobName, metav1.GetOptions{}) if err != nil { return false, err } diff --git a/test/e2e/kibana_logging.go b/test/e2e/kibana_logging.go index 9bf466e1cb..d88a328010 100644 --- a/test/e2e/kibana_logging.go +++ b/test/e2e/kibana_logging.go @@ -21,6 +21,7 @@ import ( "k8s.io/kubernetes/pkg/api" "k8s.io/kubernetes/pkg/api/v1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" "k8s.io/kubernetes/pkg/labels" "k8s.io/kubernetes/test/e2e/framework" @@ -60,7 +61,7 @@ func ClusterLevelLoggingWithKibana(f *framework.Framework) { // being run as the first e2e test just after the e2e cluster has been created. var err error for start := time.Now(); time.Since(start) < graceTime; time.Sleep(5 * time.Second) { - if _, err = s.Get("kibana-logging"); err == nil { + if _, err = s.Get("kibana-logging", metav1.GetOptions{}); err == nil { break } framework.Logf("Attempt to check for the existence of the Kibana service failed after %v", time.Since(start)) diff --git a/test/e2e/kubectl.go b/test/e2e/kubectl.go index c3cc1d664f..3a91b65ae0 100644 --- a/test/e2e/kubectl.go +++ b/test/e2e/kubectl.go @@ -205,7 +205,7 @@ var _ = framework.KubeDescribe("Kubectl alpha client", func() { framework.RunKubectlOrDie("run", sjName, "--restart=OnFailure", "--generator=scheduledjob/v2alpha1", "--schedule="+schedule, "--image="+busyboxImage, nsFlag) By("verifying the ScheduledJob " + sjName + " was created") - sj, err := c.BatchV2alpha1().CronJobs(ns).Get(sjName) + sj, err := c.BatchV2alpha1().CronJobs(ns).Get(sjName, metav1.GetOptions{}) if err != nil { framework.Failf("Failed getting ScheduledJob %s: %v", sjName, err) } @@ -242,7 +242,7 @@ var _ = framework.KubeDescribe("Kubectl alpha client", func() { framework.RunKubectlOrDie("run", cjName, "--restart=OnFailure", "--generator=cronjob/v2alpha1", "--schedule="+schedule, "--image="+busyboxImage, nsFlag) By("verifying the CronJob " + cjName + " was created") - sj, err := c.BatchV2alpha1().CronJobs(ns).Get(cjName) + sj, err := c.BatchV2alpha1().CronJobs(ns).Get(cjName, metav1.GetOptions{}) if err != nil { framework.Failf("Failed getting CronJob %s: %v", cjName, err) } @@ -756,7 +756,7 @@ var _ = framework.KubeDescribe("Kubectl client", func() { }) validateService := func(name string, servicePort int, timeout time.Duration) { err := wait.Poll(framework.Poll, timeout, func() (bool, error) { - endpoints, err := c.Core().Endpoints(ns).Get(name) + endpoints, err := c.Core().Endpoints(ns).Get(name, metav1.GetOptions{}) if err != nil { // log the real error framework.Logf("Get endpoints failed (interval %v): %v", framework.Poll, err) @@ -787,7 +787,7 @@ var _ = framework.KubeDescribe("Kubectl client", func() { }) Expect(err).NotTo(HaveOccurred()) - service, err := c.Core().Services(ns).Get(name) + service, err := c.Core().Services(ns).Get(name, metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred()) if len(service.Spec.Ports) != 1 { @@ -1016,7 +1016,7 @@ var _ = framework.KubeDescribe("Kubectl client", func() { By("running the image " + nginxImage) framework.RunKubectlOrDie("run", rcName, "--image="+nginxImage, "--generator=run/v1", nsFlag) By("verifying the rc " + rcName + " was created") - rc, err := c.Core().ReplicationControllers(ns).Get(rcName) + rc, err := c.Core().ReplicationControllers(ns).Get(rcName, metav1.GetOptions{}) if err != nil { framework.Failf("Failed getting rc %s: %v", rcName, err) } @@ -1072,7 +1072,7 @@ var _ = framework.KubeDescribe("Kubectl client", func() { By("running the image " + nginxImage) framework.RunKubectlOrDie("run", rcName, "--image="+nginxImage, "--generator=run/v1", nsFlag) By("verifying the rc " + rcName + " was created") - rc, err := c.Core().ReplicationControllers(ns).Get(rcName) + rc, err := c.Core().ReplicationControllers(ns).Get(rcName, metav1.GetOptions{}) if err != nil { framework.Failf("Failed getting rc %s: %v", rcName, err) } @@ -1118,7 +1118,7 @@ var _ = framework.KubeDescribe("Kubectl client", func() { By("running the image " + nginxImage) framework.RunKubectlOrDie("run", dName, "--image="+nginxImage, "--generator=deployment/v1beta1", nsFlag) By("verifying the deployment " + dName + " was created") - d, err := c.Extensions().Deployments(ns).Get(dName) + d, err := c.Extensions().Deployments(ns).Get(dName, metav1.GetOptions{}) if err != nil { framework.Failf("Failed getting deployment %s: %v", dName, err) } @@ -1160,7 +1160,7 @@ var _ = framework.KubeDescribe("Kubectl client", func() { By("running the image " + nginxImage) framework.RunKubectlOrDie("run", jobName, "--restart=OnFailure", "--generator=job/v1", "--image="+nginxImage, nsFlag) By("verifying the job " + jobName + " was created") - job, err := c.Batch().Jobs(ns).Get(jobName) + job, err := c.Batch().Jobs(ns).Get(jobName, metav1.GetOptions{}) if err != nil { framework.Failf("Failed getting job %s: %v", jobName, err) } @@ -1193,7 +1193,7 @@ var _ = framework.KubeDescribe("Kubectl client", func() { By("running the image " + nginxImage) framework.RunKubectlOrDie("run", podName, "--restart=Never", "--generator=run-pod/v1", "--image="+nginxImage, nsFlag) By("verifying the pod " + podName + " was created") - pod, err := c.Core().Pods(ns).Get(podName) + pod, err := c.Core().Pods(ns).Get(podName, metav1.GetOptions{}) if err != nil { framework.Failf("Failed getting pod %s: %v", podName, err) } @@ -1244,7 +1244,7 @@ var _ = framework.KubeDescribe("Kubectl client", func() { framework.RunKubectlOrDieInput(podJson, "replace", "-f", "-", nsFlag) By("verifying the pod " + podName + " has the right image " + busyboxImage) - pod, err := c.Core().Pods(ns).Get(podName) + pod, err := c.Core().Pods(ns).Get(podName, metav1.GetOptions{}) if err != nil { framework.Failf("Failed getting deployment %s: %v", podName, err) } @@ -1276,7 +1276,7 @@ var _ = framework.KubeDescribe("Kubectl client", func() { Expect(runOutput).To(ContainSubstring("stdin closed")) By("verifying the job " + jobName + " was deleted") - _, err := c.Batch().Jobs(ns).Get(jobName) + _, err := c.Batch().Jobs(ns).Get(jobName, metav1.GetOptions{}) Expect(err).To(HaveOccurred()) Expect(apierrs.IsNotFound(err)).To(BeTrue()) }) @@ -1425,7 +1425,7 @@ var _ = framework.KubeDescribe("Kubectl client", func() { framework.RunKubectlOrDie("create", "quota", quotaName, "--hard=pods=1000000,services=1000000", nsFlag) By("verifying that the quota was created") - quota, err := c.Core().ResourceQuotas(ns).Get(quotaName) + quota, err := c.Core().ResourceQuotas(ns).Get(quotaName, metav1.GetOptions{}) if err != nil { framework.Failf("Failed getting quota %s: %v", quotaName, err) } @@ -1455,7 +1455,7 @@ var _ = framework.KubeDescribe("Kubectl client", func() { framework.RunKubectlOrDie("create", "quota", quotaName, "--hard=pods=1000000", "--scopes=BestEffort,NotTerminating", nsFlag) By("verifying that the quota was created") - quota, err := c.Core().ResourceQuotas(ns).Get(quotaName) + quota, err := c.Core().ResourceQuotas(ns).Get(quotaName, metav1.GetOptions{}) if err != nil { framework.Failf("Failed getting quota %s: %v", quotaName, err) } diff --git a/test/e2e/kubelet.go b/test/e2e/kubelet.go index 02b2610414..c388737ff1 100644 --- a/test/e2e/kubelet.go +++ b/test/e2e/kubelet.go @@ -22,6 +22,7 @@ import ( "time" "k8s.io/kubernetes/pkg/api/v1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" clientset "k8s.io/kubernetes/pkg/client/clientset_generated/release_1_5" "k8s.io/kubernetes/pkg/util/sets" "k8s.io/kubernetes/pkg/util/uuid" @@ -101,7 +102,7 @@ func updateNodeLabels(c clientset.Interface, nodeNames sets.String, toAdd, toRem var node *v1.Node var err error for i := 0; i < maxRetries; i++ { - node, err = c.Core().Nodes().Get(nodeName) + node, err = c.Core().Nodes().Get(nodeName, metav1.GetOptions{}) if err != nil { framework.Logf("Error getting node %s: %v", nodeName, err) continue diff --git a/test/e2e/limit_range.go b/test/e2e/limit_range.go index 7a52af3bc9..0a2faa478d 100644 --- a/test/e2e/limit_range.go +++ b/test/e2e/limit_range.go @@ -21,6 +21,7 @@ import ( "k8s.io/kubernetes/pkg/api/resource" "k8s.io/kubernetes/pkg/api/v1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" "k8s.io/kubernetes/test/e2e/framework" . "github.com/onsi/ginkgo" @@ -46,7 +47,7 @@ var _ = framework.KubeDescribe("LimitRange", func() { Expect(err).NotTo(HaveOccurred()) By("Fetching the LimitRange to ensure it has proper values") - limitRange, err = f.ClientSet.Core().LimitRanges(f.Namespace.Name).Get(limitRange.Name) + limitRange, err = f.ClientSet.Core().LimitRanges(f.Namespace.Name).Get(limitRange.Name, metav1.GetOptions{}) expected := v1.ResourceRequirements{Requests: defaultRequest, Limits: defaultLimit} actual := v1.ResourceRequirements{Requests: limitRange.Spec.Limits[0].DefaultRequest, Limits: limitRange.Spec.Limits[0].Default} err = equalResourceRequirement(expected, actual) @@ -58,7 +59,7 @@ var _ = framework.KubeDescribe("LimitRange", func() { Expect(err).NotTo(HaveOccurred()) By("Ensuring Pod has resource requirements applied from LimitRange") - pod, err = f.ClientSet.Core().Pods(f.Namespace.Name).Get(pod.Name) + pod, err = f.ClientSet.Core().Pods(f.Namespace.Name).Get(pod.Name, metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred()) for i := range pod.Spec.Containers { err = equalResourceRequirement(expected, pod.Spec.Containers[i].Resources) @@ -75,7 +76,7 @@ var _ = framework.KubeDescribe("LimitRange", func() { Expect(err).NotTo(HaveOccurred()) By("Ensuring Pod has merged resource requirements applied from LimitRange") - pod, err = f.ClientSet.Core().Pods(f.Namespace.Name).Get(pod.Name) + pod, err = f.ClientSet.Core().Pods(f.Namespace.Name).Get(pod.Name, metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred()) // This is an interesting case, so it's worth a comment // If you specify a Limit, and no Request, the Limit will default to the Request diff --git a/test/e2e/mesos.go b/test/e2e/mesos.go index 4fcef61fb6..0e3b37cf23 100644 --- a/test/e2e/mesos.go +++ b/test/e2e/mesos.go @@ -101,7 +101,7 @@ var _ = framework.KubeDescribe("Mesos", func() { framework.ExpectNoError(err) framework.ExpectNoError(framework.WaitForPodNameRunningInNamespace(c, podName, ns)) - pod, err := c.Core().Pods(ns).Get(podName) + pod, err := c.Core().Pods(ns).Get(podName, metav1.GetOptions{}) framework.ExpectNoError(err) nodeClient := f.ClientSet.Core().Nodes() diff --git a/test/e2e/namespace.go b/test/e2e/namespace.go index 97c62323f6..6afbfe508b 100644 --- a/test/e2e/namespace.go +++ b/test/e2e/namespace.go @@ -24,6 +24,7 @@ import ( "k8s.io/kubernetes/pkg/api/errors" "k8s.io/kubernetes/pkg/api/v1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" "k8s.io/kubernetes/pkg/util/intstr" "k8s.io/kubernetes/pkg/util/wait" "k8s.io/kubernetes/test/e2e/framework" @@ -116,7 +117,7 @@ func ensurePodsAreRemovedWhenNamespaceIsDeleted(f *framework.Framework) { maxWaitSeconds := int64(60) + *pod.Spec.TerminationGracePeriodSeconds framework.ExpectNoError(wait.Poll(1*time.Second, time.Duration(maxWaitSeconds)*time.Second, func() (bool, error) { - _, err = f.ClientSet.Core().Namespaces().Get(namespace.Name) + _, err = f.ClientSet.Core().Namespaces().Get(namespace.Name, metav1.GetOptions{}) if err != nil && errors.IsNotFound(err) { return true, nil } @@ -124,7 +125,7 @@ func ensurePodsAreRemovedWhenNamespaceIsDeleted(f *framework.Framework) { })) By("Verifying there is no pod in the namespace") - _, err = f.ClientSet.Core().Pods(namespace.Name).Get(pod.Name) + _, err = f.ClientSet.Core().Pods(namespace.Name).Get(pod.Name, metav1.GetOptions{}) Expect(err).To(HaveOccurred()) } @@ -168,7 +169,7 @@ func ensureServicesAreRemovedWhenNamespaceIsDeleted(f *framework.Framework) { maxWaitSeconds := int64(60) framework.ExpectNoError(wait.Poll(1*time.Second, time.Duration(maxWaitSeconds)*time.Second, func() (bool, error) { - _, err = f.ClientSet.Core().Namespaces().Get(namespace.Name) + _, err = f.ClientSet.Core().Namespaces().Get(namespace.Name, metav1.GetOptions{}) if err != nil && errors.IsNotFound(err) { return true, nil } @@ -176,7 +177,7 @@ func ensureServicesAreRemovedWhenNamespaceIsDeleted(f *framework.Framework) { })) By("Verifying there is no service in the namespace") - _, err = f.ClientSet.Core().Services(namespace.Name).Get(service.Name) + _, err = f.ClientSet.Core().Services(namespace.Name).Get(service.Name, metav1.GetOptions{}) Expect(err).To(HaveOccurred()) } diff --git a/test/e2e/network_partition.go b/test/e2e/network_partition.go index c54a3fbf67..96099ef968 100644 --- a/test/e2e/network_partition.go +++ b/test/e2e/network_partition.go @@ -23,6 +23,7 @@ import ( "k8s.io/kubernetes/pkg/api" "k8s.io/kubernetes/pkg/api/v1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" "k8s.io/kubernetes/pkg/client/cache" clientset "k8s.io/kubernetes/pkg/client/clientset_generated/release_1_5" "k8s.io/kubernetes/pkg/fields" @@ -268,7 +269,7 @@ var _ = framework.KubeDescribe("Network Partition [Disruptive] [Slow]", func() { Expect(err).NotTo(HaveOccurred()) nodeName := pods.Items[0].Spec.NodeName - node, err := c.Core().Nodes().Get(nodeName) + node, err := c.Core().Nodes().Get(nodeName, metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred()) // This creates a temporary network partition, verifies that 'podNameToDisappear', @@ -306,7 +307,7 @@ var _ = framework.KubeDescribe("Network Partition [Disruptive] [Slow]", func() { // verify that it is really on the requested node { - pod, err := c.Core().Pods(ns).Get(additionalPod) + pod, err := c.Core().Pods(ns).Get(additionalPod, metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred()) if pod.Spec.NodeName != node.Name { framework.Logf("Pod %s found on invalid node: %s instead of %s", pod.Name, pod.Spec.NodeName, node.Name) @@ -333,7 +334,7 @@ var _ = framework.KubeDescribe("Network Partition [Disruptive] [Slow]", func() { Expect(err).NotTo(HaveOccurred()) nodeName := pods.Items[0].Spec.NodeName - node, err := c.Core().Nodes().Get(nodeName) + node, err := c.Core().Nodes().Get(nodeName, metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred()) // This creates a temporary network partition, verifies that 'podNameToDisappear', @@ -411,7 +412,7 @@ var _ = framework.KubeDescribe("Network Partition [Disruptive] [Slow]", func() { pst.waitForRunningAndReady(*ps.Spec.Replicas, ps) pod := pst.getPodList(ps).Items[0] - node, err := c.Core().Nodes().Get(pod.Spec.NodeName) + node, err := c.Core().Nodes().Get(pod.Spec.NodeName, metav1.GetOptions{}) framework.ExpectNoError(err) // Blocks outgoing network traffic on 'node'. Then verifies that 'podNameToDisappear', @@ -453,7 +454,7 @@ var _ = framework.KubeDescribe("Network Partition [Disruptive] [Slow]", func() { Expect(err).NotTo(HaveOccurred()) nodeName := pods.Items[0].Spec.NodeName - node, err := c.Core().Nodes().Get(nodeName) + node, err := c.Core().Nodes().Get(nodeName, metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred()) // This creates a temporary network partition, verifies that the job has 'parallelism' number of diff --git a/test/e2e/node_problem_detector.go b/test/e2e/node_problem_detector.go index d59227ed8c..28cdc7d8bd 100644 --- a/test/e2e/node_problem_detector.go +++ b/test/e2e/node_problem_detector.go @@ -24,6 +24,7 @@ import ( "k8s.io/kubernetes/pkg/api" "k8s.io/kubernetes/pkg/api/v1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" clientset "k8s.io/kubernetes/pkg/client/clientset_generated/release_1_5" coreclientset "k8s.io/kubernetes/pkg/client/clientset_generated/release_1_5/typed/core/v1" "k8s.io/kubernetes/pkg/fields" @@ -411,7 +412,7 @@ func verifyNoEvents(e coreclientset.EventInterface, options v1.ListOptions) erro // verifyCondition verifies specific node condition is generated, if reason and message are empty, they will not be checked func verifyCondition(n coreclientset.NodeInterface, nodeName string, condition v1.NodeConditionType, status v1.ConditionStatus, reason, message string) error { - node, err := n.Get(nodeName) + node, err := n.Get(nodeName, metav1.GetOptions{}) if err != nil { return err } diff --git a/test/e2e/nodeoutofdisk.go b/test/e2e/nodeoutofdisk.go index eafbeb22b8..e9763a6f03 100644 --- a/test/e2e/nodeoutofdisk.go +++ b/test/e2e/nodeoutofdisk.go @@ -24,6 +24,7 @@ import ( cadvisorapi "github.com/google/cadvisor/info/v1" "k8s.io/kubernetes/pkg/api/resource" "k8s.io/kubernetes/pkg/api/v1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" clientset "k8s.io/kubernetes/pkg/client/clientset_generated/release_1_5" "k8s.io/kubernetes/pkg/fields" "k8s.io/kubernetes/pkg/util/wait" @@ -98,7 +99,7 @@ var _ = framework.KubeDescribe("NodeOutOfDisk [Serial] [Flaky] [Disruptive]", fu }) It("runs out of disk space", func() { - unfilledNode, err := c.Core().Nodes().Get(unfilledNodeName) + unfilledNode, err := c.Core().Nodes().Get(unfilledNodeName, metav1.GetOptions{}) framework.ExpectNoError(err) By(fmt.Sprintf("Calculating CPU availability on node %s", unfilledNode.Name)) @@ -122,7 +123,7 @@ var _ = framework.KubeDescribe("NodeOutOfDisk [Serial] [Flaky] [Disruptive]", fu createOutOfDiskPod(c, ns, name, podCPU) framework.ExpectNoError(f.WaitForPodRunning(name)) - pod, err := podClient.Get(name) + pod, err := podClient.Get(name, metav1.GetOptions{}) framework.ExpectNoError(err) Expect(pod.Spec.NodeName).To(Equal(unfilledNodeName)) } @@ -161,7 +162,7 @@ var _ = framework.KubeDescribe("NodeOutOfDisk [Serial] [Flaky] [Disruptive]", fu By(fmt.Sprintf("Verifying that pod %s schedules on node %s", pendingPodName, recoveredNodeName)) framework.ExpectNoError(f.WaitForPodRunning(pendingPodName)) - pendingPod, err := podClient.Get(pendingPodName) + pendingPod, err := podClient.Get(pendingPodName, metav1.GetOptions{}) framework.ExpectNoError(err) Expect(pendingPod.Spec.NodeName).To(Equal(recoveredNodeName)) }) diff --git a/test/e2e/pd.go b/test/e2e/pd.go index 8c93ce7830..f808001405 100644 --- a/test/e2e/pd.go +++ b/test/e2e/pd.go @@ -724,7 +724,7 @@ func waitForPDInVolumesInUse( "Waiting for node %s's VolumesInUse Status %s PD %q", nodeName, logStr, diskName) for start := time.Now(); time.Since(start) < timeout; time.Sleep(nodeStatusPollTime) { - nodeObj, err := nodeClient.Get(string(nodeName)) + nodeObj, err := nodeClient.Get(string(nodeName), metav1.GetOptions{}) if err != nil || nodeObj == nil { framework.Logf( "Failed to fetch node object %q from API server. err=%v", diff --git a/test/e2e/persistent_volumes.go b/test/e2e/persistent_volumes.go index 18caf9ac38..9596a100c2 100644 --- a/test/e2e/persistent_volumes.go +++ b/test/e2e/persistent_volumes.go @@ -69,7 +69,7 @@ func pvPvcCleanup(c clientset.Interface, ns string, pvols pvmap, claims pvcmap) if c != nil && len(ns) > 0 { for pvcKey := range claims { - _, err := c.Core().PersistentVolumeClaims(pvcKey.Namespace).Get(pvcKey.Name) + _, err := c.Core().PersistentVolumeClaims(pvcKey.Namespace).Get(pvcKey.Name, metav1.GetOptions{}) if !apierrs.IsNotFound(err) { Expect(err).NotTo(HaveOccurred()) framework.Logf(" deleting PVC %v ...", pvcKey) @@ -81,7 +81,7 @@ func pvPvcCleanup(c clientset.Interface, ns string, pvols pvmap, claims pvcmap) } for name := range pvols { - _, err := c.Core().PersistentVolumes().Get(name) + _, err := c.Core().PersistentVolumes().Get(name, metav1.GetOptions{}) if !apierrs.IsNotFound(err) { Expect(err).NotTo(HaveOccurred()) framework.Logf(" deleting PV %v ...", name) @@ -106,7 +106,7 @@ func deletePVCandValidatePV(c clientset.Interface, ns string, pvc *v1.Persistent Expect(err).NotTo(HaveOccurred()) // Check that the PVC is really deleted. - pvc, err = c.Core().PersistentVolumeClaims(ns).Get(pvc.Name) + pvc, err = c.Core().PersistentVolumeClaims(ns).Get(pvc.Name, metav1.GetOptions{}) Expect(apierrs.IsNotFound(err)).To(BeTrue()) // Wait for the PV's phase to return to the expected Phase @@ -115,7 +115,7 @@ func deletePVCandValidatePV(c clientset.Interface, ns string, pvc *v1.Persistent Expect(err).NotTo(HaveOccurred()) // examine the pv's ClaimRef and UID and compare to expected values - pv, err = c.Core().PersistentVolumes().Get(pv.Name) + pv, err = c.Core().PersistentVolumes().Get(pv.Name, metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred()) cr := pv.Spec.ClaimRef if expctPVPhase == v1.VolumeAvailable { @@ -140,7 +140,7 @@ func deletePVCandValidatePVGroup(c clientset.Interface, ns string, pvols pvmap, var expctPVPhase v1.PersistentVolumePhase for pvName := range pvols { - pv, err := c.Core().PersistentVolumes().Get(pvName) + pv, err := c.Core().PersistentVolumes().Get(pvName, metav1.GetOptions{}) Expect(apierrs.IsNotFound(err)).To(BeFalse()) cr := pv.Spec.ClaimRef // if pv is bound then delete the pvc it is bound to @@ -151,7 +151,7 @@ func deletePVCandValidatePVGroup(c clientset.Interface, ns string, pvols pvmap, pvcKey := makePvcKey(ns, cr.Name) _, found := claims[pvcKey] Expect(found).To(BeTrue()) - pvc, err := c.Core().PersistentVolumeClaims(ns).Get(cr.Name) + pvc, err := c.Core().PersistentVolumeClaims(ns).Get(cr.Name, metav1.GetOptions{}) Expect(apierrs.IsNotFound(err)).To(BeFalse()) // what Phase do we expect the PV that was bound to the claim to @@ -305,11 +305,11 @@ func waitOnPVandPVC(c clientset.Interface, ns string, pv *v1.PersistentVolume, p Expect(err).NotTo(HaveOccurred()) // Re-get the pv and pvc objects - pv, err = c.Core().PersistentVolumes().Get(pv.Name) + pv, err = c.Core().PersistentVolumes().Get(pv.Name, metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred()) // Re-get the pvc and - pvc, err = c.Core().PersistentVolumeClaims(ns).Get(pvc.Name) + pvc, err = c.Core().PersistentVolumeClaims(ns).Get(pvc.Name, metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred()) // The pv and pvc are both bound, but to each other? @@ -343,7 +343,7 @@ func waitAndVerifyBinds(c clientset.Interface, ns string, pvols pvmap, claims pv } Expect(err).NotTo(HaveOccurred()) - pv, err := c.Core().PersistentVolumes().Get(pvName) + pv, err := c.Core().PersistentVolumes().Get(pvName, metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred()) if cr := pv.Spec.ClaimRef; cr != nil && len(cr.Name) > 0 { // Assert bound pvc is a test resource. Failing assertion could @@ -435,7 +435,7 @@ func completeMultiTest(f *framework.Framework, c clientset.Interface, ns string, // 1. verify each PV permits write access to a client pod By("Checking pod has write access to PersistentVolumes") for pvcKey := range claims { - pvc, err := c.Core().PersistentVolumeClaims(pvcKey.Namespace).Get(pvcKey.Name) + pvc, err := c.Core().PersistentVolumeClaims(pvcKey.Namespace).Get(pvcKey.Name, metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred()) if len(pvc.Spec.VolumeName) == 0 { continue // claim is not bound @@ -503,7 +503,7 @@ var _ = framework.KubeDescribe("PersistentVolumes", func() { AfterEach(func() { if c != nil && len(ns) > 0 { if pvc != nil && len(pvc.Name) > 0 { - _, err := c.Core().PersistentVolumeClaims(ns).Get(pvc.Name) + _, err := c.Core().PersistentVolumeClaims(ns).Get(pvc.Name, metav1.GetOptions{}) if !apierrs.IsNotFound(err) { Expect(err).NotTo(HaveOccurred()) framework.Logf("AfterEach: deleting PVC %v", pvc.Name) @@ -515,7 +515,7 @@ var _ = framework.KubeDescribe("PersistentVolumes", func() { pvc = nil if pv != nil && len(pv.Name) > 0 { - _, err := c.Core().PersistentVolumes().Get(pv.Name) + _, err := c.Core().PersistentVolumes().Get(pv.Name, metav1.GetOptions{}) if !apierrs.IsNotFound(err) { Expect(err).NotTo(HaveOccurred()) framework.Logf("AfterEach: deleting PV %v", pv.Name) diff --git a/test/e2e/petset.go b/test/e2e/petset.go index 99f5b1019e..e30cf4383b 100644 --- a/test/e2e/petset.go +++ b/test/e2e/petset.go @@ -240,7 +240,7 @@ var _ = framework.KubeDescribe("StatefulSet", func() { By("Verifying that the 2nd pod wont be removed if it is not running and ready") pst.confirmPetCount(2, ps, 10*time.Second) expectedPodName := ps.Name + "-1" - expectedPod, err := f.ClientSet.Core().Pods(ns).Get(expectedPodName) + expectedPod, err := f.ClientSet.Core().Pods(ns).Get(expectedPodName, metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred()) watcher, err := f.ClientSet.Core().Pods(ns).Watch(v1.SingleObject( v1.ObjectMeta{ @@ -488,7 +488,7 @@ var _ = framework.KubeDescribe("Stateful Set recreate", func() { By("waiting when pet pod " + petPodName + " will be recreated in namespace " + f.Namespace.Name + " and will be in running state") // we may catch delete event, thats why we are waiting for running phase like this, and not with watch.Until Eventually(func() error { - petPod, err := f.ClientSet.Core().Pods(f.Namespace.Name).Get(petPodName) + petPod, err := f.ClientSet.Core().Pods(f.Namespace.Name).Get(petPodName, metav1.GetOptions{}) if err != nil { return err } @@ -812,7 +812,7 @@ type verifyPodFunc func(*v1.Pod) func (p *statefulSetTester) verifyPodAtIndex(index int, ps *apps.StatefulSet, verify verifyPodFunc) { name := getPodNameAtIndex(index, ps) - pod, err := p.c.Core().Pods(ps.Namespace).Get(name) + pod, err := p.c.Core().Pods(ps.Namespace).Get(name, metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred(), fmt.Sprintf("Failed to get stateful pod %s for StatefulSet %s/%s", name, ps.Namespace, ps.Name)) verify(pod) } @@ -861,7 +861,7 @@ func (p *statefulSetTester) restart(ps *apps.StatefulSet) { func (p *statefulSetTester) update(ns, name string, update func(ps *apps.StatefulSet)) { for i := 0; i < 3; i++ { - ps, err := p.c.Apps().StatefulSets(ns).Get(name) + ps, err := p.c.Apps().StatefulSets(ns).Get(name, metav1.GetOptions{}) if err != nil { framework.Failf("failed to get statefulset %q: %v", name, err) } @@ -979,7 +979,7 @@ func (p *statefulSetTester) waitForStatus(ps *apps.StatefulSet, expectedReplicas ns, name := ps.Namespace, ps.Name pollErr := wait.PollImmediate(statefulsetPoll, statefulsetTimeout, func() (bool, error) { - psGet, err := p.c.Apps().StatefulSets(ns).Get(name) + psGet, err := p.c.Apps().StatefulSets(ns).Get(name, metav1.GetOptions{}) if err != nil { return false, err } diff --git a/test/e2e/pods.go b/test/e2e/pods.go index ee4779f8b0..d6ad33cbf6 100644 --- a/test/e2e/pods.go +++ b/test/e2e/pods.go @@ -25,6 +25,7 @@ import ( "time" "k8s.io/kubernetes/pkg/api/v1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" "k8s.io/kubernetes/pkg/labels" "k8s.io/kubernetes/pkg/util/uuid" "k8s.io/kubernetes/pkg/util/wait" @@ -100,7 +101,7 @@ var _ = framework.KubeDescribe("Pods Delete Grace Period", func() { // may be carried out immediately rather than gracefully. framework.ExpectNoError(f.WaitForPodRunning(pod.Name)) // save the running pod - pod, err = podClient.Get(pod.Name) + pod, err = podClient.Get(pod.Name, metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred(), "failed to GET scheduled pod") // start local proxy, so we can send graceful deletion over query string, rather than body parameter diff --git a/test/e2e/pre_stop.go b/test/e2e/pre_stop.go index 6c2eab0a30..dac50795fe 100644 --- a/test/e2e/pre_stop.go +++ b/test/e2e/pre_stop.go @@ -22,6 +22,7 @@ import ( "time" "k8s.io/kubernetes/pkg/api/v1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" clientset "k8s.io/kubernetes/pkg/client/clientset_generated/release_1_5" "k8s.io/kubernetes/pkg/util/wait" "k8s.io/kubernetes/test/e2e/framework" @@ -66,7 +67,7 @@ func testPreStop(c clientset.Interface, ns string) { val := "{\"Source\": \"prestop\"}" - podOut, err := c.Core().Pods(ns).Get(podDescr.Name) + podOut, err := c.Core().Pods(ns).Get(podDescr.Name, metav1.GetOptions{}) framework.ExpectNoError(err, "getting pod info") preStopDescr := &v1.Pod{ diff --git a/test/e2e/rc.go b/test/e2e/rc.go index 6f2b8bc560..055bdff87f 100644 --- a/test/e2e/rc.go +++ b/test/e2e/rc.go @@ -22,6 +22,7 @@ import ( "k8s.io/kubernetes/pkg/api/resource" "k8s.io/kubernetes/pkg/api/v1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" "k8s.io/kubernetes/pkg/controller/replication" "k8s.io/kubernetes/pkg/labels" "k8s.io/kubernetes/pkg/util/uuid" @@ -165,7 +166,7 @@ func rcConditionCheck(f *framework.Framework) { Expect(err).NotTo(HaveOccurred()) err = wait.PollImmediate(1*time.Second, 1*time.Minute, func() (bool, error) { - quota, err = c.Core().ResourceQuotas(namespace).Get(name) + quota, err = c.Core().ResourceQuotas(namespace).Get(name, metav1.GetOptions{}) if err != nil { return false, err } @@ -187,7 +188,7 @@ func rcConditionCheck(f *framework.Framework) { generation := rc.Generation conditions := rc.Status.Conditions err = wait.PollImmediate(1*time.Second, 1*time.Minute, func() (bool, error) { - rc, err = c.Core().ReplicationControllers(namespace).Get(name) + rc, err = c.Core().ReplicationControllers(namespace).Get(name, metav1.GetOptions{}) if err != nil { return false, err } @@ -216,7 +217,7 @@ func rcConditionCheck(f *framework.Framework) { generation = rc.Generation conditions = rc.Status.Conditions err = wait.PollImmediate(1*time.Second, 1*time.Minute, func() (bool, error) { - rc, err = c.Core().ReplicationControllers(namespace).Get(name) + rc, err = c.Core().ReplicationControllers(namespace).Get(name, metav1.GetOptions{}) if err != nil { return false, err } diff --git a/test/e2e/reboot.go b/test/e2e/reboot.go index 3868a194f4..5dbff88ce1 100644 --- a/test/e2e/reboot.go +++ b/test/e2e/reboot.go @@ -24,6 +24,7 @@ import ( "k8s.io/kubernetes/pkg/api" "k8s.io/kubernetes/pkg/api/v1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" clientset "k8s.io/kubernetes/pkg/client/clientset_generated/release_1_5" "k8s.io/kubernetes/pkg/fields" "k8s.io/kubernetes/pkg/labels" @@ -223,7 +224,7 @@ func rebootNode(c clientset.Interface, provider, name, rebootCmd string) bool { // Get the node initially. framework.Logf("Getting %s", name) - node, err := c.Core().Nodes().Get(name) + node, err := c.Core().Nodes().Get(name, metav1.GetOptions{}) if err != nil { framework.Logf("Couldn't get node %s", name) return false diff --git a/test/e2e/replica_set.go b/test/e2e/replica_set.go index 34fb536cb5..b36eb60bff 100644 --- a/test/e2e/replica_set.go +++ b/test/e2e/replica_set.go @@ -179,7 +179,7 @@ func rsConditionCheck(f *framework.Framework) { Expect(err).NotTo(HaveOccurred()) err = wait.PollImmediate(1*time.Second, 1*time.Minute, func() (bool, error) { - quota, err = c.Core().ResourceQuotas(namespace).Get(name) + quota, err = c.Core().ResourceQuotas(namespace).Get(name, metav1.GetOptions{}) if err != nil { return false, err } @@ -201,7 +201,7 @@ func rsConditionCheck(f *framework.Framework) { generation := rs.Generation conditions := rs.Status.Conditions err = wait.PollImmediate(1*time.Second, 1*time.Minute, func() (bool, error) { - rs, err = c.Extensions().ReplicaSets(namespace).Get(name) + rs, err = c.Extensions().ReplicaSets(namespace).Get(name, metav1.GetOptions{}) if err != nil { return false, err } @@ -231,7 +231,7 @@ func rsConditionCheck(f *framework.Framework) { generation = rs.Generation conditions = rs.Status.Conditions err = wait.PollImmediate(1*time.Second, 1*time.Minute, func() (bool, error) { - rs, err = c.Extensions().ReplicaSets(namespace).Get(name) + rs, err = c.Extensions().ReplicaSets(namespace).Get(name, metav1.GetOptions{}) if err != nil { return false, err } diff --git a/test/e2e/resize_nodes.go b/test/e2e/resize_nodes.go index f5b32d0a8f..a0f982c293 100644 --- a/test/e2e/resize_nodes.go +++ b/test/e2e/resize_nodes.go @@ -214,7 +214,7 @@ func newRCByName(c clientset.Interface, ns, name string, replicas int32, gracePe } func resizeRC(c clientset.Interface, ns, name string, replicas int32) error { - rc, err := c.Core().ReplicationControllers(ns).Get(name) + rc, err := c.Core().ReplicationControllers(ns).Get(name, metav1.GetOptions{}) if err != nil { return err } diff --git a/test/e2e/resource_quota.go b/test/e2e/resource_quota.go index 2996e55b5f..9b50f4f684 100644 --- a/test/e2e/resource_quota.go +++ b/test/e2e/resource_quota.go @@ -22,6 +22,7 @@ import ( "k8s.io/kubernetes/pkg/api/resource" "k8s.io/kubernetes/pkg/api/v1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" clientset "k8s.io/kubernetes/pkg/client/clientset_generated/release_1_5" "k8s.io/kubernetes/pkg/util/intstr" "k8s.io/kubernetes/pkg/util/wait" @@ -645,7 +646,7 @@ func deleteResourceQuota(c clientset.Interface, namespace, name string) error { // wait for resource quota status to show the expected used resources value func waitForResourceQuota(c clientset.Interface, ns, quotaName string, used v1.ResourceList) error { return wait.Poll(framework.Poll, resourceQuotaTimeout, func() (bool, error) { - resourceQuota, err := c.Core().ResourceQuotas(ns).Get(quotaName) + resourceQuota, err := c.Core().ResourceQuotas(ns).Get(quotaName, metav1.GetOptions{}) if err != nil { return false, err } diff --git a/test/e2e/restart.go b/test/e2e/restart.go index 860ef7e833..946ee68b8d 100644 --- a/test/e2e/restart.go +++ b/test/e2e/restart.go @@ -22,6 +22,7 @@ import ( "k8s.io/kubernetes/pkg/api" "k8s.io/kubernetes/pkg/api/v1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" "k8s.io/kubernetes/pkg/fields" kubepod "k8s.io/kubernetes/pkg/kubelet/pod" "k8s.io/kubernetes/pkg/labels" @@ -157,7 +158,7 @@ func restartNodes(f *framework.Framework, nodeNames []string) error { // List old boot IDs. oldBootIDs := make(map[string]string) for _, name := range nodeNames { - node, err := f.ClientSet.Core().Nodes().Get(name) + node, err := f.ClientSet.Core().Nodes().Get(name, metav1.GetOptions{}) if err != nil { return fmt.Errorf("error getting node info before reboot: %s", err) } @@ -179,7 +180,7 @@ func restartNodes(f *framework.Framework, nodeNames []string) error { // Wait for their boot IDs to change. for _, name := range nodeNames { if err := wait.Poll(30*time.Second, 5*time.Minute, func() (bool, error) { - node, err := f.ClientSet.Core().Nodes().Get(name) + node, err := f.ClientSet.Core().Nodes().Get(name, metav1.GetOptions{}) if err != nil { return false, fmt.Errorf("error getting node info after reboot: %s", err) } diff --git a/test/e2e/scheduler_predicates.go b/test/e2e/scheduler_predicates.go index 499ce469b6..d260651d7a 100644 --- a/test/e2e/scheduler_predicates.go +++ b/test/e2e/scheduler_predicates.go @@ -24,6 +24,7 @@ import ( "k8s.io/kubernetes/pkg/api/errors" "k8s.io/kubernetes/pkg/api/resource" "k8s.io/kubernetes/pkg/api/v1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" clientset "k8s.io/kubernetes/pkg/client/clientset_generated/release_1_5" "k8s.io/kubernetes/pkg/util/sets" "k8s.io/kubernetes/pkg/util/uuid" @@ -59,7 +60,7 @@ var _ = framework.KubeDescribe("SchedulerPredicates [Serial]", func() { ignoreLabels := framework.ImagePullerLabels AfterEach(func() { - rc, err := cs.Core().ReplicationControllers(ns).Get(RCName) + rc, err := cs.Core().ReplicationControllers(ns).Get(RCName, metav1.GetOptions{}) if err == nil && *(rc.Spec.Replicas) != 0 { By("Cleaning up the replication controller") err := framework.DeleteRCAndPods(f.ClientSet, f.InternalClientset, ns, RCName) @@ -284,7 +285,7 @@ var _ = framework.KubeDescribe("SchedulerPredicates [Serial]", func() { // already when the kubelet does not know about its new label yet. The // kubelet will then refuse to launch the pod. framework.ExpectNoError(framework.WaitForPodNotPending(cs, ns, labelPodName, pod.ResourceVersion)) - labelPod, err := cs.Core().Pods(ns).Get(labelPodName) + labelPod, err := cs.Core().Pods(ns).Get(labelPodName, metav1.GetOptions{}) framework.ExpectNoError(err) Expect(labelPod.Spec.NodeName).To(Equal(nodeName)) }) @@ -368,7 +369,7 @@ var _ = framework.KubeDescribe("SchedulerPredicates [Serial]", func() { // already when the kubelet does not know about its new label yet. The // kubelet will then refuse to launch the pod. framework.ExpectNoError(framework.WaitForPodNotPending(cs, ns, labelPodName, pod.ResourceVersion)) - labelPod, err := cs.Core().Pods(ns).Get(labelPodName) + labelPod, err := cs.Core().Pods(ns).Get(labelPodName, metav1.GetOptions{}) framework.ExpectNoError(err) Expect(labelPod.Spec.NodeName).To(Equal(nodeName)) }) @@ -393,7 +394,7 @@ var _ = framework.KubeDescribe("SchedulerPredicates [Serial]", func() { // already when the kubelet does not know about its new label yet. The // kubelet will then refuse to launch the pod. framework.ExpectNoError(framework.WaitForPodNotPending(cs, ns, pod.Name, "")) - labelPod, err := cs.Core().Pods(ns).Get(pod.Name) + labelPod, err := cs.Core().Pods(ns).Get(pod.Name, metav1.GetOptions{}) framework.ExpectNoError(err) Expect(labelPod.Spec.NodeName).To(Equal(nodeName)) }) @@ -499,7 +500,7 @@ var _ = framework.KubeDescribe("SchedulerPredicates [Serial]", func() { // already when the kubelet does not know about its new label yet. The // kubelet will then refuse to launch the pod. framework.ExpectNoError(framework.WaitForPodNotPending(cs, ns, labelPodName, pod.ResourceVersion)) - labelPod, err := cs.Core().Pods(ns).Get(labelPodName) + labelPod, err := cs.Core().Pods(ns).Get(labelPodName, metav1.GetOptions{}) framework.ExpectNoError(err) Expect(labelPod.Spec.NodeName).To(Equal(nodeName)) }) @@ -610,7 +611,7 @@ var _ = framework.KubeDescribe("SchedulerPredicates [Serial]", func() { // already when the kubelet does not know about its new label yet. The // kubelet will then refuse to launch the pod. framework.ExpectNoError(framework.WaitForPodNotPending(cs, ns, labelPodName, pod.ResourceVersion)) - labelPod, err := cs.Core().Pods(ns).Get(labelPodName) + labelPod, err := cs.Core().Pods(ns).Get(labelPodName, metav1.GetOptions{}) framework.ExpectNoError(err) Expect(labelPod.Spec.NodeName).To(Equal(nodeName)) }) @@ -635,7 +636,7 @@ var _ = framework.KubeDescribe("SchedulerPredicates [Serial]", func() { // already when the kubelet does not know about its new label yet. The // kubelet will then refuse to launch the pod. framework.ExpectNoError(framework.WaitForPodNotPending(cs, ns, pod.Name, pod.ResourceVersion)) - labelPod, err := cs.Core().Pods(ns).Get(pod.Name) + labelPod, err := cs.Core().Pods(ns).Get(pod.Name, metav1.GetOptions{}) framework.ExpectNoError(err) Expect(labelPod.Spec.NodeName).To(Equal(nodeName)) }) @@ -659,7 +660,7 @@ var _ = framework.KubeDescribe("SchedulerPredicates [Serial]", func() { // already when the kubelet does not know about its new label yet. The // kubelet will then refuse to launch the pod. framework.ExpectNoError(framework.WaitForPodNotPending(cs, ns, pod.Name, pod.ResourceVersion)) - labelPod, err := cs.Core().Pods(ns).Get(pod.Name) + labelPod, err := cs.Core().Pods(ns).Get(pod.Name, metav1.GetOptions{}) framework.ExpectNoError(err) Expect(labelPod.Spec.NodeName).To(Equal(nodeName)) }) @@ -711,7 +712,7 @@ var _ = framework.KubeDescribe("SchedulerPredicates [Serial]", func() { // already when the kubelet does not know about its new taint yet. The // kubelet will then refuse to launch the pod. framework.ExpectNoError(framework.WaitForPodNotPending(cs, ns, tolerationPodName, pod.ResourceVersion)) - deployedPod, err := cs.Core().Pods(ns).Get(tolerationPodName) + deployedPod, err := cs.Core().Pods(ns).Get(tolerationPodName, metav1.GetOptions{}) framework.ExpectNoError(err) Expect(deployedPod.Spec.NodeName).To(Equal(nodeName)) }) @@ -799,7 +800,7 @@ func createPausePod(f *framework.Framework, conf pausePodConfig) *v1.Pod { func runPausePod(f *framework.Framework, conf pausePodConfig) *v1.Pod { pod := createPausePod(f, conf) framework.ExpectNoError(framework.WaitForPodRunningInNamespace(f.ClientSet, pod)) - pod, err := f.ClientSet.Core().Pods(f.Namespace.Name).Get(conf.Name) + pod, err := f.ClientSet.Core().Pods(f.Namespace.Name).Get(conf.Name, metav1.GetOptions{}) framework.ExpectNoError(err) return pod } diff --git a/test/e2e/security_context.go b/test/e2e/security_context.go index 95bd93fed2..cdb240cb19 100644 --- a/test/e2e/security_context.go +++ b/test/e2e/security_context.go @@ -26,6 +26,7 @@ import ( "fmt" "k8s.io/kubernetes/pkg/api/v1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" "k8s.io/kubernetes/pkg/util/uuid" "k8s.io/kubernetes/test/e2e/framework" @@ -180,7 +181,7 @@ func testPodSELinuxLabeling(f *framework.Framework, hostIPC bool, hostPID bool) Expect(err).To(BeNil()) Expect(content).To(ContainSubstring(testContent)) - foundPod, err := f.ClientSet.Core().Pods(f.Namespace.Name).Get(pod.Name) + foundPod, err := f.ClientSet.Core().Pods(f.Namespace.Name).Get(pod.Name, metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred()) // Confirm that the file can be accessed from a second diff --git a/test/e2e/service.go b/test/e2e/service.go index 667249c33b..7da5334b12 100644 --- a/test/e2e/service.go +++ b/test/e2e/service.go @@ -33,6 +33,7 @@ import ( "k8s.io/kubernetes/pkg/api/errors" "k8s.io/kubernetes/pkg/api/v1" "k8s.io/kubernetes/pkg/api/v1/service" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" "k8s.io/kubernetes/pkg/client/clientset_generated/internalclientset" clientset "k8s.io/kubernetes/pkg/client/clientset_generated/release_1_5" "k8s.io/kubernetes/pkg/controller/endpoint" @@ -107,7 +108,7 @@ var _ = framework.KubeDescribe("Services", func() { // TODO: We get coverage of TCP/UDP and multi-port services through the DNS test. We should have a simpler test for multi-port TCP here. It("should provide secure master service [Conformance]", func() { - _, err := cs.Core().Services(v1.NamespaceDefault).Get("kubernetes") + _, err := cs.Core().Services(v1.NamespaceDefault).Get("kubernetes", metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred()) }) @@ -1123,9 +1124,9 @@ var _ = framework.KubeDescribe("Services", func() { acceptPodName := createExecPodOrFail(cs, namespace, "execpod-accept") dropPodName := createExecPodOrFail(cs, namespace, "execpod-drop") - accpetPod, err := cs.Core().Pods(namespace).Get(acceptPodName) + accpetPod, err := cs.Core().Pods(namespace).Get(acceptPodName, metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred()) - dropPod, err := cs.Core().Pods(namespace).Get(dropPodName) + dropPod, err := cs.Core().Pods(namespace).Get(dropPodName, metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred()) By("creating a pod to be part of the service " + serviceName) @@ -1350,7 +1351,7 @@ var _ = framework.KubeDescribe("ESIPP [Slow]", func() { err := cs.Core().Pods(namespace).Delete(execPodName, nil) Expect(err).NotTo(HaveOccurred()) }() - execPod, err := f.ClientSet.Core().Pods(namespace).Get(execPodName) + execPod, err := f.ClientSet.Core().Pods(namespace).Get(execPodName, metav1.GetOptions{}) framework.ExpectNoError(err) framework.Logf("Waiting up to %v wget %v", kubeProxyLagTimeout, path) @@ -1485,7 +1486,7 @@ func updateService(c clientset.Interface, namespace, serviceName string, update var service *v1.Service var err error for i := 0; i < 3; i++ { - service, err = c.Core().Services(namespace).Get(serviceName) + service, err = c.Core().Services(namespace).Get(serviceName, metav1.GetOptions{}) if err != nil { return service, err } @@ -1540,7 +1541,7 @@ func translatePodNameToUIDOrFail(c clientset.Interface, ns string, expectedEndpo portsByUID := make(PortsByPodUID) for name, portList := range expectedEndpoints { - pod, err := c.Core().Pods(ns).Get(name) + pod, err := c.Core().Pods(ns).Get(name, metav1.GetOptions{}) if err != nil { framework.Failf("failed to get pod %s, that's pretty weird. validation failed: %s", name, err) } @@ -1576,7 +1577,7 @@ func validateEndpointsOrFail(c clientset.Interface, namespace, serviceName strin By(fmt.Sprintf("waiting up to %v for service %s in namespace %s to expose endpoints %v", framework.ServiceStartTimeout, serviceName, namespace, expectedEndpoints)) i := 1 for start := time.Now(); time.Since(start) < framework.ServiceStartTimeout; time.Sleep(1 * time.Second) { - endpoints, err := c.Core().Endpoints(namespace).Get(serviceName) + endpoints, err := c.Core().Endpoints(namespace).Get(serviceName, metav1.GetOptions{}) if err != nil { framework.Logf("Get endpoints failed (%v elapsed, ignoring for 5s): %v", time.Since(start), err) continue @@ -1641,7 +1642,7 @@ func createExecPodOrFail(client clientset.Interface, ns, generateName string) st created, err := client.Core().Pods(ns).Create(execPod) Expect(err).NotTo(HaveOccurred()) err = wait.PollImmediate(framework.Poll, 5*time.Minute, func() (bool, error) { - retrievedPod, err := client.Core().Pods(execPod.Namespace).Get(created.Name) + retrievedPod, err := client.Core().Pods(execPod.Namespace).Get(created.Name, metav1.GetOptions{}) if err != nil { return false, nil } @@ -1660,7 +1661,7 @@ func createExecPodOnNode(client clientset.Interface, ns, nodeName, generateName created, err := client.Core().Pods(ns).Create(execPod) Expect(err).NotTo(HaveOccurred()) err = wait.PollImmediate(framework.Poll, 5*time.Minute, func() (bool, error) { - retrievedPod, err := client.Core().Pods(execPod.Namespace).Get(created.Name) + retrievedPod, err := client.Core().Pods(execPod.Namespace).Get(created.Name, metav1.GetOptions{}) if err != nil { return false, nil } @@ -1965,7 +1966,7 @@ func startServeHostnameService(c clientset.Interface, internalClient internalcli } sort.StringSlice(podNames).Sort() - service, err := c.Core().Services(ns).Get(name) + service, err := c.Core().Services(ns).Get(name, metav1.GetOptions{}) if err != nil { return podNames, "", err } @@ -2250,7 +2251,7 @@ func (j *ServiceTestJig) createOnlyLocalLoadBalancerService(namespace, serviceNa // endpoints of the given Service are running. func (j *ServiceTestJig) getEndpointNodes(svc *v1.Service) map[string][]string { nodes := j.getNodes(maxNodesForEndpointsTests) - endpoints, err := j.Client.Core().Endpoints(svc.Namespace).Get(svc.Name) + endpoints, err := j.Client.Core().Endpoints(svc.Namespace).Get(svc.Name, metav1.GetOptions{}) if err != nil { framework.Failf("Get endpoints for service %s/%s failed (%s)", svc.Namespace, svc.Name, err) } @@ -2287,7 +2288,7 @@ func (j *ServiceTestJig) getNodes(maxNodesForTest int) (nodes *v1.NodeList) { func (j *ServiceTestJig) waitForEndpointOnNode(namespace, serviceName, nodeName string) { err := wait.PollImmediate(framework.Poll, loadBalancerCreateTimeoutDefault, func() (bool, error) { - endpoints, err := j.Client.Core().Endpoints(namespace).Get(serviceName) + endpoints, err := j.Client.Core().Endpoints(namespace).Get(serviceName, metav1.GetOptions{}) if err != nil { framework.Logf("Get endpoints for service %s/%s failed (%s)", namespace, serviceName, err) return false, nil @@ -2349,7 +2350,7 @@ func (j *ServiceTestJig) SanityCheckService(svc *v1.Service, svcType v1.ServiceT // face of timeouts and conflicts. func (j *ServiceTestJig) UpdateService(namespace, name string, update func(*v1.Service)) (*v1.Service, error) { for i := 0; i < 3; i++ { - service, err := j.Client.Core().Services(namespace).Get(name) + service, err := j.Client.Core().Services(namespace).Get(name, metav1.GetOptions{}) if err != nil { return nil, fmt.Errorf("Failed to get Service %q: %v", name, err) } @@ -2403,7 +2404,7 @@ func (j *ServiceTestJig) WaitForLoadBalancerOrFail(namespace, name string, timeo var service *v1.Service framework.Logf("Waiting up to %v for service %q to have a LoadBalancer", timeout, name) pollFunc := func() (bool, error) { - svc, err := j.Client.Core().Services(namespace).Get(name) + svc, err := j.Client.Core().Services(namespace).Get(name, metav1.GetOptions{}) if err != nil { return false, err } @@ -2430,7 +2431,7 @@ func (j *ServiceTestJig) WaitForLoadBalancerDestroyOrFail(namespace, name string var service *v1.Service framework.Logf("Waiting up to %v for service %q to have no LoadBalancer", timeout, name) pollFunc := func() (bool, error) { - svc, err := j.Client.Core().Services(namespace).Get(name) + svc, err := j.Client.Core().Services(namespace).Get(name, metav1.GetOptions{}) if err != nil { return false, err } @@ -2706,7 +2707,7 @@ func (t *ServiceTestFixture) Cleanup() []error { for rcName := range t.rcs { By("stopping RC " + rcName + " in namespace " + t.Namespace) // First, resize the RC to 0. - old, err := t.Client.Core().ReplicationControllers(t.Namespace).Get(rcName) + old, err := t.Client.Core().ReplicationControllers(t.Namespace).Get(rcName, metav1.GetOptions{}) if err != nil { errs = append(errs, err) } @@ -2778,7 +2779,7 @@ func execSourceipTest(f *framework.Framework, c clientset.Interface, ns, nodeNam err := c.Core().Pods(ns).Delete(execPodName, nil) Expect(err).NotTo(HaveOccurred()) }() - execPod, err := f.ClientSet.Core().Pods(ns).Get(execPodName) + execPod, err := f.ClientSet.Core().Pods(ns).Get(execPodName, metav1.GetOptions{}) framework.ExpectNoError(err) var stdout string diff --git a/test/e2e/service_accounts.go b/test/e2e/service_accounts.go index 47c63e00f6..0be7cd54e8 100644 --- a/test/e2e/service_accounts.go +++ b/test/e2e/service_accounts.go @@ -22,6 +22,7 @@ import ( apierrors "k8s.io/kubernetes/pkg/api/errors" "k8s.io/kubernetes/pkg/api/v1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" "k8s.io/kubernetes/pkg/util/uuid" "k8s.io/kubernetes/pkg/util/wait" "k8s.io/kubernetes/pkg/version" @@ -42,7 +43,7 @@ var _ = framework.KubeDescribe("ServiceAccounts", func() { var secrets []v1.ObjectReference framework.ExpectNoError(wait.Poll(time.Millisecond*500, time.Second*10, func() (bool, error) { By("waiting for a single token reference") - sa, err := f.ClientSet.Core().ServiceAccounts(f.Namespace.Name).Get("default") + sa, err := f.ClientSet.Core().ServiceAccounts(f.Namespace.Name).Get("default", metav1.GetOptions{}) if apierrors.IsNotFound(err) { framework.Logf("default service account was not found") return false, nil @@ -68,7 +69,7 @@ var _ = framework.KubeDescribe("ServiceAccounts", func() { { By("ensuring the single token reference persists") time.Sleep(2 * time.Second) - sa, err := f.ClientSet.Core().ServiceAccounts(f.Namespace.Name).Get("default") + sa, err := f.ClientSet.Core().ServiceAccounts(f.Namespace.Name).Get("default", metav1.GetOptions{}) framework.ExpectNoError(err) Expect(sa.Secrets).To(Equal(secrets)) } @@ -80,7 +81,7 @@ var _ = framework.KubeDescribe("ServiceAccounts", func() { // wait for the referenced secret to be removed, and another one autocreated framework.ExpectNoError(wait.Poll(time.Millisecond*500, framework.ServiceAccountProvisionTimeout, func() (bool, error) { By("waiting for a new token reference") - sa, err := f.ClientSet.Core().ServiceAccounts(f.Namespace.Name).Get("default") + sa, err := f.ClientSet.Core().ServiceAccounts(f.Namespace.Name).Get("default", metav1.GetOptions{}) if err != nil { framework.Logf("error getting default service account: %v", err) return false, err @@ -106,7 +107,7 @@ var _ = framework.KubeDescribe("ServiceAccounts", func() { { By("ensuring the single token reference persists") time.Sleep(2 * time.Second) - sa, err := f.ClientSet.Core().ServiceAccounts(f.Namespace.Name).Get("default") + sa, err := f.ClientSet.Core().ServiceAccounts(f.Namespace.Name).Get("default", metav1.GetOptions{}) framework.ExpectNoError(err) Expect(sa.Secrets).To(Equal(secrets)) } @@ -114,7 +115,7 @@ var _ = framework.KubeDescribe("ServiceAccounts", func() { // delete the reference from the service account By("deleting the reference to the service account token") { - sa, err := f.ClientSet.Core().ServiceAccounts(f.Namespace.Name).Get("default") + sa, err := f.ClientSet.Core().ServiceAccounts(f.Namespace.Name).Get("default", metav1.GetOptions{}) framework.ExpectNoError(err) sa.Secrets = nil _, updateErr := f.ClientSet.Core().ServiceAccounts(f.Namespace.Name).Update(sa) @@ -124,7 +125,7 @@ var _ = framework.KubeDescribe("ServiceAccounts", func() { // wait for another one to be autocreated framework.ExpectNoError(wait.Poll(time.Millisecond*500, framework.ServiceAccountProvisionTimeout, func() (bool, error) { By("waiting for a new token to be created and added") - sa, err := f.ClientSet.Core().ServiceAccounts(f.Namespace.Name).Get("default") + sa, err := f.ClientSet.Core().ServiceAccounts(f.Namespace.Name).Get("default", metav1.GetOptions{}) if err != nil { framework.Logf("error getting default service account: %v", err) return false, err @@ -146,7 +147,7 @@ var _ = framework.KubeDescribe("ServiceAccounts", func() { { By("ensuring the single token reference persists") time.Sleep(2 * time.Second) - sa, err := f.ClientSet.Core().ServiceAccounts(f.Namespace.Name).Get("default") + sa, err := f.ClientSet.Core().ServiceAccounts(f.Namespace.Name).Get("default", metav1.GetOptions{}) framework.ExpectNoError(err) Expect(sa.Secrets).To(Equal(secrets)) } @@ -159,7 +160,7 @@ var _ = framework.KubeDescribe("ServiceAccounts", func() { // Standard get, update retry loop framework.ExpectNoError(wait.Poll(time.Millisecond*500, framework.ServiceAccountProvisionTimeout, func() (bool, error) { By("getting the auto-created API token") - sa, err := f.ClientSet.Core().ServiceAccounts(f.Namespace.Name).Get("default") + sa, err := f.ClientSet.Core().ServiceAccounts(f.Namespace.Name).Get("default", metav1.GetOptions{}) if apierrors.IsNotFound(err) { framework.Logf("default service account was not found") return false, nil @@ -173,7 +174,7 @@ var _ = framework.KubeDescribe("ServiceAccounts", func() { return false, nil } for _, secretRef := range sa.Secrets { - secret, err := f.ClientSet.Core().Secrets(f.Namespace.Name).Get(secretRef.Name) + secret, err := f.ClientSet.Core().Secrets(f.Namespace.Name).Get(secretRef.Name, metav1.GetOptions{}) if err != nil { framework.Logf("Error getting secret %s: %v", secretRef.Name, err) continue diff --git a/test/e2e/ubernetes_lite.go b/test/e2e/ubernetes_lite.go index fd14b563bf..2dcf22487d 100644 --- a/test/e2e/ubernetes_lite.go +++ b/test/e2e/ubernetes_lite.go @@ -150,7 +150,7 @@ func getZoneCount(c clientset.Interface) (int, error) { // Find the name of the zone in which the pod is scheduled func getZoneNameForPod(c clientset.Interface, pod v1.Pod) (string, error) { By(fmt.Sprintf("Getting zone name for pod %s, on node %s", pod.Name, pod.Spec.NodeName)) - node, err := c.Core().Nodes().Get(pod.Spec.NodeName) + node, err := c.Core().Nodes().Get(pod.Spec.NodeName, metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred()) return getZoneNameForNode(*node) } diff --git a/test/e2e/volume_provisioning.go b/test/e2e/volume_provisioning.go index 05c51b944f..e3e6fe505f 100644 --- a/test/e2e/volume_provisioning.go +++ b/test/e2e/volume_provisioning.go @@ -45,11 +45,11 @@ func testDynamicProvisioning(client clientset.Interface, claim *v1.PersistentVol By("checking the claim") // Get new copy of the claim - claim, err = client.Core().PersistentVolumeClaims(claim.Namespace).Get(claim.Name) + claim, err = client.Core().PersistentVolumeClaims(claim.Namespace).Get(claim.Name, metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred()) // Get the bound PV - pv, err := client.Core().PersistentVolumes().Get(claim.Spec.VolumeName) + pv, err := client.Core().PersistentVolumes().Get(claim.Spec.VolumeName, metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred()) // Check sizes diff --git a/test/e2e/volumes.go b/test/e2e/volumes.go index ea87712819..52fb4136d7 100644 --- a/test/e2e/volumes.go +++ b/test/e2e/volumes.go @@ -150,7 +150,7 @@ func startVolumeServer(client clientset.Interface, config VolumeTestConfig) *v1. framework.ExpectNoError(framework.WaitForPodRunningInNamespace(client, serverPod)) By("locating the server pod") - pod, err := podClient.Get(serverPod.Name) + pod, err := podClient.Get(serverPod.Name, metav1.GetOptions{}) framework.ExpectNoError(err, "Cannot locate the server pod %v: %v", serverPod.Name, err) By("sleeping a bit to give the server time to start") diff --git a/test/e2e_node/apparmor_test.go b/test/e2e_node/apparmor_test.go index 85c2d648cc..cb293b5bf8 100644 --- a/test/e2e_node/apparmor_test.go +++ b/test/e2e_node/apparmor_test.go @@ -28,6 +28,7 @@ import ( "k8s.io/kubernetes/pkg/api/errors" "k8s.io/kubernetes/pkg/api/v1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" "k8s.io/kubernetes/pkg/runtime/schema" "k8s.io/kubernetes/pkg/security/apparmor" "k8s.io/kubernetes/pkg/watch" @@ -163,7 +164,7 @@ func runAppArmorTest(f *framework.Framework, shouldRun bool, profile string) v1. }) framework.ExpectNoError(err) } - p, err := f.PodClient().Get(pod.Name) + p, err := f.PodClient().Get(pod.Name, metav1.GetOptions{}) framework.ExpectNoError(err) return p.Status } diff --git a/test/e2e_node/benchmark_util.go b/test/e2e_node/benchmark_util.go index 7bc5af60de..940966091a 100644 --- a/test/e2e_node/benchmark_util.go +++ b/test/e2e_node/benchmark_util.go @@ -126,7 +126,7 @@ func getThroughputPerfData(batchLag time.Duration, e2eLags []framework.PodLatenc // getTestNodeInfo fetches the capacity of a node from API server and returns a map of labels. func getTestNodeInfo(f *framework.Framework, testName string) map[string]string { nodeName := framework.TestContext.NodeName - node, err := f.ClientSet.Core().Nodes().Get(nodeName) + node, err := f.ClientSet.Core().Nodes().Get(nodeName, metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred()) cpu, ok := node.Status.Capacity["cpu"] diff --git a/test/e2e_node/container.go b/test/e2e_node/container.go index 55daf5e802..2b5ca99c5c 100644 --- a/test/e2e_node/container.go +++ b/test/e2e_node/container.go @@ -21,6 +21,7 @@ import ( "k8s.io/kubernetes/pkg/api/errors" "k8s.io/kubernetes/pkg/api/v1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" "k8s.io/kubernetes/pkg/util/uuid" "k8s.io/kubernetes/test/e2e/framework" ) @@ -66,7 +67,7 @@ func (cc *ConformanceContainer) Delete() error { } func (cc *ConformanceContainer) IsReady() (bool, error) { - pod, err := cc.PodClient.Get(cc.podName) + pod, err := cc.PodClient.Get(cc.podName, metav1.GetOptions{}) if err != nil { return false, err } @@ -74,7 +75,7 @@ func (cc *ConformanceContainer) IsReady() (bool, error) { } func (cc *ConformanceContainer) GetPhase() (v1.PodPhase, error) { - pod, err := cc.PodClient.Get(cc.podName) + pod, err := cc.PodClient.Get(cc.podName, metav1.GetOptions{}) if err != nil { return v1.PodUnknown, err } @@ -82,7 +83,7 @@ func (cc *ConformanceContainer) GetPhase() (v1.PodPhase, error) { } func (cc *ConformanceContainer) GetStatus() (v1.ContainerStatus, error) { - pod, err := cc.PodClient.Get(cc.podName) + pod, err := cc.PodClient.Get(cc.podName, metav1.GetOptions{}) if err != nil { return v1.ContainerStatus{}, err } @@ -94,7 +95,7 @@ func (cc *ConformanceContainer) GetStatus() (v1.ContainerStatus, error) { } func (cc *ConformanceContainer) Present() (bool, error) { - _, err := cc.PodClient.Get(cc.podName) + _, err := cc.PodClient.Get(cc.podName, metav1.GetOptions{}) if err == nil { return true, nil } diff --git a/test/e2e_node/disk_eviction_test.go b/test/e2e_node/disk_eviction_test.go index b272352a5e..09d7fc39ba 100644 --- a/test/e2e_node/disk_eviction_test.go +++ b/test/e2e_node/disk_eviction_test.go @@ -22,6 +22,7 @@ import ( "time" "k8s.io/kubernetes/pkg/api/v1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" "k8s.io/kubernetes/pkg/util/uuid" "k8s.io/kubernetes/test/e2e/framework" @@ -125,7 +126,7 @@ var _ = framework.KubeDescribe("Kubelet Eviction Manager [Serial] [Disruptive]", // The pod should be evicted. if !evictionOccurred { - podData, err := podClient.Get(busyPodName) + podData, err := podClient.Get(busyPodName, metav1.GetOptions{}) if err != nil { return err } @@ -135,7 +136,7 @@ var _ = framework.KubeDescribe("Kubelet Eviction Manager [Serial] [Disruptive]", return err } - podData, err = podClient.Get(idlePodName) + podData, err = podClient.Get(idlePodName, metav1.GetOptions{}) if err != nil { return err } @@ -170,7 +171,7 @@ var _ = framework.KubeDescribe("Kubelet Eviction Manager [Serial] [Disruptive]", } // The new pod should be able to be scheduled and run after the disk pressure is relieved. - podData, err := podClient.Get(verifyPodName) + podData, err := podClient.Get(verifyPodName, metav1.GetOptions{}) if err != nil { return err } diff --git a/test/e2e_node/garbage_collector_test.go b/test/e2e_node/garbage_collector_test.go index ce97711d02..fc56069793 100644 --- a/test/e2e_node/garbage_collector_test.go +++ b/test/e2e_node/garbage_collector_test.go @@ -22,6 +22,7 @@ import ( "time" "k8s.io/kubernetes/pkg/api/v1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" docker "k8s.io/kubernetes/pkg/kubelet/dockertools" "k8s.io/kubernetes/test/e2e/framework" @@ -149,7 +150,7 @@ func containerGCTest(f *framework.Framework, test testRun) { By("Making sure all containers restart the specified number of times") Eventually(func() error { for _, podSpec := range test.testPods { - updatedPod, err := f.ClientSet.Core().Pods(f.Namespace.Name).Get(podSpec.podName) + updatedPod, err := f.ClientSet.Core().Pods(f.Namespace.Name).Get(podSpec.podName, metav1.GetOptions{}) if err != nil { return err } diff --git a/test/e2e_node/image_id_test.go b/test/e2e_node/image_id_test.go index 3636fbed5f..8cead09863 100644 --- a/test/e2e_node/image_id_test.go +++ b/test/e2e_node/image_id_test.go @@ -18,6 +18,7 @@ package e2e_node import ( "k8s.io/kubernetes/pkg/api/v1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" "k8s.io/kubernetes/pkg/kubelet/dockertools" "k8s.io/kubernetes/test/e2e/framework" @@ -51,7 +52,7 @@ var _ = framework.KubeDescribe("ImageID", func() { framework.ExpectNoError(framework.WaitTimeoutForPodNoLongerRunningInNamespace( f.ClientSet, pod.Name, f.Namespace.Name, "", framework.PodStartTimeout)) - runningPod, err := f.PodClient().Get(pod.Name) + runningPod, err := f.PodClient().Get(pod.Name, metav1.GetOptions{}) framework.ExpectNoError(err) status := runningPod.Status diff --git a/test/e2e_node/kubelet_test.go b/test/e2e_node/kubelet_test.go index cdb6bfdbec..8bc35c6653 100644 --- a/test/e2e_node/kubelet_test.go +++ b/test/e2e_node/kubelet_test.go @@ -22,7 +22,7 @@ import ( "time" "k8s.io/kubernetes/pkg/api/v1" - apiunversioned "k8s.io/kubernetes/pkg/apis/meta/v1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" "k8s.io/kubernetes/pkg/util/uuid" "k8s.io/kubernetes/test/e2e/framework" @@ -56,7 +56,7 @@ var _ = framework.KubeDescribe("Kubelet", func() { }, }) Eventually(func() string { - sinceTime := apiunversioned.NewTime(time.Now().Add(time.Duration(-1 * time.Hour))) + sinceTime := metav1.NewTime(time.Now().Add(time.Duration(-1 * time.Hour))) rc, err := podClient.GetLogs(podName, &v1.PodLogOptions{SinceTime: &sinceTime}).Stream() if err != nil { return "" @@ -93,7 +93,7 @@ var _ = framework.KubeDescribe("Kubelet", func() { It("should have an error terminated reason", func() { Eventually(func() error { - podData, err := podClient.Get(podName) + podData, err := podClient.Get(podName, metav1.GetOptions{}) if err != nil { return err } diff --git a/test/e2e_node/log_path_test.go b/test/e2e_node/log_path_test.go index f9a64f28ae..5ec742642e 100644 --- a/test/e2e_node/log_path_test.go +++ b/test/e2e_node/log_path_test.go @@ -18,6 +18,7 @@ package e2e_node import ( "k8s.io/kubernetes/pkg/api/v1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" "k8s.io/kubernetes/pkg/kubelet" kubecontainer "k8s.io/kubernetes/pkg/kubelet/container" "k8s.io/kubernetes/test/e2e/framework" @@ -67,7 +68,7 @@ var _ = framework.KubeDescribe("ContainerLogPath", func() { framework.ExpectNoError(err, "Failed waiting for pod: %s to enter success state", logPodName) // get containerID from created Pod - createdLogPod, err := podClient.Get(logPodName) + createdLogPod, err := podClient.Get(logPodName, metav1.GetOptions{}) logConID := kubecontainer.ParseContainerID(createdLogPod.Status.ContainerStatuses[0].ContainerID) framework.ExpectNoError(err, "Failed to get pod: %s", logPodName) diff --git a/test/e2e_node/memory_eviction_test.go b/test/e2e_node/memory_eviction_test.go index 727b6b951d..5558d5f561 100644 --- a/test/e2e_node/memory_eviction_test.go +++ b/test/e2e_node/memory_eviction_test.go @@ -24,6 +24,7 @@ import ( "github.com/golang/glog" "k8s.io/kubernetes/pkg/api/resource" "k8s.io/kubernetes/pkg/api/v1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" "k8s.io/kubernetes/test/e2e/framework" . "github.com/onsi/ginkgo" @@ -150,15 +151,15 @@ var _ = framework.KubeDescribe("MemoryEviction [Slow] [Serial] [Disruptive]", fu By("polling the Status.Phase of each pod and checking for violations of the eviction order.") Eventually(func() error { - gteed, gtErr := f.ClientSet.Core().Pods(f.Namespace.Name).Get(guaranteed.Name) + gteed, gtErr := f.ClientSet.Core().Pods(f.Namespace.Name).Get(guaranteed.Name, metav1.GetOptions{}) framework.ExpectNoError(gtErr, fmt.Sprintf("getting pod %s", guaranteed.Name)) gteedPh := gteed.Status.Phase - burst, buErr := f.ClientSet.Core().Pods(f.Namespace.Name).Get(burstable.Name) + burst, buErr := f.ClientSet.Core().Pods(f.Namespace.Name).Get(burstable.Name, metav1.GetOptions{}) framework.ExpectNoError(buErr, fmt.Sprintf("getting pod %s", burstable.Name)) burstPh := burst.Status.Phase - best, beErr := f.ClientSet.Core().Pods(f.Namespace.Name).Get(besteffort.Name) + best, beErr := f.ClientSet.Core().Pods(f.Namespace.Name).Get(besteffort.Name, metav1.GetOptions{}) framework.ExpectNoError(beErr, fmt.Sprintf("getting pod %s", besteffort.Name)) bestPh := best.Status.Phase diff --git a/test/e2e_node/mirror_pod_test.go b/test/e2e_node/mirror_pod_test.go index 1b6f4e57f6..dc1a6eceb5 100644 --- a/test/e2e_node/mirror_pod_test.go +++ b/test/e2e_node/mirror_pod_test.go @@ -25,6 +25,7 @@ import ( "k8s.io/kubernetes/pkg/api/errors" "k8s.io/kubernetes/pkg/api/v1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" clientset "k8s.io/kubernetes/pkg/client/clientset_generated/release_1_5" "k8s.io/kubernetes/pkg/types" "k8s.io/kubernetes/pkg/util/uuid" @@ -57,7 +58,7 @@ var _ = framework.KubeDescribe("MirrorPod", func() { }) It("should be updated when static pod updated [Conformance]", func() { By("get mirror pod uid") - pod, err := f.ClientSet.Core().Pods(ns).Get(mirrorPodName) + pod, err := f.ClientSet.Core().Pods(ns).Get(mirrorPodName, metav1.GetOptions{}) Expect(err).ShouldNot(HaveOccurred()) uid := pod.UID @@ -72,14 +73,14 @@ var _ = framework.KubeDescribe("MirrorPod", func() { }, 2*time.Minute, time.Second*4).Should(BeNil()) By("check the mirror pod container image is updated") - pod, err = f.ClientSet.Core().Pods(ns).Get(mirrorPodName) + pod, err = f.ClientSet.Core().Pods(ns).Get(mirrorPodName, metav1.GetOptions{}) Expect(err).ShouldNot(HaveOccurred()) Expect(len(pod.Spec.Containers)).Should(Equal(1)) Expect(pod.Spec.Containers[0].Image).Should(Equal(image)) }) It("should be recreated when mirror pod gracefully deleted [Conformance]", func() { By("get mirror pod uid") - pod, err := f.ClientSet.Core().Pods(ns).Get(mirrorPodName) + pod, err := f.ClientSet.Core().Pods(ns).Get(mirrorPodName, metav1.GetOptions{}) Expect(err).ShouldNot(HaveOccurred()) uid := pod.UID @@ -94,7 +95,7 @@ var _ = framework.KubeDescribe("MirrorPod", func() { }) It("should be recreated when mirror pod forcibly deleted [Conformance]", func() { By("get mirror pod uid") - pod, err := f.ClientSet.Core().Pods(ns).Get(mirrorPodName) + pod, err := f.ClientSet.Core().Pods(ns).Get(mirrorPodName, metav1.GetOptions{}) Expect(err).ShouldNot(HaveOccurred()) uid := pod.UID @@ -156,7 +157,7 @@ func deleteStaticPod(dir, name, namespace string) error { } func checkMirrorPodDisappear(cl clientset.Interface, name, namespace string) error { - _, err := cl.Core().Pods(namespace).Get(name) + _, err := cl.Core().Pods(namespace).Get(name, metav1.GetOptions{}) if errors.IsNotFound(err) { return nil } @@ -164,7 +165,7 @@ func checkMirrorPodDisappear(cl clientset.Interface, name, namespace string) err } func checkMirrorPodRunning(cl clientset.Interface, name, namespace string) error { - pod, err := cl.Core().Pods(namespace).Get(name) + pod, err := cl.Core().Pods(namespace).Get(name, metav1.GetOptions{}) if err != nil { return fmt.Errorf("expected the mirror pod %q to appear: %v", name, err) } @@ -175,7 +176,7 @@ func checkMirrorPodRunning(cl clientset.Interface, name, namespace string) error } func checkMirrorPodRecreatedAndRunnig(cl clientset.Interface, name, namespace string, oUID types.UID) error { - pod, err := cl.Core().Pods(namespace).Get(name) + pod, err := cl.Core().Pods(namespace).Get(name, metav1.GetOptions{}) if err != nil { return fmt.Errorf("expected the mirror pod %q to appear: %v", name, err) } diff --git a/test/e2e_node/util.go b/test/e2e_node/util.go index e7c144be39..a913b51b56 100644 --- a/test/e2e_node/util.go +++ b/test/e2e_node/util.go @@ -33,6 +33,7 @@ import ( "k8s.io/kubernetes/pkg/api/v1" "k8s.io/kubernetes/pkg/apis/componentconfig" v1alpha1 "k8s.io/kubernetes/pkg/apis/componentconfig/v1alpha1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" "k8s.io/kubernetes/pkg/kubelet/api/v1alpha1/stats" // utilconfig "k8s.io/kubernetes/pkg/util/config" "k8s.io/kubernetes/test/e2e/framework" @@ -86,7 +87,7 @@ func getCurrentKubeletConfig() (*componentconfig.KubeletConfiguration, error) { // Queries the API server for a Kubelet configuration for the node described by framework.TestContext.NodeName func getCurrentKubeletConfigMap(f *framework.Framework) (*v1.ConfigMap, error) { - return f.ClientSet.Core().ConfigMaps("kube-system").Get(fmt.Sprintf("kubelet-%s", framework.TestContext.NodeName)) + return f.ClientSet.Core().ConfigMaps("kube-system").Get(fmt.Sprintf("kubelet-%s", framework.TestContext.NodeName), metav1.GetOptions{}) } // Creates or updates the configmap for KubeletConfiguration, waits for the Kubelet to restart diff --git a/test/integration/client/client_test.go b/test/integration/client/client_test.go index 6de1092efd..9ca0f328be 100644 --- a/test/integration/client/client_test.go +++ b/test/integration/client/client_test.go @@ -173,7 +173,7 @@ func TestAtomicPut(t *testing.T) { go func(l, v string) { defer wg.Done() for { - tmpRC, err := rcs.Get(rc.Name) + tmpRC, err := rcs.Get(rc.Name, metav1.GetOptions{}) if err != nil { t.Errorf("Error getting atomicRC: %v", err) continue @@ -199,7 +199,7 @@ func TestAtomicPut(t *testing.T) { }(label, value) } wg.Wait() - rc, err = rcs.Get(rc.Name) + rc, err = rcs.Get(rc.Name, metav1.GetOptions{}) if err != nil { t.Fatalf("Failed getting atomicRC after writers are complete: %v", err) } @@ -281,7 +281,7 @@ func TestPatch(t *testing.T) { if err != nil { t.Fatalf("Failed updating patchpod with patch type %s: %v", k, err) } - pod, err = pods.Get(name) + pod, err = pods.Get(name, metav1.GetOptions{}) if err != nil { t.Fatalf("Failed getting patchpod: %v", err) } @@ -294,7 +294,7 @@ func TestPatch(t *testing.T) { if err != nil { t.Fatalf("Failed updating patchpod with patch type %s: %v", k, err) } - pod, err = pods.Get(name) + pod, err = pods.Get(name, metav1.GetOptions{}) if err != nil { t.Fatalf("Failed getting patchpod: %v", err) } @@ -307,7 +307,7 @@ func TestPatch(t *testing.T) { if err != nil { t.Fatalf("Failed updating patchpod with patch type %s: %v", k, err) } - pod, err = pods.Get(name) + pod, err = pods.Get(name, metav1.GetOptions{}) if err != nil { t.Fatalf("Failed getting patchpod: %v", err) } @@ -712,7 +712,7 @@ func TestMultiWatch(t *testing.T) { for i := 0; i < watcherCount; i++ { go func(i int) { name := fmt.Sprintf("multi-watch-%v", i) - pod, err := client.Core().Pods(ns.Name).Get(name) + pod, err := client.Core().Pods(ns.Name).Get(name, metav1.GetOptions{}) if err != nil { panic(fmt.Sprintf("Couldn't get %v: %v", name, err)) } diff --git a/test/integration/evictions/evictions_test.go b/test/integration/evictions/evictions_test.go index 840862656e..4f836994ec 100644 --- a/test/integration/evictions/evictions_test.go +++ b/test/integration/evictions/evictions_test.go @@ -213,7 +213,7 @@ func TestConcurrentEvictionRequests(t *testing.T) { for i := 0; i < 10; i++ { podName := fmt.Sprintf(podNameFormat, i) - _, err := clientSet.Core().Pods(ns.Name).Get(podName) + _, err := clientSet.Core().Pods(ns.Name).Get(podName, metav1.GetOptions{}) if !errors.IsNotFound(err) { t.Errorf("Pod %q is expected to be evicted", podName) } @@ -243,7 +243,7 @@ func waitToObservePods(t *testing.T, podInformer cache.SharedIndexInformer, podN func waitPDBStable(t *testing.T, clientSet clientset.Interface, podNum int32, ns, pdbName string) { if err := wait.PollImmediate(2*time.Second, 60*time.Second, func() (bool, error) { - pdb, err := clientSet.Policy().PodDisruptionBudgets(ns).Get(pdbName) + pdb, err := clientSet.Policy().PodDisruptionBudgets(ns).Get(pdbName, metav1.GetOptions{}) if err != nil { return false, err } diff --git a/test/integration/framework/master_utils.go b/test/integration/framework/master_utils.go index 0326f96bf1..8621ebe964 100644 --- a/test/integration/framework/master_utils.go +++ b/test/integration/framework/master_utils.go @@ -36,6 +36,7 @@ import ( "k8s.io/kubernetes/pkg/apis/batch" certificates "k8s.io/kubernetes/pkg/apis/certificates/v1alpha1" extensions "k8s.io/kubernetes/pkg/apis/extensions/v1beta1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" policy "k8s.io/kubernetes/pkg/apis/policy/v1alpha1" rbac "k8s.io/kubernetes/pkg/apis/rbac/v1alpha1" storage "k8s.io/kubernetes/pkg/apis/storage/v1beta1" @@ -452,7 +453,7 @@ func ScaleRC(name, ns string, replicas int32, clientset internalclientset.Interf if err != nil { return nil, err } - scaled, err := clientset.Core().ReplicationControllers(ns).Get(name) + scaled, err := clientset.Core().ReplicationControllers(ns).Get(name, metav1.GetOptions{}) if err != nil { return nil, err } diff --git a/test/integration/garbagecollector/garbage_collector_test.go b/test/integration/garbagecollector/garbage_collector_test.go index 1061094bd0..5dad57f3f2 100644 --- a/test/integration/garbagecollector/garbage_collector_test.go +++ b/test/integration/garbagecollector/garbage_collector_test.go @@ -232,10 +232,10 @@ func TestCascadingDeletion(t *testing.T) { t.Fatalf("expect pod %s to be garbage collected, got err= %v", garbageCollectedPodName, err) } // checks the garbage collect doesn't delete pods it shouldn't delete. - if _, err := podClient.Get(independentPodName); err != nil { + if _, err := podClient.Get(independentPodName, metav1.GetOptions{}); err != nil { t.Fatal(err) } - if _, err := podClient.Get(oneValidOwnerPodName); err != nil { + if _, err := podClient.Get(oneValidOwnerPodName, metav1.GetOptions{}); err != nil { t.Fatal(err) } } diff --git a/test/integration/master/master_test.go b/test/integration/master/master_test.go index 356d0da5c5..2207ceee7a 100644 --- a/test/integration/master/master_test.go +++ b/test/integration/master/master_test.go @@ -39,6 +39,7 @@ import ( "k8s.io/kubernetes/pkg/api/v1" "k8s.io/kubernetes/pkg/apimachinery/registered" "k8s.io/kubernetes/pkg/apis/batch/v2alpha1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" clientset "k8s.io/kubernetes/pkg/client/clientset_generated/internalclientset" clienttypedv1 "k8s.io/kubernetes/pkg/client/clientset_generated/release_1_5/typed/core/v1" "k8s.io/kubernetes/pkg/client/restclient" @@ -407,7 +408,7 @@ func TestMasterService(t *testing.T) { } } if found { - ep, err := client.Core().Endpoints(api.NamespaceDefault).Get("kubernetes") + ep, err := client.Core().Endpoints(api.NamespaceDefault).Get("kubernetes", metav1.GetOptions{}) if err != nil { return false, nil } @@ -451,7 +452,7 @@ func TestServiceAlloc(t *testing.T) { // Wait until the default "kubernetes" service is created. if err = wait.Poll(250*time.Millisecond, time.Minute, func() (bool, error) { - _, err := client.Core().Services(api.NamespaceDefault).Get("kubernetes") + _, err := client.Core().Services(api.NamespaceDefault).Get("kubernetes", metav1.GetOptions{}) if err != nil && !errors.IsNotFound(err) { return false, err } @@ -595,7 +596,7 @@ func TestUpdateNodeObjects(t *testing.T) { break } - n, err := c.Nodes().Get(fmt.Sprintf("node-%d", node)) + n, err := c.Nodes().Get(fmt.Sprintf("node-%d", node), metav1.GetOptions{}) if err != nil { fmt.Printf("[%d] error after %d: %v\n", node, i, err) break diff --git a/test/integration/replicaset/replicaset_test.go b/test/integration/replicaset/replicaset_test.go index 9633156d58..03092bc781 100644 --- a/test/integration/replicaset/replicaset_test.go +++ b/test/integration/replicaset/replicaset_test.go @@ -245,7 +245,7 @@ func TestAdoption(t *testing.T) { waitToObservePods(t, podInformer, 1) go rm.Run(5, stopCh) if err := wait.Poll(10*time.Second, 60*time.Second, func() (bool, error) { - updatedPod, err := podClient.Get(pod.Name) + updatedPod, err := podClient.Get(pod.Name, metav1.GetOptions{}) if err != nil { return false, err } @@ -280,7 +280,7 @@ func createRSsPods(t *testing.T, clientSet clientset.Interface, rss []*v1beta1.R func waitRSStable(t *testing.T, clientSet clientset.Interface, rs *v1beta1.ReplicaSet, ns string) { rsClient := clientSet.Extensions().ReplicaSets(ns) if err := wait.Poll(10*time.Second, 60*time.Second, func() (bool, error) { - updatedRS, err := rsClient.Get(rs.Name) + updatedRS, err := rsClient.Get(rs.Name, metav1.GetOptions{}) if err != nil { return false, err } @@ -371,7 +371,7 @@ func TestUpdateSelectorToRemoveControllerRef(t *testing.T) { t.Fatal(err) } podClient := clientSet.Core().Pods(ns.Name) - pod2, err = podClient.Get(pod2.Name) + pod2, err = podClient.Get(pod2.Name, metav1.GetOptions{}) if err != nil { t.Fatalf("Failed to get pod2: %v", err) } @@ -414,7 +414,7 @@ func TestUpdateLabelToRemoveControllerRef(t *testing.T) { }); err != nil { t.Fatal(err) } - pod2, err = podClient.Get(pod2.Name) + pod2, err = podClient.Get(pod2.Name, metav1.GetOptions{}) if err != nil { t.Fatalf("Failed to get pod2: %v", err) } diff --git a/test/integration/replicationcontroller/replicationcontroller_test.go b/test/integration/replicationcontroller/replicationcontroller_test.go index 0cbfc0b3a1..a9fd548de8 100644 --- a/test/integration/replicationcontroller/replicationcontroller_test.go +++ b/test/integration/replicationcontroller/replicationcontroller_test.go @@ -243,7 +243,7 @@ func TestAdoption(t *testing.T) { waitToObservePods(t, podInformer, 1) go rm.Run(5, stopCh) if err := wait.Poll(10*time.Second, 60*time.Second, func() (bool, error) { - updatedPod, err := podClient.Get(pod.Name) + updatedPod, err := podClient.Get(pod.Name, metav1.GetOptions{}) if err != nil { return false, err } @@ -278,7 +278,7 @@ func createRCsPods(t *testing.T, clientSet clientset.Interface, rcs []*v1.Replic func waitRCStable(t *testing.T, clientSet clientset.Interface, rc *v1.ReplicationController, ns string) { rcClient := clientSet.Core().ReplicationControllers(ns) if err := wait.Poll(10*time.Second, 60*time.Second, func() (bool, error) { - updatedRC, err := rcClient.Get(rc.Name) + updatedRC, err := rcClient.Get(rc.Name, metav1.GetOptions{}) if err != nil { return false, err } @@ -367,7 +367,7 @@ func TestUpdateSelectorToRemoveControllerRef(t *testing.T) { t.Fatal(err) } podClient := clientSet.Core().Pods(ns.Name) - pod2, err = podClient.Get(pod2.Name) + pod2, err = podClient.Get(pod2.Name, metav1.GetOptions{}) if err != nil { t.Fatalf("Failed to get pod2: %v", err) } @@ -409,7 +409,7 @@ func TestUpdateLabelToRemoveControllerRef(t *testing.T) { }); err != nil { t.Fatal(err) } - pod2, err = podClient.Get(pod2.Name) + pod2, err = podClient.Get(pod2.Name, metav1.GetOptions{}) if err != nil { t.Fatalf("Failed to get pod2: %v", err) } diff --git a/test/integration/scheduler/extender_test.go b/test/integration/scheduler/extender_test.go index ab8c27000a..7470efb198 100644 --- a/test/integration/scheduler/extender_test.go +++ b/test/integration/scheduler/extender_test.go @@ -298,7 +298,7 @@ func DoTestPodScheduling(ns *v1.Namespace, t *testing.T, cs clientset.Interface) t.Fatalf("Failed to schedule pod: %v", err) } - if myPod, err := cs.Core().Pods(ns.Name).Get(myPod.Name); err != nil { + if myPod, err := cs.Core().Pods(ns.Name).Get(myPod.Name, metav1.GetOptions{}); err != nil { t.Fatalf("Failed to get pod: %v", err) } else if myPod.Spec.NodeName != "machine3" { t.Fatalf("Failed to schedule using extender, expected machine3, got %v", myPod.Spec.NodeName) diff --git a/test/integration/scheduler/scheduler_test.go b/test/integration/scheduler/scheduler_test.go index daf5a572f4..456013077f 100644 --- a/test/integration/scheduler/scheduler_test.go +++ b/test/integration/scheduler/scheduler_test.go @@ -76,7 +76,7 @@ func TestUnschedulableNodes(t *testing.T) { func podScheduled(c clientset.Interface, podNamespace, podName string) wait.ConditionFunc { return func() (bool, error) { - pod, err := c.Core().Pods(podNamespace).Get(podName) + pod, err := c.Core().Pods(podNamespace).Get(podName, metav1.GetOptions{}) if errors.IsNotFound(err) { return false, nil } @@ -268,7 +268,7 @@ func DoTestUnschedulableNodes(t *testing.T, cs clientset.Interface, ns *v1.Names } // Apply the schedulable modification to the node, and wait for the reflection - schedNode, err := cs.Core().Nodes().Get(unSchedNode.Name) + schedNode, err := cs.Core().Nodes().Get(unSchedNode.Name, metav1.GetOptions{}) if err != nil { t.Fatalf("Failed to get node: %v", err) } diff --git a/test/integration/serviceaccount/service_account_test.go b/test/integration/serviceaccount/service_account_test.go index 6335086e11..662078a8c0 100644 --- a/test/integration/serviceaccount/service_account_test.go +++ b/test/integration/serviceaccount/service_account_test.go @@ -35,6 +35,7 @@ import ( "k8s.io/kubernetes/pkg/api/errors" "k8s.io/kubernetes/pkg/api/v1" "k8s.io/kubernetes/pkg/apimachinery/registered" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" "k8s.io/kubernetes/pkg/auth/authenticator" "k8s.io/kubernetes/pkg/auth/authenticator/bearertoken" "k8s.io/kubernetes/pkg/auth/authorizer" @@ -139,7 +140,7 @@ func TestServiceAccountTokenAutoCreate(t *testing.T) { } // Trigger creation of a new referenced token - serviceAccount, err = c.Core().ServiceAccounts(ns).Get(name) + serviceAccount, err = c.Core().ServiceAccounts(ns).Get(name, metav1.GetOptions{}) if err != nil { t.Fatal(err) } @@ -435,13 +436,13 @@ func startServiceAccountTestServer(t *testing.T) (*clientset.Clientset, restclie func getServiceAccount(c *clientset.Clientset, ns string, name string, shouldWait bool) (*v1.ServiceAccount, error) { if !shouldWait { - return c.Core().ServiceAccounts(ns).Get(name) + return c.Core().ServiceAccounts(ns).Get(name, metav1.GetOptions{}) } var user *v1.ServiceAccount var err error err = wait.Poll(time.Second, 10*time.Second, func() (bool, error) { - user, err = c.Core().ServiceAccounts(ns).Get(name) + user, err = c.Core().ServiceAccounts(ns).Get(name, metav1.GetOptions{}) if errors.IsNotFound(err) { return false, nil } @@ -458,7 +459,7 @@ func getReferencedServiceAccountToken(c *clientset.Clientset, ns string, name st token := "" findToken := func() (bool, error) { - user, err := c.Core().ServiceAccounts(ns).Get(name) + user, err := c.Core().ServiceAccounts(ns).Get(name, metav1.GetOptions{}) if errors.IsNotFound(err) { return false, nil } @@ -467,7 +468,7 @@ func getReferencedServiceAccountToken(c *clientset.Clientset, ns string, name st } for _, ref := range user.Secrets { - secret, err := c.Core().Secrets(ns).Get(ref.Name) + secret, err := c.Core().Secrets(ns).Get(ref.Name, metav1.GetOptions{}) if errors.IsNotFound(err) { continue } diff --git a/test/integration/utils.go b/test/integration/utils.go index e8fdd3a49a..8e70e5404e 100644 --- a/test/integration/utils.go +++ b/test/integration/utils.go @@ -21,6 +21,7 @@ import ( "time" "k8s.io/kubernetes/pkg/api/errors" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" clientset "k8s.io/kubernetes/pkg/client/clientset_generated/release_1_5" coreclient "k8s.io/kubernetes/pkg/client/clientset_generated/release_1_5/typed/core/v1" "k8s.io/kubernetes/pkg/util/wait" @@ -48,7 +49,7 @@ var Code503 = map[int]bool{503: true} // WaitForPodToDisappear polls the API server if the pod has been deleted. func WaitForPodToDisappear(podClient coreclient.PodInterface, podName string, interval, timeout time.Duration) error { return wait.PollImmediate(interval, timeout, func() (bool, error) { - _, err := podClient.Get(podName) + _, err := podClient.Get(podName, metav1.GetOptions{}) if err == nil { return false, nil } else { diff --git a/test/integration/volume/persistent_volumes_test.go b/test/integration/volume/persistent_volumes_test.go index f776c0531c..7fe9b2c10b 100644 --- a/test/integration/volume/persistent_volumes_test.go +++ b/test/integration/volume/persistent_volumes_test.go @@ -275,7 +275,7 @@ func TestPersistentVolumeBindRace(t *testing.T) { waitForAnyPersistentVolumeClaimPhase(watchPVC, v1.ClaimBound) glog.V(2).Infof("TestPersistentVolumeBindRace pvc bound") - pv, err = testClient.PersistentVolumes().Get(pv.Name) + pv, err = testClient.PersistentVolumes().Get(pv.Name, metav1.GetOptions{}) if err != nil { t.Fatalf("Unexpected error getting pv: %v", err) } @@ -347,14 +347,14 @@ func TestPersistentVolumeClaimLabelSelector(t *testing.T) { waitForPersistentVolumeClaimPhase(testClient, pvc.Name, ns.Name, watchPVC, v1.ClaimBound) t.Log("claim bound") - pv, err := testClient.PersistentVolumes().Get("pv-false") + pv, err := testClient.PersistentVolumes().Get("pv-false", metav1.GetOptions{}) if err != nil { t.Fatalf("Unexpected error getting pv: %v", err) } if pv.Spec.ClaimRef != nil { t.Fatalf("False PV shouldn't be bound") } - pv, err = testClient.PersistentVolumes().Get("pv-true") + pv, err = testClient.PersistentVolumes().Get("pv-true", metav1.GetOptions{}) if err != nil { t.Fatalf("Unexpected error getting pv: %v", err) } @@ -446,14 +446,14 @@ func TestPersistentVolumeClaimLabelSelectorMatchExpressions(t *testing.T) { waitForPersistentVolumeClaimPhase(testClient, pvc.Name, ns.Name, watchPVC, v1.ClaimBound) t.Log("claim bound") - pv, err := testClient.PersistentVolumes().Get("pv-false") + pv, err := testClient.PersistentVolumes().Get("pv-false", metav1.GetOptions{}) if err != nil { t.Fatalf("Unexpected error getting pv: %v", err) } if pv.Spec.ClaimRef != nil { t.Fatalf("False PV shouldn't be bound") } - pv, err = testClient.PersistentVolumes().Get("pv-true") + pv, err = testClient.PersistentVolumes().Get("pv-true", metav1.GetOptions{}) if err != nil { t.Fatalf("Unexpected error getting pv: %v", err) } @@ -520,7 +520,7 @@ func TestPersistentVolumeMultiPVs(t *testing.T) { // only one PV is bound bound := 0 for i := 0; i < maxPVs; i++ { - pv, err := testClient.PersistentVolumes().Get(pvs[i].Name) + pv, err := testClient.PersistentVolumes().Get(pvs[i].Name, metav1.GetOptions{}) if err != nil { t.Fatalf("Unexpected error getting pv: %v", err) } @@ -614,7 +614,7 @@ func TestPersistentVolumeMultiPVsPVCs(t *testing.T) { // Modify PV i := rand.Intn(objCount) name := "pv-" + strconv.Itoa(i) - pv, err := testClient.PersistentVolumes().Get(name) + pv, err := testClient.PersistentVolumes().Get(name, metav1.GetOptions{}) if err != nil { // Silently ignore error, the PV may have be already deleted // or not exists yet. @@ -638,7 +638,7 @@ func TestPersistentVolumeMultiPVsPVCs(t *testing.T) { // Modify PVC i := rand.Intn(objCount) name := "pvc-" + strconv.Itoa(i) - pvc, err := testClient.PersistentVolumeClaims(v1.NamespaceDefault).Get(name) + pvc, err := testClient.PersistentVolumeClaims(v1.NamespaceDefault).Get(name, metav1.GetOptions{}) if err != nil { // Silently ignore error, the PVC may have be already // deleted or not exists yet. @@ -693,7 +693,7 @@ func TestPersistentVolumeMultiPVsPVCs(t *testing.T) { // check that everything is bound to something for i := 0; i < objCount; i++ { - pv, err := testClient.PersistentVolumes().Get(pvs[i].Name) + pv, err := testClient.PersistentVolumes().Get(pvs[i].Name, metav1.GetOptions{}) if err != nil { t.Fatalf("Unexpected error getting pv: %v", err) } @@ -702,7 +702,7 @@ func TestPersistentVolumeMultiPVsPVCs(t *testing.T) { } glog.V(2).Infof("PV %q is bound to PVC %q", pv.Name, pv.Spec.ClaimRef.Name) - pvc, err := testClient.PersistentVolumeClaims(ns.Name).Get(pvcs[i].Name) + pvc, err := testClient.PersistentVolumeClaims(ns.Name).Get(pvcs[i].Name, metav1.GetOptions{}) if err != nil { t.Fatalf("Unexpected error getting pvc: %v", err) } @@ -822,7 +822,7 @@ func TestPersistentVolumeControllerStartup(t *testing.T) { // check that everything is bound to something for i := 0; i < objCount; i++ { - pv, err := testClient.PersistentVolumes().Get(pvs[i].Name) + pv, err := testClient.PersistentVolumes().Get(pvs[i].Name, metav1.GetOptions{}) if err != nil { t.Fatalf("Unexpected error getting pv: %v", err) } @@ -831,7 +831,7 @@ func TestPersistentVolumeControllerStartup(t *testing.T) { } glog.V(2).Infof("PV %q is bound to PVC %q", pv.Name, pv.Spec.ClaimRef.Name) - pvc, err := testClient.PersistentVolumeClaims(ns.Name).Get(pvcs[i].Name) + pvc, err := testClient.PersistentVolumeClaims(ns.Name).Get(pvcs[i].Name, metav1.GetOptions{}) if err != nil { t.Fatalf("Unexpected error getting pvc: %v", err) } @@ -991,14 +991,14 @@ func TestPersistentVolumeMultiPVsDiffAccessModes(t *testing.T) { t.Log("claim bound") // only RWM PV is bound - pv, err := testClient.PersistentVolumes().Get("pv-rwo") + pv, err := testClient.PersistentVolumes().Get("pv-rwo", metav1.GetOptions{}) if err != nil { t.Fatalf("Unexpected error getting pv: %v", err) } if pv.Spec.ClaimRef != nil { t.Fatalf("ReadWriteOnce PV shouldn't be bound") } - pv, err = testClient.PersistentVolumes().Get("pv-rwm") + pv, err = testClient.PersistentVolumes().Get("pv-rwm", metav1.GetOptions{}) if err != nil { t.Fatalf("Unexpected error getting pv: %v", err) } @@ -1021,7 +1021,7 @@ func TestPersistentVolumeMultiPVsDiffAccessModes(t *testing.T) { func waitForPersistentVolumePhase(client *clientset.Clientset, pvName string, w watch.Interface, phase v1.PersistentVolumePhase) { // Check if the volume is already in requested phase - volume, err := client.Core().PersistentVolumes().Get(pvName) + volume, err := client.Core().PersistentVolumes().Get(pvName, metav1.GetOptions{}) if err == nil && volume.Status.Phase == phase { return } @@ -1042,7 +1042,7 @@ func waitForPersistentVolumePhase(client *clientset.Clientset, pvName string, w func waitForPersistentVolumeClaimPhase(client *clientset.Clientset, claimName, namespace string, w watch.Interface, phase v1.PersistentVolumeClaimPhase) { // Check if the claim is already in requested phase - claim, err := client.Core().PersistentVolumeClaims(namespace).Get(claimName) + claim, err := client.Core().PersistentVolumeClaims(namespace).Get(claimName, metav1.GetOptions{}) if err == nil && claim.Status.Phase == phase { return } diff --git a/test/utils/density_utils.go b/test/utils/density_utils.go index 234fee897e..c9523a279b 100644 --- a/test/utils/density_utils.go +++ b/test/utils/density_utils.go @@ -24,6 +24,7 @@ import ( "k8s.io/kubernetes/pkg/api" apierrs "k8s.io/kubernetes/pkg/api/errors" "k8s.io/kubernetes/pkg/api/v1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" clientset "k8s.io/kubernetes/pkg/client/clientset_generated/release_1_5" "github.com/golang/glog" @@ -61,7 +62,7 @@ func RemoveLabelOffNode(c clientset.Interface, nodeName string, labelKeys []stri var node *v1.Node var err error for attempt := 0; attempt < retries; attempt++ { - node, err = c.Core().Nodes().Get(nodeName) + node, err = c.Core().Nodes().Get(nodeName, metav1.GetOptions{}) if err != nil { return err } @@ -92,7 +93,7 @@ func RemoveLabelOffNode(c clientset.Interface, nodeName string, labelKeys []stri // VerifyLabelsRemoved checks if Node for given nodeName does not have any of labels from labelKeys. // Return non-nil error if it does. func VerifyLabelsRemoved(c clientset.Interface, nodeName string, labelKeys []string) error { - node, err := c.Core().Nodes().Get(nodeName) + node, err := c.Core().Nodes().Get(nodeName, metav1.GetOptions{}) if err != nil { return err } diff --git a/test/utils/runners.go b/test/utils/runners.go index 67c5469edf..19aa138a85 100644 --- a/test/utils/runners.go +++ b/test/utils/runners.go @@ -772,7 +772,7 @@ func DoPrepareNode(client clientset.Interface, node *v1.Node, strategy PrepareNo func DoCleanupNode(client clientset.Interface, nodeName string, strategy PrepareNodeStrategy) error { for attempt := 0; attempt < retries; attempt++ { - node, err := client.Core().Nodes().Get(nodeName) + node, err := client.Core().Nodes().Get(nodeName, metav1.GetOptions{}) if err != nil { return fmt.Errorf("Skipping cleanup of Node: failed to get Node %v: %v", nodeName, err) }