From a9ec31209eeea4a0e710aa53dbd5cb2c23834391 Mon Sep 17 00:00:00 2001 From: Wojciech Tyczynski Date: Wed, 7 Dec 2016 15:40:26 +0100 Subject: [PATCH] GetOptions - fix tests --- test/e2e/addon_update.go | 7 +- test/e2e/autoscaling_utils.go | 7 +- test/e2e/batch_v1_jobs.go | 7 +- test/e2e/cluster_logging_es.go | 3 +- test/e2e/cluster_logging_utils.go | 3 +- test/e2e/cluster_size_autoscaling.go | 3 +- test/e2e/common/container_probe.go | 11 ++- test/e2e/common/downwardapi_volume.go | 3 +- test/e2e/common/pods.go | 7 +- test/e2e/common/sysctl.go | 5 +- test/e2e/common/volumes.go | 2 +- test/e2e/cronjob.go | 9 +- test/e2e/daemon_set.go | 4 +- test/e2e/deployment.go | 38 ++++---- test/e2e/dns.go | 4 +- test/e2e/dns_autoscaling.go | 3 +- test/e2e/events.go | 3 +- test/e2e/examples.go | 3 +- test/e2e/federated-ingress.go | 13 +-- test/e2e/federated-namespace.go | 7 +- test/e2e/federated-secret.go | 13 +-- test/e2e/federated-service.go | 3 +- test/e2e/federation-apiserver.go | 3 +- test/e2e/federation-daemonset.go | 13 +-- test/e2e/federation-deployment.go | 10 +- test/e2e/federation-replicaset.go | 10 +- test/e2e/federation-util-14.go | 5 +- test/e2e/federation-util.go | 13 +-- test/e2e/framework/exec_util.go | 5 +- test/e2e/framework/framework.go | 3 +- test/e2e/framework/networking_utils.go | 8 +- test/e2e/framework/pods.go | 5 +- test/e2e/framework/util.go | 92 ++++++++++--------- test/e2e/garbage_collector.go | 4 +- test/e2e/ingress_utils.go | 9 +- test/e2e/job.go | 7 +- test/e2e/kibana_logging.go | 3 +- test/e2e/kubectl.go | 26 +++--- test/e2e/kubelet.go | 3 +- test/e2e/limit_range.go | 7 +- test/e2e/mesos.go | 2 +- test/e2e/namespace.go | 9 +- test/e2e/network_partition.go | 11 ++- test/e2e/node_problem_detector.go | 3 +- test/e2e/nodeoutofdisk.go | 7 +- test/e2e/pd.go | 2 +- test/e2e/persistent_volumes.go | 24 ++--- test/e2e/petset.go | 10 +- test/e2e/pods.go | 3 +- test/e2e/pre_stop.go | 3 +- test/e2e/rc.go | 7 +- test/e2e/reboot.go | 3 +- test/e2e/replica_set.go | 6 +- test/e2e/resize_nodes.go | 2 +- test/e2e/resource_quota.go | 3 +- test/e2e/restart.go | 5 +- test/e2e/scheduler_predicates.go | 21 +++-- test/e2e/security_context.go | 3 +- test/e2e/service.go | 35 +++---- test/e2e/service_accounts.go | 19 ++-- test/e2e/ubernetes_lite.go | 2 +- test/e2e/volume_provisioning.go | 4 +- test/e2e/volumes.go | 2 +- test/e2e_node/apparmor_test.go | 3 +- test/e2e_node/benchmark_util.go | 2 +- test/e2e_node/container.go | 9 +- test/e2e_node/disk_eviction_test.go | 7 +- test/e2e_node/garbage_collector_test.go | 3 +- test/e2e_node/image_id_test.go | 3 +- test/e2e_node/kubelet_test.go | 6 +- test/e2e_node/log_path_test.go | 3 +- test/e2e_node/memory_eviction_test.go | 7 +- test/e2e_node/mirror_pod_test.go | 15 +-- test/e2e_node/util.go | 3 +- test/integration/client/client_test.go | 12 +-- test/integration/evictions/evictions_test.go | 4 +- test/integration/framework/master_utils.go | 3 +- .../garbage_collector_test.go | 4 +- test/integration/master/master_test.go | 7 +- .../integration/replicaset/replicaset_test.go | 8 +- .../replicationcontroller_test.go | 8 +- test/integration/scheduler/extender_test.go | 2 +- test/integration/scheduler/scheduler_test.go | 4 +- .../serviceaccount/service_account_test.go | 11 ++- test/integration/utils.go | 3 +- .../volume/persistent_volumes_test.go | 32 +++---- test/utils/density_utils.go | 5 +- test/utils/runners.go | 2 +- 88 files changed, 398 insertions(+), 338 deletions(-) diff --git a/test/e2e/addon_update.go b/test/e2e/addon_update.go index 0ebe73af989..6f04e893d77 100644 --- a/test/e2e/addon_update.go +++ b/test/e2e/addon_update.go @@ -26,6 +26,7 @@ import ( "golang.org/x/crypto/ssh" "k8s.io/kubernetes/pkg/api/v1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" clientset "k8s.io/kubernetes/pkg/client/clientset_generated/release_1_5" "k8s.io/kubernetes/test/e2e/framework" @@ -329,11 +330,11 @@ var _ = framework.KubeDescribe("Addon update", func() { waitForReplicationControllerInAddonTest(f.ClientSet, addonNsName, "addon-test-v2", false) By("verify invalid API addons weren't created") - _, err = f.ClientSet.Core().ReplicationControllers(addonNsName).Get("invalid-addon-test-v1") + _, err = f.ClientSet.Core().ReplicationControllers(addonNsName).Get("invalid-addon-test-v1", metav1.GetOptions{}) Expect(err).To(HaveOccurred()) - _, err = f.ClientSet.Core().Services(addonNsName).Get("ivalid-addon-test") + _, err = f.ClientSet.Core().Services(addonNsName).Get("ivalid-addon-test", metav1.GetOptions{}) Expect(err).To(HaveOccurred()) - _, err = f.ClientSet.Core().Services(defaultNsName).Get("ivalid-addon-test-v2") + _, err = f.ClientSet.Core().Services(defaultNsName).Get("ivalid-addon-test-v2", metav1.GetOptions{}) Expect(err).To(HaveOccurred()) // invalid addons will be deleted by the deferred function diff --git a/test/e2e/autoscaling_utils.go b/test/e2e/autoscaling_utils.go index 41ea143ec25..1a304082359 100644 --- a/test/e2e/autoscaling_utils.go +++ b/test/e2e/autoscaling_utils.go @@ -22,6 +22,7 @@ import ( "time" "k8s.io/kubernetes/pkg/api/v1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" "k8s.io/kubernetes/pkg/client/clientset_generated/internalclientset" clientset "k8s.io/kubernetes/pkg/client/clientset_generated/release_1_5" "k8s.io/kubernetes/pkg/util/intstr" @@ -247,21 +248,21 @@ func (rc *ResourceConsumer) sendConsumeCustomMetric(delta int) { func (rc *ResourceConsumer) GetReplicas() int { switch rc.kind { case kindRC: - replicationController, err := rc.framework.ClientSet.Core().ReplicationControllers(rc.framework.Namespace.Name).Get(rc.name) + replicationController, err := rc.framework.ClientSet.Core().ReplicationControllers(rc.framework.Namespace.Name).Get(rc.name, metav1.GetOptions{}) framework.ExpectNoError(err) if replicationController == nil { framework.Failf(rcIsNil) } return int(replicationController.Status.Replicas) case kindDeployment: - deployment, err := rc.framework.ClientSet.Extensions().Deployments(rc.framework.Namespace.Name).Get(rc.name) + deployment, err := rc.framework.ClientSet.Extensions().Deployments(rc.framework.Namespace.Name).Get(rc.name, metav1.GetOptions{}) framework.ExpectNoError(err) if deployment == nil { framework.Failf(deploymentIsNil) } return int(deployment.Status.Replicas) case kindReplicaSet: - rs, err := rc.framework.ClientSet.Extensions().ReplicaSets(rc.framework.Namespace.Name).Get(rc.name) + rs, err := rc.framework.ClientSet.Extensions().ReplicaSets(rc.framework.Namespace.Name).Get(rc.name, metav1.GetOptions{}) framework.ExpectNoError(err) if rs == nil { framework.Failf(rsIsNil) diff --git a/test/e2e/batch_v1_jobs.go b/test/e2e/batch_v1_jobs.go index 82b9cde6cb9..a350894d457 100644 --- a/test/e2e/batch_v1_jobs.go +++ b/test/e2e/batch_v1_jobs.go @@ -28,6 +28,7 @@ import ( "k8s.io/kubernetes/pkg/api/v1" batchinternal "k8s.io/kubernetes/pkg/apis/batch" batch "k8s.io/kubernetes/pkg/apis/batch/v1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" clientset "k8s.io/kubernetes/pkg/client/clientset_generated/release_1_5" "k8s.io/kubernetes/pkg/kubectl" "k8s.io/kubernetes/pkg/labels" @@ -279,7 +280,7 @@ func newTestV1Job(behavior, name string, rPol v1.RestartPolicy, parallelism, com } func getV1Job(c clientset.Interface, ns, name string) (*batch.Job, error) { - return c.Batch().Jobs(ns).Get(name) + return c.Batch().Jobs(ns).Get(name, metav1.GetOptions{}) } func createV1Job(c clientset.Interface, ns string, job *batch.Job) (*batch.Job, error) { @@ -316,7 +317,7 @@ func waitForAllPodsRunningV1(c clientset.Interface, ns, jobName string, parallel // Wait for job to reach completions. func waitForV1JobFinish(c clientset.Interface, ns, jobName string, completions int32) error { return wait.Poll(framework.Poll, v1JobTimeout, func() (bool, error) { - curr, err := c.Batch().Jobs(ns).Get(jobName) + curr, err := c.Batch().Jobs(ns).Get(jobName, metav1.GetOptions{}) if err != nil { return false, err } @@ -327,7 +328,7 @@ func waitForV1JobFinish(c clientset.Interface, ns, jobName string, completions i // Wait for job fail. func waitForV1JobFail(c clientset.Interface, ns, jobName string, timeout time.Duration) error { return wait.Poll(framework.Poll, timeout, func() (bool, error) { - curr, err := c.Batch().Jobs(ns).Get(jobName) + curr, err := c.Batch().Jobs(ns).Get(jobName, metav1.GetOptions{}) if err != nil { return false, err } diff --git a/test/e2e/cluster_logging_es.go b/test/e2e/cluster_logging_es.go index e37656e64fd..3e76ba53427 100644 --- a/test/e2e/cluster_logging_es.go +++ b/test/e2e/cluster_logging_es.go @@ -25,6 +25,7 @@ import ( "k8s.io/kubernetes/pkg/api" "k8s.io/kubernetes/pkg/api/v1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" "k8s.io/kubernetes/pkg/labels" "k8s.io/kubernetes/test/e2e/framework" @@ -92,7 +93,7 @@ func checkElasticsearchReadiness(f *framework.Framework) error { // being run as the first e2e test just after the e2e cluster has been created. var err error for start := time.Now(); time.Since(start) < graceTime; time.Sleep(5 * time.Second) { - if _, err = s.Get("elasticsearch-logging"); err == nil { + if _, err = s.Get("elasticsearch-logging", metav1.GetOptions{}); err == nil { break } framework.Logf("Attempt to check for the existence of the Elasticsearch service failed after %v", time.Since(start)) diff --git a/test/e2e/cluster_logging_utils.go b/test/e2e/cluster_logging_utils.go index 950b789bf45..b28b46c820c 100644 --- a/test/e2e/cluster_logging_utils.go +++ b/test/e2e/cluster_logging_utils.go @@ -23,6 +23,7 @@ import ( "k8s.io/kubernetes/pkg/api" "k8s.io/kubernetes/pkg/api/v1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" "k8s.io/kubernetes/pkg/labels" "k8s.io/kubernetes/test/e2e/framework" ) @@ -62,7 +63,7 @@ func createSynthLogger(f *framework.Framework, linesCount int) { } func reportLogsFromFluentdPod(f *framework.Framework) error { - synthLoggerPod, err := f.PodClient().Get(synthLoggerPodName) + synthLoggerPod, err := f.PodClient().Get(synthLoggerPodName, metav1.GetOptions{}) if err != nil { return fmt.Errorf("Failed to get synth logger pod due to %v", err) } diff --git a/test/e2e/cluster_size_autoscaling.go b/test/e2e/cluster_size_autoscaling.go index 68dd136d032..8c9801255d6 100644 --- a/test/e2e/cluster_size_autoscaling.go +++ b/test/e2e/cluster_size_autoscaling.go @@ -27,6 +27,7 @@ import ( "time" "k8s.io/kubernetes/pkg/api/v1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" clientset "k8s.io/kubernetes/pkg/client/clientset_generated/release_1_5" "k8s.io/kubernetes/pkg/fields" "k8s.io/kubernetes/pkg/util/sets" @@ -240,7 +241,7 @@ var _ = framework.KubeDescribe("Cluster size autoscaling [Slow]", func() { By(fmt.Sprintf("New nodes: %v\n", newNodesSet)) registeredNodes := sets.NewString() for nodeName := range newNodesSet { - node, err := f.ClientSet.Core().Nodes().Get(nodeName) + node, err := f.ClientSet.Core().Nodes().Get(nodeName, metav1.GetOptions{}) if err == nil && node != nil { registeredNodes.Insert(nodeName) } else { diff --git a/test/e2e/common/container_probe.go b/test/e2e/common/container_probe.go index 4a693f6bde6..2d0dbe35e6e 100644 --- a/test/e2e/common/container_probe.go +++ b/test/e2e/common/container_probe.go @@ -21,6 +21,7 @@ import ( "time" "k8s.io/kubernetes/pkg/api/v1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" "k8s.io/kubernetes/pkg/util/intstr" "k8s.io/kubernetes/pkg/util/uuid" "k8s.io/kubernetes/test/e2e/framework" @@ -50,7 +51,7 @@ var _ = framework.KubeDescribe("Probing container", func() { p := podClient.Create(makePodSpec(probe.withInitialDelay().build(), nil)) f.WaitForPodReady(p.Name) - p, err := podClient.Get(p.Name) + p, err := podClient.Get(p.Name, metav1.GetOptions{}) framework.ExpectNoError(err) isReady, err := testutils.PodRunningReady(p) framework.ExpectNoError(err) @@ -76,14 +77,14 @@ var _ = framework.KubeDescribe("Probing container", func() { It("with readiness probe that fails should never be ready and never restart [Conformance]", func() { p := podClient.Create(makePodSpec(probe.withFailing().build(), nil)) Consistently(func() (bool, error) { - p, err := podClient.Get(p.Name) + p, err := podClient.Get(p.Name, metav1.GetOptions{}) if err != nil { return false, err } return v1.IsPodReady(p), nil }, 1*time.Minute, 1*time.Second).ShouldNot(BeTrue(), "pod should not be ready") - p, err := podClient.Get(p.Name) + p, err := podClient.Get(p.Name, metav1.GetOptions{}) framework.ExpectNoError(err) isReady, err := testutils.PodRunningReady(p) @@ -366,7 +367,7 @@ func runLivenessTest(f *framework.Framework, pod *v1.Pod, expectNumRestarts int, // Check the pod's current state and verify that restartCount is present. By("checking the pod's current state and verifying that restartCount is present") - pod, err := podClient.Get(pod.Name) + pod, err := podClient.Get(pod.Name, metav1.GetOptions{}) framework.ExpectNoError(err, fmt.Sprintf("getting pod %s in namespace %s", pod.Name, ns)) initialRestartCount := v1.GetExistingContainerStatus(pod.Status.ContainerStatuses, containerName).RestartCount framework.Logf("Initial restart count of pod %s is %d", pod.Name, initialRestartCount) @@ -376,7 +377,7 @@ func runLivenessTest(f *framework.Framework, pod *v1.Pod, expectNumRestarts int, lastRestartCount := initialRestartCount observedRestarts := int32(0) for start := time.Now(); time.Now().Before(deadline); time.Sleep(2 * time.Second) { - pod, err = podClient.Get(pod.Name) + pod, err = podClient.Get(pod.Name, metav1.GetOptions{}) framework.ExpectNoError(err, fmt.Sprintf("getting pod %s", pod.Name)) restartCount := v1.GetExistingContainerStatus(pod.Status.ContainerStatuses, containerName).RestartCount if restartCount != lastRestartCount { diff --git a/test/e2e/common/downwardapi_volume.go b/test/e2e/common/downwardapi_volume.go index f6c2f57138e..8e82aebb292 100644 --- a/test/e2e/common/downwardapi_volume.go +++ b/test/e2e/common/downwardapi_volume.go @@ -22,6 +22,7 @@ import ( "k8s.io/kubernetes/pkg/api/resource" "k8s.io/kubernetes/pkg/api/v1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" "k8s.io/kubernetes/pkg/util/uuid" "k8s.io/kubernetes/test/e2e/framework" @@ -133,7 +134,7 @@ var _ = framework.KubeDescribe("Downward API volume", func() { By("Creating the pod") podClient.CreateSync(pod) - pod, err := podClient.Get(pod.Name) + pod, err := podClient.Get(pod.Name, metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred(), "Failed to get pod %q", pod.Name) Eventually(func() (string, error) { diff --git a/test/e2e/common/pods.go b/test/e2e/common/pods.go index bc49a7af816..c663c2bbb1a 100644 --- a/test/e2e/common/pods.go +++ b/test/e2e/common/pods.go @@ -27,6 +27,7 @@ import ( "golang.org/x/net/websocket" "k8s.io/kubernetes/pkg/api/v1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" "k8s.io/kubernetes/pkg/kubelet" "k8s.io/kubernetes/pkg/labels" "k8s.io/kubernetes/pkg/util/intstr" @@ -54,7 +55,7 @@ func testHostIP(podClient *framework.PodClient, pod *v1.Pod) { hostIPTimeout := 2 * time.Minute t := time.Now() for { - p, err := podClient.Get(pod.Name) + p, err := podClient.Get(pod.Name, metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred(), "Failed to get pod %q", pod.Name) if p.Status.HostIP != "" { framework.Logf("Pod %s has hostIP: %s", p.Name, p.Status.HostIP) @@ -100,7 +101,7 @@ func getRestartDelay(podClient *framework.PodClient, podName string, containerNa beginTime := time.Now() for time.Since(beginTime) < (2 * maxBackOffTolerance) { // may just miss the 1st MaxContainerBackOff delay time.Sleep(time.Second) - pod, err := podClient.Get(podName) + pod, err := podClient.Get(podName, metav1.GetOptions{}) framework.ExpectNoError(err, fmt.Sprintf("getting pod %s", podName)) status, ok := v1.GetContainerStatus(pod.Status.ContainerStatuses, containerName) if !ok { @@ -201,7 +202,7 @@ var _ = framework.KubeDescribe("Pods", func() { // may be carried out immediately rather than gracefully. framework.ExpectNoError(f.WaitForPodRunning(pod.Name)) // save the running pod - pod, err = podClient.Get(pod.Name) + pod, err = podClient.Get(pod.Name, metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred(), "failed to GET scheduled pod") framework.Logf("running pod: %#v", pod) diff --git a/test/e2e/common/sysctl.go b/test/e2e/common/sysctl.go index 2dbf5228e74..1d149ddcc40 100644 --- a/test/e2e/common/sysctl.go +++ b/test/e2e/common/sysctl.go @@ -20,6 +20,7 @@ import ( "fmt" "k8s.io/kubernetes/pkg/api/v1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" "k8s.io/kubernetes/pkg/kubelet/events" "k8s.io/kubernetes/pkg/kubelet/sysctl" "k8s.io/kubernetes/pkg/util/uuid" @@ -107,7 +108,7 @@ var _ = framework.KubeDescribe("Sysctls", func() { By("Waiting for pod completion") err = f.WaitForPodNoLongerRunning(pod.Name) Expect(err).NotTo(HaveOccurred()) - pod, err = podClient.Get(pod.Name) + pod, err = podClient.Get(pod.Name, metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred()) By("Checking that the pod succeeded") @@ -148,7 +149,7 @@ var _ = framework.KubeDescribe("Sysctls", func() { By("Waiting for pod completion") err = f.WaitForPodNoLongerRunning(pod.Name) Expect(err).NotTo(HaveOccurred()) - pod, err = podClient.Get(pod.Name) + pod, err = podClient.Get(pod.Name, metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred()) By("Checking that the pod succeeded") diff --git a/test/e2e/common/volumes.go b/test/e2e/common/volumes.go index 94156b696e6..2da8ff4f2b7 100644 --- a/test/e2e/common/volumes.go +++ b/test/e2e/common/volumes.go @@ -148,7 +148,7 @@ func startVolumeServer(f *framework.Framework, config VolumeTestConfig) *v1.Pod serverPod = podClient.CreateSync(serverPod) By("locating the server pod") - pod, err := podClient.Get(serverPod.Name) + pod, err := podClient.Get(serverPod.Name, metav1.GetOptions{}) framework.ExpectNoError(err, "Cannot locate the server pod %v: %v", serverPod.Name, err) By("sleeping a bit to give the server time to start") diff --git a/test/e2e/cronjob.go b/test/e2e/cronjob.go index 3200a968b92..2b23c023780 100644 --- a/test/e2e/cronjob.go +++ b/test/e2e/cronjob.go @@ -26,6 +26,7 @@ import ( "k8s.io/kubernetes/pkg/api/v1" batchv1 "k8s.io/kubernetes/pkg/apis/batch/v1" batch "k8s.io/kubernetes/pkg/apis/batch/v2alpha1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" clientset "k8s.io/kubernetes/pkg/client/clientset_generated/release_1_5" "k8s.io/kubernetes/pkg/controller/job" "k8s.io/kubernetes/pkg/runtime/schema" @@ -235,7 +236,7 @@ func createCronJob(c clientset.Interface, ns string, cronJob *batch.CronJob) (*b } func getCronJob(c clientset.Interface, ns, name string) (*batch.CronJob, error) { - return c.BatchV2alpha1().CronJobs(ns).Get(name) + return c.BatchV2alpha1().CronJobs(ns).Get(name, metav1.GetOptions{}) } func deleteCronJob(c clientset.Interface, ns, name string) error { @@ -245,7 +246,7 @@ func deleteCronJob(c clientset.Interface, ns, name string) error { // Wait for at least given amount of active jobs. func waitForActiveJobs(c clientset.Interface, ns, cronJobName string, active int) error { return wait.Poll(framework.Poll, cronJobTimeout, func() (bool, error) { - curr, err := c.BatchV2alpha1().CronJobs(ns).Get(cronJobName) + curr, err := c.BatchV2alpha1().CronJobs(ns).Get(cronJobName, metav1.GetOptions{}) if err != nil { return false, err } @@ -256,7 +257,7 @@ func waitForActiveJobs(c clientset.Interface, ns, cronJobName string, active int // Wait for no jobs to appear. func waitForNoJobs(c clientset.Interface, ns, jobName string) error { return wait.Poll(framework.Poll, cronJobTimeout, func() (bool, error) { - curr, err := c.BatchV2alpha1().CronJobs(ns).Get(jobName) + curr, err := c.BatchV2alpha1().CronJobs(ns).Get(jobName, metav1.GetOptions{}) if err != nil { return false, err } @@ -312,7 +313,7 @@ func waitForAnyFinishedJob(c clientset.Interface, ns string) error { // checkNoUnexpectedEvents checks unexpected events didn't happen. // Currently only "UnexpectedJob" is checked. func checkNoUnexpectedEvents(c clientset.Interface, ns, cronJobName string) error { - sj, err := c.BatchV2alpha1().CronJobs(ns).Get(cronJobName) + sj, err := c.BatchV2alpha1().CronJobs(ns).Get(cronJobName, metav1.GetOptions{}) if err != nil { return fmt.Errorf("error in getting cronjob %s/%s: %v", ns, cronJobName, err) } diff --git a/test/e2e/daemon_set.go b/test/e2e/daemon_set.go index b5cda9b3573..dcfd3485dfe 100644 --- a/test/e2e/daemon_set.go +++ b/test/e2e/daemon_set.go @@ -303,7 +303,7 @@ func setDaemonSetNodeLabels(c clientset.Interface, nodeName string, labels map[s var newNode *v1.Node var newLabels map[string]string err := wait.Poll(dsRetryPeriod, dsRetryTimeout, func() (bool, error) { - node, err := nodeClient.Get(nodeName) + node, err := nodeClient.Get(nodeName, metav1.GetOptions{}) if err != nil { return false, err } @@ -385,7 +385,7 @@ func checkRunningOnNoNodes(f *framework.Framework, selector map[string]string) f } func checkDaemonStatus(f *framework.Framework, dsName string) error { - ds, err := f.ClientSet.Extensions().DaemonSets(f.Namespace.Name).Get(dsName) + ds, err := f.ClientSet.Extensions().DaemonSets(f.Namespace.Name).Get(dsName, metav1.GetOptions{}) if err != nil { return fmt.Errorf("Could not get daemon set from v1.") } diff --git a/test/e2e/deployment.go b/test/e2e/deployment.go index f8901e0da78..ca002191fc3 100644 --- a/test/e2e/deployment.go +++ b/test/e2e/deployment.go @@ -150,7 +150,7 @@ func newDeploymentRollback(name string, annotations map[string]string, revision // checkDeploymentRevision checks if the input deployment's and its new replica set's revision and images are as expected. func checkDeploymentRevision(c clientset.Interface, ns, deploymentName, revision, imageName, image string) (*extensions.Deployment, *extensions.ReplicaSet) { - deployment, err := c.Extensions().Deployments(ns).Get(deploymentName) + deployment, err := c.Extensions().Deployments(ns).Get(deploymentName, metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred()) // Check revision of the new replica set of this deployment newRS, err := deploymentutil.GetNewReplicaSet(deployment, c) @@ -180,7 +180,7 @@ func stopDeployment(c clientset.Interface, internalClient internalclientset.Inte } func stopDeploymentMaybeOverlap(c clientset.Interface, internalClient internalclientset.Interface, ns, deploymentName, overlapWith string) { - deployment, err := c.Extensions().Deployments(ns).Get(deploymentName) + deployment, err := c.Extensions().Deployments(ns).Get(deploymentName, metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred()) framework.Logf("Deleting deployment %s", deploymentName) @@ -191,7 +191,7 @@ func stopDeploymentMaybeOverlap(c clientset.Interface, internalClient internalcl Expect(err).NotTo(HaveOccurred()) framework.Logf("Ensuring deployment %s was deleted", deploymentName) - _, err = c.Extensions().Deployments(ns).Get(deployment.Name) + _, err = c.Extensions().Deployments(ns).Get(deployment.Name, metav1.GetOptions{}) Expect(err).To(HaveOccurred()) Expect(errors.IsNotFound(err)).To(BeTrue()) framework.Logf("Ensuring deployment %s's RSes were deleted", deploymentName) @@ -259,7 +259,7 @@ func testNewDeployment(f *framework.Framework) { err = framework.WaitForDeploymentStatus(c, deploy) Expect(err).NotTo(HaveOccurred()) - deployment, err := c.Extensions().Deployments(ns).Get(deploymentName) + deployment, err := c.Extensions().Deployments(ns).Get(deploymentName, metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred()) newRS, err := deploymentutil.GetNewReplicaSet(deployment, c) Expect(err).NotTo(HaveOccurred()) @@ -291,7 +291,7 @@ func testDeleteDeployment(f *framework.Framework) { err = framework.WaitForDeploymentStatus(c, deploy) Expect(err).NotTo(HaveOccurred()) - deployment, err := c.Extensions().Deployments(ns).Get(deploymentName) + deployment, err := c.Extensions().Deployments(ns).Get(deploymentName, metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred()) newRS, err := deploymentutil.GetNewReplicaSet(deployment, c) Expect(err).NotTo(HaveOccurred()) @@ -337,7 +337,7 @@ func testRollingUpdateDeployment(f *framework.Framework) { Expect(err).NotTo(HaveOccurred()) // There should be 1 old RS (nginx-controller, which is adopted) - deployment, err := c.Extensions().Deployments(ns).Get(deploymentName) + deployment, err := c.Extensions().Deployments(ns).Get(deploymentName, metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred()) _, allOldRSs, err := deploymentutil.GetOldReplicaSets(deployment, c) Expect(err).NotTo(HaveOccurred()) @@ -388,7 +388,7 @@ func testRollingUpdateDeploymentEvents(f *framework.Framework) { err = framework.WaitForDeploymentStatus(c, deploy) Expect(err).NotTo(HaveOccurred()) // Verify that the pods were scaled up and down as expected. We use events to verify that. - deployment, err := c.Extensions().Deployments(ns).Get(deploymentName) + deployment, err := c.Extensions().Deployments(ns).Get(deploymentName, metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred()) framework.WaitForEvents(c, ns, deployment, 2) events, err := c.Core().Events(ns).Search(deployment) @@ -441,7 +441,7 @@ func testRecreateDeployment(f *framework.Framework) { Expect(err).NotTo(HaveOccurred()) // Verify that the pods were scaled up and down as expected. We use events to verify that. - deployment, err := c.Extensions().Deployments(ns).Get(deploymentName) + deployment, err := c.Extensions().Deployments(ns).Get(deploymentName, metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred()) framework.WaitForEvents(c, ns, deployment, 2) events, err := c.Core().Events(ns).Search(deployment) @@ -572,7 +572,7 @@ func testRolloverDeployment(f *framework.Framework) { Expect(err).NotTo(HaveOccurred()) // Verify that the pods were scaled up and down as expected. - deployment, err := c.Extensions().Deployments(ns).Get(deploymentName) + deployment, err := c.Extensions().Deployments(ns).Get(deploymentName, metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred()) framework.Logf("Make sure deployment %q performs scaling operations", deploymentName) // Make sure the deployment starts to scale up and down replica sets by checking if its updated replicas >= 1 @@ -619,7 +619,7 @@ func testPausedDeployment(f *framework.Framework) { _, err := c.Extensions().Deployments(ns).Create(d) Expect(err).NotTo(HaveOccurred()) // Check that deployment is created fine. - deployment, err := c.Extensions().Deployments(ns).Get(deploymentName) + deployment, err := c.Extensions().Deployments(ns).Get(deploymentName, metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred()) // Verify that there is no latest state realized for the new deployment. @@ -846,7 +846,7 @@ func testRollbackDeploymentRSNoRevision(f *framework.Framework) { Expect(err).NotTo(HaveOccurred()) // Check that the replica set we created still doesn't contain revision information - rs, err = c.Extensions().ReplicaSets(ns).Get(rsName) + rs, err = c.Extensions().ReplicaSets(ns).Get(rsName, metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred()) Expect(rs.Annotations[deploymentutil.RevisionAnnotation]).Should(Equal("")) @@ -976,7 +976,7 @@ func testDeploymentLabelAdopted(f *framework.Framework) { Expect(err).NotTo(HaveOccurred()) // There should be no old RSs (overlapping RS) - deployment, err := c.Extensions().Deployments(ns).Get(deploymentName) + deployment, err := c.Extensions().Deployments(ns).Get(deploymentName, metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred()) oldRSs, allOldRSs, newRS, err := deploymentutil.GetAllReplicaSets(deployment, c) Expect(err).NotTo(HaveOccurred()) @@ -1011,7 +1011,7 @@ func testScalePausedDeployment(f *framework.Framework) { Expect(err).NotTo(HaveOccurred()) // Check that deployment is created fine. - deployment, err := c.Extensions().Deployments(ns).Get(deploymentName) + deployment, err := c.Extensions().Deployments(ns).Get(deploymentName, metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred()) err = framework.WaitForObservedDeployment(c, ns, deploymentName, deployment.Generation) @@ -1091,7 +1091,7 @@ func testScaledRolloutDeployment(f *framework.Framework) { err = framework.WaitForObservedDeployment(c, ns, deploymentName, deployment.Generation) Expect(err).NotTo(HaveOccurred()) - deployment, err = c.Extensions().Deployments(ns).Get(deploymentName) + deployment, err = c.Extensions().Deployments(ns).Get(deploymentName, metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred()) if deployment.Status.AvailableReplicas < deploymentutil.MinAvailable(deployment) { @@ -1102,7 +1102,7 @@ func testScaledRolloutDeployment(f *framework.Framework) { second, err := deploymentutil.GetNewReplicaSet(deployment, c) Expect(err).NotTo(HaveOccurred()) - first, err = c.Extensions().ReplicaSets(first.Namespace).Get(first.Name) + first, err = c.Extensions().ReplicaSets(first.Namespace).Get(first.Name, metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred()) firstCond := replicaSetHasDesiredReplicas(c.Extensions(), first) @@ -1151,7 +1151,7 @@ func testScaledRolloutDeployment(f *framework.Framework) { err = framework.WaitForObservedDeployment(c, ns, deploymentName, deployment.Generation) Expect(err).NotTo(HaveOccurred()) - deployment, err = c.Extensions().Deployments(ns).Get(deploymentName) + deployment, err = c.Extensions().Deployments(ns).Get(deploymentName, metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred()) if deployment.Status.AvailableReplicas < deploymentutil.MinAvailable(deployment) { @@ -1159,7 +1159,7 @@ func testScaledRolloutDeployment(f *framework.Framework) { } By(fmt.Sprintf("Checking that the replica sets for %q are synced", deploymentName)) - oldRs, err := c.Extensions().ReplicaSets(rs.Namespace).Get(rs.Name) + oldRs, err := c.Extensions().ReplicaSets(rs.Namespace).Get(rs.Name, metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred()) newRs, err := deploymentutil.GetNewReplicaSet(deployment, c) @@ -1459,7 +1459,7 @@ func testIterativeDeployments(f *framework.Framework) { } // unpause the deployment if we end up pausing it - deployment, err = c.Extensions().Deployments(ns).Get(deployment.Name) + deployment, err = c.Extensions().Deployments(ns).Get(deployment.Name, metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred()) if deployment.Spec.Paused { deployment, err = framework.UpdateDeploymentWithRetries(c, ns, deployment.Name, func(update *extensions.Deployment) { @@ -1480,7 +1480,7 @@ func testIterativeDeployments(f *framework.Framework) { func replicaSetHasDesiredReplicas(rsClient extensionsclient.ReplicaSetsGetter, replicaSet *extensions.ReplicaSet) wait.ConditionFunc { desiredGeneration := replicaSet.Generation return func() (bool, error) { - rs, err := rsClient.ReplicaSets(replicaSet.Namespace).Get(replicaSet.Name) + rs, err := rsClient.ReplicaSets(replicaSet.Namespace).Get(replicaSet.Name, metav1.GetOptions{}) if err != nil { return false, err } diff --git a/test/e2e/dns.go b/test/e2e/dns.go index 50cf871707f..3727f7ba6ec 100644 --- a/test/e2e/dns.go +++ b/test/e2e/dns.go @@ -237,7 +237,7 @@ func validateDNSResults(f *framework.Framework, pod *v1.Pod, fileNames []string) framework.ExpectNoError(f.WaitForPodRunning(pod.Name)) By("retrieving the pod") - pod, err := podClient.Get(pod.Name) + pod, err := podClient.Get(pod.Name, metav1.GetOptions{}) if err != nil { framework.Failf("Failed to get pod %s: %v", pod.Name, err) } @@ -266,7 +266,7 @@ func validateTargetedProbeOutput(f *framework.Framework, pod *v1.Pod, fileNames framework.ExpectNoError(f.WaitForPodRunning(pod.Name)) By("retrieving the pod") - pod, err := podClient.Get(pod.Name) + pod, err := podClient.Get(pod.Name, metav1.GetOptions{}) if err != nil { framework.Failf("Failed to get pod %s: %v", pod.Name, err) } diff --git a/test/e2e/dns_autoscaling.go b/test/e2e/dns_autoscaling.go index 4f0832849df..cd7553e71c5 100644 --- a/test/e2e/dns_autoscaling.go +++ b/test/e2e/dns_autoscaling.go @@ -26,6 +26,7 @@ import ( "k8s.io/kubernetes/pkg/api" "k8s.io/kubernetes/pkg/api/resource" "k8s.io/kubernetes/pkg/api/v1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" clientset "k8s.io/kubernetes/pkg/client/clientset_generated/release_1_5" "k8s.io/kubernetes/pkg/labels" "k8s.io/kubernetes/pkg/util/wait" @@ -230,7 +231,7 @@ func getScheduableCores(nodes []v1.Node) int64 { } func fetchDNSScalingConfigMap(c clientset.Interface) (*v1.ConfigMap, error) { - cm, err := c.Core().ConfigMaps(api.NamespaceSystem).Get(DNSAutoscalerLabelName) + cm, err := c.Core().ConfigMaps(api.NamespaceSystem).Get(DNSAutoscalerLabelName, metav1.GetOptions{}) if err != nil { return nil, err } diff --git a/test/e2e/events.go b/test/e2e/events.go index 1049bc1fd53..c174b328088 100644 --- a/test/e2e/events.go +++ b/test/e2e/events.go @@ -22,6 +22,7 @@ import ( "time" "k8s.io/kubernetes/pkg/api/v1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" "k8s.io/kubernetes/pkg/fields" "k8s.io/kubernetes/pkg/labels" "k8s.io/kubernetes/pkg/util/uuid" @@ -79,7 +80,7 @@ var _ = framework.KubeDescribe("Events", func() { Expect(len(pods.Items)).To(Equal(1)) By("retrieving the pod") - podWithUid, err := podClient.Get(pod.Name) + podWithUid, err := podClient.Get(pod.Name, metav1.GetOptions{}) if err != nil { framework.Failf("Failed to get pod: %v", err) } diff --git a/test/e2e/examples.go b/test/e2e/examples.go index 149ae66ba3d..31d161cbf52 100644 --- a/test/e2e/examples.go +++ b/test/e2e/examples.go @@ -27,6 +27,7 @@ import ( "time" "k8s.io/kubernetes/pkg/api/v1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" clientset "k8s.io/kubernetes/pkg/client/clientset_generated/release_1_5" "k8s.io/kubernetes/pkg/labels" "k8s.io/kubernetes/pkg/util/wait" @@ -398,7 +399,7 @@ var _ = framework.KubeDescribe("[Feature:Example]", func() { err := framework.WaitForPodNameRunningInNamespace(c, podName, ns) Expect(err).NotTo(HaveOccurred()) for t := time.Now(); time.Since(t) < timeout; time.Sleep(framework.Poll) { - pod, err := c.Core().Pods(ns).Get(podName) + pod, err := c.Core().Pods(ns).Get(podName, metav1.GetOptions{}) framework.ExpectNoError(err, fmt.Sprintf("getting pod %s", podName)) stat := v1.GetExistingContainerStatus(pod.Status.ContainerStatuses, podName) framework.Logf("Pod: %s, restart count:%d", stat.Name, stat.RestartCount) diff --git a/test/e2e/federated-ingress.go b/test/e2e/federated-ingress.go index 7ee43d930d2..b584ef72b01 100644 --- a/test/e2e/federated-ingress.go +++ b/test/e2e/federated-ingress.go @@ -29,6 +29,7 @@ import ( "k8s.io/kubernetes/pkg/api/errors" "k8s.io/kubernetes/pkg/api/v1" "k8s.io/kubernetes/pkg/apis/extensions/v1beta1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" kubeclientset "k8s.io/kubernetes/pkg/client/clientset_generated/release_1_5" "k8s.io/kubernetes/pkg/util/intstr" "k8s.io/kubernetes/pkg/util/wait" @@ -227,7 +228,7 @@ func verifyCascadingDeletionForIngress(clientset *fedclientset.Clientset, cluste // ingress should be present in underlying clusters unless orphanDependents is false. shouldExist := orphanDependents == nil || *orphanDependents == true for clusterName, clusterClientset := range clusters { - _, err := clusterClientset.Extensions().Ingresses(nsName).Get(ingressName) + _, err := clusterClientset.Extensions().Ingresses(nsName).Get(ingressName, metav1.GetOptions{}) if shouldExist && errors.IsNotFound(err) { errMessages = append(errMessages, fmt.Sprintf("unexpected NotFound error for ingress %s in cluster %s, expected ingress to exist", ingressName, clusterName)) } else if !shouldExist && !errors.IsNotFound(err) { @@ -247,7 +248,7 @@ func waitForIngressOrFail(clientset *kubeclientset.Clientset, namespace string, By(fmt.Sprintf("Fetching a federated ingress shard of ingress %q in namespace %q from cluster", ingress.Name, namespace)) var clusterIngress *v1beta1.Ingress err := wait.PollImmediate(framework.Poll, timeout, func() (bool, error) { - clusterIngress, err := clientset.Ingresses(namespace).Get(ingress.Name) + clusterIngress, err := clientset.Ingresses(namespace).Get(ingress.Name, metav1.GetOptions{}) if (!present) && errors.IsNotFound(err) { // We want it gone, and it's gone. By(fmt.Sprintf("Success: shard of federated ingress %q in namespace %q in cluster is absent", ingress.Name, namespace)) return true, nil // Success @@ -293,7 +294,7 @@ func waitForIngressShardsUpdatedOrFail(namespace string, ingress *v1beta1.Ingres func waitForIngressUpdateOrFail(clientset *kubeclientset.Clientset, namespace string, ingress *v1beta1.Ingress, timeout time.Duration) { By(fmt.Sprintf("Fetching a federated ingress shard of ingress %q in namespace %q from cluster", ingress.Name, namespace)) err := wait.PollImmediate(framework.Poll, timeout, func() (bool, error) { - clusterIngress, err := clientset.Ingresses(namespace).Get(ingress.Name) + clusterIngress, err := clientset.Ingresses(namespace).Get(ingress.Name, metav1.GetOptions{}) if err == nil { // We want it present, and the Get succeeded, so we're all good. if equivalentIngress(*clusterIngress, *ingress) { By(fmt.Sprintf("Success: shard of federated ingress %q in namespace %q in cluster is updated", ingress.Name, namespace)) @@ -326,7 +327,7 @@ func deleteIngressOrFail(clientset *fedclientset.Clientset, namespace string, in framework.ExpectNoError(err, "Error deleting ingress %q from namespace %q", ingressName, namespace) // Wait for the ingress to be deleted. err = wait.Poll(framework.Poll, wait.ForeverTestTimeout, func() (bool, error) { - _, err := clientset.Extensions().Ingresses(namespace).Get(ingressName) + _, err := clientset.Extensions().Ingresses(namespace).Get(ingressName, metav1.GetOptions{}) if err != nil && errors.IsNotFound(err) { return true, nil } @@ -456,7 +457,7 @@ func waitForFederatedIngressAddress(c *fedclientset.Clientset, ns, ingName strin // waitForFederatedIngressExists waits for the Ingress object exists. func waitForFederatedIngressExists(c *fedclientset.Clientset, ns, ingName string, timeout time.Duration) error { err := wait.PollImmediate(10*time.Second, timeout, func() (bool, error) { - _, err := c.Extensions().Ingresses(ns).Get(ingName) + _, err := c.Extensions().Ingresses(ns).Get(ingName, metav1.GetOptions{}) if err != nil { framework.Logf("Waiting for Ingress %v, error %v", ingName, err) return false, nil @@ -468,7 +469,7 @@ func waitForFederatedIngressExists(c *fedclientset.Clientset, ns, ingName string // getFederatedIngressAddress returns the ips/hostnames associated with the Ingress. func getFederatedIngressAddress(client *fedclientset.Clientset, ns, name string) ([]string, error) { - ing, err := client.Extensions().Ingresses(ns).Get(name) + ing, err := client.Extensions().Ingresses(ns).Get(name, metav1.GetOptions{}) if err != nil { return nil, err } diff --git a/test/e2e/federated-namespace.go b/test/e2e/federated-namespace.go index 42818380643..40acee5c0fc 100644 --- a/test/e2e/federated-namespace.go +++ b/test/e2e/federated-namespace.go @@ -26,6 +26,7 @@ import ( "k8s.io/kubernetes/pkg/api/errors" "k8s.io/kubernetes/pkg/api/v1" api_v1 "k8s.io/kubernetes/pkg/api/v1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" "k8s.io/kubernetes/pkg/util/wait" "k8s.io/kubernetes/test/e2e/framework" @@ -132,7 +133,7 @@ var _ = framework.KubeDescribe("Federation namespace [Feature:Federation]", func f.FederationClientset_1_5.Core().Namespaces().Delete) By(fmt.Sprintf("Verify that event %s was deleted as well", event.Name)) - latestEvent, err := f.FederationClientset_1_5.Core().Events(nsName).Get(event.Name) + latestEvent, err := f.FederationClientset_1_5.Core().Events(nsName).Get(event.Name, metav1.GetOptions{}) if !errors.IsNotFound(err) { framework.Failf("Event %s should have been deleted. Found: %v", event.Name, latestEvent) } @@ -150,7 +151,7 @@ func verifyNsCascadingDeletion(nsClient clientset.NamespaceInterface, clusters m By(fmt.Sprintf("Waiting for namespace %s to be created in all underlying clusters", nsName)) err := wait.Poll(5*time.Second, 2*time.Minute, func() (bool, error) { for _, cluster := range clusters { - _, err := cluster.Core().Namespaces().Get(nsName) + _, err := cluster.Core().Namespaces().Get(nsName, metav1.GetOptions{}) if err != nil && !errors.IsNotFound(err) { return false, err } @@ -170,7 +171,7 @@ func verifyNsCascadingDeletion(nsClient clientset.NamespaceInterface, clusters m // namespace should be present in underlying clusters unless orphanDependents is false. shouldExist := orphanDependents == nil || *orphanDependents == true for clusterName, clusterClientset := range clusters { - _, err := clusterClientset.Core().Namespaces().Get(nsName) + _, err := clusterClientset.Core().Namespaces().Get(nsName, metav1.GetOptions{}) if shouldExist && errors.IsNotFound(err) { errMessages = append(errMessages, fmt.Sprintf("unexpected NotFound error for namespace %s in cluster %s, expected namespace to exist", nsName, clusterName)) } else if !shouldExist && !errors.IsNotFound(err) { diff --git a/test/e2e/federated-secret.go b/test/e2e/federated-secret.go index ae765bad0d3..eb4d18e007b 100644 --- a/test/e2e/federated-secret.go +++ b/test/e2e/federated-secret.go @@ -27,6 +27,7 @@ import ( "k8s.io/kubernetes/federation/pkg/federation-controller/util" "k8s.io/kubernetes/pkg/api/errors" "k8s.io/kubernetes/pkg/api/v1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" kubeclientset "k8s.io/kubernetes/pkg/client/clientset_generated/release_1_5" "k8s.io/kubernetes/pkg/util/wait" "k8s.io/kubernetes/test/e2e/framework" @@ -116,7 +117,7 @@ func verifyCascadingDeletionForSecret(clientset *fedclientset.Clientset, cluster By(fmt.Sprintf("Waiting for secret %s to be created in all underlying clusters", secretName)) err := wait.Poll(5*time.Second, 2*time.Minute, func() (bool, error) { for _, cluster := range clusters { - _, err := cluster.Core().Secrets(nsName).Get(secretName) + _, err := cluster.Core().Secrets(nsName).Get(secretName, metav1.GetOptions{}) if err != nil { if !errors.IsNotFound(err) { return false, err @@ -136,7 +137,7 @@ func verifyCascadingDeletionForSecret(clientset *fedclientset.Clientset, cluster // secret should be present in underlying clusters unless orphanDependents is false. shouldExist := orphanDependents == nil || *orphanDependents == true for clusterName, clusterClientset := range clusters { - _, err := clusterClientset.Core().Secrets(nsName).Get(secretName) + _, err := clusterClientset.Core().Secrets(nsName).Get(secretName, metav1.GetOptions{}) if shouldExist && errors.IsNotFound(err) { errMessages = append(errMessages, fmt.Sprintf("unexpected NotFound error for secret %s in cluster %s, expected secret to exist", secretName, clusterName)) } else if !shouldExist && !errors.IsNotFound(err) { @@ -173,7 +174,7 @@ func deleteSecretOrFail(clientset *fedclientset.Clientset, nsName string, secret // Wait for the secret to be deleted. err = wait.Poll(5*time.Second, wait.ForeverTestTimeout, func() (bool, error) { - _, err := clientset.Core().Secrets(nsName).Get(secretName) + _, err := clientset.Core().Secrets(nsName).Get(secretName, metav1.GetOptions{}) if err != nil && errors.IsNotFound(err) { return true, nil } @@ -191,7 +192,7 @@ func updateSecretOrFail(clientset *fedclientset.Clientset, nsName string, secret var newSecret *v1.Secret for retryCount := 0; retryCount < MaxRetries; retryCount++ { - secret, err := clientset.Core().Secrets(nsName).Get(secretName) + secret, err := clientset.Core().Secrets(nsName).Get(secretName, metav1.GetOptions{}) if err != nil { framework.Failf("failed to get secret %q: %v", secretName, err) } @@ -223,7 +224,7 @@ func waitForSecretOrFail(clientset *kubeclientset.Clientset, nsName string, secr By(fmt.Sprintf("Fetching a federated secret shard of secret %q in namespace %q from cluster", secret.Name, nsName)) var clusterSecret *v1.Secret err := wait.PollImmediate(framework.Poll, timeout, func() (bool, error) { - clusterSecret, err := clientset.Core().Secrets(nsName).Get(secret.Name) + clusterSecret, err := clientset.Core().Secrets(nsName).Get(secret.Name, metav1.GetOptions{}) if (!present) && errors.IsNotFound(err) { // We want it gone, and it's gone. By(fmt.Sprintf("Success: shard of federated secret %q in namespace %q in cluster is absent", secret.Name, nsName)) return true, nil // Success @@ -252,7 +253,7 @@ func waitForSecretShardsUpdatedOrFail(nsName string, secret *v1.Secret, clusters func waitForSecretUpdateOrFail(clientset *kubeclientset.Clientset, nsName string, secret *v1.Secret, timeout time.Duration) { By(fmt.Sprintf("Fetching a federated secret shard of secret %q in namespace %q from cluster", secret.Name, nsName)) err := wait.PollImmediate(framework.Poll, timeout, func() (bool, error) { - clusterSecret, err := clientset.Core().Secrets(nsName).Get(secret.Name) + clusterSecret, err := clientset.Core().Secrets(nsName).Get(secret.Name, metav1.GetOptions{}) if err == nil { // We want it present, and the Get succeeded, so we're all good. if util.SecretEquivalent(*clusterSecret, *secret) { By(fmt.Sprintf("Success: shard of federated secret %q in namespace %q in cluster is updated", secret.Name, nsName)) diff --git a/test/e2e/federated-service.go b/test/e2e/federated-service.go index f232dcd3bb8..6ab87251179 100644 --- a/test/e2e/federated-service.go +++ b/test/e2e/federated-service.go @@ -25,6 +25,7 @@ import ( "k8s.io/kubernetes/pkg/api" "k8s.io/kubernetes/pkg/api/v1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" "k8s.io/kubernetes/test/e2e/framework" . "github.com/onsi/ginkgo" @@ -129,7 +130,7 @@ var _ = framework.KubeDescribe("[Feature:Federation]", func() { By(fmt.Sprintf("Deletion of service %q in namespace %q succeeded.", service.Name, nsName)) By(fmt.Sprintf("Verifying that services in underlying clusters are not deleted")) for clusterName, clusterClientset := range clusters { - _, err := clusterClientset.Core().Services(service.Namespace).Get(service.Name) + _, err := clusterClientset.Core().Services(service.Namespace).Get(service.Name, metav1.GetOptions{}) if err != nil { framework.Failf("Unexpected error in fetching service %s in cluster %s, %s", service.Name, clusterName, err) } diff --git a/test/e2e/federation-apiserver.go b/test/e2e/federation-apiserver.go index c57f69b291e..1285154a3a1 100644 --- a/test/e2e/federation-apiserver.go +++ b/test/e2e/federation-apiserver.go @@ -24,6 +24,7 @@ import ( federationapi "k8s.io/kubernetes/federation/apis/federation/v1beta1" "k8s.io/kubernetes/federation/client/clientset_generated/federation_release_1_5" "k8s.io/kubernetes/pkg/api/v1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" "k8s.io/kubernetes/pkg/util/wait" "k8s.io/kubernetes/test/e2e/framework" @@ -124,7 +125,7 @@ func newService(name, namespace string) *v1.Service { // Verify that the cluster is marked ready. func isReady(clusterName string, clientset *federation_release_1_5.Clientset) error { return wait.PollImmediate(time.Second, 5*time.Minute, func() (bool, error) { - c, err := clientset.Federation().Clusters().Get(clusterName) + c, err := clientset.Federation().Clusters().Get(clusterName, metav1.GetOptions{}) if err != nil { return false, err } diff --git a/test/e2e/federation-daemonset.go b/test/e2e/federation-daemonset.go index b5facdcaafe..389d9cb264a 100644 --- a/test/e2e/federation-daemonset.go +++ b/test/e2e/federation-daemonset.go @@ -28,6 +28,7 @@ import ( "k8s.io/kubernetes/pkg/api/errors" "k8s.io/kubernetes/pkg/api/v1" "k8s.io/kubernetes/pkg/apis/extensions/v1beta1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" kubeclientset "k8s.io/kubernetes/pkg/client/clientset_generated/release_1_5" "k8s.io/kubernetes/pkg/util/wait" "k8s.io/kubernetes/test/e2e/framework" @@ -120,7 +121,7 @@ func verifyCascadingDeletionForDS(clientset *fedclientset.Clientset, clusters ma By(fmt.Sprintf("Waiting for daemonset %s to be created in all underlying clusters", daemonsetName)) err := wait.Poll(5*time.Second, 2*time.Minute, func() (bool, error) { for _, cluster := range clusters { - _, err := cluster.Extensions().DaemonSets(nsName).Get(daemonsetName) + _, err := cluster.Extensions().DaemonSets(nsName).Get(daemonsetName, metav1.GetOptions{}) if err != nil && errors.IsNotFound(err) { return false, nil } @@ -140,7 +141,7 @@ func verifyCascadingDeletionForDS(clientset *fedclientset.Clientset, clusters ma // daemon set should be present in underlying clusters unless orphanDependents is false. shouldExist := orphanDependents == nil || *orphanDependents == true for clusterName, clusterClientset := range clusters { - _, err := clusterClientset.Extensions().DaemonSets(nsName).Get(daemonsetName) + _, err := clusterClientset.Extensions().DaemonSets(nsName).Get(daemonsetName, metav1.GetOptions{}) if shouldExist && errors.IsNotFound(err) { errMessages = append(errMessages, fmt.Sprintf("unexpected NotFound error for daemonset %s in cluster %s, expected daemonset to exist", daemonsetName, clusterName)) } else if !shouldExist && !errors.IsNotFound(err) { @@ -193,7 +194,7 @@ func deleteDaemonSetOrFail(clientset *fedclientset.Clientset, nsName string, dae // Wait for the daemonset to be deleted. err = wait.Poll(5*time.Second, wait.ForeverTestTimeout, func() (bool, error) { - _, err := clientset.Extensions().DaemonSets(nsName).Get(daemonsetName) + _, err := clientset.Extensions().DaemonSets(nsName).Get(daemonsetName, metav1.GetOptions{}) if err != nil && errors.IsNotFound(err) { return true, nil } @@ -211,7 +212,7 @@ func updateDaemonSetOrFail(clientset *fedclientset.Clientset, namespace string) var newDaemonSet *v1beta1.DaemonSet for retryCount := 0; retryCount < FederatedDaemonSetMaxRetries; retryCount++ { - daemonset, err := clientset.Extensions().DaemonSets(namespace).Get(FederatedDaemonSetName) + daemonset, err := clientset.Extensions().DaemonSets(namespace).Get(FederatedDaemonSetName, metav1.GetOptions{}) if err != nil { framework.Failf("failed to get daemonset %q: %v", FederatedDaemonSetName, err) } @@ -241,7 +242,7 @@ func waitForDaemonSetOrFail(clientset *kubeclientset.Clientset, namespace string By(fmt.Sprintf("Fetching a federated daemonset shard of daemonset %q in namespace %q from cluster", daemonset.Name, namespace)) var clusterDaemonSet *v1beta1.DaemonSet err := wait.PollImmediate(framework.Poll, timeout, func() (bool, error) { - clusterDaemonSet, err := clientset.Extensions().DaemonSets(namespace).Get(daemonset.Name) + clusterDaemonSet, err := clientset.Extensions().DaemonSets(namespace).Get(daemonset.Name, metav1.GetOptions{}) if (!present) && errors.IsNotFound(err) { // We want it gone, and it's gone. By(fmt.Sprintf("Success: shard of federated daemonset %q in namespace %q in cluster is absent", daemonset.Name, namespace)) return true, nil // Success @@ -270,7 +271,7 @@ func waitForDaemonSetShardsUpdatedOrFail(namespace string, daemonset *v1beta1.Da func waitForDaemonSetUpdateOrFail(clientset *kubeclientset.Clientset, namespace string, daemonset *v1beta1.DaemonSet, timeout time.Duration) { By(fmt.Sprintf("Fetching a federated daemonset shard of daemonset %q in namespace %q from cluster", daemonset.Name, namespace)) err := wait.PollImmediate(framework.Poll, timeout, func() (bool, error) { - clusterDaemonSet, err := clientset.Extensions().DaemonSets(namespace).Get(daemonset.Name) + clusterDaemonSet, err := clientset.Extensions().DaemonSets(namespace).Get(daemonset.Name, metav1.GetOptions{}) if err == nil { // We want it present, and the Get succeeded, so we're all good. if util.ObjectMetaAndSpecEquivalent(clusterDaemonSet, daemonset) { By(fmt.Sprintf("Success: shard of federated daemonset %q in namespace %q in cluster is updated", daemonset.Name, namespace)) diff --git a/test/e2e/federation-deployment.go b/test/e2e/federation-deployment.go index de4b8379b00..f4bfb0de2f1 100644 --- a/test/e2e/federation-deployment.go +++ b/test/e2e/federation-deployment.go @@ -154,7 +154,7 @@ func verifyCascadingDeletionForDeployment(clientset *fedclientset.Clientset, clu By(fmt.Sprintf("Waiting for deployment %s to be created in all underlying clusters", deploymentName)) err := wait.Poll(5*time.Second, 2*time.Minute, func() (bool, error) { for _, cluster := range clusters { - _, err := cluster.Extensions().Deployments(nsName).Get(deploymentName) + _, err := cluster.Extensions().Deployments(nsName).Get(deploymentName, metav1.GetOptions{}) if err != nil && errors.IsNotFound(err) { return false, nil } @@ -174,7 +174,7 @@ func verifyCascadingDeletionForDeployment(clientset *fedclientset.Clientset, clu // deployment should be present in underlying clusters unless orphanDependents is false. shouldExist := orphanDependents == nil || *orphanDependents == true for clusterName, clusterClientset := range clusters { - _, err := clusterClientset.Extensions().Deployments(nsName).Get(deploymentName) + _, err := clusterClientset.Extensions().Deployments(nsName).Get(deploymentName, metav1.GetOptions{}) if shouldExist && errors.IsNotFound(err) { errMessages = append(errMessages, fmt.Sprintf("unexpected NotFound error for deployment %s in cluster %s, expected deployment to exist", deploymentName, clusterName)) } else if !shouldExist && !errors.IsNotFound(err) { @@ -193,13 +193,13 @@ func waitForDeploymentOrFail(c *fedclientset.Clientset, namespace string, deploy func waitForDeployment(c *fedclientset.Clientset, namespace string, deploymentName string, clusters map[string]*cluster) error { err := wait.Poll(10*time.Second, FederatedDeploymentTimeout, func() (bool, error) { - fdep, err := c.Deployments(namespace).Get(deploymentName) + fdep, err := c.Deployments(namespace).Get(deploymentName, metav1.GetOptions{}) if err != nil { return false, err } specReplicas, statusReplicas := int32(0), int32(0) for _, cluster := range clusters { - dep, err := cluster.Deployments(namespace).Get(deploymentName) + dep, err := cluster.Deployments(namespace).Get(deploymentName, metav1.GetOptions{}) if err != nil && !errors.IsNotFound(err) { By(fmt.Sprintf("Failed getting deployment: %q/%q/%q, err: %v", cluster.name, namespace, deploymentName, err)) return false, err @@ -265,7 +265,7 @@ func deleteDeploymentOrFail(clientset *fedclientset.Clientset, nsName string, de // Wait for the deployment to be deleted. err = wait.Poll(5*time.Second, wait.ForeverTestTimeout, func() (bool, error) { - _, err := clientset.Extensions().Deployments(nsName).Get(deploymentName) + _, err := clientset.Extensions().Deployments(nsName).Get(deploymentName, metav1.GetOptions{}) if err != nil && errors.IsNotFound(err) { return true, nil } diff --git a/test/e2e/federation-replicaset.go b/test/e2e/federation-replicaset.go index 7e6b547c3e5..ef637d67a67 100644 --- a/test/e2e/federation-replicaset.go +++ b/test/e2e/federation-replicaset.go @@ -156,7 +156,7 @@ func verifyCascadingDeletionForReplicaSet(clientset *fedclientset.Clientset, clu By(fmt.Sprintf("Waiting for replica sets %s to be created in all underlying clusters", replicaSetName)) err := wait.Poll(5*time.Second, 2*time.Minute, func() (bool, error) { for _, cluster := range clusters { - _, err := cluster.Extensions().ReplicaSets(nsName).Get(replicaSetName) + _, err := cluster.Extensions().ReplicaSets(nsName).Get(replicaSetName, metav1.GetOptions{}) if err != nil && errors.IsNotFound(err) { return false, nil } @@ -174,7 +174,7 @@ func verifyCascadingDeletionForReplicaSet(clientset *fedclientset.Clientset, clu By(fmt.Sprintf("Verifying replica sets %s in underlying clusters", replicaSetName)) errMessages := []string{} for clusterName, clusterClientset := range clusters { - _, err := clusterClientset.Extensions().ReplicaSets(nsName).Get(replicaSetName) + _, err := clusterClientset.Extensions().ReplicaSets(nsName).Get(replicaSetName, metav1.GetOptions{}) if (orphanDependents == nil || *orphanDependents == true) && errors.IsNotFound(err) { errMessages = append(errMessages, fmt.Sprintf("unexpected NotFound error for replica set %s in cluster %s, expected replica set to exist", replicaSetName, clusterName)) } else if (orphanDependents != nil && *orphanDependents == false) && (err == nil || !errors.IsNotFound(err)) { @@ -193,13 +193,13 @@ func waitForReplicaSetOrFail(c *fedclientset.Clientset, namespace string, replic func waitForReplicaSet(c *fedclientset.Clientset, namespace string, replicaSetName string, clusters map[string]*cluster) error { err := wait.Poll(10*time.Second, FederatedReplicaSetTimeout, func() (bool, error) { - frs, err := c.ReplicaSets(namespace).Get(replicaSetName) + frs, err := c.ReplicaSets(namespace).Get(replicaSetName, metav1.GetOptions{}) if err != nil { return false, err } specReplicas, statusReplicas := int32(0), int32(0) for _, cluster := range clusters { - rs, err := cluster.ReplicaSets(namespace).Get(replicaSetName) + rs, err := cluster.ReplicaSets(namespace).Get(replicaSetName, metav1.GetOptions{}) if err != nil && !errors.IsNotFound(err) { By(fmt.Sprintf("Failed getting replicaset: %q/%q/%q, err: %v", cluster.name, namespace, replicaSetName, err)) return false, err @@ -251,7 +251,7 @@ func deleteReplicaSetOrFail(clientset *fedclientset.Clientset, nsName string, re // Wait for the replicaSet to be deleted. err = wait.Poll(5*time.Second, wait.ForeverTestTimeout, func() (bool, error) { - _, err := clientset.Extensions().ReplicaSets(nsName).Get(replicaSetName) + _, err := clientset.Extensions().ReplicaSets(nsName).Get(replicaSetName, metav1.GetOptions{}) if err != nil && errors.IsNotFound(err) { return true, nil } diff --git a/test/e2e/federation-util-14.go b/test/e2e/federation-util-14.go index a64ab984ffc..6f4c1c39765 100644 --- a/test/e2e/federation-util-14.go +++ b/test/e2e/federation-util-14.go @@ -22,13 +22,14 @@ import ( federationapi "k8s.io/kubernetes/federation/apis/federation/v1beta1" apiv1 "k8s.io/kubernetes/pkg/api/v1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" "k8s.io/kubernetes/pkg/util/wait" "k8s.io/kubernetes/test/e2e/framework" ) func createClusterObjectOrFail_14(f *framework.Framework, context *framework.E2EContext) { framework.Logf("Looking up cluster: %s", context.Name) - foundCluster, err := f.FederationClientset_1_5.Federation().Clusters().Get(context.Name) + foundCluster, err := f.FederationClientset_1_5.Federation().Clusters().Get(context.Name, metav1.GetOptions{}) if err == nil && foundCluster != nil { return } @@ -68,7 +69,7 @@ func buildClustersOrFail_14(f *framework.Framework) []*federationapi.Cluster { // Wait for all clusters to become ready for up to 5 min. if err := wait.PollImmediate(5*time.Second, 5*time.Minute, func() (bool, error) { for _, context := range contexts { - cluster, err := f.FederationClientset_1_5.Federation().Clusters().Get(context.Name) + cluster, err := f.FederationClientset_1_5.Federation().Clusters().Get(context.Name, metav1.GetOptions{}) if err != nil { return false, err } diff --git a/test/e2e/federation-util.go b/test/e2e/federation-util.go index f01694cf6f9..2961c910c03 100644 --- a/test/e2e/federation-util.go +++ b/test/e2e/federation-util.go @@ -27,6 +27,7 @@ import ( "k8s.io/kubernetes/pkg/api" "k8s.io/kubernetes/pkg/api/errors" "k8s.io/kubernetes/pkg/api/v1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" kubeclientset "k8s.io/kubernetes/pkg/client/clientset_generated/release_1_5" "k8s.io/kubernetes/pkg/client/restclient" "k8s.io/kubernetes/pkg/client/unversioned/clientcmd" @@ -85,7 +86,7 @@ func createClusterObjectOrFail(f *framework.Framework, context *framework.E2ECon } func clusterIsReadyOrFail(f *framework.Framework, context *framework.E2EContext) { - c, err := f.FederationClientset_1_5.Federation().Clusters().Get(context.Name) + c, err := f.FederationClientset_1_5.Federation().Clusters().Get(context.Name, metav1.GetOptions{}) framework.ExpectNoError(err, fmt.Sprintf("get cluster: %+v", err)) if c.ObjectMeta.Name != context.Name { framework.Failf("cluster name does not match input context: actual=%+v, expected=%+v", c, context) @@ -140,7 +141,7 @@ func createNamespaceInClusters(clusters map[string]*cluster, f *framework.Framew for name, c := range clusters { // The e2e Framework created the required namespace in federation control plane, but we need to create it in all the others, if it doesn't yet exist. // TODO(nikhiljindal): remove this once we have the namespace controller working as expected. - if _, err := c.Clientset.Core().Namespaces().Get(nsName); errors.IsNotFound(err) { + if _, err := c.Clientset.Core().Namespaces().Get(nsName, metav1.GetOptions{}); errors.IsNotFound(err) { ns := &v1.Namespace{ ObjectMeta: v1.ObjectMeta{ Name: nsName, @@ -164,7 +165,7 @@ func unregisterClusters(clusters map[string]*cluster, f *framework.Framework) { nsName := f.FederationNamespace.Name for name, c := range clusters { if c.namespaceCreated { - if _, err := c.Clientset.Core().Namespaces().Get(nsName); !errors.IsNotFound(err) { + if _, err := c.Clientset.Core().Namespaces().Get(nsName, metav1.GetOptions{}); !errors.IsNotFound(err) { err := c.Clientset.Core().Namespaces().Delete(nsName, &v1.DeleteOptions{}) framework.ExpectNoError(err, "Couldn't delete the namespace %s in cluster %q: %v", nsName, name, err) } @@ -217,7 +218,7 @@ func waitForServiceOrFail(clientset *kubeclientset.Clientset, namespace string, By(fmt.Sprintf("Fetching a federated service shard of service %q in namespace %q from cluster", service.Name, namespace)) var clusterService *v1.Service err := wait.PollImmediate(framework.Poll, timeout, func() (bool, error) { - clusterService, err := clientset.Services(namespace).Get(service.Name) + clusterService, err := clientset.Services(namespace).Get(service.Name, metav1.GetOptions{}) if (!present) && errors.IsNotFound(err) { // We want it gone, and it's gone. By(fmt.Sprintf("Success: shard of federated service %q in namespace %q in cluster is absent", service.Name, namespace)) return true, nil // Success @@ -297,7 +298,7 @@ func cleanupServiceShardsAndProviderResources(namespace string, service *v1.Serv err := wait.PollImmediate(framework.Poll, FederatedServiceTimeout, func() (bool, error) { var err error - cSvc, err = c.Clientset.Services(namespace).Get(service.Name) + cSvc, err = c.Clientset.Services(namespace).Get(service.Name, metav1.GetOptions{}) if err != nil && !errors.IsNotFound(err) { // Get failed with an error, try again. framework.Logf("Failed to find service %q in namespace %q, in cluster %q: %v. Trying again in %s", service.Name, namespace, name, err, framework.Poll) @@ -385,7 +386,7 @@ func podExitCodeDetector(f *framework.Framework, name, namespace string, code in } return func() error { - pod, err := f.ClientSet.Core().Pods(namespace).Get(name) + pod, err := f.ClientSet.Core().Pods(namespace).Get(name, metav1.GetOptions{}) if err != nil { return logerr(err) } diff --git a/test/e2e/framework/exec_util.go b/test/e2e/framework/exec_util.go index f35de2e78ca..ffe74fb7a6a 100644 --- a/test/e2e/framework/exec_util.go +++ b/test/e2e/framework/exec_util.go @@ -24,6 +24,7 @@ import ( "k8s.io/kubernetes/pkg/api" "k8s.io/kubernetes/pkg/api/v1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" "k8s.io/kubernetes/pkg/client/restclient" "k8s.io/kubernetes/pkg/client/unversioned/remotecommand" remotecommandserver "k8s.io/kubernetes/pkg/kubelet/server/remotecommand" @@ -112,14 +113,14 @@ func (f *Framework) ExecShellInContainer(podName, containerName string, cmd stri } func (f *Framework) ExecCommandInPod(podName string, cmd ...string) string { - pod, err := f.PodClient().Get(podName) + pod, err := f.PodClient().Get(podName, metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred(), "failed to get pod") Expect(pod.Spec.Containers).NotTo(BeEmpty()) return f.ExecCommandInContainer(podName, pod.Spec.Containers[0].Name, cmd...) } func (f *Framework) ExecCommandInPodWithFullOutput(podName string, cmd ...string) (string, string, error) { - pod, err := f.PodClient().Get(podName) + pod, err := f.PodClient().Get(podName, metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred(), "failed to get pod") Expect(pod.Spec.Containers).NotTo(BeEmpty()) return f.ExecCommandInContainerWithFullOutput(podName, pod.Spec.Containers[0].Name, cmd...) diff --git a/test/e2e/framework/framework.go b/test/e2e/framework/framework.go index 1130261e966..0cbc63a0903 100644 --- a/test/e2e/framework/framework.go +++ b/test/e2e/framework/framework.go @@ -34,6 +34,7 @@ import ( apierrs "k8s.io/kubernetes/pkg/api/errors" "k8s.io/kubernetes/pkg/api/v1" "k8s.io/kubernetes/pkg/apimachinery/registered" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" "k8s.io/kubernetes/pkg/client/clientset_generated/internalclientset" clientset "k8s.io/kubernetes/pkg/client/clientset_generated/release_1_5" "k8s.io/kubernetes/pkg/client/restclient" @@ -283,7 +284,7 @@ func (f *Framework) deleteFederationNs() { } // Verify that it got deleted. err := wait.PollImmediate(5*time.Second, timeout, func() (bool, error) { - if _, err := clientset.Core().Namespaces().Get(ns.Name); err != nil { + if _, err := clientset.Core().Namespaces().Get(ns.Name, metav1.GetOptions{}); err != nil { if apierrs.IsNotFound(err) { return true, nil } diff --git a/test/e2e/framework/networking_utils.go b/test/e2e/framework/networking_utils.go index 80e8d2f2fab..b7de5b9cad9 100644 --- a/test/e2e/framework/networking_utils.go +++ b/test/e2e/framework/networking_utils.go @@ -423,12 +423,12 @@ func (config *NetworkingTestConfig) createTestPods() { ExpectNoError(config.f.WaitForPodRunning(hostTestContainerPod.Name)) var err error - config.TestContainerPod, err = config.getPodClient().Get(testContainerPod.Name) + config.TestContainerPod, err = config.getPodClient().Get(testContainerPod.Name, metav1.GetOptions{}) if err != nil { Failf("Failed to retrieve %s pod: %v", testContainerPod.Name, err) } - config.HostTestContainerPod, err = config.getPodClient().Get(hostTestContainerPod.Name) + config.HostTestContainerPod, err = config.getPodClient().Get(hostTestContainerPod.Name, metav1.GetOptions{}) if err != nil { Failf("Failed to retrieve %s pod: %v", hostTestContainerPod.Name, err) } @@ -441,7 +441,7 @@ func (config *NetworkingTestConfig) createService(serviceSpec *v1.Service) *v1.S err = WaitForService(config.f.ClientSet, config.Namespace, serviceSpec.Name, true, 5*time.Second, 45*time.Second) Expect(err).NotTo(HaveOccurred(), fmt.Sprintf("error while waiting for service:%s err: %v", serviceSpec.Name, err)) - createdService, err := config.getServiceClient().Get(serviceSpec.Name) + createdService, err := config.getServiceClient().Get(serviceSpec.Name, metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred(), fmt.Sprintf("Failed to create %s service: %v", serviceSpec.Name, err)) return createdService @@ -542,7 +542,7 @@ func (config *NetworkingTestConfig) createNetProxyPods(podName string, selector runningPods := make([]*v1.Pod, 0, len(nodes)) for _, p := range createdPods { ExpectNoError(config.f.WaitForPodReady(p.Name)) - rp, err := config.getPodClient().Get(p.Name) + rp, err := config.getPodClient().Get(p.Name, metav1.GetOptions{}) ExpectNoError(err) runningPods = append(runningPods, rp) } diff --git a/test/e2e/framework/pods.go b/test/e2e/framework/pods.go index 47ea4c0edc3..efc3cbe97f5 100644 --- a/test/e2e/framework/pods.go +++ b/test/e2e/framework/pods.go @@ -24,6 +24,7 @@ import ( "k8s.io/kubernetes/pkg/api/errors" "k8s.io/kubernetes/pkg/api/v1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" v1core "k8s.io/kubernetes/pkg/client/clientset_generated/release_1_5/typed/core/v1" "k8s.io/kubernetes/pkg/labels" "k8s.io/kubernetes/pkg/util/sets" @@ -66,7 +67,7 @@ func (c *PodClient) CreateSync(pod *v1.Pod) *v1.Pod { p := c.Create(pod) ExpectNoError(c.f.WaitForPodRunning(p.Name)) // Get the newest pod after it becomes running, some status may change after pod created, such as pod ip. - p, err := c.Get(p.Name) + p, err := c.Get(p.Name, metav1.GetOptions{}) ExpectNoError(err) return p } @@ -92,7 +93,7 @@ func (c *PodClient) CreateBatch(pods []*v1.Pod) []*v1.Pod { // pod object. func (c *PodClient) Update(name string, updateFn func(pod *v1.Pod)) { ExpectNoError(wait.Poll(time.Millisecond*500, time.Second*30, func() (bool, error) { - pod, err := c.PodInterface.Get(name) + pod, err := c.PodInterface.Get(name, metav1.GetOptions{}) if err != nil { return false, fmt.Errorf("failed to get pod %q: %v", name, err) } diff --git a/test/e2e/framework/util.go b/test/e2e/framework/util.go index 07156a86cee..6a85fc4ffd4 100644 --- a/test/e2e/framework/util.go +++ b/test/e2e/framework/util.go @@ -367,7 +367,7 @@ func SkipUnlessFederated(c clientset.Interface) { federationNS = "federation" } - _, err := c.Core().Namespaces().Get(federationNS) + _, err := c.Core().Namespaces().Get(federationNS, metav1.GetOptions{}) if err != nil { if apierrs.IsNotFound(err) { Skipf("Could not find federation namespace %s: skipping federated test", federationNS) @@ -756,7 +756,7 @@ func waitForServiceAccountInNamespace(c clientset.Interface, ns, serviceAccountN func WaitForPodCondition(c clientset.Interface, ns, podName, desc string, timeout time.Duration, condition podCondition) error { Logf("Waiting up to %[1]v for pod %[2]s status to be %[3]s", timeout, podName, desc) for start := time.Now(); time.Since(start) < timeout; time.Sleep(Poll) { - pod, err := c.Core().Pods(ns).Get(podName) + pod, err := c.Core().Pods(ns).Get(podName, metav1.GetOptions{}) if err != nil { if apierrs.IsNotFound(err) { Logf("Pod %q in namespace %q disappeared. Error: %v", podName, ns, err) @@ -828,7 +828,7 @@ func WaitForFederationApiserverReady(c *federation_release_1_5.Clientset) error func WaitForPersistentVolumePhase(phase v1.PersistentVolumePhase, c clientset.Interface, pvName string, Poll, timeout time.Duration) error { Logf("Waiting up to %v for PersistentVolume %s to have phase %s", timeout, pvName, phase) for start := time.Now(); time.Since(start) < timeout; time.Sleep(Poll) { - pv, err := c.Core().PersistentVolumes().Get(pvName) + pv, err := c.Core().PersistentVolumes().Get(pvName, metav1.GetOptions{}) if err != nil { Logf("Get persistent volume %s in failed, ignoring for %v: %v", pvName, Poll, err) continue @@ -848,7 +848,7 @@ func WaitForPersistentVolumePhase(phase v1.PersistentVolumePhase, c clientset.In func WaitForPersistentVolumeDeleted(c clientset.Interface, pvName string, Poll, timeout time.Duration) error { Logf("Waiting up to %v for PersistentVolume %s to get deleted", timeout, pvName) for start := time.Now(); time.Since(start) < timeout; time.Sleep(Poll) { - pv, err := c.Core().PersistentVolumes().Get(pvName) + pv, err := c.Core().PersistentVolumes().Get(pvName, metav1.GetOptions{}) if err == nil { Logf("PersistentVolume %s found and phase=%s (%v)", pvName, pv.Status.Phase, time.Since(start)) continue @@ -868,7 +868,7 @@ func WaitForPersistentVolumeDeleted(c clientset.Interface, pvName string, Poll, func WaitForPersistentVolumeClaimPhase(phase v1.PersistentVolumeClaimPhase, c clientset.Interface, ns string, pvcName string, Poll, timeout time.Duration) error { Logf("Waiting up to %v for PersistentVolumeClaim %s to have phase %s", timeout, pvcName, phase) for start := time.Now(); time.Since(start) < timeout; time.Sleep(Poll) { - pvc, err := c.Core().PersistentVolumeClaims(ns).Get(pvcName) + pvc, err := c.Core().PersistentVolumeClaims(ns).Get(pvcName, metav1.GetOptions{}) if err != nil { Logf("Get persistent volume claim %s in failed, ignoring for %v: %v", pvcName, Poll, err) continue @@ -973,7 +973,7 @@ func deleteNS(c clientset.Interface, clientPool dynamic.ClientPool, namespace st // wait for namespace to delete or timeout. err := wait.PollImmediate(5*time.Second, timeout, func() (bool, error) { - if _, err := c.Core().Namespaces().Get(namespace); err != nil { + if _, err := c.Core().Namespaces().Get(namespace, metav1.GetOptions{}); err != nil { if apierrs.IsNotFound(err) { return true, nil } @@ -1046,7 +1046,7 @@ func logNamespaces(c clientset.Interface, namespace string) { // logNamespace logs detail about a namespace func logNamespace(c clientset.Interface, namespace string) { - ns, err := c.Core().Namespaces().Get(namespace) + ns, err := c.Core().Namespaces().Get(namespace, metav1.GetOptions{}) if err != nil { if apierrs.IsNotFound(err) { Logf("namespace: %v no longer exists", namespace) @@ -1452,7 +1452,7 @@ func WaitForRCPodToDisappear(c clientset.Interface, ns, rcName, podName string) // WaitForService waits until the service appears (exist == true), or disappears (exist == false) func WaitForService(c clientset.Interface, namespace, name string, exist bool, interval, timeout time.Duration) error { err := wait.PollImmediate(interval, timeout, func() (bool, error) { - _, err := c.Core().Services(namespace).Get(name) + _, err := c.Core().Services(namespace).Get(name, metav1.GetOptions{}) switch { case err == nil: if !exist { @@ -1507,7 +1507,7 @@ func countEndpointsNum(e *v1.Endpoints) int { // WaitForReplicationController waits until the RC appears (exist == true), or disappears (exist == false) func WaitForReplicationController(c clientset.Interface, namespace, name string, exist bool, interval, timeout time.Duration) error { err := wait.PollImmediate(interval, timeout, func() (bool, error) { - _, err := c.Core().ReplicationControllers(namespace).Get(name) + _, err := c.Core().ReplicationControllers(namespace).Get(name, metav1.GetOptions{}) if err != nil { Logf("Get ReplicationController %s in namespace %s failed (%v).", name, namespace, err) return !exist, nil @@ -1525,7 +1525,7 @@ func WaitForReplicationController(c clientset.Interface, namespace, name string, func WaitForEndpoint(c clientset.Interface, ns, name string) error { for t := time.Now(); time.Since(t) < EndpointRegisterTimeout; time.Sleep(Poll) { - endpoint, err := c.Core().Endpoints(ns).Get(name) + endpoint, err := c.Core().Endpoints(ns).Get(name, metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred()) if len(endpoint.Subsets) == 0 || len(endpoint.Subsets[0].Addresses) == 0 { Logf("Endpoint %s/%s is not ready yet", ns, name) @@ -2178,7 +2178,7 @@ func (f *Framework) MatchContainerOutput( } // Grab its logs. Get host first. - podStatus, err := podClient.Get(createdPod.Name) + podStatus, err := podClient.Get(createdPod.Name, metav1.GetOptions{}) if err != nil { return fmt.Errorf("failed to get pod status: %v", err) } @@ -2309,7 +2309,7 @@ func dumpAllNodeInfo(c clientset.Interface) { func DumpNodeDebugInfo(c clientset.Interface, nodeNames []string, logFunc func(fmt string, args ...interface{})) { for _, n := range nodeNames { logFunc("\nLogging node info for node %v", n) - node, err := c.Core().Nodes().Get(n) + node, err := c.Core().Nodes().Get(n, metav1.GetOptions{}) if err != nil { logFunc("Error getting node info %v", err) } @@ -2482,7 +2482,7 @@ func AddOrUpdateLabelOnNode(c clientset.Interface, nodeName string, labelKey, la func ExpectNodeHasLabel(c clientset.Interface, nodeName string, labelKey string, labelValue string) { By("verifying the node has the label " + labelKey + " " + labelValue) - node, err := c.Core().Nodes().Get(nodeName) + node, err := c.Core().Nodes().Get(nodeName, metav1.GetOptions{}) ExpectNoError(err) Expect(node.Labels[labelKey]).To(Equal(labelValue)) } @@ -2499,7 +2499,7 @@ func RemoveLabelOffNode(c clientset.Interface, nodeName string, labelKey string) func AddOrUpdateTaintOnNode(c clientset.Interface, nodeName string, taint v1.Taint) { for attempt := 0; attempt < UpdateRetries; attempt++ { - node, err := c.Core().Nodes().Get(nodeName) + node, err := c.Core().Nodes().Get(nodeName, metav1.GetOptions{}) ExpectNoError(err) nodeTaints, err := v1.GetTaintsFromNodeAnnotations(node.Annotations) @@ -2553,7 +2553,7 @@ func taintExists(taints []v1.Taint, taintToFind v1.Taint) bool { func ExpectNodeHasTaint(c clientset.Interface, nodeName string, taint v1.Taint) { By("verifying the node has the taint " + taint.ToString()) - node, err := c.Core().Nodes().Get(nodeName) + node, err := c.Core().Nodes().Get(nodeName, metav1.GetOptions{}) ExpectNoError(err) nodeTaints, err := v1.GetTaintsFromNodeAnnotations(node.Annotations) @@ -2586,7 +2586,7 @@ func deleteTaint(oldTaints []v1.Taint, taintToDelete v1.Taint) ([]v1.Taint, erro func RemoveTaintOffNode(c clientset.Interface, nodeName string, taint v1.Taint) { By("removing the taint " + taint.ToString() + " off the node " + nodeName) for attempt := 0; attempt < UpdateRetries; attempt++ { - node, err := c.Core().Nodes().Get(nodeName) + node, err := c.Core().Nodes().Get(nodeName, metav1.GetOptions{}) ExpectNoError(err) nodeTaints, err := v1.GetTaintsFromNodeAnnotations(node.Annotations) @@ -2622,7 +2622,7 @@ func RemoveTaintOffNode(c clientset.Interface, nodeName string, taint v1.Taint) time.Sleep(100 * time.Millisecond) } - nodeUpdated, err := c.Core().Nodes().Get(nodeName) + nodeUpdated, err := c.Core().Nodes().Get(nodeName, metav1.GetOptions{}) ExpectNoError(err) By("verifying the node doesn't have the taint " + taint.ToString()) taintsGot, err := v1.GetTaintsFromNodeAnnotations(nodeUpdated.Annotations) @@ -2774,11 +2774,11 @@ func WaitForPodsWithLabelRunningReady(c clientset.Interface, ns string, label la func getRuntimeObjectForKind(c clientset.Interface, kind schema.GroupKind, ns, name string) (runtime.Object, error) { switch kind { case api.Kind("ReplicationController"): - return c.Core().ReplicationControllers(ns).Get(name) + return c.Core().ReplicationControllers(ns).Get(name, metav1.GetOptions{}) case extensionsinternal.Kind("ReplicaSet"): - return c.Extensions().ReplicaSets(ns).Get(name) + return c.Extensions().ReplicaSets(ns).Get(name, metav1.GetOptions{}) case extensionsinternal.Kind("Deployment"): - return c.Extensions().Deployments(ns).Get(name) + return c.Extensions().Deployments(ns).Get(name, metav1.GetOptions{}) default: return nil, fmt.Errorf("Unsupported kind when getting runtime object: %v", kind) } @@ -3020,7 +3020,7 @@ func waitForPodsGone(ps *testutils.PodStore, interval, timeout time.Duration) er // Delete a ReplicaSet and all pods it spawned func DeleteReplicaSet(clientset clientset.Interface, internalClientset internalclientset.Interface, ns, name string) error { By(fmt.Sprintf("deleting ReplicaSet %s in namespace %s", name, ns)) - rc, err := clientset.Extensions().ReplicaSets(ns).Get(name) + rc, err := clientset.Extensions().ReplicaSets(ns).Get(name, metav1.GetOptions{}) if err != nil { if apierrs.IsNotFound(err) { Logf("ReplicaSet %s was already deleted: %v", name, err) @@ -3068,7 +3068,7 @@ func waitForReplicaSetPodsGone(c clientset.Interface, rs *extensions.ReplicaSet) // WaitForReadyReplicaSet waits until the replica set has all of its replicas ready. func WaitForReadyReplicaSet(c clientset.Interface, ns, name string) error { - rs, err := c.Extensions().ReplicaSets(ns).Get(name) + rs, err := c.Extensions().ReplicaSets(ns).Get(name, metav1.GetOptions{}) if err != nil { return err } @@ -3134,7 +3134,7 @@ func WaitForDeploymentStatusValid(c clientset.Interface, d *extensions.Deploymen err := wait.Poll(Poll, 5*time.Minute, func() (bool, error) { var err error - deployment, err = c.Extensions().Deployments(d.Namespace).Get(d.Name) + deployment, err = c.Extensions().Deployments(d.Namespace).Get(d.Name, metav1.GetOptions{}) if err != nil { return false, err } @@ -3206,7 +3206,7 @@ func WaitForDeploymentStatus(c clientset.Interface, d *extensions.Deployment) er err := wait.Poll(Poll, 5*time.Minute, func() (bool, error) { var err error - deployment, err = c.Extensions().Deployments(d.Namespace).Get(d.Name) + deployment, err = c.Extensions().Deployments(d.Namespace).Get(d.Name, metav1.GetOptions{}) if err != nil { return false, err } @@ -3260,7 +3260,7 @@ func WaitForDeploymentStatus(c clientset.Interface, d *extensions.Deployment) er // WaitForDeploymentUpdatedReplicasLTE waits for given deployment to be observed by the controller and has at least a number of updatedReplicas func WaitForDeploymentUpdatedReplicasLTE(c clientset.Interface, ns, deploymentName string, minUpdatedReplicas int, desiredGeneration int64) error { err := wait.Poll(Poll, 5*time.Minute, func() (bool, error) { - deployment, err := c.Extensions().Deployments(ns).Get(deploymentName) + deployment, err := c.Extensions().Deployments(ns).Get(deploymentName, metav1.GetOptions{}) if err != nil { return false, err } @@ -3279,7 +3279,7 @@ func WaitForDeploymentUpdatedReplicasLTE(c clientset.Interface, ns, deploymentNa // Note that rollback should be cleared shortly, so we only wait for 1 minute here to fail early. func WaitForDeploymentRollbackCleared(c clientset.Interface, ns, deploymentName string) error { err := wait.Poll(Poll, 1*time.Minute, func() (bool, error) { - deployment, err := c.Extensions().Deployments(ns).Get(deploymentName) + deployment, err := c.Extensions().Deployments(ns).Get(deploymentName, metav1.GetOptions{}) if err != nil { return false, err } @@ -3303,7 +3303,7 @@ func WaitForDeploymentRevisionAndImage(c clientset.Interface, ns, deploymentName var reason string err := wait.Poll(Poll, 1*time.Minute, func() (bool, error) { var err error - deployment, err = c.Extensions().Deployments(ns).Get(deploymentName) + deployment, err = c.Extensions().Deployments(ns).Get(deploymentName, metav1.GetOptions{}) if err != nil { return false, err } @@ -3360,7 +3360,7 @@ func WaitForDeploymentRevisionAndImage(c clientset.Interface, ns, deploymentName func WaitForOverlappingAnnotationMatch(c clientset.Interface, ns, deploymentName, expected string) error { return wait.Poll(Poll, 1*time.Minute, func() (bool, error) { - deployment, err := c.Extensions().Deployments(ns).Get(deploymentName) + deployment, err := c.Extensions().Deployments(ns).Get(deploymentName, metav1.GetOptions{}) if err != nil { return false, err } @@ -3373,7 +3373,7 @@ func WaitForOverlappingAnnotationMatch(c clientset.Interface, ns, deploymentName // CheckNewRSAnnotations check if the new RS's annotation is as expected func CheckNewRSAnnotations(c clientset.Interface, ns, deploymentName string, expectedAnnotations map[string]string) error { - deployment, err := c.Extensions().Deployments(ns).Get(deploymentName) + deployment, err := c.Extensions().Deployments(ns).Get(deploymentName, metav1.GetOptions{}) if err != nil { return err } @@ -3410,7 +3410,7 @@ func WaitForPodsReady(c clientset.Interface, ns, name string, minReadySeconds in // Waits for the deployment to clean up old rcs. func WaitForDeploymentOldRSsNum(c clientset.Interface, ns, deploymentName string, desiredRSNum int) error { return wait.Poll(Poll, 5*time.Minute, func() (bool, error) { - deployment, err := c.Extensions().Deployments(ns).Get(deploymentName) + deployment, err := c.Extensions().Deployments(ns).Get(deploymentName, metav1.GetOptions{}) if err != nil { return false, err } @@ -3435,13 +3435,15 @@ func logReplicaSetsOfDeployment(deployment *extensions.Deployment, allOldRSs []* } func WaitForObservedDeployment(c clientset.Interface, ns, deploymentName string, desiredGeneration int64) error { - return deploymentutil.WaitForObservedDeployment(func() (*extensions.Deployment, error) { return c.Extensions().Deployments(ns).Get(deploymentName) }, desiredGeneration, Poll, 1*time.Minute) + return deploymentutil.WaitForObservedDeployment(func() (*extensions.Deployment, error) { + return c.Extensions().Deployments(ns).Get(deploymentName, metav1.GetOptions{}) + }, desiredGeneration, Poll, 1*time.Minute) } func WaitForDeploymentWithCondition(c clientset.Interface, ns, deploymentName, reason string, condType extensions.DeploymentConditionType) error { var conditions []extensions.DeploymentCondition pollErr := wait.PollImmediate(time.Second, 1*time.Minute, func() (bool, error) { - deployment, err := c.Extensions().Deployments(ns).Get(deploymentName) + deployment, err := c.Extensions().Deployments(ns).Get(deploymentName, metav1.GetOptions{}) if err != nil { return false, err } @@ -3516,7 +3518,7 @@ func UpdateDeploymentWithRetries(c clientset.Interface, namespace, name string, deployments := c.Extensions().Deployments(namespace) var updateErr error pollErr := wait.Poll(10*time.Millisecond, 1*time.Minute, func() (bool, error) { - if deployment, err = deployments.Get(name); err != nil { + if deployment, err = deployments.Get(name, metav1.GetOptions{}); err != nil { return false, err } // Apply the update, then attempt to push it to the apiserver. @@ -3541,7 +3543,7 @@ func UpdateReplicaSetWithRetries(c clientset.Interface, namespace, name string, var updateErr error pollErr := wait.PollImmediate(10*time.Millisecond, 1*time.Minute, func() (bool, error) { var err error - if rs, err = c.Extensions().ReplicaSets(namespace).Get(name); err != nil { + if rs, err = c.Extensions().ReplicaSets(namespace).Get(name, metav1.GetOptions{}); err != nil { return false, err } // Apply the update, then attempt to push it to the apiserver. @@ -3566,7 +3568,7 @@ func UpdateReplicationControllerWithRetries(c clientset.Interface, namespace, na var updateErr error pollErr := wait.PollImmediate(10*time.Millisecond, 1*time.Minute, func() (bool, error) { var err error - if rc, err = c.Core().ReplicationControllers(namespace).Get(name); err != nil { + if rc, err = c.Core().ReplicationControllers(namespace).Get(name, metav1.GetOptions{}); err != nil { return false, err } // Apply the update, then attempt to push it to the apiserver. @@ -3590,7 +3592,7 @@ func UpdateStatefulSetWithRetries(c clientset.Interface, namespace, name string, statefulSets := c.Apps().StatefulSets(namespace) var updateErr error pollErr := wait.Poll(10*time.Millisecond, 1*time.Minute, func() (bool, error) { - if statefulSet, err = statefulSets.Get(name); err != nil { + if statefulSet, err = statefulSets.Get(name, metav1.GetOptions{}); err != nil { return false, err } // Apply the update, then attempt to push it to the apiserver. @@ -3614,7 +3616,7 @@ func UpdateJobWithRetries(c clientset.Interface, namespace, name string, applyUp jobs := c.Batch().Jobs(namespace) var updateErr error pollErr := wait.PollImmediate(10*time.Millisecond, 1*time.Minute, func() (bool, error) { - if job, err = jobs.Get(name); err != nil { + if job, err = jobs.Get(name, metav1.GetOptions{}); err != nil { return false, err } // Apply the update, then attempt to push it to the apiserver. @@ -3932,7 +3934,7 @@ func IsNodeConditionUnset(node *v1.Node, conditionType v1.NodeConditionType) boo func WaitForNodeToBe(c clientset.Interface, name string, conditionType v1.NodeConditionType, wantTrue bool, timeout time.Duration) bool { Logf("Waiting up to %v for node %s condition %s to be %t", timeout, name, conditionType, wantTrue) for start := time.Now(); time.Since(start) < timeout; time.Sleep(Poll) { - node, err := c.Core().Nodes().Get(name) + node, err := c.Core().Nodes().Get(name, metav1.GetOptions{}) if err != nil { Logf("Couldn't get node %s", name) continue @@ -4259,7 +4261,7 @@ func WaitForMasters(masterPrefix string, c clientset.Interface, size int, timeou // address. Returns an error if the node the pod is on doesn't have an External // address. func GetHostExternalAddress(client clientset.Interface, p *v1.Pod) (externalAddress string, err error) { - node, err := client.Core().Nodes().Get(p.Spec.NodeName) + node, err := client.Core().Nodes().Get(p.Spec.NodeName, metav1.GetOptions{}) if err != nil { return "", err } @@ -4335,7 +4337,7 @@ func OpenWebSocketForURL(url *url.URL, config *restclient.Config, protocols []st // getIngressAddress returns the ips/hostnames associated with the Ingress. func getIngressAddress(client clientset.Interface, ns, name string) ([]string, error) { - ing, err := client.Extensions().Ingresses(ns).Get(name) + ing, err := client.Extensions().Ingresses(ns).Get(name, metav1.GetOptions{}) if err != nil { return nil, err } @@ -4406,7 +4408,7 @@ func LookForString(expectedString string, timeout time.Duration, fn func() strin // getSvcNodePort returns the node port for the given service:port. func getSvcNodePort(client clientset.Interface, ns, name string, svcPort int) (int, error) { - svc, err := client.Core().Services(ns).Get(name) + svc, err := client.Core().Services(ns).Get(name, metav1.GetOptions{}) if err != nil { return 0, err } @@ -4471,7 +4473,7 @@ func ScaleRCByLabels(clientset clientset.Interface, internalClientset internalcl if err := ScaleRC(clientset, internalClientset, ns, name, replicas, false); err != nil { return err } - rc, err := clientset.Core().ReplicationControllers(ns).Get(name) + rc, err := clientset.Core().ReplicationControllers(ns).Get(name, metav1.GetOptions{}) if err != nil { return err } @@ -4739,7 +4741,7 @@ func LaunchWebserverPod(f *Framework, podName, nodeName string) (ip string) { _, err := podClient.Create(pod) ExpectNoError(err) ExpectNoError(f.WaitForPodRunning(podName)) - createdPod, err := podClient.Get(podName) + createdPod, err := podClient.Get(podName, metav1.GetOptions{}) ExpectNoError(err) ip = fmt.Sprintf("%s:%d", createdPod.Status.PodIP, port) Logf("Target pod IP:port is %s", ip) @@ -4799,7 +4801,7 @@ func CoreDump(dir string) { func UpdatePodWithRetries(client clientset.Interface, ns, name string, update func(*v1.Pod)) (*v1.Pod, error) { for i := 0; i < 3; i++ { - pod, err := client.Core().Pods(ns).Get(name) + pod, err := client.Core().Pods(ns).Get(name, metav1.GetOptions{}) if err != nil { return nil, fmt.Errorf("Failed to get pod %q: %v", name, err) } @@ -5061,7 +5063,7 @@ func getMaster(c clientset.Interface) Address { master := Address{} // Populate the internal IP. - eps, err := c.Core().Endpoints(v1.NamespaceDefault).Get("kubernetes") + eps, err := c.Core().Endpoints(v1.NamespaceDefault).Get("kubernetes", metav1.GetOptions{}) if err != nil { Failf("Failed to get kubernetes endpoints: %v", err) } diff --git a/test/e2e/garbage_collector.go b/test/e2e/garbage_collector.go index 7eab78fe671..e1215aac2ba 100644 --- a/test/e2e/garbage_collector.go +++ b/test/e2e/garbage_collector.go @@ -182,7 +182,7 @@ var _ = framework.KubeDescribe("Garbage collector", func() { } // wait for rc to create pods if err := wait.Poll(5*time.Second, 30*time.Second, func() (bool, error) { - rc, err := rcClient.Get(rc.Name) + rc, err := rcClient.Get(rc.Name, metav1.GetOptions{}) if err != nil { return false, fmt.Errorf("Failed to get rc: %v", err) } @@ -242,7 +242,7 @@ var _ = framework.KubeDescribe("Garbage collector", func() { } // wait for rc to create some pods if err := wait.Poll(5*time.Second, 30*time.Second, func() (bool, error) { - rc, err := rcClient.Get(rc.Name) + rc, err := rcClient.Get(rc.Name, metav1.GetOptions{}) if err != nil { return false, fmt.Errorf("Failed to get rc: %v", err) } diff --git a/test/e2e/ingress_utils.go b/test/e2e/ingress_utils.go index 807d32c3b8b..96e1e11b254 100644 --- a/test/e2e/ingress_utils.go +++ b/test/e2e/ingress_utils.go @@ -44,6 +44,7 @@ import ( "google.golang.org/api/googleapi" apierrs "k8s.io/kubernetes/pkg/api/errors" extensions "k8s.io/kubernetes/pkg/apis/extensions/v1beta1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" clientset "k8s.io/kubernetes/pkg/client/clientset_generated/release_1_5" gcecloud "k8s.io/kubernetes/pkg/cloudprovider/providers/gce" "k8s.io/kubernetes/pkg/labels" @@ -307,7 +308,7 @@ func createSecret(kubeClient clientset.Interface, ing *extensions.Ingress) (host }, } var s *v1.Secret - if s, err = kubeClient.Core().Secrets(ing.Namespace).Get(tls.SecretName); err == nil { + if s, err = kubeClient.Core().Secrets(ing.Namespace).Get(tls.SecretName, metav1.GetOptions{}); err == nil { // TODO: Retry the update. We don't really expect anything to conflict though. framework.Logf("Updating secret %v in ns %v with hosts %v for ingress %v", secret.Name, secret.Namespace, host, ing.Name) s.Data = secret.Data @@ -767,7 +768,7 @@ func (j *testJig) update(update func(ing *extensions.Ingress)) { var err error ns, name := j.ing.Namespace, j.ing.Name for i := 0; i < 3; i++ { - j.ing, err = j.client.Extensions().Ingresses(ns).Get(name) + j.ing, err = j.client.Extensions().Ingresses(ns).Get(name, metav1.GetOptions{}) if err != nil { framework.Failf("failed to get ingress %q: %v", name, err) } @@ -878,7 +879,7 @@ func ingFromManifest(fileName string) *extensions.Ingress { func (cont *GCEIngressController) getL7AddonUID() (string, error) { framework.Logf("Retrieving UID from config map: %v/%v", api.NamespaceSystem, uidConfigMap) - cm, err := cont.c.Core().ConfigMaps(api.NamespaceSystem).Get(uidConfigMap) + cm, err := cont.c.Core().ConfigMaps(api.NamespaceSystem).Get(uidConfigMap, metav1.GetOptions{}) if err != nil { return "", err } @@ -932,7 +933,7 @@ func (cont *NginxIngressController) init() { framework.Logf("initializing nginx ingress controller") framework.RunKubectlOrDie("create", "-f", mkpath("rc.yaml"), fmt.Sprintf("--namespace=%v", cont.ns)) - rc, err := cont.c.Core().ReplicationControllers(cont.ns).Get("nginx-ingress-controller") + rc, err := cont.c.Core().ReplicationControllers(cont.ns).Get("nginx-ingress-controller", metav1.GetOptions{}) framework.ExpectNoError(err) cont.rc = rc diff --git a/test/e2e/job.go b/test/e2e/job.go index 6bde390f701..70807d00712 100644 --- a/test/e2e/job.go +++ b/test/e2e/job.go @@ -24,6 +24,7 @@ import ( "k8s.io/kubernetes/pkg/api/v1" batchinternal "k8s.io/kubernetes/pkg/apis/batch" batch "k8s.io/kubernetes/pkg/apis/batch/v1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" clientset "k8s.io/kubernetes/pkg/client/clientset_generated/release_1_5" "k8s.io/kubernetes/pkg/kubectl" "k8s.io/kubernetes/pkg/labels" @@ -276,7 +277,7 @@ func newTestJob(behavior, name string, rPol v1.RestartPolicy, parallelism, compl } func getJob(c clientset.Interface, ns, name string) (*batch.Job, error) { - return c.Batch().Jobs(ns).Get(name) + return c.Batch().Jobs(ns).Get(name, metav1.GetOptions{}) } func createJob(c clientset.Interface, ns string, job *batch.Job) (*batch.Job, error) { @@ -313,7 +314,7 @@ func waitForAllPodsRunning(c clientset.Interface, ns, jobName string, parallelis // Wait for job to reach completions. func waitForJobFinish(c clientset.Interface, ns, jobName string, completions int32) error { return wait.Poll(framework.Poll, jobTimeout, func() (bool, error) { - curr, err := c.Batch().Jobs(ns).Get(jobName) + curr, err := c.Batch().Jobs(ns).Get(jobName, metav1.GetOptions{}) if err != nil { return false, err } @@ -324,7 +325,7 @@ func waitForJobFinish(c clientset.Interface, ns, jobName string, completions int // Wait for job fail. func waitForJobFail(c clientset.Interface, ns, jobName string, timeout time.Duration) error { return wait.Poll(framework.Poll, timeout, func() (bool, error) { - curr, err := c.Batch().Jobs(ns).Get(jobName) + curr, err := c.Batch().Jobs(ns).Get(jobName, metav1.GetOptions{}) if err != nil { return false, err } diff --git a/test/e2e/kibana_logging.go b/test/e2e/kibana_logging.go index 9bf466e1cb6..d88a328010d 100644 --- a/test/e2e/kibana_logging.go +++ b/test/e2e/kibana_logging.go @@ -21,6 +21,7 @@ import ( "k8s.io/kubernetes/pkg/api" "k8s.io/kubernetes/pkg/api/v1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" "k8s.io/kubernetes/pkg/labels" "k8s.io/kubernetes/test/e2e/framework" @@ -60,7 +61,7 @@ func ClusterLevelLoggingWithKibana(f *framework.Framework) { // being run as the first e2e test just after the e2e cluster has been created. var err error for start := time.Now(); time.Since(start) < graceTime; time.Sleep(5 * time.Second) { - if _, err = s.Get("kibana-logging"); err == nil { + if _, err = s.Get("kibana-logging", metav1.GetOptions{}); err == nil { break } framework.Logf("Attempt to check for the existence of the Kibana service failed after %v", time.Since(start)) diff --git a/test/e2e/kubectl.go b/test/e2e/kubectl.go index c3cc1d664f8..3a91b65ae04 100644 --- a/test/e2e/kubectl.go +++ b/test/e2e/kubectl.go @@ -205,7 +205,7 @@ var _ = framework.KubeDescribe("Kubectl alpha client", func() { framework.RunKubectlOrDie("run", sjName, "--restart=OnFailure", "--generator=scheduledjob/v2alpha1", "--schedule="+schedule, "--image="+busyboxImage, nsFlag) By("verifying the ScheduledJob " + sjName + " was created") - sj, err := c.BatchV2alpha1().CronJobs(ns).Get(sjName) + sj, err := c.BatchV2alpha1().CronJobs(ns).Get(sjName, metav1.GetOptions{}) if err != nil { framework.Failf("Failed getting ScheduledJob %s: %v", sjName, err) } @@ -242,7 +242,7 @@ var _ = framework.KubeDescribe("Kubectl alpha client", func() { framework.RunKubectlOrDie("run", cjName, "--restart=OnFailure", "--generator=cronjob/v2alpha1", "--schedule="+schedule, "--image="+busyboxImage, nsFlag) By("verifying the CronJob " + cjName + " was created") - sj, err := c.BatchV2alpha1().CronJobs(ns).Get(cjName) + sj, err := c.BatchV2alpha1().CronJobs(ns).Get(cjName, metav1.GetOptions{}) if err != nil { framework.Failf("Failed getting CronJob %s: %v", cjName, err) } @@ -756,7 +756,7 @@ var _ = framework.KubeDescribe("Kubectl client", func() { }) validateService := func(name string, servicePort int, timeout time.Duration) { err := wait.Poll(framework.Poll, timeout, func() (bool, error) { - endpoints, err := c.Core().Endpoints(ns).Get(name) + endpoints, err := c.Core().Endpoints(ns).Get(name, metav1.GetOptions{}) if err != nil { // log the real error framework.Logf("Get endpoints failed (interval %v): %v", framework.Poll, err) @@ -787,7 +787,7 @@ var _ = framework.KubeDescribe("Kubectl client", func() { }) Expect(err).NotTo(HaveOccurred()) - service, err := c.Core().Services(ns).Get(name) + service, err := c.Core().Services(ns).Get(name, metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred()) if len(service.Spec.Ports) != 1 { @@ -1016,7 +1016,7 @@ var _ = framework.KubeDescribe("Kubectl client", func() { By("running the image " + nginxImage) framework.RunKubectlOrDie("run", rcName, "--image="+nginxImage, "--generator=run/v1", nsFlag) By("verifying the rc " + rcName + " was created") - rc, err := c.Core().ReplicationControllers(ns).Get(rcName) + rc, err := c.Core().ReplicationControllers(ns).Get(rcName, metav1.GetOptions{}) if err != nil { framework.Failf("Failed getting rc %s: %v", rcName, err) } @@ -1072,7 +1072,7 @@ var _ = framework.KubeDescribe("Kubectl client", func() { By("running the image " + nginxImage) framework.RunKubectlOrDie("run", rcName, "--image="+nginxImage, "--generator=run/v1", nsFlag) By("verifying the rc " + rcName + " was created") - rc, err := c.Core().ReplicationControllers(ns).Get(rcName) + rc, err := c.Core().ReplicationControllers(ns).Get(rcName, metav1.GetOptions{}) if err != nil { framework.Failf("Failed getting rc %s: %v", rcName, err) } @@ -1118,7 +1118,7 @@ var _ = framework.KubeDescribe("Kubectl client", func() { By("running the image " + nginxImage) framework.RunKubectlOrDie("run", dName, "--image="+nginxImage, "--generator=deployment/v1beta1", nsFlag) By("verifying the deployment " + dName + " was created") - d, err := c.Extensions().Deployments(ns).Get(dName) + d, err := c.Extensions().Deployments(ns).Get(dName, metav1.GetOptions{}) if err != nil { framework.Failf("Failed getting deployment %s: %v", dName, err) } @@ -1160,7 +1160,7 @@ var _ = framework.KubeDescribe("Kubectl client", func() { By("running the image " + nginxImage) framework.RunKubectlOrDie("run", jobName, "--restart=OnFailure", "--generator=job/v1", "--image="+nginxImage, nsFlag) By("verifying the job " + jobName + " was created") - job, err := c.Batch().Jobs(ns).Get(jobName) + job, err := c.Batch().Jobs(ns).Get(jobName, metav1.GetOptions{}) if err != nil { framework.Failf("Failed getting job %s: %v", jobName, err) } @@ -1193,7 +1193,7 @@ var _ = framework.KubeDescribe("Kubectl client", func() { By("running the image " + nginxImage) framework.RunKubectlOrDie("run", podName, "--restart=Never", "--generator=run-pod/v1", "--image="+nginxImage, nsFlag) By("verifying the pod " + podName + " was created") - pod, err := c.Core().Pods(ns).Get(podName) + pod, err := c.Core().Pods(ns).Get(podName, metav1.GetOptions{}) if err != nil { framework.Failf("Failed getting pod %s: %v", podName, err) } @@ -1244,7 +1244,7 @@ var _ = framework.KubeDescribe("Kubectl client", func() { framework.RunKubectlOrDieInput(podJson, "replace", "-f", "-", nsFlag) By("verifying the pod " + podName + " has the right image " + busyboxImage) - pod, err := c.Core().Pods(ns).Get(podName) + pod, err := c.Core().Pods(ns).Get(podName, metav1.GetOptions{}) if err != nil { framework.Failf("Failed getting deployment %s: %v", podName, err) } @@ -1276,7 +1276,7 @@ var _ = framework.KubeDescribe("Kubectl client", func() { Expect(runOutput).To(ContainSubstring("stdin closed")) By("verifying the job " + jobName + " was deleted") - _, err := c.Batch().Jobs(ns).Get(jobName) + _, err := c.Batch().Jobs(ns).Get(jobName, metav1.GetOptions{}) Expect(err).To(HaveOccurred()) Expect(apierrs.IsNotFound(err)).To(BeTrue()) }) @@ -1425,7 +1425,7 @@ var _ = framework.KubeDescribe("Kubectl client", func() { framework.RunKubectlOrDie("create", "quota", quotaName, "--hard=pods=1000000,services=1000000", nsFlag) By("verifying that the quota was created") - quota, err := c.Core().ResourceQuotas(ns).Get(quotaName) + quota, err := c.Core().ResourceQuotas(ns).Get(quotaName, metav1.GetOptions{}) if err != nil { framework.Failf("Failed getting quota %s: %v", quotaName, err) } @@ -1455,7 +1455,7 @@ var _ = framework.KubeDescribe("Kubectl client", func() { framework.RunKubectlOrDie("create", "quota", quotaName, "--hard=pods=1000000", "--scopes=BestEffort,NotTerminating", nsFlag) By("verifying that the quota was created") - quota, err := c.Core().ResourceQuotas(ns).Get(quotaName) + quota, err := c.Core().ResourceQuotas(ns).Get(quotaName, metav1.GetOptions{}) if err != nil { framework.Failf("Failed getting quota %s: %v", quotaName, err) } diff --git a/test/e2e/kubelet.go b/test/e2e/kubelet.go index 02b2610414c..c388737ff1c 100644 --- a/test/e2e/kubelet.go +++ b/test/e2e/kubelet.go @@ -22,6 +22,7 @@ import ( "time" "k8s.io/kubernetes/pkg/api/v1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" clientset "k8s.io/kubernetes/pkg/client/clientset_generated/release_1_5" "k8s.io/kubernetes/pkg/util/sets" "k8s.io/kubernetes/pkg/util/uuid" @@ -101,7 +102,7 @@ func updateNodeLabels(c clientset.Interface, nodeNames sets.String, toAdd, toRem var node *v1.Node var err error for i := 0; i < maxRetries; i++ { - node, err = c.Core().Nodes().Get(nodeName) + node, err = c.Core().Nodes().Get(nodeName, metav1.GetOptions{}) if err != nil { framework.Logf("Error getting node %s: %v", nodeName, err) continue diff --git a/test/e2e/limit_range.go b/test/e2e/limit_range.go index 7a52af3bc96..0a2faa478dd 100644 --- a/test/e2e/limit_range.go +++ b/test/e2e/limit_range.go @@ -21,6 +21,7 @@ import ( "k8s.io/kubernetes/pkg/api/resource" "k8s.io/kubernetes/pkg/api/v1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" "k8s.io/kubernetes/test/e2e/framework" . "github.com/onsi/ginkgo" @@ -46,7 +47,7 @@ var _ = framework.KubeDescribe("LimitRange", func() { Expect(err).NotTo(HaveOccurred()) By("Fetching the LimitRange to ensure it has proper values") - limitRange, err = f.ClientSet.Core().LimitRanges(f.Namespace.Name).Get(limitRange.Name) + limitRange, err = f.ClientSet.Core().LimitRanges(f.Namespace.Name).Get(limitRange.Name, metav1.GetOptions{}) expected := v1.ResourceRequirements{Requests: defaultRequest, Limits: defaultLimit} actual := v1.ResourceRequirements{Requests: limitRange.Spec.Limits[0].DefaultRequest, Limits: limitRange.Spec.Limits[0].Default} err = equalResourceRequirement(expected, actual) @@ -58,7 +59,7 @@ var _ = framework.KubeDescribe("LimitRange", func() { Expect(err).NotTo(HaveOccurred()) By("Ensuring Pod has resource requirements applied from LimitRange") - pod, err = f.ClientSet.Core().Pods(f.Namespace.Name).Get(pod.Name) + pod, err = f.ClientSet.Core().Pods(f.Namespace.Name).Get(pod.Name, metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred()) for i := range pod.Spec.Containers { err = equalResourceRequirement(expected, pod.Spec.Containers[i].Resources) @@ -75,7 +76,7 @@ var _ = framework.KubeDescribe("LimitRange", func() { Expect(err).NotTo(HaveOccurred()) By("Ensuring Pod has merged resource requirements applied from LimitRange") - pod, err = f.ClientSet.Core().Pods(f.Namespace.Name).Get(pod.Name) + pod, err = f.ClientSet.Core().Pods(f.Namespace.Name).Get(pod.Name, metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred()) // This is an interesting case, so it's worth a comment // If you specify a Limit, and no Request, the Limit will default to the Request diff --git a/test/e2e/mesos.go b/test/e2e/mesos.go index 4fcef61fb63..0e3b37cf238 100644 --- a/test/e2e/mesos.go +++ b/test/e2e/mesos.go @@ -101,7 +101,7 @@ var _ = framework.KubeDescribe("Mesos", func() { framework.ExpectNoError(err) framework.ExpectNoError(framework.WaitForPodNameRunningInNamespace(c, podName, ns)) - pod, err := c.Core().Pods(ns).Get(podName) + pod, err := c.Core().Pods(ns).Get(podName, metav1.GetOptions{}) framework.ExpectNoError(err) nodeClient := f.ClientSet.Core().Nodes() diff --git a/test/e2e/namespace.go b/test/e2e/namespace.go index 97c62323f6b..6afbfe508ba 100644 --- a/test/e2e/namespace.go +++ b/test/e2e/namespace.go @@ -24,6 +24,7 @@ import ( "k8s.io/kubernetes/pkg/api/errors" "k8s.io/kubernetes/pkg/api/v1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" "k8s.io/kubernetes/pkg/util/intstr" "k8s.io/kubernetes/pkg/util/wait" "k8s.io/kubernetes/test/e2e/framework" @@ -116,7 +117,7 @@ func ensurePodsAreRemovedWhenNamespaceIsDeleted(f *framework.Framework) { maxWaitSeconds := int64(60) + *pod.Spec.TerminationGracePeriodSeconds framework.ExpectNoError(wait.Poll(1*time.Second, time.Duration(maxWaitSeconds)*time.Second, func() (bool, error) { - _, err = f.ClientSet.Core().Namespaces().Get(namespace.Name) + _, err = f.ClientSet.Core().Namespaces().Get(namespace.Name, metav1.GetOptions{}) if err != nil && errors.IsNotFound(err) { return true, nil } @@ -124,7 +125,7 @@ func ensurePodsAreRemovedWhenNamespaceIsDeleted(f *framework.Framework) { })) By("Verifying there is no pod in the namespace") - _, err = f.ClientSet.Core().Pods(namespace.Name).Get(pod.Name) + _, err = f.ClientSet.Core().Pods(namespace.Name).Get(pod.Name, metav1.GetOptions{}) Expect(err).To(HaveOccurred()) } @@ -168,7 +169,7 @@ func ensureServicesAreRemovedWhenNamespaceIsDeleted(f *framework.Framework) { maxWaitSeconds := int64(60) framework.ExpectNoError(wait.Poll(1*time.Second, time.Duration(maxWaitSeconds)*time.Second, func() (bool, error) { - _, err = f.ClientSet.Core().Namespaces().Get(namespace.Name) + _, err = f.ClientSet.Core().Namespaces().Get(namespace.Name, metav1.GetOptions{}) if err != nil && errors.IsNotFound(err) { return true, nil } @@ -176,7 +177,7 @@ func ensureServicesAreRemovedWhenNamespaceIsDeleted(f *framework.Framework) { })) By("Verifying there is no service in the namespace") - _, err = f.ClientSet.Core().Services(namespace.Name).Get(service.Name) + _, err = f.ClientSet.Core().Services(namespace.Name).Get(service.Name, metav1.GetOptions{}) Expect(err).To(HaveOccurred()) } diff --git a/test/e2e/network_partition.go b/test/e2e/network_partition.go index c54a3fbf67b..96099ef968a 100644 --- a/test/e2e/network_partition.go +++ b/test/e2e/network_partition.go @@ -23,6 +23,7 @@ import ( "k8s.io/kubernetes/pkg/api" "k8s.io/kubernetes/pkg/api/v1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" "k8s.io/kubernetes/pkg/client/cache" clientset "k8s.io/kubernetes/pkg/client/clientset_generated/release_1_5" "k8s.io/kubernetes/pkg/fields" @@ -268,7 +269,7 @@ var _ = framework.KubeDescribe("Network Partition [Disruptive] [Slow]", func() { Expect(err).NotTo(HaveOccurred()) nodeName := pods.Items[0].Spec.NodeName - node, err := c.Core().Nodes().Get(nodeName) + node, err := c.Core().Nodes().Get(nodeName, metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred()) // This creates a temporary network partition, verifies that 'podNameToDisappear', @@ -306,7 +307,7 @@ var _ = framework.KubeDescribe("Network Partition [Disruptive] [Slow]", func() { // verify that it is really on the requested node { - pod, err := c.Core().Pods(ns).Get(additionalPod) + pod, err := c.Core().Pods(ns).Get(additionalPod, metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred()) if pod.Spec.NodeName != node.Name { framework.Logf("Pod %s found on invalid node: %s instead of %s", pod.Name, pod.Spec.NodeName, node.Name) @@ -333,7 +334,7 @@ var _ = framework.KubeDescribe("Network Partition [Disruptive] [Slow]", func() { Expect(err).NotTo(HaveOccurred()) nodeName := pods.Items[0].Spec.NodeName - node, err := c.Core().Nodes().Get(nodeName) + node, err := c.Core().Nodes().Get(nodeName, metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred()) // This creates a temporary network partition, verifies that 'podNameToDisappear', @@ -411,7 +412,7 @@ var _ = framework.KubeDescribe("Network Partition [Disruptive] [Slow]", func() { pst.waitForRunningAndReady(*ps.Spec.Replicas, ps) pod := pst.getPodList(ps).Items[0] - node, err := c.Core().Nodes().Get(pod.Spec.NodeName) + node, err := c.Core().Nodes().Get(pod.Spec.NodeName, metav1.GetOptions{}) framework.ExpectNoError(err) // Blocks outgoing network traffic on 'node'. Then verifies that 'podNameToDisappear', @@ -453,7 +454,7 @@ var _ = framework.KubeDescribe("Network Partition [Disruptive] [Slow]", func() { Expect(err).NotTo(HaveOccurred()) nodeName := pods.Items[0].Spec.NodeName - node, err := c.Core().Nodes().Get(nodeName) + node, err := c.Core().Nodes().Get(nodeName, metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred()) // This creates a temporary network partition, verifies that the job has 'parallelism' number of diff --git a/test/e2e/node_problem_detector.go b/test/e2e/node_problem_detector.go index d59227ed8c5..28cdc7d8bd4 100644 --- a/test/e2e/node_problem_detector.go +++ b/test/e2e/node_problem_detector.go @@ -24,6 +24,7 @@ import ( "k8s.io/kubernetes/pkg/api" "k8s.io/kubernetes/pkg/api/v1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" clientset "k8s.io/kubernetes/pkg/client/clientset_generated/release_1_5" coreclientset "k8s.io/kubernetes/pkg/client/clientset_generated/release_1_5/typed/core/v1" "k8s.io/kubernetes/pkg/fields" @@ -411,7 +412,7 @@ func verifyNoEvents(e coreclientset.EventInterface, options v1.ListOptions) erro // verifyCondition verifies specific node condition is generated, if reason and message are empty, they will not be checked func verifyCondition(n coreclientset.NodeInterface, nodeName string, condition v1.NodeConditionType, status v1.ConditionStatus, reason, message string) error { - node, err := n.Get(nodeName) + node, err := n.Get(nodeName, metav1.GetOptions{}) if err != nil { return err } diff --git a/test/e2e/nodeoutofdisk.go b/test/e2e/nodeoutofdisk.go index eafbeb22b8e..e9763a6f032 100644 --- a/test/e2e/nodeoutofdisk.go +++ b/test/e2e/nodeoutofdisk.go @@ -24,6 +24,7 @@ import ( cadvisorapi "github.com/google/cadvisor/info/v1" "k8s.io/kubernetes/pkg/api/resource" "k8s.io/kubernetes/pkg/api/v1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" clientset "k8s.io/kubernetes/pkg/client/clientset_generated/release_1_5" "k8s.io/kubernetes/pkg/fields" "k8s.io/kubernetes/pkg/util/wait" @@ -98,7 +99,7 @@ var _ = framework.KubeDescribe("NodeOutOfDisk [Serial] [Flaky] [Disruptive]", fu }) It("runs out of disk space", func() { - unfilledNode, err := c.Core().Nodes().Get(unfilledNodeName) + unfilledNode, err := c.Core().Nodes().Get(unfilledNodeName, metav1.GetOptions{}) framework.ExpectNoError(err) By(fmt.Sprintf("Calculating CPU availability on node %s", unfilledNode.Name)) @@ -122,7 +123,7 @@ var _ = framework.KubeDescribe("NodeOutOfDisk [Serial] [Flaky] [Disruptive]", fu createOutOfDiskPod(c, ns, name, podCPU) framework.ExpectNoError(f.WaitForPodRunning(name)) - pod, err := podClient.Get(name) + pod, err := podClient.Get(name, metav1.GetOptions{}) framework.ExpectNoError(err) Expect(pod.Spec.NodeName).To(Equal(unfilledNodeName)) } @@ -161,7 +162,7 @@ var _ = framework.KubeDescribe("NodeOutOfDisk [Serial] [Flaky] [Disruptive]", fu By(fmt.Sprintf("Verifying that pod %s schedules on node %s", pendingPodName, recoveredNodeName)) framework.ExpectNoError(f.WaitForPodRunning(pendingPodName)) - pendingPod, err := podClient.Get(pendingPodName) + pendingPod, err := podClient.Get(pendingPodName, metav1.GetOptions{}) framework.ExpectNoError(err) Expect(pendingPod.Spec.NodeName).To(Equal(recoveredNodeName)) }) diff --git a/test/e2e/pd.go b/test/e2e/pd.go index 8c93ce7830d..f808001405b 100644 --- a/test/e2e/pd.go +++ b/test/e2e/pd.go @@ -724,7 +724,7 @@ func waitForPDInVolumesInUse( "Waiting for node %s's VolumesInUse Status %s PD %q", nodeName, logStr, diskName) for start := time.Now(); time.Since(start) < timeout; time.Sleep(nodeStatusPollTime) { - nodeObj, err := nodeClient.Get(string(nodeName)) + nodeObj, err := nodeClient.Get(string(nodeName), metav1.GetOptions{}) if err != nil || nodeObj == nil { framework.Logf( "Failed to fetch node object %q from API server. err=%v", diff --git a/test/e2e/persistent_volumes.go b/test/e2e/persistent_volumes.go index 18caf9ac386..9596a100c2a 100644 --- a/test/e2e/persistent_volumes.go +++ b/test/e2e/persistent_volumes.go @@ -69,7 +69,7 @@ func pvPvcCleanup(c clientset.Interface, ns string, pvols pvmap, claims pvcmap) if c != nil && len(ns) > 0 { for pvcKey := range claims { - _, err := c.Core().PersistentVolumeClaims(pvcKey.Namespace).Get(pvcKey.Name) + _, err := c.Core().PersistentVolumeClaims(pvcKey.Namespace).Get(pvcKey.Name, metav1.GetOptions{}) if !apierrs.IsNotFound(err) { Expect(err).NotTo(HaveOccurred()) framework.Logf(" deleting PVC %v ...", pvcKey) @@ -81,7 +81,7 @@ func pvPvcCleanup(c clientset.Interface, ns string, pvols pvmap, claims pvcmap) } for name := range pvols { - _, err := c.Core().PersistentVolumes().Get(name) + _, err := c.Core().PersistentVolumes().Get(name, metav1.GetOptions{}) if !apierrs.IsNotFound(err) { Expect(err).NotTo(HaveOccurred()) framework.Logf(" deleting PV %v ...", name) @@ -106,7 +106,7 @@ func deletePVCandValidatePV(c clientset.Interface, ns string, pvc *v1.Persistent Expect(err).NotTo(HaveOccurred()) // Check that the PVC is really deleted. - pvc, err = c.Core().PersistentVolumeClaims(ns).Get(pvc.Name) + pvc, err = c.Core().PersistentVolumeClaims(ns).Get(pvc.Name, metav1.GetOptions{}) Expect(apierrs.IsNotFound(err)).To(BeTrue()) // Wait for the PV's phase to return to the expected Phase @@ -115,7 +115,7 @@ func deletePVCandValidatePV(c clientset.Interface, ns string, pvc *v1.Persistent Expect(err).NotTo(HaveOccurred()) // examine the pv's ClaimRef and UID and compare to expected values - pv, err = c.Core().PersistentVolumes().Get(pv.Name) + pv, err = c.Core().PersistentVolumes().Get(pv.Name, metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred()) cr := pv.Spec.ClaimRef if expctPVPhase == v1.VolumeAvailable { @@ -140,7 +140,7 @@ func deletePVCandValidatePVGroup(c clientset.Interface, ns string, pvols pvmap, var expctPVPhase v1.PersistentVolumePhase for pvName := range pvols { - pv, err := c.Core().PersistentVolumes().Get(pvName) + pv, err := c.Core().PersistentVolumes().Get(pvName, metav1.GetOptions{}) Expect(apierrs.IsNotFound(err)).To(BeFalse()) cr := pv.Spec.ClaimRef // if pv is bound then delete the pvc it is bound to @@ -151,7 +151,7 @@ func deletePVCandValidatePVGroup(c clientset.Interface, ns string, pvols pvmap, pvcKey := makePvcKey(ns, cr.Name) _, found := claims[pvcKey] Expect(found).To(BeTrue()) - pvc, err := c.Core().PersistentVolumeClaims(ns).Get(cr.Name) + pvc, err := c.Core().PersistentVolumeClaims(ns).Get(cr.Name, metav1.GetOptions{}) Expect(apierrs.IsNotFound(err)).To(BeFalse()) // what Phase do we expect the PV that was bound to the claim to @@ -305,11 +305,11 @@ func waitOnPVandPVC(c clientset.Interface, ns string, pv *v1.PersistentVolume, p Expect(err).NotTo(HaveOccurred()) // Re-get the pv and pvc objects - pv, err = c.Core().PersistentVolumes().Get(pv.Name) + pv, err = c.Core().PersistentVolumes().Get(pv.Name, metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred()) // Re-get the pvc and - pvc, err = c.Core().PersistentVolumeClaims(ns).Get(pvc.Name) + pvc, err = c.Core().PersistentVolumeClaims(ns).Get(pvc.Name, metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred()) // The pv and pvc are both bound, but to each other? @@ -343,7 +343,7 @@ func waitAndVerifyBinds(c clientset.Interface, ns string, pvols pvmap, claims pv } Expect(err).NotTo(HaveOccurred()) - pv, err := c.Core().PersistentVolumes().Get(pvName) + pv, err := c.Core().PersistentVolumes().Get(pvName, metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred()) if cr := pv.Spec.ClaimRef; cr != nil && len(cr.Name) > 0 { // Assert bound pvc is a test resource. Failing assertion could @@ -435,7 +435,7 @@ func completeMultiTest(f *framework.Framework, c clientset.Interface, ns string, // 1. verify each PV permits write access to a client pod By("Checking pod has write access to PersistentVolumes") for pvcKey := range claims { - pvc, err := c.Core().PersistentVolumeClaims(pvcKey.Namespace).Get(pvcKey.Name) + pvc, err := c.Core().PersistentVolumeClaims(pvcKey.Namespace).Get(pvcKey.Name, metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred()) if len(pvc.Spec.VolumeName) == 0 { continue // claim is not bound @@ -503,7 +503,7 @@ var _ = framework.KubeDescribe("PersistentVolumes", func() { AfterEach(func() { if c != nil && len(ns) > 0 { if pvc != nil && len(pvc.Name) > 0 { - _, err := c.Core().PersistentVolumeClaims(ns).Get(pvc.Name) + _, err := c.Core().PersistentVolumeClaims(ns).Get(pvc.Name, metav1.GetOptions{}) if !apierrs.IsNotFound(err) { Expect(err).NotTo(HaveOccurred()) framework.Logf("AfterEach: deleting PVC %v", pvc.Name) @@ -515,7 +515,7 @@ var _ = framework.KubeDescribe("PersistentVolumes", func() { pvc = nil if pv != nil && len(pv.Name) > 0 { - _, err := c.Core().PersistentVolumes().Get(pv.Name) + _, err := c.Core().PersistentVolumes().Get(pv.Name, metav1.GetOptions{}) if !apierrs.IsNotFound(err) { Expect(err).NotTo(HaveOccurred()) framework.Logf("AfterEach: deleting PV %v", pv.Name) diff --git a/test/e2e/petset.go b/test/e2e/petset.go index 99f5b1019ee..e30cf4383bc 100644 --- a/test/e2e/petset.go +++ b/test/e2e/petset.go @@ -240,7 +240,7 @@ var _ = framework.KubeDescribe("StatefulSet", func() { By("Verifying that the 2nd pod wont be removed if it is not running and ready") pst.confirmPetCount(2, ps, 10*time.Second) expectedPodName := ps.Name + "-1" - expectedPod, err := f.ClientSet.Core().Pods(ns).Get(expectedPodName) + expectedPod, err := f.ClientSet.Core().Pods(ns).Get(expectedPodName, metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred()) watcher, err := f.ClientSet.Core().Pods(ns).Watch(v1.SingleObject( v1.ObjectMeta{ @@ -488,7 +488,7 @@ var _ = framework.KubeDescribe("Stateful Set recreate", func() { By("waiting when pet pod " + petPodName + " will be recreated in namespace " + f.Namespace.Name + " and will be in running state") // we may catch delete event, thats why we are waiting for running phase like this, and not with watch.Until Eventually(func() error { - petPod, err := f.ClientSet.Core().Pods(f.Namespace.Name).Get(petPodName) + petPod, err := f.ClientSet.Core().Pods(f.Namespace.Name).Get(petPodName, metav1.GetOptions{}) if err != nil { return err } @@ -812,7 +812,7 @@ type verifyPodFunc func(*v1.Pod) func (p *statefulSetTester) verifyPodAtIndex(index int, ps *apps.StatefulSet, verify verifyPodFunc) { name := getPodNameAtIndex(index, ps) - pod, err := p.c.Core().Pods(ps.Namespace).Get(name) + pod, err := p.c.Core().Pods(ps.Namespace).Get(name, metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred(), fmt.Sprintf("Failed to get stateful pod %s for StatefulSet %s/%s", name, ps.Namespace, ps.Name)) verify(pod) } @@ -861,7 +861,7 @@ func (p *statefulSetTester) restart(ps *apps.StatefulSet) { func (p *statefulSetTester) update(ns, name string, update func(ps *apps.StatefulSet)) { for i := 0; i < 3; i++ { - ps, err := p.c.Apps().StatefulSets(ns).Get(name) + ps, err := p.c.Apps().StatefulSets(ns).Get(name, metav1.GetOptions{}) if err != nil { framework.Failf("failed to get statefulset %q: %v", name, err) } @@ -979,7 +979,7 @@ func (p *statefulSetTester) waitForStatus(ps *apps.StatefulSet, expectedReplicas ns, name := ps.Namespace, ps.Name pollErr := wait.PollImmediate(statefulsetPoll, statefulsetTimeout, func() (bool, error) { - psGet, err := p.c.Apps().StatefulSets(ns).Get(name) + psGet, err := p.c.Apps().StatefulSets(ns).Get(name, metav1.GetOptions{}) if err != nil { return false, err } diff --git a/test/e2e/pods.go b/test/e2e/pods.go index ee4779f8b00..d6ad33cbf62 100644 --- a/test/e2e/pods.go +++ b/test/e2e/pods.go @@ -25,6 +25,7 @@ import ( "time" "k8s.io/kubernetes/pkg/api/v1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" "k8s.io/kubernetes/pkg/labels" "k8s.io/kubernetes/pkg/util/uuid" "k8s.io/kubernetes/pkg/util/wait" @@ -100,7 +101,7 @@ var _ = framework.KubeDescribe("Pods Delete Grace Period", func() { // may be carried out immediately rather than gracefully. framework.ExpectNoError(f.WaitForPodRunning(pod.Name)) // save the running pod - pod, err = podClient.Get(pod.Name) + pod, err = podClient.Get(pod.Name, metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred(), "failed to GET scheduled pod") // start local proxy, so we can send graceful deletion over query string, rather than body parameter diff --git a/test/e2e/pre_stop.go b/test/e2e/pre_stop.go index 6c2eab0a30d..dac50795fe8 100644 --- a/test/e2e/pre_stop.go +++ b/test/e2e/pre_stop.go @@ -22,6 +22,7 @@ import ( "time" "k8s.io/kubernetes/pkg/api/v1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" clientset "k8s.io/kubernetes/pkg/client/clientset_generated/release_1_5" "k8s.io/kubernetes/pkg/util/wait" "k8s.io/kubernetes/test/e2e/framework" @@ -66,7 +67,7 @@ func testPreStop(c clientset.Interface, ns string) { val := "{\"Source\": \"prestop\"}" - podOut, err := c.Core().Pods(ns).Get(podDescr.Name) + podOut, err := c.Core().Pods(ns).Get(podDescr.Name, metav1.GetOptions{}) framework.ExpectNoError(err, "getting pod info") preStopDescr := &v1.Pod{ diff --git a/test/e2e/rc.go b/test/e2e/rc.go index 6f2b8bc560b..055bdff87fb 100644 --- a/test/e2e/rc.go +++ b/test/e2e/rc.go @@ -22,6 +22,7 @@ import ( "k8s.io/kubernetes/pkg/api/resource" "k8s.io/kubernetes/pkg/api/v1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" "k8s.io/kubernetes/pkg/controller/replication" "k8s.io/kubernetes/pkg/labels" "k8s.io/kubernetes/pkg/util/uuid" @@ -165,7 +166,7 @@ func rcConditionCheck(f *framework.Framework) { Expect(err).NotTo(HaveOccurred()) err = wait.PollImmediate(1*time.Second, 1*time.Minute, func() (bool, error) { - quota, err = c.Core().ResourceQuotas(namespace).Get(name) + quota, err = c.Core().ResourceQuotas(namespace).Get(name, metav1.GetOptions{}) if err != nil { return false, err } @@ -187,7 +188,7 @@ func rcConditionCheck(f *framework.Framework) { generation := rc.Generation conditions := rc.Status.Conditions err = wait.PollImmediate(1*time.Second, 1*time.Minute, func() (bool, error) { - rc, err = c.Core().ReplicationControllers(namespace).Get(name) + rc, err = c.Core().ReplicationControllers(namespace).Get(name, metav1.GetOptions{}) if err != nil { return false, err } @@ -216,7 +217,7 @@ func rcConditionCheck(f *framework.Framework) { generation = rc.Generation conditions = rc.Status.Conditions err = wait.PollImmediate(1*time.Second, 1*time.Minute, func() (bool, error) { - rc, err = c.Core().ReplicationControllers(namespace).Get(name) + rc, err = c.Core().ReplicationControllers(namespace).Get(name, metav1.GetOptions{}) if err != nil { return false, err } diff --git a/test/e2e/reboot.go b/test/e2e/reboot.go index 3868a194f41..5dbff88ce1c 100644 --- a/test/e2e/reboot.go +++ b/test/e2e/reboot.go @@ -24,6 +24,7 @@ import ( "k8s.io/kubernetes/pkg/api" "k8s.io/kubernetes/pkg/api/v1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" clientset "k8s.io/kubernetes/pkg/client/clientset_generated/release_1_5" "k8s.io/kubernetes/pkg/fields" "k8s.io/kubernetes/pkg/labels" @@ -223,7 +224,7 @@ func rebootNode(c clientset.Interface, provider, name, rebootCmd string) bool { // Get the node initially. framework.Logf("Getting %s", name) - node, err := c.Core().Nodes().Get(name) + node, err := c.Core().Nodes().Get(name, metav1.GetOptions{}) if err != nil { framework.Logf("Couldn't get node %s", name) return false diff --git a/test/e2e/replica_set.go b/test/e2e/replica_set.go index 34fb536cb51..b36eb60bffe 100644 --- a/test/e2e/replica_set.go +++ b/test/e2e/replica_set.go @@ -179,7 +179,7 @@ func rsConditionCheck(f *framework.Framework) { Expect(err).NotTo(HaveOccurred()) err = wait.PollImmediate(1*time.Second, 1*time.Minute, func() (bool, error) { - quota, err = c.Core().ResourceQuotas(namespace).Get(name) + quota, err = c.Core().ResourceQuotas(namespace).Get(name, metav1.GetOptions{}) if err != nil { return false, err } @@ -201,7 +201,7 @@ func rsConditionCheck(f *framework.Framework) { generation := rs.Generation conditions := rs.Status.Conditions err = wait.PollImmediate(1*time.Second, 1*time.Minute, func() (bool, error) { - rs, err = c.Extensions().ReplicaSets(namespace).Get(name) + rs, err = c.Extensions().ReplicaSets(namespace).Get(name, metav1.GetOptions{}) if err != nil { return false, err } @@ -231,7 +231,7 @@ func rsConditionCheck(f *framework.Framework) { generation = rs.Generation conditions = rs.Status.Conditions err = wait.PollImmediate(1*time.Second, 1*time.Minute, func() (bool, error) { - rs, err = c.Extensions().ReplicaSets(namespace).Get(name) + rs, err = c.Extensions().ReplicaSets(namespace).Get(name, metav1.GetOptions{}) if err != nil { return false, err } diff --git a/test/e2e/resize_nodes.go b/test/e2e/resize_nodes.go index f5b32d0a8f6..a0f982c293b 100644 --- a/test/e2e/resize_nodes.go +++ b/test/e2e/resize_nodes.go @@ -214,7 +214,7 @@ func newRCByName(c clientset.Interface, ns, name string, replicas int32, gracePe } func resizeRC(c clientset.Interface, ns, name string, replicas int32) error { - rc, err := c.Core().ReplicationControllers(ns).Get(name) + rc, err := c.Core().ReplicationControllers(ns).Get(name, metav1.GetOptions{}) if err != nil { return err } diff --git a/test/e2e/resource_quota.go b/test/e2e/resource_quota.go index 2996e55b5fd..9b50f4f6846 100644 --- a/test/e2e/resource_quota.go +++ b/test/e2e/resource_quota.go @@ -22,6 +22,7 @@ import ( "k8s.io/kubernetes/pkg/api/resource" "k8s.io/kubernetes/pkg/api/v1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" clientset "k8s.io/kubernetes/pkg/client/clientset_generated/release_1_5" "k8s.io/kubernetes/pkg/util/intstr" "k8s.io/kubernetes/pkg/util/wait" @@ -645,7 +646,7 @@ func deleteResourceQuota(c clientset.Interface, namespace, name string) error { // wait for resource quota status to show the expected used resources value func waitForResourceQuota(c clientset.Interface, ns, quotaName string, used v1.ResourceList) error { return wait.Poll(framework.Poll, resourceQuotaTimeout, func() (bool, error) { - resourceQuota, err := c.Core().ResourceQuotas(ns).Get(quotaName) + resourceQuota, err := c.Core().ResourceQuotas(ns).Get(quotaName, metav1.GetOptions{}) if err != nil { return false, err } diff --git a/test/e2e/restart.go b/test/e2e/restart.go index 860ef7e833b..946ee68b8d2 100644 --- a/test/e2e/restart.go +++ b/test/e2e/restart.go @@ -22,6 +22,7 @@ import ( "k8s.io/kubernetes/pkg/api" "k8s.io/kubernetes/pkg/api/v1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" "k8s.io/kubernetes/pkg/fields" kubepod "k8s.io/kubernetes/pkg/kubelet/pod" "k8s.io/kubernetes/pkg/labels" @@ -157,7 +158,7 @@ func restartNodes(f *framework.Framework, nodeNames []string) error { // List old boot IDs. oldBootIDs := make(map[string]string) for _, name := range nodeNames { - node, err := f.ClientSet.Core().Nodes().Get(name) + node, err := f.ClientSet.Core().Nodes().Get(name, metav1.GetOptions{}) if err != nil { return fmt.Errorf("error getting node info before reboot: %s", err) } @@ -179,7 +180,7 @@ func restartNodes(f *framework.Framework, nodeNames []string) error { // Wait for their boot IDs to change. for _, name := range nodeNames { if err := wait.Poll(30*time.Second, 5*time.Minute, func() (bool, error) { - node, err := f.ClientSet.Core().Nodes().Get(name) + node, err := f.ClientSet.Core().Nodes().Get(name, metav1.GetOptions{}) if err != nil { return false, fmt.Errorf("error getting node info after reboot: %s", err) } diff --git a/test/e2e/scheduler_predicates.go b/test/e2e/scheduler_predicates.go index 499ce469b61..d260651d7ae 100644 --- a/test/e2e/scheduler_predicates.go +++ b/test/e2e/scheduler_predicates.go @@ -24,6 +24,7 @@ import ( "k8s.io/kubernetes/pkg/api/errors" "k8s.io/kubernetes/pkg/api/resource" "k8s.io/kubernetes/pkg/api/v1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" clientset "k8s.io/kubernetes/pkg/client/clientset_generated/release_1_5" "k8s.io/kubernetes/pkg/util/sets" "k8s.io/kubernetes/pkg/util/uuid" @@ -59,7 +60,7 @@ var _ = framework.KubeDescribe("SchedulerPredicates [Serial]", func() { ignoreLabels := framework.ImagePullerLabels AfterEach(func() { - rc, err := cs.Core().ReplicationControllers(ns).Get(RCName) + rc, err := cs.Core().ReplicationControllers(ns).Get(RCName, metav1.GetOptions{}) if err == nil && *(rc.Spec.Replicas) != 0 { By("Cleaning up the replication controller") err := framework.DeleteRCAndPods(f.ClientSet, f.InternalClientset, ns, RCName) @@ -284,7 +285,7 @@ var _ = framework.KubeDescribe("SchedulerPredicates [Serial]", func() { // already when the kubelet does not know about its new label yet. The // kubelet will then refuse to launch the pod. framework.ExpectNoError(framework.WaitForPodNotPending(cs, ns, labelPodName, pod.ResourceVersion)) - labelPod, err := cs.Core().Pods(ns).Get(labelPodName) + labelPod, err := cs.Core().Pods(ns).Get(labelPodName, metav1.GetOptions{}) framework.ExpectNoError(err) Expect(labelPod.Spec.NodeName).To(Equal(nodeName)) }) @@ -368,7 +369,7 @@ var _ = framework.KubeDescribe("SchedulerPredicates [Serial]", func() { // already when the kubelet does not know about its new label yet. The // kubelet will then refuse to launch the pod. framework.ExpectNoError(framework.WaitForPodNotPending(cs, ns, labelPodName, pod.ResourceVersion)) - labelPod, err := cs.Core().Pods(ns).Get(labelPodName) + labelPod, err := cs.Core().Pods(ns).Get(labelPodName, metav1.GetOptions{}) framework.ExpectNoError(err) Expect(labelPod.Spec.NodeName).To(Equal(nodeName)) }) @@ -393,7 +394,7 @@ var _ = framework.KubeDescribe("SchedulerPredicates [Serial]", func() { // already when the kubelet does not know about its new label yet. The // kubelet will then refuse to launch the pod. framework.ExpectNoError(framework.WaitForPodNotPending(cs, ns, pod.Name, "")) - labelPod, err := cs.Core().Pods(ns).Get(pod.Name) + labelPod, err := cs.Core().Pods(ns).Get(pod.Name, metav1.GetOptions{}) framework.ExpectNoError(err) Expect(labelPod.Spec.NodeName).To(Equal(nodeName)) }) @@ -499,7 +500,7 @@ var _ = framework.KubeDescribe("SchedulerPredicates [Serial]", func() { // already when the kubelet does not know about its new label yet. The // kubelet will then refuse to launch the pod. framework.ExpectNoError(framework.WaitForPodNotPending(cs, ns, labelPodName, pod.ResourceVersion)) - labelPod, err := cs.Core().Pods(ns).Get(labelPodName) + labelPod, err := cs.Core().Pods(ns).Get(labelPodName, metav1.GetOptions{}) framework.ExpectNoError(err) Expect(labelPod.Spec.NodeName).To(Equal(nodeName)) }) @@ -610,7 +611,7 @@ var _ = framework.KubeDescribe("SchedulerPredicates [Serial]", func() { // already when the kubelet does not know about its new label yet. The // kubelet will then refuse to launch the pod. framework.ExpectNoError(framework.WaitForPodNotPending(cs, ns, labelPodName, pod.ResourceVersion)) - labelPod, err := cs.Core().Pods(ns).Get(labelPodName) + labelPod, err := cs.Core().Pods(ns).Get(labelPodName, metav1.GetOptions{}) framework.ExpectNoError(err) Expect(labelPod.Spec.NodeName).To(Equal(nodeName)) }) @@ -635,7 +636,7 @@ var _ = framework.KubeDescribe("SchedulerPredicates [Serial]", func() { // already when the kubelet does not know about its new label yet. The // kubelet will then refuse to launch the pod. framework.ExpectNoError(framework.WaitForPodNotPending(cs, ns, pod.Name, pod.ResourceVersion)) - labelPod, err := cs.Core().Pods(ns).Get(pod.Name) + labelPod, err := cs.Core().Pods(ns).Get(pod.Name, metav1.GetOptions{}) framework.ExpectNoError(err) Expect(labelPod.Spec.NodeName).To(Equal(nodeName)) }) @@ -659,7 +660,7 @@ var _ = framework.KubeDescribe("SchedulerPredicates [Serial]", func() { // already when the kubelet does not know about its new label yet. The // kubelet will then refuse to launch the pod. framework.ExpectNoError(framework.WaitForPodNotPending(cs, ns, pod.Name, pod.ResourceVersion)) - labelPod, err := cs.Core().Pods(ns).Get(pod.Name) + labelPod, err := cs.Core().Pods(ns).Get(pod.Name, metav1.GetOptions{}) framework.ExpectNoError(err) Expect(labelPod.Spec.NodeName).To(Equal(nodeName)) }) @@ -711,7 +712,7 @@ var _ = framework.KubeDescribe("SchedulerPredicates [Serial]", func() { // already when the kubelet does not know about its new taint yet. The // kubelet will then refuse to launch the pod. framework.ExpectNoError(framework.WaitForPodNotPending(cs, ns, tolerationPodName, pod.ResourceVersion)) - deployedPod, err := cs.Core().Pods(ns).Get(tolerationPodName) + deployedPod, err := cs.Core().Pods(ns).Get(tolerationPodName, metav1.GetOptions{}) framework.ExpectNoError(err) Expect(deployedPod.Spec.NodeName).To(Equal(nodeName)) }) @@ -799,7 +800,7 @@ func createPausePod(f *framework.Framework, conf pausePodConfig) *v1.Pod { func runPausePod(f *framework.Framework, conf pausePodConfig) *v1.Pod { pod := createPausePod(f, conf) framework.ExpectNoError(framework.WaitForPodRunningInNamespace(f.ClientSet, pod)) - pod, err := f.ClientSet.Core().Pods(f.Namespace.Name).Get(conf.Name) + pod, err := f.ClientSet.Core().Pods(f.Namespace.Name).Get(conf.Name, metav1.GetOptions{}) framework.ExpectNoError(err) return pod } diff --git a/test/e2e/security_context.go b/test/e2e/security_context.go index 95bd93fed2b..cdb240cb19d 100644 --- a/test/e2e/security_context.go +++ b/test/e2e/security_context.go @@ -26,6 +26,7 @@ import ( "fmt" "k8s.io/kubernetes/pkg/api/v1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" "k8s.io/kubernetes/pkg/util/uuid" "k8s.io/kubernetes/test/e2e/framework" @@ -180,7 +181,7 @@ func testPodSELinuxLabeling(f *framework.Framework, hostIPC bool, hostPID bool) Expect(err).To(BeNil()) Expect(content).To(ContainSubstring(testContent)) - foundPod, err := f.ClientSet.Core().Pods(f.Namespace.Name).Get(pod.Name) + foundPod, err := f.ClientSet.Core().Pods(f.Namespace.Name).Get(pod.Name, metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred()) // Confirm that the file can be accessed from a second diff --git a/test/e2e/service.go b/test/e2e/service.go index 667249c33b5..7da5334b12a 100644 --- a/test/e2e/service.go +++ b/test/e2e/service.go @@ -33,6 +33,7 @@ import ( "k8s.io/kubernetes/pkg/api/errors" "k8s.io/kubernetes/pkg/api/v1" "k8s.io/kubernetes/pkg/api/v1/service" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" "k8s.io/kubernetes/pkg/client/clientset_generated/internalclientset" clientset "k8s.io/kubernetes/pkg/client/clientset_generated/release_1_5" "k8s.io/kubernetes/pkg/controller/endpoint" @@ -107,7 +108,7 @@ var _ = framework.KubeDescribe("Services", func() { // TODO: We get coverage of TCP/UDP and multi-port services through the DNS test. We should have a simpler test for multi-port TCP here. It("should provide secure master service [Conformance]", func() { - _, err := cs.Core().Services(v1.NamespaceDefault).Get("kubernetes") + _, err := cs.Core().Services(v1.NamespaceDefault).Get("kubernetes", metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred()) }) @@ -1123,9 +1124,9 @@ var _ = framework.KubeDescribe("Services", func() { acceptPodName := createExecPodOrFail(cs, namespace, "execpod-accept") dropPodName := createExecPodOrFail(cs, namespace, "execpod-drop") - accpetPod, err := cs.Core().Pods(namespace).Get(acceptPodName) + accpetPod, err := cs.Core().Pods(namespace).Get(acceptPodName, metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred()) - dropPod, err := cs.Core().Pods(namespace).Get(dropPodName) + dropPod, err := cs.Core().Pods(namespace).Get(dropPodName, metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred()) By("creating a pod to be part of the service " + serviceName) @@ -1350,7 +1351,7 @@ var _ = framework.KubeDescribe("ESIPP [Slow]", func() { err := cs.Core().Pods(namespace).Delete(execPodName, nil) Expect(err).NotTo(HaveOccurred()) }() - execPod, err := f.ClientSet.Core().Pods(namespace).Get(execPodName) + execPod, err := f.ClientSet.Core().Pods(namespace).Get(execPodName, metav1.GetOptions{}) framework.ExpectNoError(err) framework.Logf("Waiting up to %v wget %v", kubeProxyLagTimeout, path) @@ -1485,7 +1486,7 @@ func updateService(c clientset.Interface, namespace, serviceName string, update var service *v1.Service var err error for i := 0; i < 3; i++ { - service, err = c.Core().Services(namespace).Get(serviceName) + service, err = c.Core().Services(namespace).Get(serviceName, metav1.GetOptions{}) if err != nil { return service, err } @@ -1540,7 +1541,7 @@ func translatePodNameToUIDOrFail(c clientset.Interface, ns string, expectedEndpo portsByUID := make(PortsByPodUID) for name, portList := range expectedEndpoints { - pod, err := c.Core().Pods(ns).Get(name) + pod, err := c.Core().Pods(ns).Get(name, metav1.GetOptions{}) if err != nil { framework.Failf("failed to get pod %s, that's pretty weird. validation failed: %s", name, err) } @@ -1576,7 +1577,7 @@ func validateEndpointsOrFail(c clientset.Interface, namespace, serviceName strin By(fmt.Sprintf("waiting up to %v for service %s in namespace %s to expose endpoints %v", framework.ServiceStartTimeout, serviceName, namespace, expectedEndpoints)) i := 1 for start := time.Now(); time.Since(start) < framework.ServiceStartTimeout; time.Sleep(1 * time.Second) { - endpoints, err := c.Core().Endpoints(namespace).Get(serviceName) + endpoints, err := c.Core().Endpoints(namespace).Get(serviceName, metav1.GetOptions{}) if err != nil { framework.Logf("Get endpoints failed (%v elapsed, ignoring for 5s): %v", time.Since(start), err) continue @@ -1641,7 +1642,7 @@ func createExecPodOrFail(client clientset.Interface, ns, generateName string) st created, err := client.Core().Pods(ns).Create(execPod) Expect(err).NotTo(HaveOccurred()) err = wait.PollImmediate(framework.Poll, 5*time.Minute, func() (bool, error) { - retrievedPod, err := client.Core().Pods(execPod.Namespace).Get(created.Name) + retrievedPod, err := client.Core().Pods(execPod.Namespace).Get(created.Name, metav1.GetOptions{}) if err != nil { return false, nil } @@ -1660,7 +1661,7 @@ func createExecPodOnNode(client clientset.Interface, ns, nodeName, generateName created, err := client.Core().Pods(ns).Create(execPod) Expect(err).NotTo(HaveOccurred()) err = wait.PollImmediate(framework.Poll, 5*time.Minute, func() (bool, error) { - retrievedPod, err := client.Core().Pods(execPod.Namespace).Get(created.Name) + retrievedPod, err := client.Core().Pods(execPod.Namespace).Get(created.Name, metav1.GetOptions{}) if err != nil { return false, nil } @@ -1965,7 +1966,7 @@ func startServeHostnameService(c clientset.Interface, internalClient internalcli } sort.StringSlice(podNames).Sort() - service, err := c.Core().Services(ns).Get(name) + service, err := c.Core().Services(ns).Get(name, metav1.GetOptions{}) if err != nil { return podNames, "", err } @@ -2250,7 +2251,7 @@ func (j *ServiceTestJig) createOnlyLocalLoadBalancerService(namespace, serviceNa // endpoints of the given Service are running. func (j *ServiceTestJig) getEndpointNodes(svc *v1.Service) map[string][]string { nodes := j.getNodes(maxNodesForEndpointsTests) - endpoints, err := j.Client.Core().Endpoints(svc.Namespace).Get(svc.Name) + endpoints, err := j.Client.Core().Endpoints(svc.Namespace).Get(svc.Name, metav1.GetOptions{}) if err != nil { framework.Failf("Get endpoints for service %s/%s failed (%s)", svc.Namespace, svc.Name, err) } @@ -2287,7 +2288,7 @@ func (j *ServiceTestJig) getNodes(maxNodesForTest int) (nodes *v1.NodeList) { func (j *ServiceTestJig) waitForEndpointOnNode(namespace, serviceName, nodeName string) { err := wait.PollImmediate(framework.Poll, loadBalancerCreateTimeoutDefault, func() (bool, error) { - endpoints, err := j.Client.Core().Endpoints(namespace).Get(serviceName) + endpoints, err := j.Client.Core().Endpoints(namespace).Get(serviceName, metav1.GetOptions{}) if err != nil { framework.Logf("Get endpoints for service %s/%s failed (%s)", namespace, serviceName, err) return false, nil @@ -2349,7 +2350,7 @@ func (j *ServiceTestJig) SanityCheckService(svc *v1.Service, svcType v1.ServiceT // face of timeouts and conflicts. func (j *ServiceTestJig) UpdateService(namespace, name string, update func(*v1.Service)) (*v1.Service, error) { for i := 0; i < 3; i++ { - service, err := j.Client.Core().Services(namespace).Get(name) + service, err := j.Client.Core().Services(namespace).Get(name, metav1.GetOptions{}) if err != nil { return nil, fmt.Errorf("Failed to get Service %q: %v", name, err) } @@ -2403,7 +2404,7 @@ func (j *ServiceTestJig) WaitForLoadBalancerOrFail(namespace, name string, timeo var service *v1.Service framework.Logf("Waiting up to %v for service %q to have a LoadBalancer", timeout, name) pollFunc := func() (bool, error) { - svc, err := j.Client.Core().Services(namespace).Get(name) + svc, err := j.Client.Core().Services(namespace).Get(name, metav1.GetOptions{}) if err != nil { return false, err } @@ -2430,7 +2431,7 @@ func (j *ServiceTestJig) WaitForLoadBalancerDestroyOrFail(namespace, name string var service *v1.Service framework.Logf("Waiting up to %v for service %q to have no LoadBalancer", timeout, name) pollFunc := func() (bool, error) { - svc, err := j.Client.Core().Services(namespace).Get(name) + svc, err := j.Client.Core().Services(namespace).Get(name, metav1.GetOptions{}) if err != nil { return false, err } @@ -2706,7 +2707,7 @@ func (t *ServiceTestFixture) Cleanup() []error { for rcName := range t.rcs { By("stopping RC " + rcName + " in namespace " + t.Namespace) // First, resize the RC to 0. - old, err := t.Client.Core().ReplicationControllers(t.Namespace).Get(rcName) + old, err := t.Client.Core().ReplicationControllers(t.Namespace).Get(rcName, metav1.GetOptions{}) if err != nil { errs = append(errs, err) } @@ -2778,7 +2779,7 @@ func execSourceipTest(f *framework.Framework, c clientset.Interface, ns, nodeNam err := c.Core().Pods(ns).Delete(execPodName, nil) Expect(err).NotTo(HaveOccurred()) }() - execPod, err := f.ClientSet.Core().Pods(ns).Get(execPodName) + execPod, err := f.ClientSet.Core().Pods(ns).Get(execPodName, metav1.GetOptions{}) framework.ExpectNoError(err) var stdout string diff --git a/test/e2e/service_accounts.go b/test/e2e/service_accounts.go index 47c63e00f68..0be7cd54e8e 100644 --- a/test/e2e/service_accounts.go +++ b/test/e2e/service_accounts.go @@ -22,6 +22,7 @@ import ( apierrors "k8s.io/kubernetes/pkg/api/errors" "k8s.io/kubernetes/pkg/api/v1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" "k8s.io/kubernetes/pkg/util/uuid" "k8s.io/kubernetes/pkg/util/wait" "k8s.io/kubernetes/pkg/version" @@ -42,7 +43,7 @@ var _ = framework.KubeDescribe("ServiceAccounts", func() { var secrets []v1.ObjectReference framework.ExpectNoError(wait.Poll(time.Millisecond*500, time.Second*10, func() (bool, error) { By("waiting for a single token reference") - sa, err := f.ClientSet.Core().ServiceAccounts(f.Namespace.Name).Get("default") + sa, err := f.ClientSet.Core().ServiceAccounts(f.Namespace.Name).Get("default", metav1.GetOptions{}) if apierrors.IsNotFound(err) { framework.Logf("default service account was not found") return false, nil @@ -68,7 +69,7 @@ var _ = framework.KubeDescribe("ServiceAccounts", func() { { By("ensuring the single token reference persists") time.Sleep(2 * time.Second) - sa, err := f.ClientSet.Core().ServiceAccounts(f.Namespace.Name).Get("default") + sa, err := f.ClientSet.Core().ServiceAccounts(f.Namespace.Name).Get("default", metav1.GetOptions{}) framework.ExpectNoError(err) Expect(sa.Secrets).To(Equal(secrets)) } @@ -80,7 +81,7 @@ var _ = framework.KubeDescribe("ServiceAccounts", func() { // wait for the referenced secret to be removed, and another one autocreated framework.ExpectNoError(wait.Poll(time.Millisecond*500, framework.ServiceAccountProvisionTimeout, func() (bool, error) { By("waiting for a new token reference") - sa, err := f.ClientSet.Core().ServiceAccounts(f.Namespace.Name).Get("default") + sa, err := f.ClientSet.Core().ServiceAccounts(f.Namespace.Name).Get("default", metav1.GetOptions{}) if err != nil { framework.Logf("error getting default service account: %v", err) return false, err @@ -106,7 +107,7 @@ var _ = framework.KubeDescribe("ServiceAccounts", func() { { By("ensuring the single token reference persists") time.Sleep(2 * time.Second) - sa, err := f.ClientSet.Core().ServiceAccounts(f.Namespace.Name).Get("default") + sa, err := f.ClientSet.Core().ServiceAccounts(f.Namespace.Name).Get("default", metav1.GetOptions{}) framework.ExpectNoError(err) Expect(sa.Secrets).To(Equal(secrets)) } @@ -114,7 +115,7 @@ var _ = framework.KubeDescribe("ServiceAccounts", func() { // delete the reference from the service account By("deleting the reference to the service account token") { - sa, err := f.ClientSet.Core().ServiceAccounts(f.Namespace.Name).Get("default") + sa, err := f.ClientSet.Core().ServiceAccounts(f.Namespace.Name).Get("default", metav1.GetOptions{}) framework.ExpectNoError(err) sa.Secrets = nil _, updateErr := f.ClientSet.Core().ServiceAccounts(f.Namespace.Name).Update(sa) @@ -124,7 +125,7 @@ var _ = framework.KubeDescribe("ServiceAccounts", func() { // wait for another one to be autocreated framework.ExpectNoError(wait.Poll(time.Millisecond*500, framework.ServiceAccountProvisionTimeout, func() (bool, error) { By("waiting for a new token to be created and added") - sa, err := f.ClientSet.Core().ServiceAccounts(f.Namespace.Name).Get("default") + sa, err := f.ClientSet.Core().ServiceAccounts(f.Namespace.Name).Get("default", metav1.GetOptions{}) if err != nil { framework.Logf("error getting default service account: %v", err) return false, err @@ -146,7 +147,7 @@ var _ = framework.KubeDescribe("ServiceAccounts", func() { { By("ensuring the single token reference persists") time.Sleep(2 * time.Second) - sa, err := f.ClientSet.Core().ServiceAccounts(f.Namespace.Name).Get("default") + sa, err := f.ClientSet.Core().ServiceAccounts(f.Namespace.Name).Get("default", metav1.GetOptions{}) framework.ExpectNoError(err) Expect(sa.Secrets).To(Equal(secrets)) } @@ -159,7 +160,7 @@ var _ = framework.KubeDescribe("ServiceAccounts", func() { // Standard get, update retry loop framework.ExpectNoError(wait.Poll(time.Millisecond*500, framework.ServiceAccountProvisionTimeout, func() (bool, error) { By("getting the auto-created API token") - sa, err := f.ClientSet.Core().ServiceAccounts(f.Namespace.Name).Get("default") + sa, err := f.ClientSet.Core().ServiceAccounts(f.Namespace.Name).Get("default", metav1.GetOptions{}) if apierrors.IsNotFound(err) { framework.Logf("default service account was not found") return false, nil @@ -173,7 +174,7 @@ var _ = framework.KubeDescribe("ServiceAccounts", func() { return false, nil } for _, secretRef := range sa.Secrets { - secret, err := f.ClientSet.Core().Secrets(f.Namespace.Name).Get(secretRef.Name) + secret, err := f.ClientSet.Core().Secrets(f.Namespace.Name).Get(secretRef.Name, metav1.GetOptions{}) if err != nil { framework.Logf("Error getting secret %s: %v", secretRef.Name, err) continue diff --git a/test/e2e/ubernetes_lite.go b/test/e2e/ubernetes_lite.go index fd14b563bfd..2dcf22487da 100644 --- a/test/e2e/ubernetes_lite.go +++ b/test/e2e/ubernetes_lite.go @@ -150,7 +150,7 @@ func getZoneCount(c clientset.Interface) (int, error) { // Find the name of the zone in which the pod is scheduled func getZoneNameForPod(c clientset.Interface, pod v1.Pod) (string, error) { By(fmt.Sprintf("Getting zone name for pod %s, on node %s", pod.Name, pod.Spec.NodeName)) - node, err := c.Core().Nodes().Get(pod.Spec.NodeName) + node, err := c.Core().Nodes().Get(pod.Spec.NodeName, metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred()) return getZoneNameForNode(*node) } diff --git a/test/e2e/volume_provisioning.go b/test/e2e/volume_provisioning.go index 05c51b944ff..e3e6fe505f0 100644 --- a/test/e2e/volume_provisioning.go +++ b/test/e2e/volume_provisioning.go @@ -45,11 +45,11 @@ func testDynamicProvisioning(client clientset.Interface, claim *v1.PersistentVol By("checking the claim") // Get new copy of the claim - claim, err = client.Core().PersistentVolumeClaims(claim.Namespace).Get(claim.Name) + claim, err = client.Core().PersistentVolumeClaims(claim.Namespace).Get(claim.Name, metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred()) // Get the bound PV - pv, err := client.Core().PersistentVolumes().Get(claim.Spec.VolumeName) + pv, err := client.Core().PersistentVolumes().Get(claim.Spec.VolumeName, metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred()) // Check sizes diff --git a/test/e2e/volumes.go b/test/e2e/volumes.go index ea877128191..52fb4136d7b 100644 --- a/test/e2e/volumes.go +++ b/test/e2e/volumes.go @@ -150,7 +150,7 @@ func startVolumeServer(client clientset.Interface, config VolumeTestConfig) *v1. framework.ExpectNoError(framework.WaitForPodRunningInNamespace(client, serverPod)) By("locating the server pod") - pod, err := podClient.Get(serverPod.Name) + pod, err := podClient.Get(serverPod.Name, metav1.GetOptions{}) framework.ExpectNoError(err, "Cannot locate the server pod %v: %v", serverPod.Name, err) By("sleeping a bit to give the server time to start") diff --git a/test/e2e_node/apparmor_test.go b/test/e2e_node/apparmor_test.go index 85c2d648cc8..cb293b5bf8a 100644 --- a/test/e2e_node/apparmor_test.go +++ b/test/e2e_node/apparmor_test.go @@ -28,6 +28,7 @@ import ( "k8s.io/kubernetes/pkg/api/errors" "k8s.io/kubernetes/pkg/api/v1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" "k8s.io/kubernetes/pkg/runtime/schema" "k8s.io/kubernetes/pkg/security/apparmor" "k8s.io/kubernetes/pkg/watch" @@ -163,7 +164,7 @@ func runAppArmorTest(f *framework.Framework, shouldRun bool, profile string) v1. }) framework.ExpectNoError(err) } - p, err := f.PodClient().Get(pod.Name) + p, err := f.PodClient().Get(pod.Name, metav1.GetOptions{}) framework.ExpectNoError(err) return p.Status } diff --git a/test/e2e_node/benchmark_util.go b/test/e2e_node/benchmark_util.go index 7bc5af60ded..940966091aa 100644 --- a/test/e2e_node/benchmark_util.go +++ b/test/e2e_node/benchmark_util.go @@ -126,7 +126,7 @@ func getThroughputPerfData(batchLag time.Duration, e2eLags []framework.PodLatenc // getTestNodeInfo fetches the capacity of a node from API server and returns a map of labels. func getTestNodeInfo(f *framework.Framework, testName string) map[string]string { nodeName := framework.TestContext.NodeName - node, err := f.ClientSet.Core().Nodes().Get(nodeName) + node, err := f.ClientSet.Core().Nodes().Get(nodeName, metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred()) cpu, ok := node.Status.Capacity["cpu"] diff --git a/test/e2e_node/container.go b/test/e2e_node/container.go index 55daf5e8022..2b5ca99c5c3 100644 --- a/test/e2e_node/container.go +++ b/test/e2e_node/container.go @@ -21,6 +21,7 @@ import ( "k8s.io/kubernetes/pkg/api/errors" "k8s.io/kubernetes/pkg/api/v1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" "k8s.io/kubernetes/pkg/util/uuid" "k8s.io/kubernetes/test/e2e/framework" ) @@ -66,7 +67,7 @@ func (cc *ConformanceContainer) Delete() error { } func (cc *ConformanceContainer) IsReady() (bool, error) { - pod, err := cc.PodClient.Get(cc.podName) + pod, err := cc.PodClient.Get(cc.podName, metav1.GetOptions{}) if err != nil { return false, err } @@ -74,7 +75,7 @@ func (cc *ConformanceContainer) IsReady() (bool, error) { } func (cc *ConformanceContainer) GetPhase() (v1.PodPhase, error) { - pod, err := cc.PodClient.Get(cc.podName) + pod, err := cc.PodClient.Get(cc.podName, metav1.GetOptions{}) if err != nil { return v1.PodUnknown, err } @@ -82,7 +83,7 @@ func (cc *ConformanceContainer) GetPhase() (v1.PodPhase, error) { } func (cc *ConformanceContainer) GetStatus() (v1.ContainerStatus, error) { - pod, err := cc.PodClient.Get(cc.podName) + pod, err := cc.PodClient.Get(cc.podName, metav1.GetOptions{}) if err != nil { return v1.ContainerStatus{}, err } @@ -94,7 +95,7 @@ func (cc *ConformanceContainer) GetStatus() (v1.ContainerStatus, error) { } func (cc *ConformanceContainer) Present() (bool, error) { - _, err := cc.PodClient.Get(cc.podName) + _, err := cc.PodClient.Get(cc.podName, metav1.GetOptions{}) if err == nil { return true, nil } diff --git a/test/e2e_node/disk_eviction_test.go b/test/e2e_node/disk_eviction_test.go index b272352a5e6..09d7fc39baa 100644 --- a/test/e2e_node/disk_eviction_test.go +++ b/test/e2e_node/disk_eviction_test.go @@ -22,6 +22,7 @@ import ( "time" "k8s.io/kubernetes/pkg/api/v1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" "k8s.io/kubernetes/pkg/util/uuid" "k8s.io/kubernetes/test/e2e/framework" @@ -125,7 +126,7 @@ var _ = framework.KubeDescribe("Kubelet Eviction Manager [Serial] [Disruptive]", // The pod should be evicted. if !evictionOccurred { - podData, err := podClient.Get(busyPodName) + podData, err := podClient.Get(busyPodName, metav1.GetOptions{}) if err != nil { return err } @@ -135,7 +136,7 @@ var _ = framework.KubeDescribe("Kubelet Eviction Manager [Serial] [Disruptive]", return err } - podData, err = podClient.Get(idlePodName) + podData, err = podClient.Get(idlePodName, metav1.GetOptions{}) if err != nil { return err } @@ -170,7 +171,7 @@ var _ = framework.KubeDescribe("Kubelet Eviction Manager [Serial] [Disruptive]", } // The new pod should be able to be scheduled and run after the disk pressure is relieved. - podData, err := podClient.Get(verifyPodName) + podData, err := podClient.Get(verifyPodName, metav1.GetOptions{}) if err != nil { return err } diff --git a/test/e2e_node/garbage_collector_test.go b/test/e2e_node/garbage_collector_test.go index ce97711d02f..fc560697931 100644 --- a/test/e2e_node/garbage_collector_test.go +++ b/test/e2e_node/garbage_collector_test.go @@ -22,6 +22,7 @@ import ( "time" "k8s.io/kubernetes/pkg/api/v1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" docker "k8s.io/kubernetes/pkg/kubelet/dockertools" "k8s.io/kubernetes/test/e2e/framework" @@ -149,7 +150,7 @@ func containerGCTest(f *framework.Framework, test testRun) { By("Making sure all containers restart the specified number of times") Eventually(func() error { for _, podSpec := range test.testPods { - updatedPod, err := f.ClientSet.Core().Pods(f.Namespace.Name).Get(podSpec.podName) + updatedPod, err := f.ClientSet.Core().Pods(f.Namespace.Name).Get(podSpec.podName, metav1.GetOptions{}) if err != nil { return err } diff --git a/test/e2e_node/image_id_test.go b/test/e2e_node/image_id_test.go index 3636fbed5f4..8cead09863a 100644 --- a/test/e2e_node/image_id_test.go +++ b/test/e2e_node/image_id_test.go @@ -18,6 +18,7 @@ package e2e_node import ( "k8s.io/kubernetes/pkg/api/v1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" "k8s.io/kubernetes/pkg/kubelet/dockertools" "k8s.io/kubernetes/test/e2e/framework" @@ -51,7 +52,7 @@ var _ = framework.KubeDescribe("ImageID", func() { framework.ExpectNoError(framework.WaitTimeoutForPodNoLongerRunningInNamespace( f.ClientSet, pod.Name, f.Namespace.Name, "", framework.PodStartTimeout)) - runningPod, err := f.PodClient().Get(pod.Name) + runningPod, err := f.PodClient().Get(pod.Name, metav1.GetOptions{}) framework.ExpectNoError(err) status := runningPod.Status diff --git a/test/e2e_node/kubelet_test.go b/test/e2e_node/kubelet_test.go index cdb6bfdbecf..8bc35c66533 100644 --- a/test/e2e_node/kubelet_test.go +++ b/test/e2e_node/kubelet_test.go @@ -22,7 +22,7 @@ import ( "time" "k8s.io/kubernetes/pkg/api/v1" - apiunversioned "k8s.io/kubernetes/pkg/apis/meta/v1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" "k8s.io/kubernetes/pkg/util/uuid" "k8s.io/kubernetes/test/e2e/framework" @@ -56,7 +56,7 @@ var _ = framework.KubeDescribe("Kubelet", func() { }, }) Eventually(func() string { - sinceTime := apiunversioned.NewTime(time.Now().Add(time.Duration(-1 * time.Hour))) + sinceTime := metav1.NewTime(time.Now().Add(time.Duration(-1 * time.Hour))) rc, err := podClient.GetLogs(podName, &v1.PodLogOptions{SinceTime: &sinceTime}).Stream() if err != nil { return "" @@ -93,7 +93,7 @@ var _ = framework.KubeDescribe("Kubelet", func() { It("should have an error terminated reason", func() { Eventually(func() error { - podData, err := podClient.Get(podName) + podData, err := podClient.Get(podName, metav1.GetOptions{}) if err != nil { return err } diff --git a/test/e2e_node/log_path_test.go b/test/e2e_node/log_path_test.go index f9a64f28ae0..5ec742642e3 100644 --- a/test/e2e_node/log_path_test.go +++ b/test/e2e_node/log_path_test.go @@ -18,6 +18,7 @@ package e2e_node import ( "k8s.io/kubernetes/pkg/api/v1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" "k8s.io/kubernetes/pkg/kubelet" kubecontainer "k8s.io/kubernetes/pkg/kubelet/container" "k8s.io/kubernetes/test/e2e/framework" @@ -67,7 +68,7 @@ var _ = framework.KubeDescribe("ContainerLogPath", func() { framework.ExpectNoError(err, "Failed waiting for pod: %s to enter success state", logPodName) // get containerID from created Pod - createdLogPod, err := podClient.Get(logPodName) + createdLogPod, err := podClient.Get(logPodName, metav1.GetOptions{}) logConID := kubecontainer.ParseContainerID(createdLogPod.Status.ContainerStatuses[0].ContainerID) framework.ExpectNoError(err, "Failed to get pod: %s", logPodName) diff --git a/test/e2e_node/memory_eviction_test.go b/test/e2e_node/memory_eviction_test.go index 727b6b951d1..5558d5f5615 100644 --- a/test/e2e_node/memory_eviction_test.go +++ b/test/e2e_node/memory_eviction_test.go @@ -24,6 +24,7 @@ import ( "github.com/golang/glog" "k8s.io/kubernetes/pkg/api/resource" "k8s.io/kubernetes/pkg/api/v1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" "k8s.io/kubernetes/test/e2e/framework" . "github.com/onsi/ginkgo" @@ -150,15 +151,15 @@ var _ = framework.KubeDescribe("MemoryEviction [Slow] [Serial] [Disruptive]", fu By("polling the Status.Phase of each pod and checking for violations of the eviction order.") Eventually(func() error { - gteed, gtErr := f.ClientSet.Core().Pods(f.Namespace.Name).Get(guaranteed.Name) + gteed, gtErr := f.ClientSet.Core().Pods(f.Namespace.Name).Get(guaranteed.Name, metav1.GetOptions{}) framework.ExpectNoError(gtErr, fmt.Sprintf("getting pod %s", guaranteed.Name)) gteedPh := gteed.Status.Phase - burst, buErr := f.ClientSet.Core().Pods(f.Namespace.Name).Get(burstable.Name) + burst, buErr := f.ClientSet.Core().Pods(f.Namespace.Name).Get(burstable.Name, metav1.GetOptions{}) framework.ExpectNoError(buErr, fmt.Sprintf("getting pod %s", burstable.Name)) burstPh := burst.Status.Phase - best, beErr := f.ClientSet.Core().Pods(f.Namespace.Name).Get(besteffort.Name) + best, beErr := f.ClientSet.Core().Pods(f.Namespace.Name).Get(besteffort.Name, metav1.GetOptions{}) framework.ExpectNoError(beErr, fmt.Sprintf("getting pod %s", besteffort.Name)) bestPh := best.Status.Phase diff --git a/test/e2e_node/mirror_pod_test.go b/test/e2e_node/mirror_pod_test.go index 1b6f4e57f68..dc1a6eceb53 100644 --- a/test/e2e_node/mirror_pod_test.go +++ b/test/e2e_node/mirror_pod_test.go @@ -25,6 +25,7 @@ import ( "k8s.io/kubernetes/pkg/api/errors" "k8s.io/kubernetes/pkg/api/v1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" clientset "k8s.io/kubernetes/pkg/client/clientset_generated/release_1_5" "k8s.io/kubernetes/pkg/types" "k8s.io/kubernetes/pkg/util/uuid" @@ -57,7 +58,7 @@ var _ = framework.KubeDescribe("MirrorPod", func() { }) It("should be updated when static pod updated [Conformance]", func() { By("get mirror pod uid") - pod, err := f.ClientSet.Core().Pods(ns).Get(mirrorPodName) + pod, err := f.ClientSet.Core().Pods(ns).Get(mirrorPodName, metav1.GetOptions{}) Expect(err).ShouldNot(HaveOccurred()) uid := pod.UID @@ -72,14 +73,14 @@ var _ = framework.KubeDescribe("MirrorPod", func() { }, 2*time.Minute, time.Second*4).Should(BeNil()) By("check the mirror pod container image is updated") - pod, err = f.ClientSet.Core().Pods(ns).Get(mirrorPodName) + pod, err = f.ClientSet.Core().Pods(ns).Get(mirrorPodName, metav1.GetOptions{}) Expect(err).ShouldNot(HaveOccurred()) Expect(len(pod.Spec.Containers)).Should(Equal(1)) Expect(pod.Spec.Containers[0].Image).Should(Equal(image)) }) It("should be recreated when mirror pod gracefully deleted [Conformance]", func() { By("get mirror pod uid") - pod, err := f.ClientSet.Core().Pods(ns).Get(mirrorPodName) + pod, err := f.ClientSet.Core().Pods(ns).Get(mirrorPodName, metav1.GetOptions{}) Expect(err).ShouldNot(HaveOccurred()) uid := pod.UID @@ -94,7 +95,7 @@ var _ = framework.KubeDescribe("MirrorPod", func() { }) It("should be recreated when mirror pod forcibly deleted [Conformance]", func() { By("get mirror pod uid") - pod, err := f.ClientSet.Core().Pods(ns).Get(mirrorPodName) + pod, err := f.ClientSet.Core().Pods(ns).Get(mirrorPodName, metav1.GetOptions{}) Expect(err).ShouldNot(HaveOccurred()) uid := pod.UID @@ -156,7 +157,7 @@ func deleteStaticPod(dir, name, namespace string) error { } func checkMirrorPodDisappear(cl clientset.Interface, name, namespace string) error { - _, err := cl.Core().Pods(namespace).Get(name) + _, err := cl.Core().Pods(namespace).Get(name, metav1.GetOptions{}) if errors.IsNotFound(err) { return nil } @@ -164,7 +165,7 @@ func checkMirrorPodDisappear(cl clientset.Interface, name, namespace string) err } func checkMirrorPodRunning(cl clientset.Interface, name, namespace string) error { - pod, err := cl.Core().Pods(namespace).Get(name) + pod, err := cl.Core().Pods(namespace).Get(name, metav1.GetOptions{}) if err != nil { return fmt.Errorf("expected the mirror pod %q to appear: %v", name, err) } @@ -175,7 +176,7 @@ func checkMirrorPodRunning(cl clientset.Interface, name, namespace string) error } func checkMirrorPodRecreatedAndRunnig(cl clientset.Interface, name, namespace string, oUID types.UID) error { - pod, err := cl.Core().Pods(namespace).Get(name) + pod, err := cl.Core().Pods(namespace).Get(name, metav1.GetOptions{}) if err != nil { return fmt.Errorf("expected the mirror pod %q to appear: %v", name, err) } diff --git a/test/e2e_node/util.go b/test/e2e_node/util.go index e7c144be391..a913b51b56e 100644 --- a/test/e2e_node/util.go +++ b/test/e2e_node/util.go @@ -33,6 +33,7 @@ import ( "k8s.io/kubernetes/pkg/api/v1" "k8s.io/kubernetes/pkg/apis/componentconfig" v1alpha1 "k8s.io/kubernetes/pkg/apis/componentconfig/v1alpha1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" "k8s.io/kubernetes/pkg/kubelet/api/v1alpha1/stats" // utilconfig "k8s.io/kubernetes/pkg/util/config" "k8s.io/kubernetes/test/e2e/framework" @@ -86,7 +87,7 @@ func getCurrentKubeletConfig() (*componentconfig.KubeletConfiguration, error) { // Queries the API server for a Kubelet configuration for the node described by framework.TestContext.NodeName func getCurrentKubeletConfigMap(f *framework.Framework) (*v1.ConfigMap, error) { - return f.ClientSet.Core().ConfigMaps("kube-system").Get(fmt.Sprintf("kubelet-%s", framework.TestContext.NodeName)) + return f.ClientSet.Core().ConfigMaps("kube-system").Get(fmt.Sprintf("kubelet-%s", framework.TestContext.NodeName), metav1.GetOptions{}) } // Creates or updates the configmap for KubeletConfiguration, waits for the Kubelet to restart diff --git a/test/integration/client/client_test.go b/test/integration/client/client_test.go index 6de1092efd2..9ca0f328bee 100644 --- a/test/integration/client/client_test.go +++ b/test/integration/client/client_test.go @@ -173,7 +173,7 @@ func TestAtomicPut(t *testing.T) { go func(l, v string) { defer wg.Done() for { - tmpRC, err := rcs.Get(rc.Name) + tmpRC, err := rcs.Get(rc.Name, metav1.GetOptions{}) if err != nil { t.Errorf("Error getting atomicRC: %v", err) continue @@ -199,7 +199,7 @@ func TestAtomicPut(t *testing.T) { }(label, value) } wg.Wait() - rc, err = rcs.Get(rc.Name) + rc, err = rcs.Get(rc.Name, metav1.GetOptions{}) if err != nil { t.Fatalf("Failed getting atomicRC after writers are complete: %v", err) } @@ -281,7 +281,7 @@ func TestPatch(t *testing.T) { if err != nil { t.Fatalf("Failed updating patchpod with patch type %s: %v", k, err) } - pod, err = pods.Get(name) + pod, err = pods.Get(name, metav1.GetOptions{}) if err != nil { t.Fatalf("Failed getting patchpod: %v", err) } @@ -294,7 +294,7 @@ func TestPatch(t *testing.T) { if err != nil { t.Fatalf("Failed updating patchpod with patch type %s: %v", k, err) } - pod, err = pods.Get(name) + pod, err = pods.Get(name, metav1.GetOptions{}) if err != nil { t.Fatalf("Failed getting patchpod: %v", err) } @@ -307,7 +307,7 @@ func TestPatch(t *testing.T) { if err != nil { t.Fatalf("Failed updating patchpod with patch type %s: %v", k, err) } - pod, err = pods.Get(name) + pod, err = pods.Get(name, metav1.GetOptions{}) if err != nil { t.Fatalf("Failed getting patchpod: %v", err) } @@ -712,7 +712,7 @@ func TestMultiWatch(t *testing.T) { for i := 0; i < watcherCount; i++ { go func(i int) { name := fmt.Sprintf("multi-watch-%v", i) - pod, err := client.Core().Pods(ns.Name).Get(name) + pod, err := client.Core().Pods(ns.Name).Get(name, metav1.GetOptions{}) if err != nil { panic(fmt.Sprintf("Couldn't get %v: %v", name, err)) } diff --git a/test/integration/evictions/evictions_test.go b/test/integration/evictions/evictions_test.go index 840862656e4..4f836994ecd 100644 --- a/test/integration/evictions/evictions_test.go +++ b/test/integration/evictions/evictions_test.go @@ -213,7 +213,7 @@ func TestConcurrentEvictionRequests(t *testing.T) { for i := 0; i < 10; i++ { podName := fmt.Sprintf(podNameFormat, i) - _, err := clientSet.Core().Pods(ns.Name).Get(podName) + _, err := clientSet.Core().Pods(ns.Name).Get(podName, metav1.GetOptions{}) if !errors.IsNotFound(err) { t.Errorf("Pod %q is expected to be evicted", podName) } @@ -243,7 +243,7 @@ func waitToObservePods(t *testing.T, podInformer cache.SharedIndexInformer, podN func waitPDBStable(t *testing.T, clientSet clientset.Interface, podNum int32, ns, pdbName string) { if err := wait.PollImmediate(2*time.Second, 60*time.Second, func() (bool, error) { - pdb, err := clientSet.Policy().PodDisruptionBudgets(ns).Get(pdbName) + pdb, err := clientSet.Policy().PodDisruptionBudgets(ns).Get(pdbName, metav1.GetOptions{}) if err != nil { return false, err } diff --git a/test/integration/framework/master_utils.go b/test/integration/framework/master_utils.go index 0326f96bf1a..8621ebe9641 100644 --- a/test/integration/framework/master_utils.go +++ b/test/integration/framework/master_utils.go @@ -36,6 +36,7 @@ import ( "k8s.io/kubernetes/pkg/apis/batch" certificates "k8s.io/kubernetes/pkg/apis/certificates/v1alpha1" extensions "k8s.io/kubernetes/pkg/apis/extensions/v1beta1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" policy "k8s.io/kubernetes/pkg/apis/policy/v1alpha1" rbac "k8s.io/kubernetes/pkg/apis/rbac/v1alpha1" storage "k8s.io/kubernetes/pkg/apis/storage/v1beta1" @@ -452,7 +453,7 @@ func ScaleRC(name, ns string, replicas int32, clientset internalclientset.Interf if err != nil { return nil, err } - scaled, err := clientset.Core().ReplicationControllers(ns).Get(name) + scaled, err := clientset.Core().ReplicationControllers(ns).Get(name, metav1.GetOptions{}) if err != nil { return nil, err } diff --git a/test/integration/garbagecollector/garbage_collector_test.go b/test/integration/garbagecollector/garbage_collector_test.go index 1061094bd06..5dad57f3f21 100644 --- a/test/integration/garbagecollector/garbage_collector_test.go +++ b/test/integration/garbagecollector/garbage_collector_test.go @@ -232,10 +232,10 @@ func TestCascadingDeletion(t *testing.T) { t.Fatalf("expect pod %s to be garbage collected, got err= %v", garbageCollectedPodName, err) } // checks the garbage collect doesn't delete pods it shouldn't delete. - if _, err := podClient.Get(independentPodName); err != nil { + if _, err := podClient.Get(independentPodName, metav1.GetOptions{}); err != nil { t.Fatal(err) } - if _, err := podClient.Get(oneValidOwnerPodName); err != nil { + if _, err := podClient.Get(oneValidOwnerPodName, metav1.GetOptions{}); err != nil { t.Fatal(err) } } diff --git a/test/integration/master/master_test.go b/test/integration/master/master_test.go index 356d0da5c54..2207ceee7a2 100644 --- a/test/integration/master/master_test.go +++ b/test/integration/master/master_test.go @@ -39,6 +39,7 @@ import ( "k8s.io/kubernetes/pkg/api/v1" "k8s.io/kubernetes/pkg/apimachinery/registered" "k8s.io/kubernetes/pkg/apis/batch/v2alpha1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" clientset "k8s.io/kubernetes/pkg/client/clientset_generated/internalclientset" clienttypedv1 "k8s.io/kubernetes/pkg/client/clientset_generated/release_1_5/typed/core/v1" "k8s.io/kubernetes/pkg/client/restclient" @@ -407,7 +408,7 @@ func TestMasterService(t *testing.T) { } } if found { - ep, err := client.Core().Endpoints(api.NamespaceDefault).Get("kubernetes") + ep, err := client.Core().Endpoints(api.NamespaceDefault).Get("kubernetes", metav1.GetOptions{}) if err != nil { return false, nil } @@ -451,7 +452,7 @@ func TestServiceAlloc(t *testing.T) { // Wait until the default "kubernetes" service is created. if err = wait.Poll(250*time.Millisecond, time.Minute, func() (bool, error) { - _, err := client.Core().Services(api.NamespaceDefault).Get("kubernetes") + _, err := client.Core().Services(api.NamespaceDefault).Get("kubernetes", metav1.GetOptions{}) if err != nil && !errors.IsNotFound(err) { return false, err } @@ -595,7 +596,7 @@ func TestUpdateNodeObjects(t *testing.T) { break } - n, err := c.Nodes().Get(fmt.Sprintf("node-%d", node)) + n, err := c.Nodes().Get(fmt.Sprintf("node-%d", node), metav1.GetOptions{}) if err != nil { fmt.Printf("[%d] error after %d: %v\n", node, i, err) break diff --git a/test/integration/replicaset/replicaset_test.go b/test/integration/replicaset/replicaset_test.go index 9633156d587..03092bc7819 100644 --- a/test/integration/replicaset/replicaset_test.go +++ b/test/integration/replicaset/replicaset_test.go @@ -245,7 +245,7 @@ func TestAdoption(t *testing.T) { waitToObservePods(t, podInformer, 1) go rm.Run(5, stopCh) if err := wait.Poll(10*time.Second, 60*time.Second, func() (bool, error) { - updatedPod, err := podClient.Get(pod.Name) + updatedPod, err := podClient.Get(pod.Name, metav1.GetOptions{}) if err != nil { return false, err } @@ -280,7 +280,7 @@ func createRSsPods(t *testing.T, clientSet clientset.Interface, rss []*v1beta1.R func waitRSStable(t *testing.T, clientSet clientset.Interface, rs *v1beta1.ReplicaSet, ns string) { rsClient := clientSet.Extensions().ReplicaSets(ns) if err := wait.Poll(10*time.Second, 60*time.Second, func() (bool, error) { - updatedRS, err := rsClient.Get(rs.Name) + updatedRS, err := rsClient.Get(rs.Name, metav1.GetOptions{}) if err != nil { return false, err } @@ -371,7 +371,7 @@ func TestUpdateSelectorToRemoveControllerRef(t *testing.T) { t.Fatal(err) } podClient := clientSet.Core().Pods(ns.Name) - pod2, err = podClient.Get(pod2.Name) + pod2, err = podClient.Get(pod2.Name, metav1.GetOptions{}) if err != nil { t.Fatalf("Failed to get pod2: %v", err) } @@ -414,7 +414,7 @@ func TestUpdateLabelToRemoveControllerRef(t *testing.T) { }); err != nil { t.Fatal(err) } - pod2, err = podClient.Get(pod2.Name) + pod2, err = podClient.Get(pod2.Name, metav1.GetOptions{}) if err != nil { t.Fatalf("Failed to get pod2: %v", err) } diff --git a/test/integration/replicationcontroller/replicationcontroller_test.go b/test/integration/replicationcontroller/replicationcontroller_test.go index 0cbfc0b3a1a..a9fd548de8d 100644 --- a/test/integration/replicationcontroller/replicationcontroller_test.go +++ b/test/integration/replicationcontroller/replicationcontroller_test.go @@ -243,7 +243,7 @@ func TestAdoption(t *testing.T) { waitToObservePods(t, podInformer, 1) go rm.Run(5, stopCh) if err := wait.Poll(10*time.Second, 60*time.Second, func() (bool, error) { - updatedPod, err := podClient.Get(pod.Name) + updatedPod, err := podClient.Get(pod.Name, metav1.GetOptions{}) if err != nil { return false, err } @@ -278,7 +278,7 @@ func createRCsPods(t *testing.T, clientSet clientset.Interface, rcs []*v1.Replic func waitRCStable(t *testing.T, clientSet clientset.Interface, rc *v1.ReplicationController, ns string) { rcClient := clientSet.Core().ReplicationControllers(ns) if err := wait.Poll(10*time.Second, 60*time.Second, func() (bool, error) { - updatedRC, err := rcClient.Get(rc.Name) + updatedRC, err := rcClient.Get(rc.Name, metav1.GetOptions{}) if err != nil { return false, err } @@ -367,7 +367,7 @@ func TestUpdateSelectorToRemoveControllerRef(t *testing.T) { t.Fatal(err) } podClient := clientSet.Core().Pods(ns.Name) - pod2, err = podClient.Get(pod2.Name) + pod2, err = podClient.Get(pod2.Name, metav1.GetOptions{}) if err != nil { t.Fatalf("Failed to get pod2: %v", err) } @@ -409,7 +409,7 @@ func TestUpdateLabelToRemoveControllerRef(t *testing.T) { }); err != nil { t.Fatal(err) } - pod2, err = podClient.Get(pod2.Name) + pod2, err = podClient.Get(pod2.Name, metav1.GetOptions{}) if err != nil { t.Fatalf("Failed to get pod2: %v", err) } diff --git a/test/integration/scheduler/extender_test.go b/test/integration/scheduler/extender_test.go index ab8c27000a6..7470efb1983 100644 --- a/test/integration/scheduler/extender_test.go +++ b/test/integration/scheduler/extender_test.go @@ -298,7 +298,7 @@ func DoTestPodScheduling(ns *v1.Namespace, t *testing.T, cs clientset.Interface) t.Fatalf("Failed to schedule pod: %v", err) } - if myPod, err := cs.Core().Pods(ns.Name).Get(myPod.Name); err != nil { + if myPod, err := cs.Core().Pods(ns.Name).Get(myPod.Name, metav1.GetOptions{}); err != nil { t.Fatalf("Failed to get pod: %v", err) } else if myPod.Spec.NodeName != "machine3" { t.Fatalf("Failed to schedule using extender, expected machine3, got %v", myPod.Spec.NodeName) diff --git a/test/integration/scheduler/scheduler_test.go b/test/integration/scheduler/scheduler_test.go index daf5a572f42..456013077f3 100644 --- a/test/integration/scheduler/scheduler_test.go +++ b/test/integration/scheduler/scheduler_test.go @@ -76,7 +76,7 @@ func TestUnschedulableNodes(t *testing.T) { func podScheduled(c clientset.Interface, podNamespace, podName string) wait.ConditionFunc { return func() (bool, error) { - pod, err := c.Core().Pods(podNamespace).Get(podName) + pod, err := c.Core().Pods(podNamespace).Get(podName, metav1.GetOptions{}) if errors.IsNotFound(err) { return false, nil } @@ -268,7 +268,7 @@ func DoTestUnschedulableNodes(t *testing.T, cs clientset.Interface, ns *v1.Names } // Apply the schedulable modification to the node, and wait for the reflection - schedNode, err := cs.Core().Nodes().Get(unSchedNode.Name) + schedNode, err := cs.Core().Nodes().Get(unSchedNode.Name, metav1.GetOptions{}) if err != nil { t.Fatalf("Failed to get node: %v", err) } diff --git a/test/integration/serviceaccount/service_account_test.go b/test/integration/serviceaccount/service_account_test.go index 6335086e119..662078a8c08 100644 --- a/test/integration/serviceaccount/service_account_test.go +++ b/test/integration/serviceaccount/service_account_test.go @@ -35,6 +35,7 @@ import ( "k8s.io/kubernetes/pkg/api/errors" "k8s.io/kubernetes/pkg/api/v1" "k8s.io/kubernetes/pkg/apimachinery/registered" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" "k8s.io/kubernetes/pkg/auth/authenticator" "k8s.io/kubernetes/pkg/auth/authenticator/bearertoken" "k8s.io/kubernetes/pkg/auth/authorizer" @@ -139,7 +140,7 @@ func TestServiceAccountTokenAutoCreate(t *testing.T) { } // Trigger creation of a new referenced token - serviceAccount, err = c.Core().ServiceAccounts(ns).Get(name) + serviceAccount, err = c.Core().ServiceAccounts(ns).Get(name, metav1.GetOptions{}) if err != nil { t.Fatal(err) } @@ -435,13 +436,13 @@ func startServiceAccountTestServer(t *testing.T) (*clientset.Clientset, restclie func getServiceAccount(c *clientset.Clientset, ns string, name string, shouldWait bool) (*v1.ServiceAccount, error) { if !shouldWait { - return c.Core().ServiceAccounts(ns).Get(name) + return c.Core().ServiceAccounts(ns).Get(name, metav1.GetOptions{}) } var user *v1.ServiceAccount var err error err = wait.Poll(time.Second, 10*time.Second, func() (bool, error) { - user, err = c.Core().ServiceAccounts(ns).Get(name) + user, err = c.Core().ServiceAccounts(ns).Get(name, metav1.GetOptions{}) if errors.IsNotFound(err) { return false, nil } @@ -458,7 +459,7 @@ func getReferencedServiceAccountToken(c *clientset.Clientset, ns string, name st token := "" findToken := func() (bool, error) { - user, err := c.Core().ServiceAccounts(ns).Get(name) + user, err := c.Core().ServiceAccounts(ns).Get(name, metav1.GetOptions{}) if errors.IsNotFound(err) { return false, nil } @@ -467,7 +468,7 @@ func getReferencedServiceAccountToken(c *clientset.Clientset, ns string, name st } for _, ref := range user.Secrets { - secret, err := c.Core().Secrets(ns).Get(ref.Name) + secret, err := c.Core().Secrets(ns).Get(ref.Name, metav1.GetOptions{}) if errors.IsNotFound(err) { continue } diff --git a/test/integration/utils.go b/test/integration/utils.go index e8fdd3a49ab..8e70e5404e7 100644 --- a/test/integration/utils.go +++ b/test/integration/utils.go @@ -21,6 +21,7 @@ import ( "time" "k8s.io/kubernetes/pkg/api/errors" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" clientset "k8s.io/kubernetes/pkg/client/clientset_generated/release_1_5" coreclient "k8s.io/kubernetes/pkg/client/clientset_generated/release_1_5/typed/core/v1" "k8s.io/kubernetes/pkg/util/wait" @@ -48,7 +49,7 @@ var Code503 = map[int]bool{503: true} // WaitForPodToDisappear polls the API server if the pod has been deleted. func WaitForPodToDisappear(podClient coreclient.PodInterface, podName string, interval, timeout time.Duration) error { return wait.PollImmediate(interval, timeout, func() (bool, error) { - _, err := podClient.Get(podName) + _, err := podClient.Get(podName, metav1.GetOptions{}) if err == nil { return false, nil } else { diff --git a/test/integration/volume/persistent_volumes_test.go b/test/integration/volume/persistent_volumes_test.go index f776c0531c0..7fe9b2c10b6 100644 --- a/test/integration/volume/persistent_volumes_test.go +++ b/test/integration/volume/persistent_volumes_test.go @@ -275,7 +275,7 @@ func TestPersistentVolumeBindRace(t *testing.T) { waitForAnyPersistentVolumeClaimPhase(watchPVC, v1.ClaimBound) glog.V(2).Infof("TestPersistentVolumeBindRace pvc bound") - pv, err = testClient.PersistentVolumes().Get(pv.Name) + pv, err = testClient.PersistentVolumes().Get(pv.Name, metav1.GetOptions{}) if err != nil { t.Fatalf("Unexpected error getting pv: %v", err) } @@ -347,14 +347,14 @@ func TestPersistentVolumeClaimLabelSelector(t *testing.T) { waitForPersistentVolumeClaimPhase(testClient, pvc.Name, ns.Name, watchPVC, v1.ClaimBound) t.Log("claim bound") - pv, err := testClient.PersistentVolumes().Get("pv-false") + pv, err := testClient.PersistentVolumes().Get("pv-false", metav1.GetOptions{}) if err != nil { t.Fatalf("Unexpected error getting pv: %v", err) } if pv.Spec.ClaimRef != nil { t.Fatalf("False PV shouldn't be bound") } - pv, err = testClient.PersistentVolumes().Get("pv-true") + pv, err = testClient.PersistentVolumes().Get("pv-true", metav1.GetOptions{}) if err != nil { t.Fatalf("Unexpected error getting pv: %v", err) } @@ -446,14 +446,14 @@ func TestPersistentVolumeClaimLabelSelectorMatchExpressions(t *testing.T) { waitForPersistentVolumeClaimPhase(testClient, pvc.Name, ns.Name, watchPVC, v1.ClaimBound) t.Log("claim bound") - pv, err := testClient.PersistentVolumes().Get("pv-false") + pv, err := testClient.PersistentVolumes().Get("pv-false", metav1.GetOptions{}) if err != nil { t.Fatalf("Unexpected error getting pv: %v", err) } if pv.Spec.ClaimRef != nil { t.Fatalf("False PV shouldn't be bound") } - pv, err = testClient.PersistentVolumes().Get("pv-true") + pv, err = testClient.PersistentVolumes().Get("pv-true", metav1.GetOptions{}) if err != nil { t.Fatalf("Unexpected error getting pv: %v", err) } @@ -520,7 +520,7 @@ func TestPersistentVolumeMultiPVs(t *testing.T) { // only one PV is bound bound := 0 for i := 0; i < maxPVs; i++ { - pv, err := testClient.PersistentVolumes().Get(pvs[i].Name) + pv, err := testClient.PersistentVolumes().Get(pvs[i].Name, metav1.GetOptions{}) if err != nil { t.Fatalf("Unexpected error getting pv: %v", err) } @@ -614,7 +614,7 @@ func TestPersistentVolumeMultiPVsPVCs(t *testing.T) { // Modify PV i := rand.Intn(objCount) name := "pv-" + strconv.Itoa(i) - pv, err := testClient.PersistentVolumes().Get(name) + pv, err := testClient.PersistentVolumes().Get(name, metav1.GetOptions{}) if err != nil { // Silently ignore error, the PV may have be already deleted // or not exists yet. @@ -638,7 +638,7 @@ func TestPersistentVolumeMultiPVsPVCs(t *testing.T) { // Modify PVC i := rand.Intn(objCount) name := "pvc-" + strconv.Itoa(i) - pvc, err := testClient.PersistentVolumeClaims(v1.NamespaceDefault).Get(name) + pvc, err := testClient.PersistentVolumeClaims(v1.NamespaceDefault).Get(name, metav1.GetOptions{}) if err != nil { // Silently ignore error, the PVC may have be already // deleted or not exists yet. @@ -693,7 +693,7 @@ func TestPersistentVolumeMultiPVsPVCs(t *testing.T) { // check that everything is bound to something for i := 0; i < objCount; i++ { - pv, err := testClient.PersistentVolumes().Get(pvs[i].Name) + pv, err := testClient.PersistentVolumes().Get(pvs[i].Name, metav1.GetOptions{}) if err != nil { t.Fatalf("Unexpected error getting pv: %v", err) } @@ -702,7 +702,7 @@ func TestPersistentVolumeMultiPVsPVCs(t *testing.T) { } glog.V(2).Infof("PV %q is bound to PVC %q", pv.Name, pv.Spec.ClaimRef.Name) - pvc, err := testClient.PersistentVolumeClaims(ns.Name).Get(pvcs[i].Name) + pvc, err := testClient.PersistentVolumeClaims(ns.Name).Get(pvcs[i].Name, metav1.GetOptions{}) if err != nil { t.Fatalf("Unexpected error getting pvc: %v", err) } @@ -822,7 +822,7 @@ func TestPersistentVolumeControllerStartup(t *testing.T) { // check that everything is bound to something for i := 0; i < objCount; i++ { - pv, err := testClient.PersistentVolumes().Get(pvs[i].Name) + pv, err := testClient.PersistentVolumes().Get(pvs[i].Name, metav1.GetOptions{}) if err != nil { t.Fatalf("Unexpected error getting pv: %v", err) } @@ -831,7 +831,7 @@ func TestPersistentVolumeControllerStartup(t *testing.T) { } glog.V(2).Infof("PV %q is bound to PVC %q", pv.Name, pv.Spec.ClaimRef.Name) - pvc, err := testClient.PersistentVolumeClaims(ns.Name).Get(pvcs[i].Name) + pvc, err := testClient.PersistentVolumeClaims(ns.Name).Get(pvcs[i].Name, metav1.GetOptions{}) if err != nil { t.Fatalf("Unexpected error getting pvc: %v", err) } @@ -991,14 +991,14 @@ func TestPersistentVolumeMultiPVsDiffAccessModes(t *testing.T) { t.Log("claim bound") // only RWM PV is bound - pv, err := testClient.PersistentVolumes().Get("pv-rwo") + pv, err := testClient.PersistentVolumes().Get("pv-rwo", metav1.GetOptions{}) if err != nil { t.Fatalf("Unexpected error getting pv: %v", err) } if pv.Spec.ClaimRef != nil { t.Fatalf("ReadWriteOnce PV shouldn't be bound") } - pv, err = testClient.PersistentVolumes().Get("pv-rwm") + pv, err = testClient.PersistentVolumes().Get("pv-rwm", metav1.GetOptions{}) if err != nil { t.Fatalf("Unexpected error getting pv: %v", err) } @@ -1021,7 +1021,7 @@ func TestPersistentVolumeMultiPVsDiffAccessModes(t *testing.T) { func waitForPersistentVolumePhase(client *clientset.Clientset, pvName string, w watch.Interface, phase v1.PersistentVolumePhase) { // Check if the volume is already in requested phase - volume, err := client.Core().PersistentVolumes().Get(pvName) + volume, err := client.Core().PersistentVolumes().Get(pvName, metav1.GetOptions{}) if err == nil && volume.Status.Phase == phase { return } @@ -1042,7 +1042,7 @@ func waitForPersistentVolumePhase(client *clientset.Clientset, pvName string, w func waitForPersistentVolumeClaimPhase(client *clientset.Clientset, claimName, namespace string, w watch.Interface, phase v1.PersistentVolumeClaimPhase) { // Check if the claim is already in requested phase - claim, err := client.Core().PersistentVolumeClaims(namespace).Get(claimName) + claim, err := client.Core().PersistentVolumeClaims(namespace).Get(claimName, metav1.GetOptions{}) if err == nil && claim.Status.Phase == phase { return } diff --git a/test/utils/density_utils.go b/test/utils/density_utils.go index 234fee897e5..c9523a279bc 100644 --- a/test/utils/density_utils.go +++ b/test/utils/density_utils.go @@ -24,6 +24,7 @@ import ( "k8s.io/kubernetes/pkg/api" apierrs "k8s.io/kubernetes/pkg/api/errors" "k8s.io/kubernetes/pkg/api/v1" + metav1 "k8s.io/kubernetes/pkg/apis/meta/v1" clientset "k8s.io/kubernetes/pkg/client/clientset_generated/release_1_5" "github.com/golang/glog" @@ -61,7 +62,7 @@ func RemoveLabelOffNode(c clientset.Interface, nodeName string, labelKeys []stri var node *v1.Node var err error for attempt := 0; attempt < retries; attempt++ { - node, err = c.Core().Nodes().Get(nodeName) + node, err = c.Core().Nodes().Get(nodeName, metav1.GetOptions{}) if err != nil { return err } @@ -92,7 +93,7 @@ func RemoveLabelOffNode(c clientset.Interface, nodeName string, labelKeys []stri // VerifyLabelsRemoved checks if Node for given nodeName does not have any of labels from labelKeys. // Return non-nil error if it does. func VerifyLabelsRemoved(c clientset.Interface, nodeName string, labelKeys []string) error { - node, err := c.Core().Nodes().Get(nodeName) + node, err := c.Core().Nodes().Get(nodeName, metav1.GetOptions{}) if err != nil { return err } diff --git a/test/utils/runners.go b/test/utils/runners.go index 67c5469edfb..19aa138a85c 100644 --- a/test/utils/runners.go +++ b/test/utils/runners.go @@ -772,7 +772,7 @@ func DoPrepareNode(client clientset.Interface, node *v1.Node, strategy PrepareNo func DoCleanupNode(client clientset.Interface, nodeName string, strategy PrepareNodeStrategy) error { for attempt := 0; attempt < retries; attempt++ { - node, err := client.Core().Nodes().Get(nodeName) + node, err := client.Core().Nodes().Get(nodeName, metav1.GetOptions{}) if err != nil { return fmt.Errorf("Skipping cleanup of Node: failed to get Node %v: %v", nodeName, err) }