diff --git a/cmd/kube-apiserver/app/testing/testserver.go b/cmd/kube-apiserver/app/testing/testserver.go index d1a26dc7592..5638a0e0afe 100644 --- a/cmd/kube-apiserver/app/testing/testserver.go +++ b/cmd/kube-apiserver/app/testing/testserver.go @@ -106,7 +106,7 @@ func StartTestServer(t *testing.T) (result *restclient.Config, tearDownForCaller default: } - result := client.CoreV1Client.RESTClient().Get().AbsPath("/healthz").Do() + result := client.CoreV1().RESTClient().Get().AbsPath("/healthz").Do() status := 0 result.StatusCode(&status) if status == 200 { diff --git a/cmd/kubeadm/app/phases/markmaster/markmaster_test.go b/cmd/kubeadm/app/phases/markmaster/markmaster_test.go index ea82a45a804..02dda93a06b 100644 --- a/cmd/kubeadm/app/phases/markmaster/markmaster_test.go +++ b/cmd/kubeadm/app/phases/markmaster/markmaster_test.go @@ -27,7 +27,6 @@ import ( "k8s.io/api/core/v1" metav1 "k8s.io/apimachinery/pkg/apis/meta/v1" clientset "k8s.io/client-go/kubernetes" - corev1 "k8s.io/client-go/kubernetes/typed/core/v1" restclient "k8s.io/client-go/rest" kubeadmconstants "k8s.io/kubernetes/cmd/kubeadm/app/constants" kubeletapis "k8s.io/kubernetes/pkg/kubelet/apis" @@ -121,7 +120,7 @@ func TestMarkMaster(t *testing.T) { })) defer s.Close() - cs, err := clientsetFromTestServer(s) + cs, err := clientset.NewForConfig(&restclient.Config{Host: s.URL}) if err != nil { t.Fatalf("MarkMaster(%s): unexpected error building clientset: %v", tc.name, err) } @@ -137,15 +136,6 @@ func TestMarkMaster(t *testing.T) { } } -func clientsetFromTestServer(s *httptest.Server) (*clientset.Clientset, error) { - rc := &restclient.Config{Host: s.URL} - c, err := corev1.NewForConfig(rc) - if err != nil { - return nil, err - } - return &clientset.Clientset{CoreV1Client: c}, nil -} - func toString(r io.Reader) string { buf := new(bytes.Buffer) buf.ReadFrom(r) diff --git a/cmd/kubelet/app/server.go b/cmd/kubelet/app/server.go index b005383289a..77cf0c37e7e 100644 --- a/cmd/kubelet/app/server.go +++ b/cmd/kubelet/app/server.go @@ -203,14 +203,14 @@ func getRemoteKubeletConfig(s *options.KubeletServer, kubeDeps *kubelet.Dependen return nil, err } // look for kubelet- configmap from "kube-system" - configmap, err := kubeClient.CoreV1Client.ConfigMaps("kube-system").Get(fmt.Sprintf("kubelet-%s", nodename), metav1.GetOptions{}) + configmap, err := kubeClient.CoreV1().ConfigMaps("kube-system").Get(fmt.Sprintf("kubelet-%s", nodename), metav1.GetOptions{}) if err != nil { return nil, err } return configmap, nil } // No cloud provider yet, so can't get the nodename via Cloud.Instances().CurrentNodeName(hostname), try just using the hostname - configmap, err := kubeClient.CoreV1Client.ConfigMaps("kube-system").Get(fmt.Sprintf("kubelet-%s", hostname), metav1.GetOptions{}) + configmap, err := kubeClient.CoreV1().ConfigMaps("kube-system").Get(fmt.Sprintf("kubelet-%s", hostname), metav1.GetOptions{}) if err != nil { return nil, fmt.Errorf("cloud provider was nil, and attempt to use hostname to find config resulted in: %v", err) } diff --git a/cmd/kubemark/hollow-node.go b/cmd/kubemark/hollow-node.go index 025a954f72e..a6f7d307475 100644 --- a/cmd/kubemark/hollow-node.go +++ b/cmd/kubemark/hollow-node.go @@ -147,7 +147,7 @@ func main() { hollowProxy, err := kubemark.NewHollowProxyOrDie( config.NodeName, internalClientset, - client.Core(), + client.CoreV1(), iptInterface, sysctl, execer, diff --git a/pkg/kubectl/cmd/testing/fake.go b/pkg/kubectl/cmd/testing/fake.go index 0f66cafa143..b69bc1123ef 100644 --- a/pkg/kubectl/cmd/testing/fake.go +++ b/pkg/kubectl/cmd/testing/fake.go @@ -587,17 +587,17 @@ func (f *fakeAPIFactory) ClientSet() (internalclientset.Interface, error) { // version. fakeClient := f.tf.Client.(*fake.RESTClient) clientset := internalclientset.NewForConfigOrDie(f.tf.ClientConfig) - clientset.CoreClient.RESTClient().(*restclient.RESTClient).Client = fakeClient.Client - clientset.AuthenticationClient.RESTClient().(*restclient.RESTClient).Client = fakeClient.Client - clientset.AuthorizationClient.RESTClient().(*restclient.RESTClient).Client = fakeClient.Client - clientset.AutoscalingClient.RESTClient().(*restclient.RESTClient).Client = fakeClient.Client - clientset.BatchClient.RESTClient().(*restclient.RESTClient).Client = fakeClient.Client - clientset.CertificatesClient.RESTClient().(*restclient.RESTClient).Client = fakeClient.Client - clientset.ExtensionsClient.RESTClient().(*restclient.RESTClient).Client = fakeClient.Client - clientset.RbacClient.RESTClient().(*restclient.RESTClient).Client = fakeClient.Client - clientset.StorageClient.RESTClient().(*restclient.RESTClient).Client = fakeClient.Client - clientset.AppsClient.RESTClient().(*restclient.RESTClient).Client = fakeClient.Client - clientset.PolicyClient.RESTClient().(*restclient.RESTClient).Client = fakeClient.Client + clientset.Core().RESTClient().(*restclient.RESTClient).Client = fakeClient.Client + clientset.Authentication().RESTClient().(*restclient.RESTClient).Client = fakeClient.Client + clientset.Authorization().RESTClient().(*restclient.RESTClient).Client = fakeClient.Client + clientset.Autoscaling().RESTClient().(*restclient.RESTClient).Client = fakeClient.Client + clientset.Batch().RESTClient().(*restclient.RESTClient).Client = fakeClient.Client + clientset.Certificates().RESTClient().(*restclient.RESTClient).Client = fakeClient.Client + clientset.Extensions().RESTClient().(*restclient.RESTClient).Client = fakeClient.Client + clientset.Rbac().RESTClient().(*restclient.RESTClient).Client = fakeClient.Client + clientset.Storage().RESTClient().(*restclient.RESTClient).Client = fakeClient.Client + clientset.Apps().RESTClient().(*restclient.RESTClient).Client = fakeClient.Client + clientset.Policy().RESTClient().(*restclient.RESTClient).Client = fakeClient.Client clientset.DiscoveryClient.RESTClient().(*restclient.RESTClient).Client = fakeClient.Client return clientset, f.tf.Err } diff --git a/pkg/kubectl/rolling_updater_test.go b/pkg/kubectl/rolling_updater_test.go index 922104a3c9a..2070df1d554 100644 --- a/pkg/kubectl/rolling_updater_test.go +++ b/pkg/kubectl/rolling_updater_test.go @@ -1501,7 +1501,7 @@ func TestUpdateRcWithRetries(t *testing.T) { clientset := internalclientset.New(restClient) if rc, err := updateRcWithRetries( - clientset, "default", rc, func(c *api.ReplicationController) { + clientset.Core(), "default", rc, func(c *api.ReplicationController) { c.Spec.Selector["baz"] = "foobar" }); err != nil { t.Errorf("unexpected error: %v", err) diff --git a/staging/src/k8s.io/apiextensions-apiserver/pkg/apiserver/apiserver.go b/staging/src/k8s.io/apiextensions-apiserver/pkg/apiserver/apiserver.go index 253f8523a5c..e178bd140a6 100644 --- a/staging/src/k8s.io/apiextensions-apiserver/pkg/apiserver/apiserver.go +++ b/staging/src/k8s.io/apiextensions-apiserver/pkg/apiserver/apiserver.go @@ -184,10 +184,10 @@ func (c completedConfig) New(delegationTarget genericapiserver.DelegationTarget) } crdController := NewDiscoveryController(s.Informers.Apiextensions().InternalVersion().CustomResourceDefinitions(), versionDiscoveryHandler, groupDiscoveryHandler, c.GenericConfig.RequestContextMapper) - namingController := status.NewNamingConditionController(s.Informers.Apiextensions().InternalVersion().CustomResourceDefinitions(), crdClient) + namingController := status.NewNamingConditionController(s.Informers.Apiextensions().InternalVersion().CustomResourceDefinitions(), crdClient.Apiextensions()) finalizingController := finalizer.NewCRDFinalizer( s.Informers.Apiextensions().InternalVersion().CustomResourceDefinitions(), - crdClient, + crdClient.Apiextensions(), crdHandler, ) diff --git a/test/e2e/apps/deployment.go b/test/e2e/apps/deployment.go index 6434fa0160d..95ce897699d 100644 --- a/test/e2e/apps/deployment.go +++ b/test/e2e/apps/deployment.go @@ -132,7 +132,7 @@ func failureTrap(c clientset.Interface, ns string) { d := deployments.Items[i] framework.Logf(spew.Sprintf("Deployment %q:\n%+v\n", d.Name, d)) - _, allOldRSs, newRS, err := deploymentutil.GetAllReplicaSets(&d, c) + _, allOldRSs, newRS, err := deploymentutil.GetAllReplicaSets(&d, c.ExtensionsV1beta1()) if err != nil { framework.Logf("Could not list ReplicaSets for Deployment %q: %v", d.Name, err) return @@ -887,7 +887,7 @@ func testDeploymentLabelAdopted(f *framework.Framework) { // There should be no old RSs (overlapping RS) deployment, err := c.Extensions().Deployments(ns).Get(deploymentName, metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred()) - oldRSs, allOldRSs, newRS, err := deploymentutil.GetAllReplicaSets(deployment, c) + oldRSs, allOldRSs, newRS, err := deploymentutil.GetAllReplicaSets(deployment, c.ExtensionsV1beta1()) Expect(err).NotTo(HaveOccurred()) Expect(len(oldRSs)).Should(Equal(0)) Expect(len(allOldRSs)).Should(Equal(0)) @@ -1035,7 +1035,7 @@ func testScaledRolloutDeployment(f *framework.Framework) { framework.Logf("Waiting for deployment status to sync (current available: %d, minimum available: %d)", deployment.Status.AvailableReplicas, deploymentutil.MinAvailable(deployment)) Expect(framework.WaitForDeploymentStatusValid(c, deployment)).NotTo(HaveOccurred()) - oldRSs, _, rs, err := deploymentutil.GetAllReplicaSets(deployment, c) + oldRSs, _, rs, err := deploymentutil.GetAllReplicaSets(deployment, c.ExtensionsV1beta1()) Expect(err).NotTo(HaveOccurred()) for _, rs := range append(oldRSs, rs) { @@ -1095,7 +1095,7 @@ func testScaledRolloutDeployment(f *framework.Framework) { framework.Logf("Waiting for deployment status to sync (current available: %d, minimum available: %d)", deployment.Status.AvailableReplicas, deploymentutil.MinAvailable(deployment)) Expect(framework.WaitForDeploymentStatusValid(c, deployment)).NotTo(HaveOccurred()) - oldRSs, _, rs, err = deploymentutil.GetAllReplicaSets(deployment, c) + oldRSs, _, rs, err = deploymentutil.GetAllReplicaSets(deployment, c.ExtensionsV1beta1()) Expect(err).NotTo(HaveOccurred()) for _, rs := range append(oldRSs, rs) { diff --git a/test/e2e/apps/disruption.go b/test/e2e/apps/disruption.go index f9dd59f4450..0a7db3b0557 100644 --- a/test/e2e/apps/disruption.go +++ b/test/e2e/apps/disruption.go @@ -163,7 +163,7 @@ var _ = SIGDescribe("DisruptionController", func() { // Locate a running pod. var pod v1.Pod err := wait.PollImmediate(framework.Poll, schedulingTimeout, func() (bool, error) { - podList, err := cs.Pods(ns).List(metav1.ListOptions{}) + podList, err := cs.CoreV1().Pods(ns).List(metav1.ListOptions{}) if err != nil { return false, err } @@ -191,7 +191,7 @@ var _ = SIGDescribe("DisruptionController", func() { // this gives the controller enough time to have truly set the status. time.Sleep(timeout) - err = cs.Pods(ns).Evict(e) + err = cs.CoreV1().Pods(ns).Evict(e) Expect(err).Should(MatchError("Cannot evict pod as it would violate the pod's disruption budget.")) } else { // Only wait for running pods in the "allow" case @@ -202,7 +202,7 @@ var _ = SIGDescribe("DisruptionController", func() { // Since disruptionAllowed starts out false, if an eviction is ever allowed, // that means the controller is working. err = wait.PollImmediate(framework.Poll, timeout, func() (bool, error) { - err = cs.Pods(ns).Evict(e) + err = cs.CoreV1().Pods(ns).Evict(e) if err != nil { return false, nil } else { @@ -264,7 +264,7 @@ func createPodsOrDie(cs *kubernetes.Clientset, ns string, n int) { }, } - _, err := cs.Pods(ns).Create(pod) + _, err := cs.CoreV1().Pods(ns).Create(pod) framework.ExpectNoError(err, "Creating pod %q in namespace %q", pod.Name, ns) } } diff --git a/test/e2e/upgrades/apps/deployments.go b/test/e2e/upgrades/apps/deployments.go index b4089272d9c..b796b16317d 100644 --- a/test/e2e/upgrades/apps/deployments.go +++ b/test/e2e/upgrades/apps/deployments.go @@ -141,7 +141,7 @@ func (t *DeploymentUpgradeTest) Test(f *framework.Framework, done <-chan struct{ framework.ExpectNoError(framework.WaitForDeploymentStatus(c, deployment)) By(fmt.Sprintf("Checking that replica sets for deployment %q are the same as prior to the upgrade", t.updatedD.Name)) - _, allOldRSs, newRS, err := deploymentutil.GetAllReplicaSets(t.updatedD, c) + _, allOldRSs, newRS, err := deploymentutil.GetAllReplicaSets(t.updatedD, c.ExtensionsV1beta1()) framework.ExpectNoError(err) if newRS == nil { By(t.spewDeploymentAndReplicaSets(newRS, allOldRSs)) diff --git a/test/e2e_federation/ingress.go b/test/e2e_federation/ingress.go index d0043ca4247..bc77721a22e 100644 --- a/test/e2e_federation/ingress.go +++ b/test/e2e_federation/ingress.go @@ -327,7 +327,7 @@ func waitForIngressOrFail(clientset *kubeclientset.Clientset, namespace string, var clusterIngress *v1beta1.Ingress err := wait.PollImmediate(framework.Poll, timeout, func() (bool, error) { var err error - clusterIngress, err = clientset.Ingresses(namespace).Get(ingress.Name, metav1.GetOptions{}) + clusterIngress, err = clientset.ExtensionsV1beta1().Ingresses(namespace).Get(ingress.Name, metav1.GetOptions{}) if (!present) && errors.IsNotFound(err) { // We want it gone, and it's gone. By(fmt.Sprintf("Success: shard of federated ingress %q in namespace %q in cluster is absent", ingress.Name, namespace)) return true, nil // Success @@ -367,7 +367,7 @@ func waitForIngressShardsUpdatedOrFail(namespace string, ingress *v1beta1.Ingres func waitForIngressUpdateOrFail(clientset *kubeclientset.Clientset, namespace string, ingress *v1beta1.Ingress, timeout time.Duration) { By(fmt.Sprintf("Fetching a federated ingress shard of ingress %q in namespace %q from cluster", ingress.Name, namespace)) err := wait.PollImmediate(framework.Poll, timeout, func() (bool, error) { - clusterIngress, err := clientset.Ingresses(namespace).Get(ingress.Name, metav1.GetOptions{}) + clusterIngress, err := clientset.ExtensionsV1beta1().Ingresses(namespace).Get(ingress.Name, metav1.GetOptions{}) if err == nil { // We want it present, and the Get succeeded, so we're all good. if equivalentIngress(*clusterIngress, *ingress) { By(fmt.Sprintf("Success: shard of federated ingress %q in namespace %q in cluster is updated", ingress.Name, namespace)) @@ -394,7 +394,7 @@ func deleteIngressOrFail(clientset *fedclientset.Clientset, namespace string, in if clientset == nil || len(namespace) == 0 || len(ingressName) == 0 { Fail(fmt.Sprintf("Internal error: invalid parameters passed to deleteIngressOrFail: clientset: %v, namespace: %v, ingress: %v", clientset, namespace, ingressName)) } - err := clientset.Ingresses(namespace).Delete(ingressName, &metav1.DeleteOptions{OrphanDependents: orphanDependents}) + err := clientset.ExtensionsV1beta1().Ingresses(namespace).Delete(ingressName, &metav1.DeleteOptions{OrphanDependents: orphanDependents}) framework.ExpectNoError(err, "Error deleting ingress %q from namespace %q", ingressName, namespace) // Wait for the ingress to be deleted. err = wait.Poll(framework.Poll, FederatedIngressDeleteTimeout, func() (bool, error) { @@ -414,7 +414,7 @@ func deleteClusterIngressOrFail(clusterName string, clientset *kubeclientset.Cli if clientset == nil || len(namespace) == 0 || len(ingressName) == 0 { Fail(fmt.Sprintf("Internal error: invalid parameters passed to deleteClusterIngressOrFail: cluster: %q, clientset: %v, namespace: %v, ingress: %v", clusterName, clientset, namespace, ingressName)) } - err := clientset.Ingresses(namespace).Delete(ingressName, metav1.NewDeleteOptions(0)) + err := clientset.ExtensionsV1beta1().Ingresses(namespace).Delete(ingressName, metav1.NewDeleteOptions(0)) framework.ExpectNoError(err, "Error deleting cluster ingress %q/%q from cluster %q", namespace, ingressName, clusterName) } diff --git a/test/e2e_federation/replicaset.go b/test/e2e_federation/replicaset.go index 45bc9802097..6df94ee85e3 100644 --- a/test/e2e_federation/replicaset.go +++ b/test/e2e_federation/replicaset.go @@ -311,7 +311,7 @@ func waitForReplicaSetOrFail(c *fedclientset.Clientset, namespace string, replic func waitForReplicaSet(c *fedclientset.Clientset, namespace string, replicaSetName string, clusters fedframework.ClusterSlice, expect map[string]int32) error { framework.Logf("waitForReplicaSet: %s/%s; clusters: %v; expect: %v", namespace, replicaSetName, clusters, expect) err := wait.Poll(10*time.Second, fedframework.FederatedDefaultTestTimeout, func() (bool, error) { - frs, err := c.ReplicaSets(namespace).Get(replicaSetName, metav1.GetOptions{}) + frs, err := c.ExtensionsV1beta1().ReplicaSets(namespace).Get(replicaSetName, metav1.GetOptions{}) if err != nil { return false, err } @@ -392,7 +392,7 @@ func updateReplicaSetOrFail(clientset *fedclientset.Clientset, replicaset *v1bet } By(fmt.Sprintf("Updating federation replicaset %q in namespace %q", replicaset.Name, namespace)) - newRS, err := clientset.ReplicaSets(namespace).Update(replicaset) + newRS, err := clientset.ExtensionsV1beta1().ReplicaSets(namespace).Update(replicaset) framework.ExpectNoError(err, "Updating replicaset %q in namespace %q", replicaset.Name, namespace) By(fmt.Sprintf("Successfully updated federation replicaset %q in namespace %q", replicaset.Name, namespace)) diff --git a/test/e2e_federation/service.go b/test/e2e_federation/service.go index a7940675c1e..6dda940c687 100644 --- a/test/e2e_federation/service.go +++ b/test/e2e_federation/service.go @@ -68,7 +68,7 @@ var _ = framework.KubeDescribe("Federated Services [Feature:Federation]", func() By(fmt.Sprintf("Creation of service %q in namespace %q succeeded. Deleting service.", service.Name, nsName)) // Cleanup - err := f.FederationClientset.Services(nsName).Delete(service.Name, &metav1.DeleteOptions{}) + err := f.FederationClientset.CoreV1().Services(nsName).Delete(service.Name, &metav1.DeleteOptions{}) framework.ExpectNoError(err, "Error deleting service %q in namespace %q", service.Name, service.Namespace) By(fmt.Sprintf("Deletion of service %q in namespace %q succeeded.", service.Name, nsName)) }) @@ -113,7 +113,7 @@ var _ = framework.KubeDescribe("Federated Services [Feature:Federation]", func() service = createServiceOrFail(f.FederationClientset, nsName, FederatedServiceName) defer func() { // Cleanup By(fmt.Sprintf("Deleting service %q in namespace %q", service.Name, nsName)) - err := f.FederationClientset.Services(nsName).Delete(service.Name, &metav1.DeleteOptions{}) + err := f.FederationClientset.CoreV1().Services(nsName).Delete(service.Name, &metav1.DeleteOptions{}) framework.ExpectNoError(err, "Error deleting service %q in namespace %q", service.Name, nsName) }() By(fmt.Sprintf("Wait for service shards to be created in all clusters for service \"%s/%s\"", nsName, service.Name)) @@ -151,7 +151,7 @@ var _ = framework.KubeDescribe("Federated Services [Feature:Federation]", func() defer func() { // Cleanup By(fmt.Sprintf("Deleting service %q in namespace %q", service.Name, nsName)) - err := f.FederationClientset.Services(nsName).Delete(service.Name, &metav1.DeleteOptions{}) + err := f.FederationClientset.CoreV1().Services(nsName).Delete(service.Name, &metav1.DeleteOptions{}) framework.ExpectNoError(err, "Error deleting service %q in namespace %q", service.Name, nsName) }() By(fmt.Sprintf("Wait for service shards to be created in all clusters for service \"%s/%s\"", nsName, service.Name)) @@ -258,7 +258,7 @@ var _ = framework.KubeDescribe("Federated Services [Feature:Federation]", func() By("Verified that DNS rules are working as expected") By("Deleting the service to verify that DNS rules still work") - err := f.FederationClientset.Services(nsName).Delete(FederatedServiceName, &metav1.DeleteOptions{}) + err := f.FederationClientset.CoreV1().Services(nsName).Delete(FederatedServiceName, &metav1.DeleteOptions{}) framework.ExpectNoError(err, "Error deleting service %q in namespace %q", service.Name, service.Namespace) // Service is deleted, unset the test block-global service variable. service = nil @@ -359,16 +359,16 @@ func verifyCascadingDeletionForService(clientset *fedclientset.Clientset, cluste } func updateServiceOrFail(clientset *fedclientset.Clientset, namespace, name string) *v1.Service { - service, err := clientset.Services(namespace).Get(name, metav1.GetOptions{}) + service, err := clientset.CoreV1().Services(namespace).Get(name, metav1.GetOptions{}) framework.ExpectNoError(err, "Getting service %q in namespace %q", name, namespace) service.Spec.Selector["name"] = "update-demo" - newService, err := clientset.Services(namespace).Update(service) + newService, err := clientset.CoreV1().Services(namespace).Update(service) By(fmt.Sprintf("Successfully updated federated service %q in namespace %q", name, namespace)) return newService } func deleteServiceShard(c *fedframework.Cluster, namespace, service string) error { - err := c.Clientset.Services(namespace).Delete(service, &metav1.DeleteOptions{}) + err := c.Clientset.CoreV1().Services(namespace).Delete(service, &metav1.DeleteOptions{}) if err != nil && !errors.IsNotFound(err) { framework.Logf("Failed to delete service %q in namespace %q, in cluster %q", service, namespace, c.Name) return err diff --git a/test/e2e_federation/util.go b/test/e2e_federation/util.go index 7e7735d9501..0ddf52c0da5 100644 --- a/test/e2e_federation/util.go +++ b/test/e2e_federation/util.go @@ -88,7 +88,7 @@ func createClusterObjectOrFail(f *fedframework.Framework, context *fedframework. func waitForServiceOrFail(clientset *kubeclientset.Clientset, namespace string, service *v1.Service, present bool, timeout time.Duration) { By(fmt.Sprintf("Fetching a federated service shard of service %q in namespace %q from cluster", service.Name, namespace)) err := wait.PollImmediate(framework.Poll, timeout, func() (bool, error) { - clusterService, err := clientset.Services(namespace).Get(service.Name, metav1.GetOptions{}) + clusterService, err := clientset.CoreV1().Services(namespace).Get(service.Name, metav1.GetOptions{}) if (!present) && errors.IsNotFound(err) { // We want it gone, and it's gone. By(fmt.Sprintf("Success: shard of federated service %q in namespace %q in cluster is absent", service.Name, namespace)) return true, nil // Success @@ -141,7 +141,7 @@ func createService(clientset *fedclientset.Clientset, namespace, name string) (* } By(fmt.Sprintf("Trying to create service %q in namespace %q", service.Name, namespace)) - return clientset.Services(namespace).Create(service) + return clientset.CoreV1().Services(namespace).Create(service) } func createLBService(clientset *fedclientset.Clientset, namespace, name string) (*v1.Service, error) { @@ -178,7 +178,7 @@ func createLBService(clientset *fedclientset.Clientset, namespace, name string) } By(fmt.Sprintf("Trying to create service %q in namespace %q", service.Name, namespace)) - return clientset.Services(namespace).Create(service) + return clientset.CoreV1().Services(namespace).Create(service) } func createServiceOrFail(clientset *fedclientset.Clientset, namespace, name string) *v1.Service { @@ -200,7 +200,7 @@ func deleteServiceOrFail(clientset *fedclientset.Clientset, namespace string, se Fail(fmt.Sprintf("Internal error: invalid parameters passed to deleteServiceOrFail: clientset: %v, namespace: %v, service: %v", clientset, namespace, serviceName)) } framework.Logf("Deleting service %q in namespace %v", serviceName, namespace) - err := clientset.Services(namespace).Delete(serviceName, &metav1.DeleteOptions{OrphanDependents: orphanDependents}) + err := clientset.CoreV1().Services(namespace).Delete(serviceName, &metav1.DeleteOptions{OrphanDependents: orphanDependents}) framework.ExpectNoError(err, "Error deleting service %q from namespace %q", serviceName, namespace) // Wait for the service to be deleted. err = wait.Poll(5*time.Second, fedframework.FederatedDefaultTestTimeout, func() (bool, error) { @@ -224,7 +224,7 @@ func cleanupServiceShardsAndProviderResources(namespace string, service *v1.Serv err := wait.PollImmediate(framework.Poll, fedframework.FederatedDefaultTestTimeout, func() (bool, error) { var err error - cSvc, err = c.Clientset.Services(namespace).Get(service.Name, metav1.GetOptions{}) + cSvc, err = c.Clientset.CoreV1().Services(namespace).Get(service.Name, metav1.GetOptions{}) if err != nil && !errors.IsNotFound(err) { // Get failed with an error, try again. framework.Logf("Failed to find service %q in namespace %q, in cluster %q: %v. Trying again in %s", service.Name, namespace, name, err, framework.Poll) @@ -260,7 +260,7 @@ func cleanupServiceShardsAndProviderResources(namespace string, service *v1.Serv func cleanupServiceShard(clientset *kubeclientset.Clientset, clusterName, namespace string, service *v1.Service, timeout time.Duration) error { err := wait.PollImmediate(framework.Poll, timeout, func() (bool, error) { - err := clientset.Services(namespace).Delete(service.Name, &metav1.DeleteOptions{}) + err := clientset.CoreV1().Services(namespace).Delete(service.Name, &metav1.DeleteOptions{}) if err != nil && !errors.IsNotFound(err) { // Deletion failed with an error, try again. framework.Logf("Failed to delete service %q in namespace %q, in cluster %q", service.Name, namespace, clusterName) diff --git a/test/e2e_node/e2e_node_suite_test.go b/test/e2e_node/e2e_node_suite_test.go index 9a228bb4fc2..dbef75738fa 100644 --- a/test/e2e_node/e2e_node_suite_test.go +++ b/test/e2e_node/e2e_node_suite_test.go @@ -265,7 +265,7 @@ func updateTestContext() error { // getNode gets node object from the apiserver. func getNode(c *clientset.Clientset) (*v1.Node, error) { - nodes, err := c.Nodes().List(metav1.ListOptions{}) + nodes, err := c.CoreV1().Nodes().List(metav1.ListOptions{}) Expect(err).NotTo(HaveOccurred(), "should be able to list nodes.") if nodes == nil { return nil, fmt.Errorf("the node list is nil.") diff --git a/test/integration/master/master_test.go b/test/integration/master/master_test.go index d7378662ef3..dd7f5fd0dc4 100644 --- a/test/integration/master/master_test.go +++ b/test/integration/master/master_test.go @@ -99,7 +99,7 @@ func TestKubernetesService(t *testing.T) { _, _, closeFn := framework.RunAMaster(config) defer closeFn() coreClient := clientset.NewForConfigOrDie(config.GenericConfig.LoopbackClientConfig) - if _, err := coreClient.Services(metav1.NamespaceDefault).Get("kubernetes", metav1.GetOptions{}); err != nil { + if _, err := coreClient.Core().Services(metav1.NamespaceDefault).Get("kubernetes", metav1.GetOptions{}); err != nil { t.Fatalf("Expected kubernetes service to exists, got: %v", err) } } diff --git a/test/integration/volume/persistent_volumes_test.go b/test/integration/volume/persistent_volumes_test.go index d79ffbe918f..b50efd829a8 100644 --- a/test/integration/volume/persistent_volumes_test.go +++ b/test/integration/volume/persistent_volumes_test.go @@ -127,13 +127,13 @@ func TestPersistentVolumeRecycler(t *testing.T) { pv := createPV("fake-pv-recycler", "/tmp/foo", "10G", []v1.PersistentVolumeAccessMode{v1.ReadWriteOnce}, v1.PersistentVolumeReclaimRecycle) pvc := createPVC("fake-pvc-recycler", ns.Name, "5G", []v1.PersistentVolumeAccessMode{v1.ReadWriteOnce}, "") - _, err := testClient.PersistentVolumes().Create(pv) + _, err := testClient.CoreV1().PersistentVolumes().Create(pv) if err != nil { t.Errorf("Failed to create PersistentVolume: %v", err) } glog.V(2).Infof("TestPersistentVolumeRecycler pvc created") - _, err = testClient.PersistentVolumeClaims(ns.Name).Create(pvc) + _, err = testClient.CoreV1().PersistentVolumeClaims(ns.Name).Create(pvc) if err != nil { t.Errorf("Failed to create PersistentVolumeClaim: %v", err) } @@ -146,7 +146,7 @@ func TestPersistentVolumeRecycler(t *testing.T) { glog.V(2).Infof("TestPersistentVolumeRecycler pvc bound") // deleting a claim releases the volume, after which it can be recycled - if err := testClient.PersistentVolumeClaims(ns.Name).Delete(pvc.Name, nil); err != nil { + if err := testClient.CoreV1().PersistentVolumeClaims(ns.Name).Delete(pvc.Name, nil); err != nil { t.Errorf("error deleting claim %s", pvc.Name) } glog.V(2).Infof("TestPersistentVolumeRecycler pvc deleted") @@ -182,12 +182,12 @@ func TestPersistentVolumeDeleter(t *testing.T) { pv := createPV("fake-pv-deleter", "/tmp/foo", "10G", []v1.PersistentVolumeAccessMode{v1.ReadWriteOnce}, v1.PersistentVolumeReclaimDelete) pvc := createPVC("fake-pvc-deleter", ns.Name, "5G", []v1.PersistentVolumeAccessMode{v1.ReadWriteOnce}, "") - _, err := testClient.PersistentVolumes().Create(pv) + _, err := testClient.CoreV1().PersistentVolumes().Create(pv) if err != nil { t.Errorf("Failed to create PersistentVolume: %v", err) } glog.V(2).Infof("TestPersistentVolumeDeleter pv created") - _, err = testClient.PersistentVolumeClaims(ns.Name).Create(pvc) + _, err = testClient.CoreV1().PersistentVolumeClaims(ns.Name).Create(pvc) if err != nil { t.Errorf("Failed to create PersistentVolumeClaim: %v", err) } @@ -198,7 +198,7 @@ func TestPersistentVolumeDeleter(t *testing.T) { glog.V(2).Infof("TestPersistentVolumeDeleter pvc bound") // deleting a claim releases the volume, after which it can be recycled - if err := testClient.PersistentVolumeClaims(ns.Name).Delete(pvc.Name, nil); err != nil { + if err := testClient.CoreV1().PersistentVolumeClaims(ns.Name).Delete(pvc.Name, nil); err != nil { t.Errorf("error deleting claim %s", pvc.Name) } glog.V(2).Infof("TestPersistentVolumeDeleter pvc deleted") @@ -248,7 +248,7 @@ func TestPersistentVolumeBindRace(t *testing.T) { clone, _ := api.Scheme.DeepCopy(pvc) newPvc, _ := clone.(*v1.PersistentVolumeClaim) newPvc.ObjectMeta = metav1.ObjectMeta{Name: fmt.Sprintf("fake-pvc-race-%d", counter)} - claim, err := testClient.PersistentVolumeClaims(ns.Name).Create(newPvc) + claim, err := testClient.CoreV1().PersistentVolumeClaims(ns.Name).Create(newPvc) if err != nil { t.Fatalf("Error creating newPvc: %v", err) } @@ -266,7 +266,7 @@ func TestPersistentVolumeBindRace(t *testing.T) { pv.Spec.ClaimRef = claimRef pv.Spec.ClaimRef.UID = "" - pv, err = testClient.PersistentVolumes().Create(pv) + pv, err = testClient.CoreV1().PersistentVolumes().Create(pv) if err != nil { t.Fatalf("Unexpected error creating pv: %v", err) } @@ -277,7 +277,7 @@ func TestPersistentVolumeBindRace(t *testing.T) { waitForAnyPersistentVolumeClaimPhase(watchPVC, v1.ClaimBound) glog.V(2).Infof("TestPersistentVolumeBindRace pvc bound") - pv, err = testClient.PersistentVolumes().Get(pv.Name, metav1.GetOptions{}) + pv, err = testClient.CoreV1().PersistentVolumes().Get(pv.Name, metav1.GetOptions{}) if err != nil { t.Fatalf("Unexpected error getting pv: %v", err) } @@ -323,11 +323,11 @@ func TestPersistentVolumeClaimLabelSelector(t *testing.T) { pv_true.ObjectMeta.SetLabels(map[string]string{"foo": "true"}) pv_false.ObjectMeta.SetLabels(map[string]string{"foo": "false"}) - _, err = testClient.PersistentVolumes().Create(pv_true) + _, err = testClient.CoreV1().PersistentVolumes().Create(pv_true) if err != nil { t.Fatalf("Failed to create PersistentVolume: %v", err) } - _, err = testClient.PersistentVolumes().Create(pv_false) + _, err = testClient.CoreV1().PersistentVolumes().Create(pv_false) if err != nil { t.Fatalf("Failed to create PersistentVolume: %v", err) } @@ -339,7 +339,7 @@ func TestPersistentVolumeClaimLabelSelector(t *testing.T) { }, } - _, err = testClient.PersistentVolumeClaims(ns.Name).Create(pvc) + _, err = testClient.CoreV1().PersistentVolumeClaims(ns.Name).Create(pvc) if err != nil { t.Fatalf("Failed to create PersistentVolumeClaim: %v", err) } @@ -350,14 +350,14 @@ func TestPersistentVolumeClaimLabelSelector(t *testing.T) { waitForPersistentVolumeClaimPhase(testClient, pvc.Name, ns.Name, watchPVC, v1.ClaimBound) t.Log("claim bound") - pv, err := testClient.PersistentVolumes().Get("pv-false", metav1.GetOptions{}) + pv, err := testClient.CoreV1().PersistentVolumes().Get("pv-false", metav1.GetOptions{}) if err != nil { t.Fatalf("Unexpected error getting pv: %v", err) } if pv.Spec.ClaimRef != nil { t.Fatalf("False PV shouldn't be bound") } - pv, err = testClient.PersistentVolumes().Get("pv-true", metav1.GetOptions{}) + pv, err = testClient.CoreV1().PersistentVolumes().Get("pv-true", metav1.GetOptions{}) if err != nil { t.Fatalf("Unexpected error getting pv: %v", err) } @@ -404,11 +404,11 @@ func TestPersistentVolumeClaimLabelSelectorMatchExpressions(t *testing.T) { pv_true.ObjectMeta.SetLabels(map[string]string{"foo": "valA", "bar": ""}) pv_false.ObjectMeta.SetLabels(map[string]string{"foo": "valB", "baz": ""}) - _, err = testClient.PersistentVolumes().Create(pv_true) + _, err = testClient.CoreV1().PersistentVolumes().Create(pv_true) if err != nil { t.Fatalf("Failed to create PersistentVolume: %v", err) } - _, err = testClient.PersistentVolumes().Create(pv_false) + _, err = testClient.CoreV1().PersistentVolumes().Create(pv_false) if err != nil { t.Fatalf("Failed to create PersistentVolume: %v", err) } @@ -439,7 +439,7 @@ func TestPersistentVolumeClaimLabelSelectorMatchExpressions(t *testing.T) { }, } - _, err = testClient.PersistentVolumeClaims(ns.Name).Create(pvc) + _, err = testClient.CoreV1().PersistentVolumeClaims(ns.Name).Create(pvc) if err != nil { t.Fatalf("Failed to create PersistentVolumeClaim: %v", err) } @@ -450,14 +450,14 @@ func TestPersistentVolumeClaimLabelSelectorMatchExpressions(t *testing.T) { waitForPersistentVolumeClaimPhase(testClient, pvc.Name, ns.Name, watchPVC, v1.ClaimBound) t.Log("claim bound") - pv, err := testClient.PersistentVolumes().Get("pv-false", metav1.GetOptions{}) + pv, err := testClient.CoreV1().PersistentVolumes().Get("pv-false", metav1.GetOptions{}) if err != nil { t.Fatalf("Unexpected error getting pv: %v", err) } if pv.Spec.ClaimRef != nil { t.Fatalf("False PV shouldn't be bound") } - pv, err = testClient.PersistentVolumes().Get("pv-true", metav1.GetOptions{}) + pv, err = testClient.CoreV1().PersistentVolumes().Get("pv-true", metav1.GetOptions{}) if err != nil { t.Fatalf("Unexpected error getting pv: %v", err) } @@ -502,7 +502,7 @@ func TestPersistentVolumeMultiPVs(t *testing.T) { pvc := createPVC("pvc-2", ns.Name, strconv.Itoa(maxPVs/2)+"G", []v1.PersistentVolumeAccessMode{v1.ReadWriteOnce}, "") for i := 0; i < maxPVs; i++ { - _, err := testClient.PersistentVolumes().Create(pvs[i]) + _, err := testClient.CoreV1().PersistentVolumes().Create(pvs[i]) if err != nil { t.Errorf("Failed to create PersistentVolume %d: %v", i, err) } @@ -510,7 +510,7 @@ func TestPersistentVolumeMultiPVs(t *testing.T) { } t.Log("volumes created") - _, err := testClient.PersistentVolumeClaims(ns.Name).Create(pvc) + _, err := testClient.CoreV1().PersistentVolumeClaims(ns.Name).Create(pvc) if err != nil { t.Errorf("Failed to create PersistentVolumeClaim: %v", err) } @@ -525,7 +525,7 @@ func TestPersistentVolumeMultiPVs(t *testing.T) { // only one PV is bound bound := 0 for i := 0; i < maxPVs; i++ { - pv, err := testClient.PersistentVolumes().Get(pvs[i].Name, metav1.GetOptions{}) + pv, err := testClient.CoreV1().PersistentVolumes().Get(pvs[i].Name, metav1.GetOptions{}) if err != nil { t.Fatalf("Unexpected error getting pv: %v", err) } @@ -550,7 +550,7 @@ func TestPersistentVolumeMultiPVs(t *testing.T) { } // deleting a claim releases the volume - if err := testClient.PersistentVolumeClaims(ns.Name).Delete(pvc.Name, nil); err != nil { + if err := testClient.CoreV1().PersistentVolumeClaims(ns.Name).Delete(pvc.Name, nil); err != nil { t.Errorf("error deleting claim %s", pvc.Name) } t.Log("claim deleted") @@ -599,7 +599,7 @@ func TestPersistentVolumeMultiPVsPVCs(t *testing.T) { // with >3000 volumes. go func() { for i := 0; i < objCount; i++ { - _, _ = testClient.PersistentVolumes().Create(pvs[i]) + _, _ = testClient.CoreV1().PersistentVolumes().Create(pvs[i]) } }() // Wait for them to get Available @@ -620,7 +620,7 @@ func TestPersistentVolumeMultiPVsPVCs(t *testing.T) { // Modify PV i := rand.Intn(objCount) name := "pv-" + strconv.Itoa(i) - pv, err := testClient.PersistentVolumes().Get(name, metav1.GetOptions{}) + pv, err := testClient.CoreV1().PersistentVolumes().Get(name, metav1.GetOptions{}) if err != nil { // Silently ignore error, the PV may have be already deleted // or not exists yet. @@ -632,7 +632,7 @@ func TestPersistentVolumeMultiPVsPVCs(t *testing.T) { } else { pv.Annotations["TestAnnotation"] = fmt.Sprint(rand.Int()) } - _, err = testClient.PersistentVolumes().Update(pv) + _, err = testClient.CoreV1().PersistentVolumes().Update(pv) if err != nil { // Silently ignore error, the PV may have been updated by // the controller. @@ -644,7 +644,7 @@ func TestPersistentVolumeMultiPVsPVCs(t *testing.T) { // Modify PVC i := rand.Intn(objCount) name := "pvc-" + strconv.Itoa(i) - pvc, err := testClient.PersistentVolumeClaims(metav1.NamespaceDefault).Get(name, metav1.GetOptions{}) + pvc, err := testClient.CoreV1().PersistentVolumeClaims(metav1.NamespaceDefault).Get(name, metav1.GetOptions{}) if err != nil { // Silently ignore error, the PVC may have be already // deleted or not exists yet. @@ -656,7 +656,7 @@ func TestPersistentVolumeMultiPVsPVCs(t *testing.T) { } else { pvc.Annotations["TestAnnotation"] = fmt.Sprint(rand.Int()) } - _, err = testClient.PersistentVolumeClaims(metav1.NamespaceDefault).Update(pvc) + _, err = testClient.CoreV1().PersistentVolumeClaims(metav1.NamespaceDefault).Update(pvc) if err != nil { // Silently ignore error, the PVC may have been updated by // the controller. @@ -679,7 +679,7 @@ func TestPersistentVolumeMultiPVsPVCs(t *testing.T) { // Create the claims, again in a separate goroutine. go func() { for i := 0; i < objCount; i++ { - _, _ = testClient.PersistentVolumeClaims(ns.Name).Create(pvcs[i]) + _, _ = testClient.CoreV1().PersistentVolumeClaims(ns.Name).Create(pvcs[i]) } }() @@ -699,7 +699,7 @@ func TestPersistentVolumeMultiPVsPVCs(t *testing.T) { // check that everything is bound to something for i := 0; i < objCount; i++ { - pv, err := testClient.PersistentVolumes().Get(pvs[i].Name, metav1.GetOptions{}) + pv, err := testClient.CoreV1().PersistentVolumes().Get(pvs[i].Name, metav1.GetOptions{}) if err != nil { t.Fatalf("Unexpected error getting pv: %v", err) } @@ -708,7 +708,7 @@ func TestPersistentVolumeMultiPVsPVCs(t *testing.T) { } glog.V(2).Infof("PV %q is bound to PVC %q", pv.Name, pv.Spec.ClaimRef.Name) - pvc, err := testClient.PersistentVolumeClaims(ns.Name).Get(pvcs[i].Name, metav1.GetOptions{}) + pvc, err := testClient.CoreV1().PersistentVolumeClaims(ns.Name).Get(pvcs[i].Name, metav1.GetOptions{}) if err != nil { t.Fatalf("Unexpected error getting pvc: %v", err) } @@ -748,13 +748,13 @@ func TestPersistentVolumeControllerStartup(t *testing.T) { pvc := createPVC(pvcName, ns.Name, "1G", []v1.PersistentVolumeAccessMode{v1.ReadWriteOnce}, "") pvc.Annotations = map[string]string{"annBindCompleted": ""} pvc.Spec.VolumeName = pvName - newPVC, err := testClient.PersistentVolumeClaims(ns.Name).Create(pvc) + newPVC, err := testClient.CoreV1().PersistentVolumeClaims(ns.Name).Create(pvc) if err != nil { t.Fatalf("Cannot create claim %q: %v", pvc.Name, err) } // Save Bound status as a separate transaction newPVC.Status.Phase = v1.ClaimBound - newPVC, err = testClient.PersistentVolumeClaims(ns.Name).UpdateStatus(newPVC) + newPVC, err = testClient.CoreV1().PersistentVolumeClaims(ns.Name).UpdateStatus(newPVC) if err != nil { t.Fatalf("Cannot update claim status %q: %v", pvc.Name, err) } @@ -772,13 +772,13 @@ func TestPersistentVolumeControllerStartup(t *testing.T) { return } pv.Spec.ClaimRef = claimRef - newPV, err := testClient.PersistentVolumes().Create(pv) + newPV, err := testClient.CoreV1().PersistentVolumes().Create(pv) if err != nil { t.Fatalf("Cannot create volume %q: %v", pv.Name, err) } // Save Bound status as a separate transaction newPV.Status.Phase = v1.VolumeBound - newPV, err = testClient.PersistentVolumes().UpdateStatus(newPV) + newPV, err = testClient.CoreV1().PersistentVolumes().UpdateStatus(newPV) if err != nil { t.Fatalf("Cannot update volume status %q: %v", pv.Name, err) } @@ -829,7 +829,7 @@ func TestPersistentVolumeControllerStartup(t *testing.T) { // check that everything is bound to something for i := 0; i < objCount; i++ { - pv, err := testClient.PersistentVolumes().Get(pvs[i].Name, metav1.GetOptions{}) + pv, err := testClient.CoreV1().PersistentVolumes().Get(pvs[i].Name, metav1.GetOptions{}) if err != nil { t.Fatalf("Unexpected error getting pv: %v", err) } @@ -838,7 +838,7 @@ func TestPersistentVolumeControllerStartup(t *testing.T) { } glog.V(2).Infof("PV %q is bound to PVC %q", pv.Name, pv.Spec.ClaimRef.Name) - pvc, err := testClient.PersistentVolumeClaims(ns.Name).Get(pvcs[i].Name, metav1.GetOptions{}) + pvc, err := testClient.CoreV1().PersistentVolumeClaims(ns.Name).Get(pvcs[i].Name, metav1.GetOptions{}) if err != nil { t.Fatalf("Unexpected error getting pvc: %v", err) } @@ -895,7 +895,7 @@ func TestPersistentVolumeProvisionMultiPVCs(t *testing.T) { // early. It gets stuck with >3000 claims. go func() { for i := 0; i < objCount; i++ { - _, _ = testClient.PersistentVolumeClaims(ns.Name).Create(pvcs[i]) + _, _ = testClient.CoreV1().PersistentVolumeClaims(ns.Name).Create(pvcs[i]) } }() @@ -907,7 +907,7 @@ func TestPersistentVolumeProvisionMultiPVCs(t *testing.T) { glog.V(2).Infof("TestPersistentVolumeProvisionMultiPVCs: claims are bound") // check that we have enough bound PVs - pvList, err := testClient.PersistentVolumes().List(metav1.ListOptions{}) + pvList, err := testClient.CoreV1().PersistentVolumes().List(metav1.ListOptions{}) if err != nil { t.Fatalf("Failed to list volumes: %s", err) } @@ -924,13 +924,13 @@ func TestPersistentVolumeProvisionMultiPVCs(t *testing.T) { // Delete the claims for i := 0; i < objCount; i++ { - _ = testClient.PersistentVolumeClaims(ns.Name).Delete(pvcs[i].Name, nil) + _ = testClient.CoreV1().PersistentVolumeClaims(ns.Name).Delete(pvcs[i].Name, nil) } // Wait for the PVs to get deleted by listing remaining volumes // (delete events were unreliable) for { - volumes, err := testClient.PersistentVolumes().List(metav1.ListOptions{}) + volumes, err := testClient.CoreV1().PersistentVolumes().List(metav1.ListOptions{}) if err != nil { t.Fatalf("Failed to list volumes: %v", err) } @@ -974,17 +974,17 @@ func TestPersistentVolumeMultiPVsDiffAccessModes(t *testing.T) { pvc := createPVC("pvc-rwm", ns.Name, "5G", []v1.PersistentVolumeAccessMode{v1.ReadWriteMany}, "") - _, err := testClient.PersistentVolumes().Create(pv_rwm) + _, err := testClient.CoreV1().PersistentVolumes().Create(pv_rwm) if err != nil { t.Errorf("Failed to create PersistentVolume: %v", err) } - _, err = testClient.PersistentVolumes().Create(pv_rwo) + _, err = testClient.CoreV1().PersistentVolumes().Create(pv_rwo) if err != nil { t.Errorf("Failed to create PersistentVolume: %v", err) } t.Log("volumes created") - _, err = testClient.PersistentVolumeClaims(ns.Name).Create(pvc) + _, err = testClient.CoreV1().PersistentVolumeClaims(ns.Name).Create(pvc) if err != nil { t.Errorf("Failed to create PersistentVolumeClaim: %v", err) } @@ -997,14 +997,14 @@ func TestPersistentVolumeMultiPVsDiffAccessModes(t *testing.T) { t.Log("claim bound") // only RWM PV is bound - pv, err := testClient.PersistentVolumes().Get("pv-rwo", metav1.GetOptions{}) + pv, err := testClient.CoreV1().PersistentVolumes().Get("pv-rwo", metav1.GetOptions{}) if err != nil { t.Fatalf("Unexpected error getting pv: %v", err) } if pv.Spec.ClaimRef != nil { t.Fatalf("ReadWriteOnce PV shouldn't be bound") } - pv, err = testClient.PersistentVolumes().Get("pv-rwm", metav1.GetOptions{}) + pv, err = testClient.CoreV1().PersistentVolumes().Get("pv-rwm", metav1.GetOptions{}) if err != nil { t.Fatalf("Unexpected error getting pv: %v", err) } @@ -1016,7 +1016,7 @@ func TestPersistentVolumeMultiPVsDiffAccessModes(t *testing.T) { } // deleting a claim releases the volume - if err := testClient.PersistentVolumeClaims(ns.Name).Delete(pvc.Name, nil); err != nil { + if err := testClient.CoreV1().PersistentVolumeClaims(ns.Name).Delete(pvc.Name, nil); err != nil { t.Errorf("error deleting claim %s", pvc.Name) } t.Log("claim deleted") @@ -1142,11 +1142,11 @@ func createClients(ns *v1.Namespace, t *testing.T, s *httptest.Server, syncPerio t.Fatalf("Failed to construct PersistentVolumes: %v", err) } - watchPV, err := testClient.PersistentVolumes().Watch(metav1.ListOptions{}) + watchPV, err := testClient.CoreV1().PersistentVolumes().Watch(metav1.ListOptions{}) if err != nil { t.Fatalf("Failed to watch PersistentVolumes: %v", err) } - watchPVC, err := testClient.PersistentVolumeClaims(ns.Name).Watch(metav1.ListOptions{}) + watchPVC, err := testClient.CoreV1().PersistentVolumeClaims(ns.Name).Watch(metav1.ListOptions{}) if err != nil { t.Fatalf("Failed to watch PersistentVolumeClaims: %v", err) } diff --git a/test/soak/serve_hostnames/serve_hostnames.go b/test/soak/serve_hostnames/serve_hostnames.go index 88653c2d592..62883b71d43 100644 --- a/test/soak/serve_hostnames/serve_hostnames.go +++ b/test/soak/serve_hostnames/serve_hostnames.go @@ -92,7 +92,7 @@ func main() { var nodes *v1.NodeList for start := time.Now(); time.Since(start) < nodeListTimeout; time.Sleep(2 * time.Second) { - nodes, err = client.Nodes().List(metav1.ListOptions{}) + nodes, err = client.CoreV1().Nodes().List(metav1.ListOptions{}) if err == nil { break } @@ -114,7 +114,7 @@ func main() { queries := *queriesAverage * len(nodes.Items) * *podsPerNode // Create the namespace - got, err := client.Namespaces().Create(&v1.Namespace{ObjectMeta: metav1.ObjectMeta{GenerateName: "serve-hostnames-"}}) + got, err := client.CoreV1().Namespaces().Create(&v1.Namespace{ObjectMeta: metav1.ObjectMeta{GenerateName: "serve-hostnames-"}}) if err != nil { glog.Fatalf("Failed to create namespace: %v", err) } @@ -125,7 +125,7 @@ func main() { } else { // wait until the namespace disappears for i := 0; i < int(namespaceDeleteTimeout/time.Second); i++ { - if _, err := client.Namespaces().Get(ns, metav1.GetOptions{}); err != nil { + if _, err := client.CoreV1().Namespaces().Get(ns, metav1.GetOptions{}); err != nil { if errors.IsNotFound(err) { return } @@ -142,7 +142,7 @@ func main() { var svc *v1.Service for start := time.Now(); time.Since(start) < serviceCreateTimeout; time.Sleep(2 * time.Second) { t := time.Now() - svc, err = client.Services(ns).Create(&v1.Service{ + svc, err = client.CoreV1().Services(ns).Create(&v1.Service{ ObjectMeta: metav1.ObjectMeta{ Name: "serve-hostnames", Labels: map[string]string{ @@ -175,7 +175,7 @@ func main() { glog.Infof("Cleaning up service %s/serve-hostnames", ns) // Make several attempts to delete the service. for start := time.Now(); time.Since(start) < deleteTimeout; time.Sleep(1 * time.Second) { - if err := client.Services(ns).Delete(svc.Name, nil); err == nil { + if err := client.CoreV1().Services(ns).Delete(svc.Name, nil); err == nil { return } glog.Warningf("After %v unable to delete service %s/%s: %v", time.Since(start), ns, svc.Name, err) @@ -192,7 +192,7 @@ func main() { for start := time.Now(); time.Since(start) < podCreateTimeout; time.Sleep(2 * time.Second) { glog.Infof("Creating pod %s/%s on node %s", ns, podName, node.Name) t := time.Now() - _, err = client.Pods(ns).Create(&v1.Pod{ + _, err = client.CoreV1().Pods(ns).Create(&v1.Pod{ ObjectMeta: metav1.ObjectMeta{ Name: podName, Labels: map[string]string{ @@ -228,7 +228,7 @@ func main() { // Make several attempts to delete the pods. for _, podName := range podNames { for start := time.Now(); time.Since(start) < deleteTimeout; time.Sleep(1 * time.Second) { - if err = client.Pods(ns).Delete(podName, nil); err == nil { + if err = client.CoreV1().Pods(ns).Delete(podName, nil); err == nil { break } glog.Warningf("After %v failed to delete pod %s/%s: %v", time.Since(start), ns, podName, err) @@ -240,7 +240,7 @@ func main() { for _, podName := range podNames { var pod *v1.Pod for start := time.Now(); time.Since(start) < podStartTimeout; time.Sleep(5 * time.Second) { - pod, err = client.Pods(ns).Get(podName, metav1.GetOptions{}) + pod, err = client.CoreV1().Pods(ns).Get(podName, metav1.GetOptions{}) if err != nil { glog.Warningf("Get pod %s/%s failed, ignoring for %v: %v", ns, podName, err, podStartTimeout) continue