From 1470df7a052eeed03c7ca794cb988971da98c687 Mon Sep 17 00:00:00 2001 From: Andrew Sy Kim Date: Fri, 22 Mar 2019 12:09:07 -0700 Subject: [PATCH] remove usages of internal clientset in e2e framework Signed-off-by: Andrew Sy Kim --- test/e2e/apps/BUILD | 1 + test/e2e/apps/cronjob.go | 4 +- test/e2e/apps/daemon_restart.go | 13 ++-- test/e2e/autoscaling/autoscaling_timer.go | 2 +- .../cluster_autoscaler_scalability.go | 32 ++++----- .../autoscaling/cluster_size_autoscaling.go | 71 +++++++++---------- .../autoscaling/horizontal_pod_autoscaling.go | 2 +- test/e2e/common/BUILD | 1 - test/e2e/common/autoscaling_utils.go | 44 ++++++------ test/e2e/framework/BUILD | 1 - test/e2e/framework/framework.go | 8 +-- test/e2e/framework/service_util.go | 4 +- test/e2e/framework/util.go | 8 --- .../instrumentation/monitoring/prometheus.go | 6 +- .../instrumentation/monitoring/stackdriver.go | 2 +- test/e2e/network/BUILD | 1 - test/e2e/network/proxy.go | 13 ++-- test/e2e/network/service.go | 25 +++---- test/e2e/network/service_latency.go | 13 ++-- test/e2e/node/kubelet.go | 13 ++-- test/e2e/node/kubelet_perf.go | 11 ++- test/e2e/scalability/BUILD | 1 - test/e2e/scalability/density.go | 34 ++++----- test/e2e/scalability/load.go | 25 +++---- .../equivalence_cache_predicates.go | 17 +++-- test/e2e/scheduling/predicates.go | 15 ++-- .../upgrades/horizontal_pod_autoscalers.go | 1 - test/utils/BUILD | 1 - test/utils/runners.go | 12 ---- 29 files changed, 158 insertions(+), 223 deletions(-) diff --git a/test/e2e/apps/BUILD b/test/e2e/apps/BUILD index 75b06a53607..ca13a062568 100644 --- a/test/e2e/apps/BUILD +++ b/test/e2e/apps/BUILD @@ -57,6 +57,7 @@ go_library( "//staging/src/k8s.io/apimachinery/pkg/util/wait:go_default_library", "//staging/src/k8s.io/apimachinery/pkg/watch:go_default_library", "//staging/src/k8s.io/client-go/kubernetes:go_default_library", + "//staging/src/k8s.io/client-go/kubernetes/scheme:go_default_library", "//staging/src/k8s.io/client-go/tools/cache:go_default_library", "//staging/src/k8s.io/client-go/tools/watch:go_default_library", "//test/e2e/common:go_default_library", diff --git a/test/e2e/apps/cronjob.go b/test/e2e/apps/cronjob.go index 9be96fffbe1..20bb632d70e 100644 --- a/test/e2e/apps/cronjob.go +++ b/test/e2e/apps/cronjob.go @@ -30,7 +30,7 @@ import ( metav1 "k8s.io/apimachinery/pkg/apis/meta/v1" "k8s.io/apimachinery/pkg/util/wait" clientset "k8s.io/client-go/kubernetes" - "k8s.io/kubernetes/pkg/api/legacyscheme" + "k8s.io/client-go/kubernetes/scheme" batchinternal "k8s.io/kubernetes/pkg/apis/batch" "k8s.io/kubernetes/pkg/controller/job" "k8s.io/kubernetes/test/e2e/framework" @@ -464,7 +464,7 @@ func waitForEventWithReason(c clientset.Interface, ns, cronJobName string, reaso if err != nil { return false, err } - events, err := c.CoreV1().Events(ns).Search(legacyscheme.Scheme, sj) + events, err := c.CoreV1().Events(ns).Search(scheme.Scheme, sj) if err != nil { return false, err } diff --git a/test/e2e/apps/daemon_restart.go b/test/e2e/apps/daemon_restart.go index e37a2bda03f..97077476167 100644 --- a/test/e2e/apps/daemon_restart.go +++ b/test/e2e/apps/daemon_restart.go @@ -199,13 +199,12 @@ var _ = SIGDescribe("DaemonRestart [Disruptive]", func() { // All the restart tests need an rc and a watch on pods of the rc. // Additionally some of them might scale the rc during the test. config = testutils.RCConfig{ - Client: f.ClientSet, - InternalClient: f.InternalClientset, - Name: rcName, - Namespace: ns, - Image: imageutils.GetPauseImageName(), - Replicas: numPods, - CreatedPods: &[]*v1.Pod{}, + Client: f.ClientSet, + Name: rcName, + Namespace: ns, + Image: imageutils.GetPauseImageName(), + Replicas: numPods, + CreatedPods: &[]*v1.Pod{}, } Expect(framework.RunRC(config)).NotTo(HaveOccurred()) replacePods(*config.CreatedPods, existingPods) diff --git a/test/e2e/autoscaling/autoscaling_timer.go b/test/e2e/autoscaling/autoscaling_timer.go index ab2581f739b..6a21f626737 100644 --- a/test/e2e/autoscaling/autoscaling_timer.go +++ b/test/e2e/autoscaling/autoscaling_timer.go @@ -93,7 +93,7 @@ var _ = SIGDescribe("[Feature:ClusterSizeAutoscalingScaleUp] [Slow] Autoscaling" nodeMemoryMB := (&nodeMemoryBytes).Value() / 1024 / 1024 memRequestMB := nodeMemoryMB / 10 // Ensure each pod takes not more than 10% of node's allocatable memory. replicas := 1 - resourceConsumer := common.NewDynamicResourceConsumer("resource-consumer", f.Namespace.Name, common.KindDeployment, replicas, 0, 0, 0, cpuRequestMillis, memRequestMB, f.ClientSet, f.InternalClientset, f.ScalesGetter) + resourceConsumer := common.NewDynamicResourceConsumer("resource-consumer", f.Namespace.Name, common.KindDeployment, replicas, 0, 0, 0, cpuRequestMillis, memRequestMB, f.ClientSet, f.ScalesGetter) defer resourceConsumer.CleanUp() resourceConsumer.WaitForReplicas(replicas, 1*time.Minute) // Should finish ~immediately, so 1 minute is more than enough. diff --git a/test/e2e/autoscaling/cluster_autoscaler_scalability.go b/test/e2e/autoscaling/cluster_autoscaler_scalability.go index 8832cf503f6..5c42a68b877 100644 --- a/test/e2e/autoscaling/cluster_autoscaler_scalability.go +++ b/test/e2e/autoscaling/cluster_autoscaler_scalability.go @@ -408,14 +408,13 @@ func simpleScaleUpTest(f *framework.Framework, config *scaleUpTestConfig) func() func reserveMemoryRCConfig(f *framework.Framework, id string, replicas, megabytes int, timeout time.Duration) *testutils.RCConfig { return &testutils.RCConfig{ - Client: f.ClientSet, - InternalClient: f.InternalClientset, - Name: id, - Namespace: f.Namespace.Name, - Timeout: timeout, - Image: imageutils.GetPauseImageName(), - Replicas: replicas, - MemRequest: int64(1024 * 1024 * megabytes / replicas), + Client: f.ClientSet, + Name: id, + Namespace: f.Namespace.Name, + Timeout: timeout, + Image: imageutils.GetPauseImageName(), + Replicas: replicas, + MemRequest: int64(1024 * 1024 * megabytes / replicas), } } @@ -468,15 +467,14 @@ func createHostPortPodsWithMemory(f *framework.Framework, id string, replicas, p By(fmt.Sprintf("Running RC which reserves host port and memory")) request := int64(1024 * 1024 * megabytes / replicas) config := &testutils.RCConfig{ - Client: f.ClientSet, - InternalClient: f.InternalClientset, - Name: id, - Namespace: f.Namespace.Name, - Timeout: timeout, - Image: imageutils.GetPauseImageName(), - Replicas: replicas, - HostPorts: map[string]int{"port1": port}, - MemRequest: request, + Client: f.ClientSet, + Name: id, + Namespace: f.Namespace.Name, + Timeout: timeout, + Image: imageutils.GetPauseImageName(), + Replicas: replicas, + HostPorts: map[string]int{"port1": port}, + MemRequest: request, } err := framework.RunRC(*config) framework.ExpectNoError(err) diff --git a/test/e2e/autoscaling/cluster_size_autoscaling.go b/test/e2e/autoscaling/cluster_size_autoscaling.go index 94de8427564..7158b97236c 100644 --- a/test/e2e/autoscaling/cluster_size_autoscaling.go +++ b/test/e2e/autoscaling/cluster_size_autoscaling.go @@ -1322,7 +1322,6 @@ func reserveMemory(f *framework.Framework, id string, replicas, megabytes int, e request := int64(1024 * 1024 * megabytes / replicas) config := &testutils.RCConfig{ Client: f.ClientSet, - InternalClient: f.InternalClientset, Name: id, Namespace: f.Namespace.Name, Timeout: timeout, @@ -1564,15 +1563,14 @@ func ScheduleAnySingleGpuPod(f *framework.Framework, id string) error { // ScheduleGpuPod schedules a pod which requires a given number of gpus of given type func ScheduleGpuPod(f *framework.Framework, id string, gpuType string, gpuLimit int64) error { config := &testutils.RCConfig{ - Client: f.ClientSet, - InternalClient: f.InternalClientset, - Name: id, - Namespace: f.Namespace.Name, - Timeout: 3 * scaleUpTimeout, // spinning up GPU node is slow - Image: imageutils.GetPauseImageName(), - Replicas: 1, - GpuLimit: gpuLimit, - Labels: map[string]string{"requires-gpu": "yes"}, + Client: f.ClientSet, + Name: id, + Namespace: f.Namespace.Name, + Timeout: 3 * scaleUpTimeout, // spinning up GPU node is slow + Image: imageutils.GetPauseImageName(), + Replicas: 1, + GpuLimit: gpuLimit, + Labels: map[string]string{"requires-gpu": "yes"}, } if gpuType != "" { @@ -1589,15 +1587,14 @@ func ScheduleGpuPod(f *framework.Framework, id string, gpuType string, gpuLimit // Create an RC running a given number of pods with anti-affinity func runAntiAffinityPods(f *framework.Framework, namespace string, pods int, id string, podLabels, antiAffinityLabels map[string]string) error { config := &testutils.RCConfig{ - Affinity: buildAntiAffinity(antiAffinityLabels), - Client: f.ClientSet, - InternalClient: f.InternalClientset, - Name: id, - Namespace: namespace, - Timeout: scaleUpTimeout, - Image: imageutils.GetPauseImageName(), - Replicas: pods, - Labels: podLabels, + Affinity: buildAntiAffinity(antiAffinityLabels), + Client: f.ClientSet, + Name: id, + Namespace: namespace, + Timeout: scaleUpTimeout, + Image: imageutils.GetPauseImageName(), + Replicas: pods, + Labels: podLabels, } err := framework.RunRC(*config) if err != nil { @@ -1612,16 +1609,15 @@ func runAntiAffinityPods(f *framework.Framework, namespace string, pods int, id func runVolumeAntiAffinityPods(f *framework.Framework, namespace string, pods int, id string, podLabels, antiAffinityLabels map[string]string, volumes []v1.Volume) error { config := &testutils.RCConfig{ - Affinity: buildAntiAffinity(antiAffinityLabels), - Volumes: volumes, - Client: f.ClientSet, - InternalClient: f.InternalClientset, - Name: id, - Namespace: namespace, - Timeout: scaleUpTimeout, - Image: imageutils.GetPauseImageName(), - Replicas: pods, - Labels: podLabels, + Affinity: buildAntiAffinity(antiAffinityLabels), + Volumes: volumes, + Client: f.ClientSet, + Name: id, + Namespace: namespace, + Timeout: scaleUpTimeout, + Image: imageutils.GetPauseImageName(), + Replicas: pods, + Labels: podLabels, } err := framework.RunRC(*config) if err != nil { @@ -1695,15 +1691,14 @@ func runReplicatedPodOnEachNode(f *framework.Framework, nodes []v1.Node, namespa } } config := &testutils.RCConfig{ - Client: f.ClientSet, - InternalClient: f.InternalClientset, - Name: id, - Namespace: namespace, - Timeout: defaultTimeout, - Image: imageutils.GetPauseImageName(), - Replicas: 0, - Labels: labels, - MemRequest: memRequest, + Client: f.ClientSet, + Name: id, + Namespace: namespace, + Timeout: defaultTimeout, + Image: imageutils.GetPauseImageName(), + Replicas: 0, + Labels: labels, + MemRequest: memRequest, } err := framework.RunRC(*config) if err != nil { diff --git a/test/e2e/autoscaling/horizontal_pod_autoscaling.go b/test/e2e/autoscaling/horizontal_pod_autoscaling.go index bdba3f73923..00737474245 100644 --- a/test/e2e/autoscaling/horizontal_pod_autoscaling.go +++ b/test/e2e/autoscaling/horizontal_pod_autoscaling.go @@ -116,7 +116,7 @@ type HPAScaleTest struct { // TODO The use of 3 states is arbitrary, we could eventually make this test handle "n" states once this test stabilizes. func (scaleTest *HPAScaleTest) run(name string, kind schema.GroupVersionKind, rc *common.ResourceConsumer, f *framework.Framework) { const timeToWait = 15 * time.Minute - rc = common.NewDynamicResourceConsumer(name, f.Namespace.Name, kind, scaleTest.initPods, scaleTest.totalInitialCPUUsage, 0, 0, scaleTest.perPodCPURequest, 200, f.ClientSet, f.InternalClientset, f.ScalesGetter) + rc = common.NewDynamicResourceConsumer(name, f.Namespace.Name, kind, scaleTest.initPods, scaleTest.totalInitialCPUUsage, 0, 0, scaleTest.perPodCPURequest, 200, f.ClientSet, f.ScalesGetter) defer rc.CleanUp() hpa := common.CreateCPUHorizontalPodAutoscaler(rc, scaleTest.targetCPUUtilizationPercent, scaleTest.minPods, scaleTest.maxPods) defer common.DeleteHorizontalPodAutoscaler(rc, hpa.Name) diff --git a/test/e2e/common/BUILD b/test/e2e/common/BUILD index 2b0d983a552..d1149253304 100644 --- a/test/e2e/common/BUILD +++ b/test/e2e/common/BUILD @@ -45,7 +45,6 @@ go_library( deps = [ "//pkg/api/v1/pod:go_default_library", "//pkg/apis/core:go_default_library", - "//pkg/client/clientset_generated/internalclientset:go_default_library", "//pkg/client/conditions:go_default_library", "//pkg/kubelet:go_default_library", "//pkg/kubelet/events:go_default_library", diff --git a/test/e2e/common/autoscaling_utils.go b/test/e2e/common/autoscaling_utils.go index 817e963b0f6..ff6fbf89c94 100644 --- a/test/e2e/common/autoscaling_utils.go +++ b/test/e2e/common/autoscaling_utils.go @@ -31,7 +31,6 @@ import ( "k8s.io/apimachinery/pkg/util/wait" clientset "k8s.io/client-go/kubernetes" api "k8s.io/kubernetes/pkg/apis/core" - "k8s.io/kubernetes/pkg/client/clientset_generated/internalclientset" "k8s.io/kubernetes/test/e2e/framework" testutils "k8s.io/kubernetes/test/utils" @@ -86,7 +85,6 @@ type ResourceConsumer struct { kind schema.GroupVersionKind nsName string clientSet clientset.Interface - internalClientset *internalclientset.Clientset scaleClient scaleclient.ScalesGetter cpu chan int mem chan int @@ -106,20 +104,20 @@ func GetResourceConsumerImage() string { return resourceConsumerImage } -func NewDynamicResourceConsumer(name, nsName string, kind schema.GroupVersionKind, replicas, initCPUTotal, initMemoryTotal, initCustomMetric int, cpuLimit, memLimit int64, clientset clientset.Interface, internalClientset *internalclientset.Clientset, scaleClient scaleclient.ScalesGetter) *ResourceConsumer { +func NewDynamicResourceConsumer(name, nsName string, kind schema.GroupVersionKind, replicas, initCPUTotal, initMemoryTotal, initCustomMetric int, cpuLimit, memLimit int64, clientset clientset.Interface, scaleClient scaleclient.ScalesGetter) *ResourceConsumer { return newResourceConsumer(name, nsName, kind, replicas, initCPUTotal, initMemoryTotal, initCustomMetric, dynamicConsumptionTimeInSeconds, - dynamicRequestSizeInMillicores, dynamicRequestSizeInMegabytes, dynamicRequestSizeCustomMetric, cpuLimit, memLimit, clientset, internalClientset, scaleClient, nil, nil) + dynamicRequestSizeInMillicores, dynamicRequestSizeInMegabytes, dynamicRequestSizeCustomMetric, cpuLimit, memLimit, clientset, scaleClient, nil, nil) } // TODO this still defaults to replication controller -func NewStaticResourceConsumer(name, nsName string, replicas, initCPUTotal, initMemoryTotal, initCustomMetric int, cpuLimit, memLimit int64, clientset clientset.Interface, internalClientset *internalclientset.Clientset, scaleClient scaleclient.ScalesGetter) *ResourceConsumer { +func NewStaticResourceConsumer(name, nsName string, replicas, initCPUTotal, initMemoryTotal, initCustomMetric int, cpuLimit, memLimit int64, clientset clientset.Interface, scaleClient scaleclient.ScalesGetter) *ResourceConsumer { return newResourceConsumer(name, nsName, KindRC, replicas, initCPUTotal, initMemoryTotal, initCustomMetric, staticConsumptionTimeInSeconds, - initCPUTotal/replicas, initMemoryTotal/replicas, initCustomMetric/replicas, cpuLimit, memLimit, clientset, internalClientset, scaleClient, nil, nil) + initCPUTotal/replicas, initMemoryTotal/replicas, initCustomMetric/replicas, cpuLimit, memLimit, clientset, scaleClient, nil, nil) } -func NewMetricExporter(name, nsName string, podAnnotations, serviceAnnotations map[string]string, metricValue int, clientset clientset.Interface, internalClientset *internalclientset.Clientset, scaleClient scaleclient.ScalesGetter) *ResourceConsumer { +func NewMetricExporter(name, nsName string, podAnnotations, serviceAnnotations map[string]string, metricValue int, clientset clientset.Interface, scaleClient scaleclient.ScalesGetter) *ResourceConsumer { return newResourceConsumer(name, nsName, KindDeployment, 1, 0, 0, metricValue, dynamicConsumptionTimeInSeconds, - dynamicRequestSizeInMillicores, dynamicRequestSizeInMegabytes, dynamicRequestSizeCustomMetric, 100, 100, clientset, internalClientset, scaleClient, podAnnotations, serviceAnnotations) + dynamicRequestSizeInMillicores, dynamicRequestSizeInMegabytes, dynamicRequestSizeCustomMetric, 100, 100, clientset, scaleClient, podAnnotations, serviceAnnotations) } /* @@ -130,21 +128,20 @@ memLimit argument is in megabytes, memLimit is a maximum amount of memory that c cpuLimit argument is in millicores, cpuLimit is a maximum amount of cpu that can be consumed by a single pod */ func newResourceConsumer(name, nsName string, kind schema.GroupVersionKind, replicas, initCPUTotal, initMemoryTotal, initCustomMetric, consumptionTimeInSeconds, requestSizeInMillicores, - requestSizeInMegabytes int, requestSizeCustomMetric int, cpuLimit, memLimit int64, clientset clientset.Interface, internalClientset *internalclientset.Clientset, scaleClient scaleclient.ScalesGetter, podAnnotations, serviceAnnotations map[string]string) *ResourceConsumer { + requestSizeInMegabytes int, requestSizeCustomMetric int, cpuLimit, memLimit int64, clientset clientset.Interface, scaleClient scaleclient.ScalesGetter, podAnnotations, serviceAnnotations map[string]string) *ResourceConsumer { if podAnnotations == nil { podAnnotations = make(map[string]string) } if serviceAnnotations == nil { serviceAnnotations = make(map[string]string) } - runServiceAndWorkloadForResourceConsumer(clientset, internalClientset, nsName, name, kind, replicas, cpuLimit, memLimit, podAnnotations, serviceAnnotations) + runServiceAndWorkloadForResourceConsumer(clientset, nsName, name, kind, replicas, cpuLimit, memLimit, podAnnotations, serviceAnnotations) rc := &ResourceConsumer{ name: name, controllerName: name + "-ctrl", kind: kind, nsName: nsName, clientSet: clientset, - internalClientset: internalClientset, scaleClient: scaleClient, cpu: make(chan int), mem: make(chan int), @@ -436,7 +433,7 @@ func (rc *ResourceConsumer) CleanUp() { framework.ExpectNoError(rc.clientSet.CoreV1().Services(rc.nsName).Delete(rc.controllerName, nil)) } -func runServiceAndWorkloadForResourceConsumer(c clientset.Interface, internalClient internalclientset.Interface, ns, name string, kind schema.GroupVersionKind, replicas int, cpuLimitMillis, memLimitMb int64, podAnnotations, serviceAnnotations map[string]string) { +func runServiceAndWorkloadForResourceConsumer(c clientset.Interface, ns, name string, kind schema.GroupVersionKind, replicas int, cpuLimitMillis, memLimitMb int64, podAnnotations, serviceAnnotations map[string]string) { By(fmt.Sprintf("Running consuming RC %s via %s with %v replicas", name, kind, replicas)) _, err := c.CoreV1().Services(ns).Create(&v1.Service{ ObjectMeta: metav1.ObjectMeta{ @@ -457,18 +454,17 @@ func runServiceAndWorkloadForResourceConsumer(c clientset.Interface, internalCli framework.ExpectNoError(err) rcConfig := testutils.RCConfig{ - Client: c, - InternalClient: internalClient, - Image: resourceConsumerImage, - Name: name, - Namespace: ns, - Timeout: timeoutRC, - Replicas: replicas, - CpuRequest: cpuLimitMillis, - CpuLimit: cpuLimitMillis, - MemRequest: memLimitMb * 1024 * 1024, // MemLimit is in bytes - MemLimit: memLimitMb * 1024 * 1024, - Annotations: podAnnotations, + Client: c, + Image: resourceConsumerImage, + Name: name, + Namespace: ns, + Timeout: timeoutRC, + Replicas: replicas, + CpuRequest: cpuLimitMillis, + CpuLimit: cpuLimitMillis, + MemRequest: memLimitMb * 1024 * 1024, // MemLimit is in bytes + MemLimit: memLimitMb * 1024 * 1024, + Annotations: podAnnotations, } switch kind { diff --git a/test/e2e/framework/BUILD b/test/e2e/framework/BUILD index 523b3c7ff8c..56f7543b37a 100644 --- a/test/e2e/framework/BUILD +++ b/test/e2e/framework/BUILD @@ -47,7 +47,6 @@ go_library( "//pkg/apis/batch:go_default_library", "//pkg/apis/core:go_default_library", "//pkg/apis/extensions:go_default_library", - "//pkg/client/clientset_generated/internalclientset:go_default_library", "//pkg/client/conditions:go_default_library", "//pkg/controller:go_default_library", "//pkg/controller/deployment/util:go_default_library", diff --git a/test/e2e/framework/framework.go b/test/e2e/framework/framework.go index 2defe4ae9df..c76c4535108 100644 --- a/test/e2e/framework/framework.go +++ b/test/e2e/framework/framework.go @@ -47,7 +47,6 @@ import ( scaleclient "k8s.io/client-go/scale" aggregatorclient "k8s.io/kube-aggregator/pkg/client/clientset_generated/clientset" "k8s.io/kubernetes/pkg/api/legacyscheme" - "k8s.io/kubernetes/pkg/client/clientset_generated/internalclientset" "k8s.io/kubernetes/test/e2e/framework/metrics" testutils "k8s.io/kubernetes/test/utils" @@ -75,9 +74,8 @@ type Framework struct { ClientSet clientset.Interface KubemarkExternalClusterClientSet clientset.Interface - InternalClientset *internalclientset.Clientset - AggregatorClient *aggregatorclient.Clientset - DynamicClient dynamic.Interface + AggregatorClient *aggregatorclient.Clientset + DynamicClient dynamic.Interface ScalesGetter scaleclient.ScalesGetter @@ -180,8 +178,6 @@ func (f *Framework) BeforeEach() { } f.ClientSet, err = clientset.NewForConfig(config) ExpectNoError(err) - f.InternalClientset, err = internalclientset.NewForConfig(config) - ExpectNoError(err) f.AggregatorClient, err = aggregatorclient.NewForConfig(config) ExpectNoError(err) f.DynamicClient, err = dynamic.NewForConfig(config) diff --git a/test/e2e/framework/service_util.go b/test/e2e/framework/service_util.go index b98ddba72e8..088934bffff 100644 --- a/test/e2e/framework/service_util.go +++ b/test/e2e/framework/service_util.go @@ -39,7 +39,6 @@ import ( clientset "k8s.io/client-go/kubernetes" "k8s.io/client-go/util/retry" api "k8s.io/kubernetes/pkg/apis/core" - "k8s.io/kubernetes/pkg/client/clientset_generated/internalclientset" "k8s.io/kubernetes/pkg/registry/core/service/portallocator" testutils "k8s.io/kubernetes/test/utils" imageutils "k8s.io/kubernetes/test/utils/image" @@ -1321,7 +1320,7 @@ func ValidateEndpointsOrFail(c clientset.Interface, namespace, serviceName strin // StartServeHostnameService creates a replication controller that serves its // hostname and a service on top of it. -func StartServeHostnameService(c clientset.Interface, internalClient internalclientset.Interface, svc *v1.Service, ns string, replicas int) ([]string, string, error) { +func StartServeHostnameService(c clientset.Interface, svc *v1.Service, ns string, replicas int) ([]string, string, error) { podNames := make([]string, replicas) name := svc.ObjectMeta.Name By("creating service " + name + " in namespace " + ns) @@ -1334,7 +1333,6 @@ func StartServeHostnameService(c clientset.Interface, internalClient internalcli maxContainerFailures := 0 config := testutils.RCConfig{ Client: c, - InternalClient: internalClient, Image: ServeHostnameImage, Name: name, Namespace: ns, diff --git a/test/e2e/framework/util.go b/test/e2e/framework/util.go index 9dde191297b..0dd59ba1499 100644 --- a/test/e2e/framework/util.go +++ b/test/e2e/framework/util.go @@ -80,7 +80,6 @@ import ( batchinternal "k8s.io/kubernetes/pkg/apis/batch" api "k8s.io/kubernetes/pkg/apis/core" extensionsinternal "k8s.io/kubernetes/pkg/apis/extensions" - "k8s.io/kubernetes/pkg/client/clientset_generated/internalclientset" "k8s.io/kubernetes/pkg/client/conditions" "k8s.io/kubernetes/pkg/controller" nodectlr "k8s.io/kubernetes/pkg/controller/nodelifecycle" @@ -2133,13 +2132,6 @@ func LoadConfig() (*restclient.Config, error) { return clientcmd.NewDefaultClientConfig(*c, &clientcmd.ConfigOverrides{ClusterInfo: clientcmdapi.Cluster{Server: TestContext.Host}}).ClientConfig() } -func LoadInternalClientset() (*internalclientset.Clientset, error) { - config, err := LoadConfig() - if err != nil { - return nil, fmt.Errorf("error creating client: %v", err.Error()) - } - return internalclientset.NewForConfig(config) -} func LoadClientset() (*clientset.Clientset, error) { config, err := LoadConfig() diff --git a/test/e2e/instrumentation/monitoring/prometheus.go b/test/e2e/instrumentation/monitoring/prometheus.go index b5727f1f0e0..6ecd493da8d 100644 --- a/test/e2e/instrumentation/monitoring/prometheus.go +++ b/test/e2e/instrumentation/monitoring/prometheus.go @@ -105,7 +105,7 @@ func prometheusPodCustomMetricQuery(namespace, podNamePrefix string) string { func consumeCPUResources(f *framework.Framework, consumerName string, cpuUsage int) *common.ResourceConsumer { return common.NewDynamicResourceConsumer(consumerName, f.Namespace.Name, common.KindDeployment, 1, cpuUsage, - memoryUsed, 0, int64(cpuUsage), memoryLimit, f.ClientSet, f.InternalClientset, f.ScalesGetter) + memoryUsed, 0, int64(cpuUsage), memoryLimit, f.ClientSet, f.ScalesGetter) } func exportCustomMetricFromPod(f *framework.Framework, consumerName string, metricValue int) *common.ResourceConsumer { @@ -114,7 +114,7 @@ func exportCustomMetricFromPod(f *framework.Framework, consumerName string, metr "prometheus.io/path": "/metrics", "prometheus.io/port": "8080", } - return common.NewMetricExporter(consumerName, f.Namespace.Name, podAnnotations, nil, metricValue, f.ClientSet, f.InternalClientset, f.ScalesGetter) + return common.NewMetricExporter(consumerName, f.Namespace.Name, podAnnotations, nil, metricValue, f.ClientSet, f.ScalesGetter) } func exportCustomMetricFromService(f *framework.Framework, consumerName string, metricValue int) *common.ResourceConsumer { @@ -123,7 +123,7 @@ func exportCustomMetricFromService(f *framework.Framework, consumerName string, "prometheus.io/path": "/metrics", "prometheus.io/port": "8080", } - return common.NewMetricExporter(consumerName, f.Namespace.Name, nil, serviceAnnotations, metricValue, f.ClientSet, f.InternalClientset, f.ScalesGetter) + return common.NewMetricExporter(consumerName, f.Namespace.Name, nil, serviceAnnotations, metricValue, f.ClientSet, f.ScalesGetter) } func validateMetricAvailableForAllNodes(c clientset.Interface, metric string, expectedNodesNames []string) error { diff --git a/test/e2e/instrumentation/monitoring/stackdriver.go b/test/e2e/instrumentation/monitoring/stackdriver.go index f668aef8581..1d7a7cd1129 100644 --- a/test/e2e/instrumentation/monitoring/stackdriver.go +++ b/test/e2e/instrumentation/monitoring/stackdriver.go @@ -101,7 +101,7 @@ func testStackdriverMonitoring(f *framework.Framework, pods, allPodsCPU int, per framework.ExpectNoError(err) - rc := common.NewDynamicResourceConsumer(rcName, f.Namespace.Name, common.KindDeployment, pods, allPodsCPU, memoryUsed, 0, perPodCPU, memoryLimit, f.ClientSet, f.InternalClientset, f.ScalesGetter) + rc := common.NewDynamicResourceConsumer(rcName, f.Namespace.Name, common.KindDeployment, pods, allPodsCPU, memoryUsed, 0, perPodCPU, memoryLimit, f.ClientSet, f.ScalesGetter) defer rc.CleanUp() rc.WaitForReplicas(pods, 15*time.Minute) diff --git a/test/e2e/network/BUILD b/test/e2e/network/BUILD index 8df831586ee..fdba7173ec5 100644 --- a/test/e2e/network/BUILD +++ b/test/e2e/network/BUILD @@ -32,7 +32,6 @@ go_library( importpath = "k8s.io/kubernetes/test/e2e/network", deps = [ "//pkg/apis/core:go_default_library", - "//pkg/client/clientset_generated/internalclientset:go_default_library", "//pkg/cloudprovider/providers/gce:go_default_library", "//pkg/controller/endpoint:go_default_library", "//pkg/master/ports:go_default_library", diff --git a/test/e2e/network/proxy.go b/test/e2e/network/proxy.go index 05cf2fcbee6..2e27463f86a 100644 --- a/test/e2e/network/proxy.go +++ b/test/e2e/network/proxy.go @@ -122,13 +122,12 @@ var _ = SIGDescribe("Proxy", func() { By("starting an echo server on multiple ports") pods := []*v1.Pod{} cfg := testutils.RCConfig{ - Client: f.ClientSet, - InternalClient: f.InternalClientset, - Image: imageutils.GetE2EImage(imageutils.Porter), - Name: service.Name, - Namespace: f.Namespace.Name, - Replicas: 1, - PollInterval: time.Second, + Client: f.ClientSet, + Image: imageutils.GetE2EImage(imageutils.Porter), + Name: service.Name, + Namespace: f.Namespace.Name, + Replicas: 1, + PollInterval: time.Second, Env: map[string]string{ "SERVE_PORT_80": `test`, "SERVE_PORT_1080": `test`, diff --git a/test/e2e/network/service.go b/test/e2e/network/service.go index 55f743992da..0f55fea22ae 100644 --- a/test/e2e/network/service.go +++ b/test/e2e/network/service.go @@ -35,7 +35,6 @@ import ( "k8s.io/apimachinery/pkg/util/wait" clientset "k8s.io/client-go/kubernetes" cloudprovider "k8s.io/cloud-provider" - "k8s.io/kubernetes/pkg/client/clientset_generated/internalclientset" gcecloud "k8s.io/kubernetes/pkg/cloudprovider/providers/gce" "k8s.io/kubernetes/pkg/controller/endpoint" "k8s.io/kubernetes/test/e2e/framework" @@ -80,12 +79,10 @@ var _ = SIGDescribe("Services", func() { f := framework.NewDefaultFramework("services") var cs clientset.Interface - var internalClientset internalclientset.Interface serviceLBNames := []string{} BeforeEach(func() { cs = f.ClientSet - internalClientset = f.InternalClientset }) AfterEach(func() { @@ -322,10 +319,10 @@ var _ = SIGDescribe("Services", func() { numPods, servicePort := 3, defaultServeHostnameServicePort By("creating service1 in namespace " + ns) - podNames1, svc1IP, err := framework.StartServeHostnameService(cs, internalClientset, getServeHostnameService("service1"), ns, numPods) + podNames1, svc1IP, err := framework.StartServeHostnameService(cs, getServeHostnameService("service1"), ns, numPods) Expect(err).NotTo(HaveOccurred(), "failed to create replication controller with service: %s in the namespace: %s", svc1IP, ns) By("creating service2 in namespace " + ns) - podNames2, svc2IP, err := framework.StartServeHostnameService(cs, internalClientset, getServeHostnameService("service2"), ns, numPods) + podNames2, svc2IP, err := framework.StartServeHostnameService(cs, getServeHostnameService("service2"), ns, numPods) Expect(err).NotTo(HaveOccurred(), "failed to create replication controller with service: %s in the namespace: %s", svc2IP, ns) hosts, err := framework.NodeSSHHosts(cs) @@ -352,7 +349,7 @@ var _ = SIGDescribe("Services", func() { // Start another service and verify both are up. By("creating service3 in namespace " + ns) - podNames3, svc3IP, err := framework.StartServeHostnameService(cs, internalClientset, getServeHostnameService("service3"), ns, numPods) + podNames3, svc3IP, err := framework.StartServeHostnameService(cs, getServeHostnameService("service3"), ns, numPods) Expect(err).NotTo(HaveOccurred(), "failed to create replication controller with service: %s in the namespace: %s", svc3IP, ns) if svc2IP == svc3IP { @@ -379,13 +376,13 @@ var _ = SIGDescribe("Services", func() { defer func() { framework.ExpectNoError(framework.StopServeHostnameService(f.ClientSet, ns, svc1)) }() - podNames1, svc1IP, err := framework.StartServeHostnameService(cs, internalClientset, getServeHostnameService(svc1), ns, numPods) + podNames1, svc1IP, err := framework.StartServeHostnameService(cs, getServeHostnameService(svc1), ns, numPods) Expect(err).NotTo(HaveOccurred(), "failed to create replication controller with service: %s in the namespace: %s", svc1IP, ns) defer func() { framework.ExpectNoError(framework.StopServeHostnameService(f.ClientSet, ns, svc2)) }() - podNames2, svc2IP, err := framework.StartServeHostnameService(cs, internalClientset, getServeHostnameService(svc2), ns, numPods) + podNames2, svc2IP, err := framework.StartServeHostnameService(cs, getServeHostnameService(svc2), ns, numPods) Expect(err).NotTo(HaveOccurred(), "failed to create replication controller with service: %s in the namespace: %s", svc2IP, ns) if svc1IP == svc2IP { @@ -432,7 +429,7 @@ var _ = SIGDescribe("Services", func() { defer func() { framework.ExpectNoError(framework.StopServeHostnameService(f.ClientSet, ns, "service1")) }() - podNames1, svc1IP, err := framework.StartServeHostnameService(cs, internalClientset, getServeHostnameService("service1"), ns, numPods) + podNames1, svc1IP, err := framework.StartServeHostnameService(cs, getServeHostnameService("service1"), ns, numPods) Expect(err).NotTo(HaveOccurred(), "failed to create replication controller with service: %s in the namespace: %s", svc1IP, ns) hosts, err := framework.NodeSSHHosts(cs) @@ -459,7 +456,7 @@ var _ = SIGDescribe("Services", func() { defer func() { framework.ExpectNoError(framework.StopServeHostnameService(f.ClientSet, ns, "service2")) }() - podNames2, svc2IP, err := framework.StartServeHostnameService(cs, internalClientset, getServeHostnameService("service2"), ns, numPods) + podNames2, svc2IP, err := framework.StartServeHostnameService(cs, getServeHostnameService("service2"), ns, numPods) Expect(err).NotTo(HaveOccurred(), "failed to create replication controller with service: %s in the namespace: %s", svc2IP, ns) if svc1IP == svc2IP { @@ -1743,12 +1740,12 @@ var _ = SIGDescribe("Services", func() { By("creating service-disabled in namespace " + ns) svcDisabled := getServeHostnameService("service-disabled") svcDisabled.ObjectMeta.Labels = serviceProxyNameLabels - _, svcDisabledIP, err := framework.StartServeHostnameService(cs, internalClientset, svcDisabled, ns, numPods) + _, svcDisabledIP, err := framework.StartServeHostnameService(cs, svcDisabled, ns, numPods) Expect(err).NotTo(HaveOccurred(), "failed to create replication controller with service: %s in the namespace: %s", svcDisabledIP, ns) By("creating service in namespace " + ns) svcToggled := getServeHostnameService("service") - podToggledNames, svcToggledIP, err := framework.StartServeHostnameService(cs, internalClientset, svcToggled, ns, numPods) + podToggledNames, svcToggledIP, err := framework.StartServeHostnameService(cs, svcToggled, ns, numPods) Expect(err).NotTo(HaveOccurred(), "failed to create replication controller with service: %s in the namespace: %s", svcToggledIP, ns) jig := framework.NewServiceTestJig(cs, svcToggled.ObjectMeta.Name) @@ -2211,7 +2208,7 @@ func execAffinityTestForNonLBService(f *framework.Framework, cs clientset.Interf By("creating service in namespace " + ns) serviceType := svc.Spec.Type svc.Spec.SessionAffinity = v1.ServiceAffinityClientIP - _, _, err := framework.StartServeHostnameService(cs, f.InternalClientset, svc, ns, numPods) + _, _, err := framework.StartServeHostnameService(cs, svc, ns, numPods) Expect(err).NotTo(HaveOccurred(), "failed to create replication controller with service in the namespace: %s", ns) defer func() { framework.StopServeHostnameService(cs, ns, serviceName) @@ -2262,7 +2259,7 @@ func execAffinityTestForLBService(f *framework.Framework, cs clientset.Interface By("creating service in namespace " + ns) svc.Spec.SessionAffinity = v1.ServiceAffinityClientIP - _, _, err := framework.StartServeHostnameService(cs, f.InternalClientset, svc, ns, numPods) + _, _, err := framework.StartServeHostnameService(cs, svc, ns, numPods) Expect(err).NotTo(HaveOccurred(), "failed to create replication controller with service in the namespace: %s", ns) jig := framework.NewServiceTestJig(cs, serviceName) By("waiting for loadbalancer for service " + ns + "/" + serviceName) diff --git a/test/e2e/network/service_latency.go b/test/e2e/network/service_latency.go index ff565bb515f..97492c698a6 100644 --- a/test/e2e/network/service_latency.go +++ b/test/e2e/network/service_latency.go @@ -128,13 +128,12 @@ var _ = SIGDescribe("Service endpoints latency", func() { func runServiceLatencies(f *framework.Framework, inParallel, total int, acceptableFailureRatio float32) (output []time.Duration, err error) { cfg := testutils.RCConfig{ - Client: f.ClientSet, - InternalClient: f.InternalClientset, - Image: imageutils.GetPauseImageName(), - Name: "svc-latency-rc", - Namespace: f.Namespace.Name, - Replicas: 1, - PollInterval: time.Second, + Client: f.ClientSet, + Image: imageutils.GetPauseImageName(), + Name: "svc-latency-rc", + Namespace: f.Namespace.Name, + Replicas: 1, + PollInterval: time.Second, } if err := framework.RunRC(cfg); err != nil { return nil, err diff --git a/test/e2e/node/kubelet.go b/test/e2e/node/kubelet.go index 7c14dc71455..620c05da34b 100644 --- a/test/e2e/node/kubelet.go +++ b/test/e2e/node/kubelet.go @@ -315,13 +315,12 @@ var _ = SIGDescribe("kubelet", func() { rcName := fmt.Sprintf("cleanup%d-%s", totalPods, string(uuid.NewUUID())) Expect(framework.RunRC(testutils.RCConfig{ - Client: f.ClientSet, - InternalClient: f.InternalClientset, - Name: rcName, - Namespace: f.Namespace.Name, - Image: imageutils.GetPauseImageName(), - Replicas: totalPods, - NodeSelector: nodeLabels, + Client: f.ClientSet, + Name: rcName, + Namespace: f.Namespace.Name, + Image: imageutils.GetPauseImageName(), + Replicas: totalPods, + NodeSelector: nodeLabels, })).NotTo(HaveOccurred()) // Perform a sanity check so that we know all desired pods are // running on the nodes according to kubelet. The timeout is set to diff --git a/test/e2e/node/kubelet_perf.go b/test/e2e/node/kubelet_perf.go index 98b8a9cd6b5..e61251301f4 100644 --- a/test/e2e/node/kubelet_perf.go +++ b/test/e2e/node/kubelet_perf.go @@ -70,12 +70,11 @@ func runResourceTrackingTest(f *framework.Framework, podsPerNode int, nodeNames // TODO: Use a more realistic workload Expect(framework.RunRC(testutils.RCConfig{ - Client: f.ClientSet, - InternalClient: f.InternalClientset, - Name: rcName, - Namespace: f.Namespace.Name, - Image: imageutils.GetPauseImageName(), - Replicas: totalPods, + Client: f.ClientSet, + Name: rcName, + Namespace: f.Namespace.Name, + Image: imageutils.GetPauseImageName(), + Replicas: totalPods, })).NotTo(HaveOccurred()) // Log once and flush the stats. diff --git a/test/e2e/scalability/BUILD b/test/e2e/scalability/BUILD index 7e9f5a1ddd6..4087d385571 100644 --- a/test/e2e/scalability/BUILD +++ b/test/e2e/scalability/BUILD @@ -15,7 +15,6 @@ go_library( "//pkg/apis/batch:go_default_library", "//pkg/apis/core:go_default_library", "//pkg/apis/extensions:go_default_library", - "//pkg/client/clientset_generated/internalclientset:go_default_library", "//staging/src/k8s.io/api/core/v1:go_default_library", "//staging/src/k8s.io/apimachinery/pkg/api/resource:go_default_library", "//staging/src/k8s.io/apimachinery/pkg/apis/meta/v1:go_default_library", diff --git a/test/e2e/scalability/density.go b/test/e2e/scalability/density.go index 5db463bceff..03f76981f98 100644 --- a/test/e2e/scalability/density.go +++ b/test/e2e/scalability/density.go @@ -42,7 +42,6 @@ import ( "k8s.io/kubernetes/pkg/apis/batch" api "k8s.io/kubernetes/pkg/apis/core" "k8s.io/kubernetes/pkg/apis/extensions" - "k8s.io/kubernetes/pkg/client/clientset_generated/internalclientset" "k8s.io/kubernetes/test/e2e/framework" "k8s.io/kubernetes/test/e2e/framework/timer" testutils "k8s.io/kubernetes/test/utils" @@ -70,12 +69,11 @@ var MaxMissingPodStartupMeasurements = 0 var nodeCount = 0 type DensityTestConfig struct { - Configs []testutils.RunObjectConfig - ClientSets []clientset.Interface - InternalClientsets []internalclientset.Interface - ScaleClients []scaleclient.ScalesGetter - PollInterval time.Duration - PodCount int + Configs []testutils.RunObjectConfig + ClientSets []clientset.Interface + ScaleClients []scaleclient.ScalesGetter + PollInterval time.Duration + PodCount int // What kind of resource we want to create kind schema.GroupKind SecretConfigs []*testutils.SecretConfig @@ -644,7 +642,7 @@ var _ = SIGDescribe("Density", func() { } timeout += 3 * time.Minute // createClients is defined in load.go - clients, internalClients, scalesClients, err := createClients(numberOfCollections) + clients, scalesClients, err := createClients(numberOfCollections) framework.ExpectNoError(err) for i := 0; i < numberOfCollections; i++ { nsName := namespaces[i].Name @@ -675,7 +673,6 @@ var _ = SIGDescribe("Density", func() { name := fmt.Sprintf("density%v-%v-%v", totalPods, i, uuid) baseConfig := &testutils.RCConfig{ Client: clients[i], - InternalClient: internalClients[i], ScalesGetter: scalesClients[i], Image: imageutils.GetPauseImageName(), Name: name, @@ -722,18 +719,17 @@ var _ = SIGDescribe("Density", func() { } // Single client is running out of http2 connections in delete phase, hence we need more. - clients, internalClients, scalesClients, err = createClients(2) + clients, scalesClients, err = createClients(2) framework.ExpectNoError(err) dConfig := DensityTestConfig{ - ClientSets: clients, - InternalClientsets: internalClients, - ScaleClients: scalesClients, - Configs: configs, - PodCount: totalPods, - PollInterval: DensityPollInterval, - kind: itArg.kind, - SecretConfigs: secretConfigs, - ConfigMapConfigs: configMapConfigs, + ClientSets: clients, + ScaleClients: scalesClients, + Configs: configs, + PodCount: totalPods, + PollInterval: DensityPollInterval, + kind: itArg.kind, + SecretConfigs: secretConfigs, + ConfigMapConfigs: configMapConfigs, } for i := 0; i < itArg.daemonsPerNode; i++ { diff --git a/test/e2e/scalability/load.go b/test/e2e/scalability/load.go index 43ebd350bef..9224607afaf 100644 --- a/test/e2e/scalability/load.go +++ b/test/e2e/scalability/load.go @@ -49,7 +49,6 @@ import ( "k8s.io/kubernetes/pkg/apis/batch" api "k8s.io/kubernetes/pkg/apis/core" "k8s.io/kubernetes/pkg/apis/extensions" - "k8s.io/kubernetes/pkg/client/clientset_generated/internalclientset" "k8s.io/kubernetes/test/e2e/framework" "k8s.io/kubernetes/test/e2e/framework/timer" testutils "k8s.io/kubernetes/test/utils" @@ -341,9 +340,8 @@ var _ = SIGDescribe("Load capacity", func() { } }) -func createClients(numberOfClients int) ([]clientset.Interface, []internalclientset.Interface, []scaleclient.ScalesGetter, error) { +func createClients(numberOfClients int) ([]clientset.Interface, []scaleclient.ScalesGetter, error) { clients := make([]clientset.Interface, numberOfClients) - internalClients := make([]internalclientset.Interface, numberOfClients) scalesClients := make([]scaleclient.ScalesGetter, numberOfClients) for i := 0; i < numberOfClients; i++ { @@ -361,11 +359,11 @@ func createClients(numberOfClients int) ([]clientset.Interface, []internalclient // each client here. transportConfig, err := config.TransportConfig() if err != nil { - return nil, nil, nil, err + return nil, nil, err } tlsConfig, err := transport.TLSConfigFor(transportConfig) if err != nil { - return nil, nil, nil, err + return nil, nil, err } config.Transport = utilnet.SetTransportDefaults(&http.Transport{ Proxy: http.ProxyFromEnvironment, @@ -387,14 +385,9 @@ func createClients(numberOfClients int) ([]clientset.Interface, []internalclient c, err := clientset.NewForConfig(config) if err != nil { - return nil, nil, nil, err + return nil, nil, err } clients[i] = c - internalClient, err := internalclientset.NewForConfig(config) - if err != nil { - return nil, nil, nil, err - } - internalClients[i] = internalClient // create scale client, if GroupVersion or NegotiatedSerializer are not set // assign default values - these fields are mandatory (required by RESTClientFor). @@ -406,11 +399,11 @@ func createClients(numberOfClients int) ([]clientset.Interface, []internalclient } restClient, err := restclient.RESTClientFor(config) if err != nil { - return nil, nil, nil, err + return nil, nil, err } discoClient, err := discovery.NewDiscoveryClientForConfig(config) if err != nil { - return nil, nil, nil, err + return nil, nil, err } cachedDiscoClient := cacheddiscovery.NewMemCacheClient(discoClient) restMapper := restmapper.NewDeferredDiscoveryRESTMapper(cachedDiscoClient) @@ -418,7 +411,7 @@ func createClients(numberOfClients int) ([]clientset.Interface, []internalclient resolver := scaleclient.NewDiscoveryScaleKindResolver(cachedDiscoClient) scalesClients[i] = scaleclient.New(restClient, restMapper, dynamic.LegacyAPIPathResolverFunc, resolver) } - return clients, internalClients, scalesClients, nil + return clients, scalesClients, nil } func computePodCounts(total int) (int, int, int) { @@ -478,12 +471,11 @@ func generateConfigs( // Create a number of clients to better simulate real usecase // where not everyone is using exactly the same client. rcsPerClient := 20 - clients, internalClients, scalesClients, err := createClients((len(configs) + rcsPerClient - 1) / rcsPerClient) + clients, scalesClients, err := createClients((len(configs) + rcsPerClient - 1) / rcsPerClient) framework.ExpectNoError(err) for i := 0; i < len(configs); i++ { configs[i].SetClient(clients[i%len(clients)]) - configs[i].SetInternalClient(internalClients[i%len(internalClients)]) configs[i].SetScalesClient(scalesClients[i%len(clients)]) } for i := 0; i < len(secretConfigs); i++ { @@ -542,7 +534,6 @@ func GenerateConfigsForGroup( baseConfig := &testutils.RCConfig{ Client: nil, // this will be overwritten later - InternalClient: nil, // this will be overwritten later Name: groupName + "-" + strconv.Itoa(i), Namespace: namespace, Timeout: UnreadyNodeToleration, diff --git a/test/e2e/scheduling/equivalence_cache_predicates.go b/test/e2e/scheduling/equivalence_cache_predicates.go index 27f0eaf43c3..c0105e78daf 100644 --- a/test/e2e/scheduling/equivalence_cache_predicates.go +++ b/test/e2e/scheduling/equivalence_cache_predicates.go @@ -272,15 +272,14 @@ func CreateNodeSelectorPods(f *framework.Framework, id string, replicas int, nod By(fmt.Sprintf("Running RC which reserves host port and defines node selector")) config := &testutils.RCConfig{ - Client: f.ClientSet, - InternalClient: f.InternalClientset, - Name: id, - Namespace: f.Namespace.Name, - Timeout: defaultTimeout, - Image: imageutils.GetPauseImageName(), - Replicas: replicas, - HostPorts: map[string]int{"port1": 4321}, - NodeSelector: nodeSelector, + Client: f.ClientSet, + Name: id, + Namespace: f.Namespace.Name, + Timeout: defaultTimeout, + Image: imageutils.GetPauseImageName(), + Replicas: replicas, + HostPorts: map[string]int{"port1": 4321}, + NodeSelector: nodeSelector, } err := framework.RunRC(*config) if expectRunning { diff --git a/test/e2e/scheduling/predicates.go b/test/e2e/scheduling/predicates.go index dd9ebd36f1d..2efc175726e 100644 --- a/test/e2e/scheduling/predicates.go +++ b/test/e2e/scheduling/predicates.go @@ -787,14 +787,13 @@ func getNodeThatCanRunPodWithoutToleration(f *framework.Framework) string { func CreateHostPortPods(f *framework.Framework, id string, replicas int, expectRunning bool) { By(fmt.Sprintf("Running RC which reserves host port")) config := &testutils.RCConfig{ - Client: f.ClientSet, - InternalClient: f.InternalClientset, - Name: id, - Namespace: f.Namespace.Name, - Timeout: defaultTimeout, - Image: imageutils.GetPauseImageName(), - Replicas: replicas, - HostPorts: map[string]int{"port1": 4321}, + Client: f.ClientSet, + Name: id, + Namespace: f.Namespace.Name, + Timeout: defaultTimeout, + Image: imageutils.GetPauseImageName(), + Replicas: replicas, + HostPorts: map[string]int{"port1": 4321}, } err := framework.RunRC(*config) if expectRunning { diff --git a/test/e2e/upgrades/horizontal_pod_autoscalers.go b/test/e2e/upgrades/horizontal_pod_autoscalers.go index aaad13f1715..9ca6618962b 100644 --- a/test/e2e/upgrades/horizontal_pod_autoscalers.go +++ b/test/e2e/upgrades/horizontal_pod_autoscalers.go @@ -49,7 +49,6 @@ func (t *HPAUpgradeTest) Setup(f *framework.Framework) { 500, /* cpuLimit */ 200, /* memLimit */ f.ClientSet, - f.InternalClientset, f.ScalesGetter) t.hpa = common.CreateCPUHorizontalPodAutoscaler( t.rc, diff --git a/test/utils/BUILD b/test/utils/BUILD index 90b00e25e3f..b4233c8c831 100644 --- a/test/utils/BUILD +++ b/test/utils/BUILD @@ -30,7 +30,6 @@ go_library( "//pkg/apis/batch:go_default_library", "//pkg/apis/core:go_default_library", "//pkg/apis/extensions:go_default_library", - "//pkg/client/clientset_generated/internalclientset:go_default_library", "//pkg/controller/deployment/util:go_default_library", "//pkg/kubectl:go_default_library", "//pkg/util/labels:go_default_library", diff --git a/test/utils/runners.go b/test/utils/runners.go index dc04d6d2ad3..74fad6c1fd2 100644 --- a/test/utils/runners.go +++ b/test/utils/runners.go @@ -45,7 +45,6 @@ import ( batchinternal "k8s.io/kubernetes/pkg/apis/batch" api "k8s.io/kubernetes/pkg/apis/core" extensionsinternal "k8s.io/kubernetes/pkg/apis/extensions" - "k8s.io/kubernetes/pkg/client/clientset_generated/internalclientset" "k8s.io/klog" ) @@ -99,10 +98,8 @@ type RunObjectConfig interface { GetNamespace() string GetKind() schema.GroupKind GetClient() clientset.Interface - GetInternalClient() internalclientset.Interface GetScalesGetter() scaleclient.ScalesGetter SetClient(clientset.Interface) - SetInternalClient(internalclientset.Interface) SetScalesClient(scaleclient.ScalesGetter) GetReplicas() int GetLabelValue(string) (string, bool) @@ -112,7 +109,6 @@ type RunObjectConfig interface { type RCConfig struct { Affinity *v1.Affinity Client clientset.Interface - InternalClient internalclientset.Interface ScalesGetter scaleclient.ScalesGetter Image string Command []string @@ -527,10 +523,6 @@ func (config *RCConfig) GetClient() clientset.Interface { return config.Client } -func (config *RCConfig) GetInternalClient() internalclientset.Interface { - return config.InternalClient -} - func (config *RCConfig) GetScalesGetter() scaleclient.ScalesGetter { return config.ScalesGetter } @@ -539,10 +531,6 @@ func (config *RCConfig) SetClient(c clientset.Interface) { config.Client = c } -func (config *RCConfig) SetInternalClient(c internalclientset.Interface) { - config.InternalClient = c -} - func (config *RCConfig) SetScalesClient(getter scaleclient.ScalesGetter) { config.ScalesGetter = getter }