From bcddb157e4a737f29ce9d9dbd22880f2001474b0 Mon Sep 17 00:00:00 2001 From: "Madhusudan.C.S" Date: Thu, 1 Dec 2016 15:27:42 -0800 Subject: [PATCH] [Federation] Separate the cleanup phases of service and service shards so that service shards can be cleaned up even after the service is deleted elsewhere. This separation is necessary because "Federated Service DNS should be able to discover a federated service" e2e test recently added a case where it deletes the service from federation but not the shards from the underlying clusters. Because of the way cleanup was implemented in the AfterEach block currently, we did not cleanup any of the underlying shards. However, separating the two phases of the cleanup needs this separation. --- test/e2e/federated-service.go | 39 +++++++++++++++++++++++++++++------ 1 file changed, 33 insertions(+), 6 deletions(-) diff --git a/test/e2e/federated-service.go b/test/e2e/federated-service.go index ae465a1c372..8320565f322 100644 --- a/test/e2e/federated-service.go +++ b/test/e2e/federated-service.go @@ -24,6 +24,7 @@ import ( "time" "k8s.io/kubernetes/pkg/api/v1" + "k8s.io/kubernetes/pkg/conversion" "k8s.io/kubernetes/test/e2e/framework" . "github.com/onsi/ginkgo" @@ -139,7 +140,8 @@ var _ = framework.KubeDescribe("[Feature:Federation]", func() { var _ = Describe("DNS", func() { var ( - service *v1.Service + service *v1.Service + serviceShard *v1.Service ) BeforeEach(func() { @@ -168,8 +170,25 @@ var _ = framework.KubeDescribe("[Feature:Federation]", func() { } createBackendPodsOrFail(clusters, nsName, FederatedServicePodName) + service = createServiceOrFail(f.FederationClientset_1_5, nsName, FederatedServiceName) - waitForServiceShardsOrFail(nsName, service, clusters) + obj, err := conversion.NewCloner().DeepCopy(service) + // Cloning shouldn't fail. On the off-chance it does, we should + // shallow copy service to serviceShard before failing. + if err != nil { + serviceCopy := *service + serviceShard = &serviceCopy + framework.ExpectNoError(err, fmt.Sprintf("Error in deep copying service %q", service.Name)) + } + var ok bool + serviceShard, ok = obj.(*v1.Service) + if !ok { + serviceCopy := *service + serviceShard = &serviceCopy + framework.ExpectNoError(err, fmt.Sprintf("Unexpected service object copied %T", obj)) + } + + waitForServiceShardsOrFail(nsName, serviceShard, clusters) }) AfterEach(func() { @@ -180,14 +199,19 @@ var _ = framework.KubeDescribe("[Feature:Federation]", func() { if service != nil { deleteServiceOrFail(f.FederationClientset_1_5, nsName, service.Name) - - By(fmt.Sprintf("Deleting service shards and their provider resources in underlying clusters for service %q in namespace %q", service.Name, nsName)) - cleanupServiceShardsAndProviderResources(nsName, service, clusters) - service = nil } else { By("No service to delete. Service is nil") } + + if serviceShard != nil { + By(fmt.Sprintf("Deleting service shards and their provider resources in underlying clusters for service %q in namespace %q", service.Name, nsName)) + cleanupServiceShardsAndProviderResources(nsName, service, clusters) + serviceShard = nil + } else { + By("No service shards to delete. `serviceShard` is nil") + } + // Delete the kube-dns config map from all clusters. for clusterName, cluster := range clusters { By(fmt.Sprintf("Deleting kube dns config map from cluster: %s", clusterName)) @@ -218,6 +242,9 @@ var _ = framework.KubeDescribe("[Feature:Federation]", func() { By("Deleting the service to verify that DNS rules still work") err := f.FederationClientset_1_5.Services(nsName).Delete(FederatedServiceName, &v1.DeleteOptions{}) framework.ExpectNoError(err, "Error deleting service %q in namespace %q", service.Name, service.Namespace) + // Service is deleted, unset the test block-global service variable. + service = nil + for i, DNSName := range svcDNSNames { discoverService(f, DNSName, true, "federated-service-e2e-discovery-pod-"+strconv.Itoa(i)) }