diff --git a/test/e2e/common/pods.go b/test/e2e/common/pods.go index c2e7af9f452..e8d5c02224a 100644 --- a/test/e2e/common/pods.go +++ b/test/e2e/common/pods.go @@ -67,7 +67,7 @@ func testHostIP(podClient *framework.PodClient, pod *v1.Pod) { t := time.Now() for { p, err := podClient.Get(pod.Name, metav1.GetOptions{}) - gomega.Expect(err).NotTo(gomega.HaveOccurred(), "Failed to get pod %q", pod.Name) + framework.ExpectNoError(err, "Failed to get pod %q", pod.Name) if p.Status.HostIP != "" { e2elog.Logf("Pod %s has hostIP: %s", p.Name, p.Status.HostIP) break @@ -218,7 +218,7 @@ var _ = framework.KubeDescribe("Pods", func() { selector := labels.SelectorFromSet(labels.Set(map[string]string{"time": value})) options := metav1.ListOptions{LabelSelector: selector.String()} pods, err := podClient.List(options) - gomega.Expect(err).NotTo(gomega.HaveOccurred(), "failed to query for pods") + framework.ExpectNoError(err, "failed to query for pods") gomega.Expect(len(pods.Items)).To(gomega.Equal(0)) options = metav1.ListOptions{ LabelSelector: selector.String(), @@ -256,7 +256,7 @@ var _ = framework.KubeDescribe("Pods", func() { selector = labels.SelectorFromSet(labels.Set(map[string]string{"time": value})) options = metav1.ListOptions{LabelSelector: selector.String()} pods, err = podClient.List(options) - gomega.Expect(err).NotTo(gomega.HaveOccurred(), "failed to query for pods") + framework.ExpectNoError(err, "failed to query for pods") gomega.Expect(len(pods.Items)).To(gomega.Equal(1)) ginkgo.By("verifying pod creation was observed") @@ -279,11 +279,11 @@ var _ = framework.KubeDescribe("Pods", func() { framework.ExpectNoError(f.WaitForPodRunning(pod.Name)) // save the running pod pod, err = podClient.Get(pod.Name, metav1.GetOptions{}) - gomega.Expect(err).NotTo(gomega.HaveOccurred(), "failed to GET scheduled pod") + framework.ExpectNoError(err, "failed to GET scheduled pod") ginkgo.By("deleting the pod gracefully") err = podClient.Delete(pod.Name, metav1.NewDeleteOptions(30)) - gomega.Expect(err).NotTo(gomega.HaveOccurred(), "failed to delete pod") + framework.ExpectNoError(err, "failed to delete pod") ginkgo.By("verifying the kubelet observed the termination notice") gomega.Expect(wait.Poll(time.Second*5, time.Second*30, func() (bool, error) { @@ -335,7 +335,7 @@ var _ = framework.KubeDescribe("Pods", func() { selector = labels.SelectorFromSet(labels.Set(map[string]string{"time": value})) options = metav1.ListOptions{LabelSelector: selector.String()} pods, err = podClient.List(options) - gomega.Expect(err).NotTo(gomega.HaveOccurred(), "failed to query for pods") + framework.ExpectNoError(err, "failed to query for pods") gomega.Expect(len(pods.Items)).To(gomega.Equal(0)) }) @@ -373,7 +373,7 @@ var _ = framework.KubeDescribe("Pods", func() { selector := labels.SelectorFromSet(labels.Set(map[string]string{"time": value})) options := metav1.ListOptions{LabelSelector: selector.String()} pods, err := podClient.List(options) - gomega.Expect(err).NotTo(gomega.HaveOccurred(), "failed to query for pods") + framework.ExpectNoError(err, "failed to query for pods") gomega.Expect(len(pods.Items)).To(gomega.Equal(1)) ginkgo.By("updating the pod") @@ -388,7 +388,7 @@ var _ = framework.KubeDescribe("Pods", func() { selector = labels.SelectorFromSet(labels.Set(map[string]string{"time": value})) options = metav1.ListOptions{LabelSelector: selector.String()} pods, err = podClient.List(options) - gomega.Expect(err).NotTo(gomega.HaveOccurred(), "failed to query for pods") + framework.ExpectNoError(err, "failed to query for pods") gomega.Expect(len(pods.Items)).To(gomega.Equal(1)) e2elog.Logf("Pod update OK") }) @@ -427,7 +427,7 @@ var _ = framework.KubeDescribe("Pods", func() { selector := labels.SelectorFromSet(labels.Set(map[string]string{"time": value})) options := metav1.ListOptions{LabelSelector: selector.String()} pods, err := podClient.List(options) - gomega.Expect(err).NotTo(gomega.HaveOccurred(), "failed to query for pods") + framework.ExpectNoError(err, "failed to query for pods") gomega.Expect(len(pods.Items)).To(gomega.Equal(1)) ginkgo.By("updating the pod") @@ -491,7 +491,7 @@ var _ = framework.KubeDescribe("Pods", func() { }, } _, err := f.ClientSet.CoreV1().Services(f.Namespace.Name).Create(svc) - gomega.Expect(err).NotTo(gomega.HaveOccurred(), "failed to create service") + framework.ExpectNoError(err, "failed to create service") // Make a client pod that verifies that it has the service environment variables. podName := "client-envvars-" + string(uuid.NewUUID()) @@ -538,7 +538,7 @@ var _ = framework.KubeDescribe("Pods", func() { */ framework.ConformanceIt("should support remote command execution over websockets [NodeConformance]", func() { config, err := framework.LoadConfig() - gomega.Expect(err).NotTo(gomega.HaveOccurred(), "unable to get base config") + framework.ExpectNoError(err, "unable to get base config") ginkgo.By("creating the pod") name := "pod-exec-websocket-" + string(uuid.NewUUID()) @@ -620,7 +620,7 @@ var _ = framework.KubeDescribe("Pods", func() { */ framework.ConformanceIt("should support retrieving logs from the container over websockets [NodeConformance]", func() { config, err := framework.LoadConfig() - gomega.Expect(err).NotTo(gomega.HaveOccurred(), "unable to get base config") + framework.ExpectNoError(err, "unable to get base config") ginkgo.By("creating the pod") name := "pod-logs-websocket-" + string(uuid.NewUUID()) @@ -798,14 +798,15 @@ var _ = framework.KubeDescribe("Pods", func() { } validatePodReadiness := func(expectReady bool) { - gomega.Expect(wait.Poll(time.Second, maxReadyStatusUpdateTolerance, func() (bool, error) { + err := wait.Poll(time.Second, maxReadyStatusUpdateTolerance, func() (bool, error) { podReady := podClient.PodIsReady(podName) res := expectReady == podReady if !res { e2elog.Logf("Expect the Ready condition of pod %q to be %v, but got %v", podName, expectReady, podReady) } return res, nil - })).NotTo(gomega.HaveOccurred()) + }) + framework.ExpectNoError(err) } ginkgo.By("submitting the pod to kubernetes") diff --git a/test/e2e/common/runtime.go b/test/e2e/common/runtime.go index ac92014d641..d51da57ee75 100644 --- a/test/e2e/common/runtime.go +++ b/test/e2e/common/runtime.go @@ -114,7 +114,7 @@ while true; do sleep 1; done Expect(terminateContainer.IsReady()).Should(Equal(testCase.Ready)) status, err := terminateContainer.GetStatus() - Expect(err).ShouldNot(HaveOccurred()) + framework.ExpectNoError(err) By(fmt.Sprintf("Container '%s': should get the expected 'State'", testContainer.Name)) Expect(GetContainerState(status.State)).To(Equal(testCase.State)) @@ -148,7 +148,7 @@ while true; do sleep 1; done By("get the container status") status, err := c.GetStatus() - Expect(err).NotTo(HaveOccurred()) + framework.ExpectNoError(err) By("the container should be terminated") Expect(GetContainerState(status.State)).To(Equal(ContainerStateTerminated)) @@ -286,7 +286,7 @@ while true; do sleep 1; done secret.Name = "image-pull-secret-" + string(uuid.NewUUID()) By("create image pull secret") _, err := f.ClientSet.CoreV1().Secrets(f.Namespace.Name).Create(secret) - Expect(err).NotTo(HaveOccurred()) + framework.ExpectNoError(err) defer f.ClientSet.CoreV1().Secrets(f.Namespace.Name).Delete(secret.Name, nil) container.ImagePullSecrets = []string{secret.Name} } diff --git a/test/e2e/common/sysctl.go b/test/e2e/common/sysctl.go index 2d8ef72a6c7..88f6c79bc72 100644 --- a/test/e2e/common/sysctl.go +++ b/test/e2e/common/sysctl.go @@ -77,7 +77,7 @@ var _ = framework.KubeDescribe("Sysctls [NodeFeature:Sysctls]", func() { // failed pods without running containers. This would create a race as the pod // might have already been deleted here. ev, err := f.PodClient().WaitForErrorEventOrSuccess(pod) - Expect(err).NotTo(HaveOccurred()) + framework.ExpectNoError(err) if ev != nil && ev.Reason == sysctl.UnsupportedReason { framework.Skipf("No sysctl support in Docker <1.12") } @@ -85,16 +85,16 @@ var _ = framework.KubeDescribe("Sysctls [NodeFeature:Sysctls]", func() { By("Waiting for pod completion") err = f.WaitForPodNoLongerRunning(pod.Name) - Expect(err).NotTo(HaveOccurred()) + framework.ExpectNoError(err) pod, err = podClient.Get(pod.Name, metav1.GetOptions{}) - Expect(err).NotTo(HaveOccurred()) + framework.ExpectNoError(err) By("Checking that the pod succeeded") Expect(pod.Status.Phase).To(Equal(v1.PodSucceeded)) By("Getting logs from the pod") log, err := framework.GetPodLogs(f.ClientSet, f.Namespace.Name, pod.Name, pod.Spec.Containers[0].Name) - Expect(err).NotTo(HaveOccurred()) + framework.ExpectNoError(err) By("Checking that the sysctl is actually updated") Expect(log).To(ContainSubstring("kernel.shm_rmid_forced = 1")) @@ -120,7 +120,7 @@ var _ = framework.KubeDescribe("Sysctls [NodeFeature:Sysctls]", func() { // failed pods without running containers. This would create a race as the pod // might have already been deleted here. ev, err := f.PodClient().WaitForErrorEventOrSuccess(pod) - Expect(err).NotTo(HaveOccurred()) + framework.ExpectNoError(err) if ev != nil && ev.Reason == sysctl.UnsupportedReason { framework.Skipf("No sysctl support in Docker <1.12") } @@ -128,16 +128,16 @@ var _ = framework.KubeDescribe("Sysctls [NodeFeature:Sysctls]", func() { By("Waiting for pod completion") err = f.WaitForPodNoLongerRunning(pod.Name) - Expect(err).NotTo(HaveOccurred()) + framework.ExpectNoError(err) pod, err = podClient.Get(pod.Name, metav1.GetOptions{}) - Expect(err).NotTo(HaveOccurred()) + framework.ExpectNoError(err) By("Checking that the pod succeeded") Expect(pod.Status.Phase).To(Equal(v1.PodSucceeded)) By("Getting logs from the pod") log, err := framework.GetPodLogs(f.ClientSet, f.Namespace.Name, pod.Name, pod.Spec.Containers[0].Name) - Expect(err).NotTo(HaveOccurred()) + framework.ExpectNoError(err) By("Checking that the sysctl is actually updated") Expect(log).To(ContainSubstring("kernel.shm_rmid_forced = 1")) @@ -197,7 +197,7 @@ var _ = framework.KubeDescribe("Sysctls [NodeFeature:Sysctls]", func() { // failed pods without running containers. This would create a race as the pod // might have already been deleted here. ev, err := f.PodClient().WaitForErrorEventOrSuccess(pod) - Expect(err).NotTo(HaveOccurred()) + framework.ExpectNoError(err) if ev != nil && ev.Reason == sysctl.UnsupportedReason { framework.Skipf("No sysctl support in Docker <1.12") }