diff --git a/test/e2e/apimachinery/aggregator.go b/test/e2e/apimachinery/aggregator.go index 9a6b194ca8f..2f2569aa1fc 100644 --- a/test/e2e/apimachinery/aggregator.go +++ b/test/e2e/apimachinery/aggregator.go @@ -28,7 +28,7 @@ import ( appsv1 "k8s.io/api/apps/v1" v1 "k8s.io/api/core/v1" rbacv1 "k8s.io/api/rbac/v1" - apierrs "k8s.io/apimachinery/pkg/api/errors" + apierrors "k8s.io/apimachinery/pkg/api/errors" metav1 "k8s.io/apimachinery/pkg/apis/meta/v1" "k8s.io/apimachinery/pkg/apis/meta/v1/unstructured" unstructuredv1 "k8s.io/apimachinery/pkg/apis/meta/v1/unstructured" @@ -367,7 +367,7 @@ func TestSampleAPIServer(f *framework.Framework, aggrclient *aggregatorclient.Cl request.SetHeader("Accept", "application/json") _, err := request.DoRaw() if err != nil { - status, ok := err.(*apierrs.StatusError) + status, ok := err.(*apierrors.StatusError) if !ok { return false, err } diff --git a/test/e2e/auth/pod_security_policy.go b/test/e2e/auth/pod_security_policy.go index 0fe3d7e30cd..f0929f69bde 100644 --- a/test/e2e/auth/pod_security_policy.go +++ b/test/e2e/auth/pod_security_policy.go @@ -22,7 +22,7 @@ import ( v1 "k8s.io/api/core/v1" policyv1beta1 "k8s.io/api/policy/v1beta1" rbacv1 "k8s.io/api/rbac/v1" - apierrs "k8s.io/apimachinery/pkg/api/errors" + apierrors "k8s.io/apimachinery/pkg/api/errors" metav1 "k8s.io/apimachinery/pkg/apis/meta/v1" "k8s.io/apimachinery/pkg/runtime/schema" "k8s.io/apiserver/pkg/authentication/serviceaccount" @@ -122,7 +122,7 @@ var _ = SIGDescribe("PodSecurityPolicy", func() { func expectForbidden(err error) { framework.ExpectError(err, "should be forbidden") - framework.ExpectEqual(apierrs.IsForbidden(err), true, "should be forbidden error") + framework.ExpectEqual(apierrors.IsForbidden(err), true, "should be forbidden error") } func testPrivilegedPods(tester func(pod *v1.Pod)) { diff --git a/test/e2e/cloud/nodes.go b/test/e2e/cloud/nodes.go index 76b9bb331c0..f0eb8511201 100644 --- a/test/e2e/cloud/nodes.go +++ b/test/e2e/cloud/nodes.go @@ -20,7 +20,7 @@ import ( "time" v1 "k8s.io/api/core/v1" - apierrs "k8s.io/apimachinery/pkg/api/errors" + apierrors "k8s.io/apimachinery/pkg/api/errors" metav1 "k8s.io/apimachinery/pkg/apis/meta/v1" clientset "k8s.io/client-go/kubernetes" "k8s.io/kubernetes/test/e2e/framework" @@ -67,7 +67,7 @@ var _ = SIGDescribe("[Feature:CloudProvider][Disruptive] Nodes", func() { _, err = c.CoreV1().Nodes().Get(nodeToDelete.Name, metav1.GetOptions{}) if err == nil { framework.Failf("node %q still exists when it should be deleted", nodeToDelete.Name) - } else if !apierrs.IsNotFound(err) { + } else if !apierrors.IsNotFound(err) { framework.Failf("failed to get node %q err: %q", nodeToDelete.Name, err) } diff --git a/test/e2e/common/runtimeclass.go b/test/e2e/common/runtimeclass.go index 53aead2277f..8087bedee3f 100644 --- a/test/e2e/common/runtimeclass.go +++ b/test/e2e/common/runtimeclass.go @@ -22,7 +22,7 @@ import ( v1 "k8s.io/api/core/v1" "k8s.io/apimachinery/pkg/api/errors" - apierrs "k8s.io/apimachinery/pkg/api/errors" + apierrors "k8s.io/apimachinery/pkg/api/errors" metav1 "k8s.io/apimachinery/pkg/apis/meta/v1" "k8s.io/apimachinery/pkg/fields" "k8s.io/apimachinery/pkg/util/wait" @@ -124,7 +124,7 @@ func expectPodRejection(f *framework.Framework, pod *v1.Pod) { } else { _, err := f.ClientSet.CoreV1().Pods(f.Namespace.Name).Create(pod) framework.ExpectError(err, "should be forbidden") - framework.ExpectEqual(apierrs.IsForbidden(err), true, "should be forbidden error") + framework.ExpectEqual(apierrors.IsForbidden(err), true, "should be forbidden error") } } diff --git a/test/e2e/framework/ingress/ingress_utils.go b/test/e2e/framework/ingress/ingress_utils.go index 3593cfdc3f3..88bc504d93c 100644 --- a/test/e2e/framework/ingress/ingress_utils.go +++ b/test/e2e/framework/ingress/ingress_utils.go @@ -41,7 +41,7 @@ import ( appsv1 "k8s.io/api/apps/v1" v1 "k8s.io/api/core/v1" networkingv1beta1 "k8s.io/api/networking/v1beta1" - apierrs "k8s.io/apimachinery/pkg/api/errors" + apierrors "k8s.io/apimachinery/pkg/api/errors" metav1 "k8s.io/apimachinery/pkg/apis/meta/v1" "k8s.io/apimachinery/pkg/labels" "k8s.io/apimachinery/pkg/util/intstr" @@ -522,7 +522,7 @@ func (j *TestJig) Update(update func(ing *networkingv1beta1.Ingress)) { framework.DescribeIng(j.Ingress.Namespace) return } - if !apierrs.IsConflict(err) && !apierrs.IsServerTimeout(err) { + if !apierrors.IsConflict(err) && !apierrors.IsServerTimeout(err) { framework.Failf("failed to update ingress %s/%s: %v", ns, name, err) } } diff --git a/test/e2e/framework/node/wait_test.go b/test/e2e/framework/node/wait_test.go index 050c4661143..7ecf37e7754 100644 --- a/test/e2e/framework/node/wait_test.go +++ b/test/e2e/framework/node/wait_test.go @@ -21,7 +21,7 @@ import ( "testing" v1 "k8s.io/api/core/v1" - apierrs "k8s.io/apimachinery/pkg/api/errors" + apierrors "k8s.io/apimachinery/pkg/api/errors" metav1 "k8s.io/apimachinery/pkg/apis/meta/v1" "k8s.io/apimachinery/pkg/runtime" "k8s.io/client-go/kubernetes/fake" @@ -157,7 +157,7 @@ func TestCheckReadyForTests(t *testing.T) { expectedErr: "Forced error", }, { desc: "Retryable errors from node list are reported but still return false", - nodeListErr: apierrs.NewTimeoutError("Retryable error", 10), + nodeListErr: apierrors.NewTimeoutError("Retryable error", 10), expected: false, }, } diff --git a/test/e2e/framework/pod/delete.go b/test/e2e/framework/pod/delete.go index 5810615e686..a4a2bfcdf26 100644 --- a/test/e2e/framework/pod/delete.go +++ b/test/e2e/framework/pod/delete.go @@ -23,7 +23,7 @@ import ( "github.com/onsi/ginkgo" v1 "k8s.io/api/core/v1" - apierrs "k8s.io/apimachinery/pkg/api/errors" + apierrors "k8s.io/apimachinery/pkg/api/errors" clientset "k8s.io/client-go/kubernetes" e2elog "k8s.io/kubernetes/test/e2e/framework/log" ) @@ -55,7 +55,7 @@ func DeletePodWithWaitByName(c clientset.Interface, podName, podNamespace string e2elog.Logf("Deleting pod %q in namespace %q", podName, podNamespace) err := c.CoreV1().Pods(podNamespace).Delete(podName, nil) if err != nil { - if apierrs.IsNotFound(err) { + if apierrors.IsNotFound(err) { return nil // assume pod was already deleted } return fmt.Errorf("pod Delete API error: %v", err) diff --git a/test/e2e/framework/pod/wait.go b/test/e2e/framework/pod/wait.go index 409d736054b..41ab4601afe 100644 --- a/test/e2e/framework/pod/wait.go +++ b/test/e2e/framework/pod/wait.go @@ -27,7 +27,7 @@ import ( "github.com/onsi/ginkgo" v1 "k8s.io/api/core/v1" - apierrs "k8s.io/apimachinery/pkg/api/errors" + apierrors "k8s.io/apimachinery/pkg/api/errors" metav1 "k8s.io/apimachinery/pkg/apis/meta/v1" "k8s.io/apimachinery/pkg/labels" "k8s.io/apimachinery/pkg/runtime/schema" @@ -213,7 +213,7 @@ func WaitForPodCondition(c clientset.Interface, ns, podName, desc string, timeou for start := time.Now(); time.Since(start) < timeout; time.Sleep(poll) { pod, err := c.CoreV1().Pods(ns).Get(podName, metav1.GetOptions{}) if err != nil { - if apierrs.IsNotFound(err) { + if apierrors.IsNotFound(err) { e2elog.Logf("Pod %q in namespace %q not found. Error: %v", podName, ns, err) return err } @@ -387,7 +387,7 @@ func WaitForPodSuccessInNamespaceSlow(c clientset.Interface, podName string, nam func WaitForPodNotFoundInNamespace(c clientset.Interface, podName, ns string, timeout time.Duration) error { return wait.PollImmediate(poll, timeout, func() (bool, error) { _, err := c.CoreV1().Pods(ns).Get(podName, metav1.GetOptions{}) - if apierrs.IsNotFound(err) { + if apierrors.IsNotFound(err) { return true, nil // done } if err != nil { diff --git a/test/e2e/framework/psp.go b/test/e2e/framework/psp.go index 06210c668d2..3518b106c2b 100644 --- a/test/e2e/framework/psp.go +++ b/test/e2e/framework/psp.go @@ -23,7 +23,7 @@ import ( v1 "k8s.io/api/core/v1" policyv1beta1 "k8s.io/api/policy/v1beta1" rbacv1 "k8s.io/api/rbac/v1" - apierrs "k8s.io/apimachinery/pkg/api/errors" + apierrors "k8s.io/apimachinery/pkg/api/errors" metav1 "k8s.io/apimachinery/pkg/apis/meta/v1" "k8s.io/apimachinery/pkg/runtime/schema" "k8s.io/apiserver/pkg/authentication/serviceaccount" @@ -111,7 +111,7 @@ func CreatePrivilegedPSPBinding(kubeClient clientset.Interface, namespace string privilegedPSPOnce.Do(func() { _, err := kubeClient.PolicyV1beta1().PodSecurityPolicies().Get( podSecurityPolicyPrivileged, metav1.GetOptions{}) - if !apierrs.IsNotFound(err) { + if !apierrors.IsNotFound(err) { // Privileged PSP was already created. ExpectNoError(err, "Failed to get PodSecurityPolicy %s", podSecurityPolicyPrivileged) return @@ -119,7 +119,7 @@ func CreatePrivilegedPSPBinding(kubeClient clientset.Interface, namespace string psp := privilegedPSP(podSecurityPolicyPrivileged) _, err = kubeClient.PolicyV1beta1().PodSecurityPolicies().Create(psp) - if !apierrs.IsAlreadyExists(err) { + if !apierrors.IsAlreadyExists(err) { ExpectNoError(err, "Failed to create PSP %s", podSecurityPolicyPrivileged) } @@ -134,7 +134,7 @@ func CreatePrivilegedPSPBinding(kubeClient clientset.Interface, namespace string Verbs: []string{"use"}, }}, }) - if !apierrs.IsAlreadyExists(err) { + if !apierrors.IsAlreadyExists(err) { ExpectNoError(err, "Failed to create PSP role") } } diff --git a/test/e2e/framework/pv/pv.go b/test/e2e/framework/pv/pv.go index 109edf52ee7..901cc8576ba 100644 --- a/test/e2e/framework/pv/pv.go +++ b/test/e2e/framework/pv/pv.go @@ -22,7 +22,7 @@ import ( "github.com/onsi/ginkgo" v1 "k8s.io/api/core/v1" - apierrs "k8s.io/apimachinery/pkg/api/errors" + apierrors "k8s.io/apimachinery/pkg/api/errors" "k8s.io/apimachinery/pkg/api/resource" metav1 "k8s.io/apimachinery/pkg/apis/meta/v1" "k8s.io/apimachinery/pkg/labels" @@ -186,7 +186,7 @@ func DeletePersistentVolume(c clientset.Interface, pvName string) error { if c != nil && len(pvName) > 0 { framework.Logf("Deleting PersistentVolume %q", pvName) err := c.CoreV1().PersistentVolumes().Delete(pvName, nil) - if err != nil && !apierrs.IsNotFound(err) { + if err != nil && !apierrors.IsNotFound(err) { return fmt.Errorf("PV Delete API error: %v", err) } } @@ -198,7 +198,7 @@ func DeletePersistentVolumeClaim(c clientset.Interface, pvcName string, ns strin if c != nil && len(pvcName) > 0 { framework.Logf("Deleting PersistentVolumeClaim %q", pvcName) err := c.CoreV1().PersistentVolumeClaims(ns).Delete(pvcName, nil) - if err != nil && !apierrs.IsNotFound(err) { + if err != nil && !apierrors.IsNotFound(err) { return fmt.Errorf("PVC Delete API error: %v", err) } } @@ -275,10 +275,10 @@ func DeletePVCandValidatePVGroup(c clientset.Interface, ns string, pvols PVMap, if err = DeletePVCandValidatePV(c, ns, pvc, pv, expectPVPhase); err != nil { return err } - } else if !apierrs.IsNotFound(err) { + } else if !apierrors.IsNotFound(err) { return fmt.Errorf("PVC Get API error: %v", err) } - // delete pvckey from map even if apierrs.IsNotFound above is true and thus the + // delete pvckey from map even if apierrors.IsNotFound above is true and thus the // claim was not actually deleted here delete(claims, pvcKey) deletedPVCs++ diff --git a/test/e2e/framework/skip.go b/test/e2e/framework/skip.go index 4608f4c64f2..e25c6afe684 100644 --- a/test/e2e/framework/skip.go +++ b/test/e2e/framework/skip.go @@ -19,7 +19,7 @@ package framework import ( "fmt" - apierrs "k8s.io/apimachinery/pkg/api/errors" + apierrors "k8s.io/apimachinery/pkg/api/errors" metav1 "k8s.io/apimachinery/pkg/apis/meta/v1" "k8s.io/apimachinery/pkg/runtime/schema" utilversion "k8s.io/apimachinery/pkg/util/version" @@ -65,7 +65,7 @@ func SkipIfMissingResource(dynamicClient dynamic.Interface, gvr schema.GroupVers _, err := resourceClient.List(metav1.ListOptions{}) if err != nil { // not all resources support list, so we ignore those - if apierrs.IsMethodNotSupported(err) || apierrs.IsNotFound(err) || apierrs.IsForbidden(err) { + if apierrors.IsMethodNotSupported(err) || apierrors.IsNotFound(err) || apierrors.IsForbidden(err) { skipInternalf(1, "Could not find %s resource, skipping test: %#v", gvr, err) } Failf("Unexpected error getting %v: %v", gvr, err) diff --git a/test/e2e/framework/statefulset/rest.go b/test/e2e/framework/statefulset/rest.go index cec76eb88bc..9d18f85b03d 100644 --- a/test/e2e/framework/statefulset/rest.go +++ b/test/e2e/framework/statefulset/rest.go @@ -24,7 +24,7 @@ import ( appsv1 "k8s.io/api/apps/v1" v1 "k8s.io/api/core/v1" - apierrs "k8s.io/apimachinery/pkg/api/errors" + apierrors "k8s.io/apimachinery/pkg/api/errors" metav1 "k8s.io/apimachinery/pkg/apis/meta/v1" "k8s.io/apimachinery/pkg/labels" "k8s.io/apimachinery/pkg/util/sets" @@ -256,7 +256,7 @@ func update(c clientset.Interface, ns, name string, update func(ss *appsv1.State if err == nil { return ss } - if !apierrs.IsConflict(err) && !apierrs.IsServerTimeout(err) { + if !apierrors.IsConflict(err) && !apierrors.IsServerTimeout(err) { e2efwk.Failf("failed to update statefulset %q: %v", name, err) } } diff --git a/test/e2e/framework/util.go b/test/e2e/framework/util.go index ef919e7d740..30418b98971 100644 --- a/test/e2e/framework/util.go +++ b/test/e2e/framework/util.go @@ -46,7 +46,7 @@ import ( appsv1 "k8s.io/api/apps/v1" v1 "k8s.io/api/core/v1" - apierrs "k8s.io/apimachinery/pkg/api/errors" + apierrors "k8s.io/apimachinery/pkg/api/errors" metav1 "k8s.io/apimachinery/pkg/apis/meta/v1" "k8s.io/apimachinery/pkg/fields" "k8s.io/apimachinery/pkg/labels" @@ -346,7 +346,7 @@ func WaitForPersistentVolumeDeleted(c clientset.Interface, pvName string, Poll, Logf("PersistentVolume %s found and phase=%s (%v)", pvName, pv.Status.Phase, time.Since(start)) continue } - if apierrs.IsNotFound(err) { + if apierrors.IsNotFound(err) { Logf("PersistentVolume %s was removed", pvName) return nil } @@ -365,7 +365,7 @@ func findAvailableNamespaceName(baseName string, c clientset.Interface) (string, // Already taken return false, nil } - if apierrs.IsNotFound(err) { + if apierrors.IsNotFound(err) { return true, nil } Logf("Unexpected error while getting namespace: %v", err) @@ -470,7 +470,7 @@ func WaitForService(c clientset.Interface, namespace, name string, exist bool, i case err == nil: Logf("Service %s in namespace %s found.", name, namespace) return exist, nil - case apierrs.IsNotFound(err): + case apierrors.IsNotFound(err): Logf("Service %s in namespace %s disappeared.", name, namespace) return !exist, nil case !testutils.IsRetryableAPIError(err): @@ -1190,7 +1190,7 @@ func DeleteResourceAndWaitForGC(c clientset.Interface, kind schema.GroupKind, ns rtObject, err := e2eresource.GetRuntimeObjectForKind(c, kind, ns, name) if err != nil { - if apierrs.IsNotFound(err) { + if apierrors.IsNotFound(err) { Logf("%v %s not found: %v", kind, name, err) return nil } diff --git a/test/e2e/framework/volume/fixtures.go b/test/e2e/framework/volume/fixtures.go index 5f505670088..cf57ed2a1a0 100644 --- a/test/e2e/framework/volume/fixtures.go +++ b/test/e2e/framework/volume/fixtures.go @@ -46,7 +46,7 @@ import ( "time" v1 "k8s.io/api/core/v1" - apierrs "k8s.io/apimachinery/pkg/api/errors" + apierrors "k8s.io/apimachinery/pkg/api/errors" metav1 "k8s.io/apimachinery/pkg/apis/meta/v1" "k8s.io/apimachinery/pkg/labels" clientset "k8s.io/client-go/kubernetes" @@ -372,7 +372,7 @@ func StartVolumeServer(client clientset.Interface, config TestConfig) *v1.Pod { serverPod, err := podClient.Create(serverPod) // ok if the server pod already exists. TODO: make this controllable by callers if err != nil { - if apierrs.IsAlreadyExists(err) { + if apierrors.IsAlreadyExists(err) { framework.Logf("Ignore \"already-exists\" error, re-get pod...") ginkgo.By(fmt.Sprintf("re-getting the %q server pod", serverPodName)) serverPod, err = podClient.Get(serverPodName, metav1.GetOptions{}) diff --git a/test/e2e/kubectl/kubectl.go b/test/e2e/kubectl/kubectl.go index e3f0cfd88ca..fe0992e7f59 100644 --- a/test/e2e/kubectl/kubectl.go +++ b/test/e2e/kubectl/kubectl.go @@ -45,7 +45,7 @@ import ( v1 "k8s.io/api/core/v1" rbacv1 "k8s.io/api/rbac/v1" apiextensionsv1 "k8s.io/apiextensions-apiserver/pkg/apis/apiextensions/v1" - apierrs "k8s.io/apimachinery/pkg/api/errors" + apierrors "k8s.io/apimachinery/pkg/api/errors" "k8s.io/apimachinery/pkg/api/meta" "k8s.io/apimachinery/pkg/api/resource" metav1 "k8s.io/apimachinery/pkg/apis/meta/v1" @@ -1325,9 +1325,9 @@ metadata: framework.Logf("Get endpoints failed (interval %v): %v", framework.Poll, err) // if the error is API not found or could not find default credentials or TLS handshake timeout, try again - if apierrs.IsNotFound(err) || - apierrs.IsUnauthorized(err) || - apierrs.IsServerTimeout(err) { + if apierrors.IsNotFound(err) || + apierrors.IsUnauthorized(err) || + apierrors.IsServerTimeout(err) { err = nil } return false, err @@ -1969,7 +1969,7 @@ metadata: ginkgo.By("verifying the job " + jobName + " was deleted") _, err = c.BatchV1().Jobs(ns).Get(jobName, metav1.GetOptions{}) framework.ExpectError(err) - framework.ExpectEqual(apierrs.IsNotFound(err), true) + framework.ExpectEqual(apierrors.IsNotFound(err), true) }) }) diff --git a/test/e2e/lifecycle/bootstrap/util.go b/test/e2e/lifecycle/bootstrap/util.go index 661a18692de..f397b87bd49 100644 --- a/test/e2e/lifecycle/bootstrap/util.go +++ b/test/e2e/lifecycle/bootstrap/util.go @@ -22,8 +22,8 @@ import ( "errors" "time" - v1 "k8s.io/api/core/v1" - apierrs "k8s.io/apimachinery/pkg/api/errors" + "k8s.io/api/core/v1" + apierrors "k8s.io/apimachinery/pkg/api/errors" metav1 "k8s.io/apimachinery/pkg/apis/meta/v1" "k8s.io/apimachinery/pkg/util/wait" clientset "k8s.io/client-go/kubernetes" @@ -139,7 +139,7 @@ func WaitForBootstrapTokenSecretToDisappear(c clientset.Interface, tokenID strin return wait.Poll(framework.Poll, 1*time.Minute, func() (bool, error) { _, err := c.CoreV1().Secrets(metav1.NamespaceSystem).Get(bootstrapapi.BootstrapTokenSecretPrefix+tokenID, metav1.GetOptions{}) - if apierrs.IsNotFound(err) { + if apierrors.IsNotFound(err) { return true, nil } return false, nil @@ -150,7 +150,7 @@ func WaitForBootstrapTokenSecretToDisappear(c clientset.Interface, tokenID strin func WaitForBootstrapTokenSecretNotDisappear(c clientset.Interface, tokenID string, t time.Duration) error { err := wait.Poll(framework.Poll, t, func() (bool, error) { secret, err := c.CoreV1().Secrets(metav1.NamespaceSystem).Get(bootstrapapi.BootstrapTokenSecretPrefix+tokenID, metav1.GetOptions{}) - if apierrs.IsNotFound(err) { + if apierrors.IsNotFound(err) { return true, errors.New("secret not exists") } if secret != nil { diff --git a/test/e2e/node/runtimeclass.go b/test/e2e/node/runtimeclass.go index 0224bb3be9c..0a95b1f2842 100644 --- a/test/e2e/node/runtimeclass.go +++ b/test/e2e/node/runtimeclass.go @@ -21,7 +21,7 @@ import ( v1 "k8s.io/api/core/v1" nodev1beta1 "k8s.io/api/node/v1beta1" - apierrs "k8s.io/apimachinery/pkg/api/errors" + apierrors "k8s.io/apimachinery/pkg/api/errors" metav1 "k8s.io/apimachinery/pkg/apis/meta/v1" runtimeclasstest "k8s.io/kubernetes/pkg/kubelet/runtimeclass/testing" "k8s.io/kubernetes/test/e2e/framework" @@ -55,7 +55,7 @@ var _ = ginkgo.Describe("[sig-node] RuntimeClass", func() { } _, err = f.ClientSet.CoreV1().Pods(f.Namespace.Name).Create(pod) framework.ExpectError(err, "should be forbidden") - framework.ExpectEqual(apierrs.IsForbidden(err), true, "should be forbidden error") + framework.ExpectEqual(apierrors.IsForbidden(err), true, "should be forbidden error") }) ginkgo.It("should run a Pod requesting a RuntimeClass with scheduling [NodeFeature:RuntimeHandler] [Disruptive] ", func() { diff --git a/test/e2e/scheduling/priorities.go b/test/e2e/scheduling/priorities.go index aa290c17a4d..926b2137134 100644 --- a/test/e2e/scheduling/priorities.go +++ b/test/e2e/scheduling/priorities.go @@ -28,7 +28,7 @@ import ( _ "github.com/stretchr/testify/assert" v1 "k8s.io/api/core/v1" - apierrs "k8s.io/apimachinery/pkg/api/errors" + apierrors "k8s.io/apimachinery/pkg/api/errors" "k8s.io/apimachinery/pkg/api/resource" metav1 "k8s.io/apimachinery/pkg/apis/meta/v1" "k8s.io/apimachinery/pkg/util/uuid" @@ -84,7 +84,7 @@ func addOrUpdateAvoidPodOnNode(c clientset.Interface, nodeName string, avoidPods node.Annotations[v1.PreferAvoidPodsAnnotationKey] = string(taintsData) _, err = c.CoreV1().Nodes().Update(node) if err != nil { - if !apierrs.IsConflict(err) { + if !apierrors.IsConflict(err) { framework.ExpectNoError(err) } else { framework.Logf("Conflict when trying to add/update avoidPods %v to %v with error %v", avoidPods, nodeName, err) @@ -113,7 +113,7 @@ func removeAvoidPodsOffNode(c clientset.Interface, nodeName string) { delete(node.Annotations, v1.PreferAvoidPodsAnnotationKey) _, err = c.CoreV1().Nodes().Update(node) if err != nil { - if !apierrs.IsConflict(err) { + if !apierrors.IsConflict(err) { framework.ExpectNoError(err) } else { framework.Logf("Conflict when trying to remove avoidPods to %v", nodeName) diff --git a/test/e2e/storage/flexvolume.go b/test/e2e/storage/flexvolume.go index f047b00577c..98d6e3c0bff 100644 --- a/test/e2e/storage/flexvolume.go +++ b/test/e2e/storage/flexvolume.go @@ -25,7 +25,7 @@ import ( "github.com/onsi/ginkgo" v1 "k8s.io/api/core/v1" - apierrs "k8s.io/apimachinery/pkg/api/errors" + apierrors "k8s.io/apimachinery/pkg/api/errors" clientset "k8s.io/client-go/kubernetes" "k8s.io/kubernetes/test/e2e/framework" e2enode "k8s.io/kubernetes/test/e2e/framework/node" @@ -193,7 +193,7 @@ var _ = utils.SIGDescribe("Flexvolumes", func() { testFlexVolume(driverInstallAs, config, f) ginkgo.By("waiting for flex client pod to terminate") - if err := f.WaitForPodTerminated(config.Prefix+"-client", ""); !apierrs.IsNotFound(err) { + if err := f.WaitForPodTerminated(config.Prefix+"-client", ""); !apierrors.IsNotFound(err) { framework.ExpectNoError(err, "Failed to wait client pod terminated: %v", err) } @@ -213,7 +213,7 @@ var _ = utils.SIGDescribe("Flexvolumes", func() { testFlexVolume(driverInstallAs, config, f) ginkgo.By("waiting for flex client pod to terminate") - if err := f.WaitForPodTerminated(config.Prefix+"-client", ""); !apierrs.IsNotFound(err) { + if err := f.WaitForPodTerminated(config.Prefix+"-client", ""); !apierrors.IsNotFound(err) { framework.ExpectNoError(err, "Failed to wait client pod terminated: %v", err) } diff --git a/test/e2e/storage/pvc_protection.go b/test/e2e/storage/pvc_protection.go index a3364777248..211e2c7d874 100644 --- a/test/e2e/storage/pvc_protection.go +++ b/test/e2e/storage/pvc_protection.go @@ -22,8 +22,8 @@ import ( "fmt" "time" - v1 "k8s.io/api/core/v1" - apierrs "k8s.io/apimachinery/pkg/api/errors" + "k8s.io/api/core/v1" + apierrors "k8s.io/apimachinery/pkg/api/errors" metav1 "k8s.io/apimachinery/pkg/apis/meta/v1" clientset "k8s.io/client-go/kubernetes" "k8s.io/kubernetes/pkg/util/slice" @@ -41,7 +41,7 @@ func waitForPersistentVolumeClaimDeleted(c clientset.Interface, ns string, pvcNa for start := time.Now(); time.Since(start) < timeout; time.Sleep(Poll) { _, err := c.CoreV1().PersistentVolumeClaims(ns).Get(pvcName, metav1.GetOptions{}) if err != nil { - if apierrs.IsNotFound(err) { + if apierrors.IsNotFound(err) { framework.Logf("Claim %q in namespace %q doesn't exist in the system", pvcName, ns) return nil } diff --git a/test/e2e/storage/testsuites/base.go b/test/e2e/storage/testsuites/base.go index 7fceaf84620..59ea6b45ce4 100644 --- a/test/e2e/storage/testsuites/base.go +++ b/test/e2e/storage/testsuites/base.go @@ -30,11 +30,10 @@ import ( v1 "k8s.io/api/core/v1" storagev1 "k8s.io/api/storage/v1" - apierrs "k8s.io/apimachinery/pkg/api/errors" + apierrors "k8s.io/apimachinery/pkg/api/errors" "k8s.io/apimachinery/pkg/api/resource" metav1 "k8s.io/apimachinery/pkg/apis/meta/v1" "k8s.io/apimachinery/pkg/apis/meta/v1/unstructured" - apierrors "k8s.io/apimachinery/pkg/util/errors" utilerrors "k8s.io/apimachinery/pkg/util/errors" "k8s.io/apimachinery/pkg/util/sets" clientset "k8s.io/client-go/kubernetes" @@ -317,7 +316,7 @@ func (r *VolumeResource) CleanupResource() error { cleanUpErrs = append(cleanUpErrs, errors.Wrap(err, "Failed to delete Volume")) } } - return apierrors.NewAggregate(cleanUpErrs) + return utilerrors.NewAggregate(cleanUpErrs) } func createPVCPV( @@ -409,7 +408,7 @@ func isDelayedBinding(sc *storagev1.StorageClass) bool { // deleteStorageClass deletes the passed in StorageClass and catches errors other than "Not Found" func deleteStorageClass(cs clientset.Interface, className string) error { err := cs.StorageV1().StorageClasses().Delete(className, nil) - if err != nil && !apierrs.IsNotFound(err) { + if err != nil && !apierrors.IsNotFound(err) { return err } return nil diff --git a/test/e2e/storage/testsuites/provisioning.go b/test/e2e/storage/testsuites/provisioning.go index 8b74641d1bc..5a6d81b6358 100644 --- a/test/e2e/storage/testsuites/provisioning.go +++ b/test/e2e/storage/testsuites/provisioning.go @@ -25,7 +25,7 @@ import ( v1 "k8s.io/api/core/v1" storagev1 "k8s.io/api/storage/v1" - apierrs "k8s.io/apimachinery/pkg/api/errors" + apierrors "k8s.io/apimachinery/pkg/api/errors" "k8s.io/apimachinery/pkg/api/resource" metav1 "k8s.io/apimachinery/pkg/apis/meta/v1" "k8s.io/apimachinery/pkg/apis/meta/v1/unstructured" @@ -247,7 +247,7 @@ func (t StorageClassTest) TestDynamicProvisioning() *v1.PersistentVolume { _, err = client.StorageV1().StorageClasses().Create(class) // The "should provision storage with snapshot data source" test already has created the class. // TODO: make class creation optional and remove the IsAlreadyExists exception - framework.ExpectEqual(err == nil || apierrs.IsAlreadyExists(err), true) + framework.ExpectEqual(err == nil || apierrors.IsAlreadyExists(err), true) class, err = client.StorageV1().StorageClasses().Get(class.Name, metav1.GetOptions{}) framework.ExpectNoError(err) defer func() { @@ -263,7 +263,7 @@ func (t StorageClassTest) TestDynamicProvisioning() *v1.PersistentVolume { framework.Logf("deleting claim %q/%q", claim.Namespace, claim.Name) // typically this claim has already been deleted err = client.CoreV1().PersistentVolumeClaims(claim.Namespace).Delete(claim.Name, nil) - if err != nil && !apierrs.IsNotFound(err) { + if err != nil && !apierrors.IsNotFound(err) { framework.Failf("Error deleting claim %q. Error: %v", claim.Name, err) } }() @@ -670,13 +670,13 @@ func prepareSnapshotDataSourceForProvisioning( cleanupFunc := func() { framework.Logf("deleting snapshot %q/%q", snapshot.GetNamespace(), snapshot.GetName()) err = dynamicClient.Resource(snapshotGVR).Namespace(updatedClaim.Namespace).Delete(snapshot.GetName(), nil) - if err != nil && !apierrs.IsNotFound(err) { + if err != nil && !apierrors.IsNotFound(err) { framework.Failf("Error deleting snapshot %q. Error: %v", snapshot.GetName(), err) } framework.Logf("deleting initClaim %q/%q", updatedClaim.Namespace, updatedClaim.Name) err = client.CoreV1().PersistentVolumeClaims(updatedClaim.Namespace).Delete(updatedClaim.Name, nil) - if err != nil && !apierrs.IsNotFound(err) { + if err != nil && !apierrors.IsNotFound(err) { framework.Failf("Error deleting initClaim %q. Error: %v", updatedClaim.Name, err) } @@ -718,7 +718,7 @@ func preparePVCDataSourceForProvisioning( cleanupFunc := func() { framework.Logf("deleting source PVC %q/%q", sourcePVC.Namespace, sourcePVC.Name) err = client.CoreV1().PersistentVolumeClaims(sourcePVC.Namespace).Delete(sourcePVC.Name, nil) - if err != nil && !apierrs.IsNotFound(err) { + if err != nil && !apierrors.IsNotFound(err) { framework.Failf("Error deleting source PVC %q. Error: %v", sourcePVC.Name, err) } } diff --git a/test/e2e/storage/testsuites/snapshottable.go b/test/e2e/storage/testsuites/snapshottable.go index 2a1af497263..efdc61a780f 100644 --- a/test/e2e/storage/testsuites/snapshottable.go +++ b/test/e2e/storage/testsuites/snapshottable.go @@ -23,7 +23,7 @@ import ( "github.com/onsi/ginkgo" v1 "k8s.io/api/core/v1" - apierrs "k8s.io/apimachinery/pkg/api/errors" + apierrors "k8s.io/apimachinery/pkg/api/errors" metav1 "k8s.io/apimachinery/pkg/apis/meta/v1" "k8s.io/apimachinery/pkg/runtime/schema" "k8s.io/client-go/dynamic" @@ -143,7 +143,7 @@ func (s *snapshottableTestSuite) DefineTests(driver TestDriver, pattern testpatt framework.Logf("deleting claim %q/%q", pvc.Namespace, pvc.Name) // typically this claim has already been deleted err = cs.CoreV1().PersistentVolumeClaims(pvc.Namespace).Delete(pvc.Name, nil) - if err != nil && !apierrs.IsNotFound(err) { + if err != nil && !apierrors.IsNotFound(err) { framework.Failf("Error deleting claim %q. Error: %v", pvc.Name, err) } }() @@ -182,7 +182,7 @@ func (s *snapshottableTestSuite) DefineTests(driver TestDriver, pattern testpatt framework.Logf("deleting snapshot %q/%q", snapshot.GetNamespace(), snapshot.GetName()) // typically this snapshot has already been deleted err = dc.Resource(snapshotGVR).Namespace(snapshot.GetNamespace()).Delete(snapshot.GetName(), nil) - if err != nil && !apierrs.IsNotFound(err) { + if err != nil && !apierrors.IsNotFound(err) { framework.Failf("Error deleting snapshot %q. Error: %v", pvc.Name, err) } }() diff --git a/test/e2e/storage/utils/create.go b/test/e2e/storage/utils/create.go index 97625159b20..1d3651f4673 100644 --- a/test/e2e/storage/utils/create.go +++ b/test/e2e/storage/utils/create.go @@ -29,7 +29,7 @@ import ( rbacv1 "k8s.io/api/rbac/v1" storagev1 "k8s.io/api/storage/v1" storagev1beta1 "k8s.io/api/storage/v1beta1" - apierrs "k8s.io/apimachinery/pkg/api/errors" + apierrors "k8s.io/apimachinery/pkg/api/errors" metav1 "k8s.io/apimachinery/pkg/apis/meta/v1" "k8s.io/apimachinery/pkg/runtime" "k8s.io/apimachinery/pkg/runtime/schema" @@ -155,7 +155,7 @@ func CreateItems(f *framework.Framework, items ...interface{}) (func(), error) { // command line flags, because they would also start to apply // to non-namespaced items. for _, destructor := range destructors { - if err := destructor(); err != nil && !apierrs.IsNotFound(err) { + if err := destructor(); err != nil && !apierrors.IsNotFound(err) { framework.Logf("deleting failed: %s", err) } } diff --git a/test/e2e/storage/utils/utils.go b/test/e2e/storage/utils/utils.go index 7fb582976c9..7b7c69e5ba5 100644 --- a/test/e2e/storage/utils/utils.go +++ b/test/e2e/storage/utils/utils.go @@ -30,7 +30,7 @@ import ( v1 "k8s.io/api/core/v1" rbacv1 "k8s.io/api/rbac/v1" - apierrs "k8s.io/apimachinery/pkg/api/errors" + apierrors "k8s.io/apimachinery/pkg/api/errors" metav1 "k8s.io/apimachinery/pkg/apis/meta/v1" "k8s.io/apimachinery/pkg/util/sets" "k8s.io/apimachinery/pkg/util/wait" @@ -581,7 +581,7 @@ func PrivilegedTestPSPClusterRoleBinding(client clientset.Interface, roleBindingClient.Delete(binding.GetName(), &metav1.DeleteOptions{}) err := wait.Poll(2*time.Second, 2*time.Minute, func() (bool, error) { _, err := roleBindingClient.Get(binding.GetName(), metav1.GetOptions{}) - return apierrs.IsNotFound(err), nil + return apierrors.IsNotFound(err), nil }) framework.ExpectNoError(err, "Timed out waiting for deletion: %v", err) diff --git a/test/e2e/storage/volume_provisioning.go b/test/e2e/storage/volume_provisioning.go index c4f0e5f81c1..29f7ec8e325 100644 --- a/test/e2e/storage/volume_provisioning.go +++ b/test/e2e/storage/volume_provisioning.go @@ -33,7 +33,7 @@ import ( rbacv1 "k8s.io/api/rbac/v1" storagev1 "k8s.io/api/storage/v1" storagev1beta1 "k8s.io/api/storage/v1beta1" - apierrs "k8s.io/apimachinery/pkg/api/errors" + apierrors "k8s.io/apimachinery/pkg/api/errors" metav1 "k8s.io/apimachinery/pkg/apis/meta/v1" "k8s.io/apimachinery/pkg/runtime/schema" "k8s.io/apimachinery/pkg/types" @@ -810,7 +810,7 @@ var _ = utils.SIGDescribe("Dynamic Provisioning", func() { defer func() { framework.Logf("deleting claim %q/%q", claim.Namespace, claim.Name) err = c.CoreV1().PersistentVolumeClaims(claim.Namespace).Delete(claim.Name, nil) - if err != nil && !apierrs.IsNotFound(err) { + if err != nil && !apierrors.IsNotFound(err) { framework.Failf("Error deleting claim %q. Error: %v", claim.Name, err) } }() @@ -1039,7 +1039,7 @@ func waitForProvisionedVolumesDeleted(c clientset.Interface, scName string) ([]* // deleteStorageClass deletes the passed in StorageClass and catches errors other than "Not Found" func deleteStorageClass(c clientset.Interface, className string) { err := c.StorageV1().StorageClasses().Delete(className, nil) - if err != nil && !apierrs.IsNotFound(err) { + if err != nil && !apierrors.IsNotFound(err) { framework.ExpectNoError(err) } } diff --git a/test/e2e/storage/vsphere/pv_reclaimpolicy.go b/test/e2e/storage/vsphere/pv_reclaimpolicy.go index a63a48973ee..264024e9d8e 100644 --- a/test/e2e/storage/vsphere/pv_reclaimpolicy.go +++ b/test/e2e/storage/vsphere/pv_reclaimpolicy.go @@ -22,7 +22,7 @@ import ( "github.com/onsi/ginkgo" v1 "k8s.io/api/core/v1" - apierrs "k8s.io/apimachinery/pkg/api/errors" + apierrors "k8s.io/apimachinery/pkg/api/errors" metav1 "k8s.io/apimachinery/pkg/apis/meta/v1" clientset "k8s.io/client-go/kubernetes" "k8s.io/kubernetes/test/e2e/framework" @@ -246,7 +246,7 @@ func deletePVCAfterBind(c clientset.Interface, ns string, pvc *v1.PersistentVolu ginkgo.By("delete pvc") framework.ExpectNoError(e2epv.DeletePersistentVolumeClaim(c, pvc.Name, ns), "Failed to delete PVC ", pvc.Name) _, err = c.CoreV1().PersistentVolumeClaims(ns).Get(pvc.Name, metav1.GetOptions{}) - if !apierrs.IsNotFound(err) { + if !apierrors.IsNotFound(err) { framework.ExpectNoError(err) } } diff --git a/test/e2e/storage/vsphere/vsphere_statefulsets.go b/test/e2e/storage/vsphere/vsphere_statefulsets.go index 40ecb10276a..88c9e257c29 100644 --- a/test/e2e/storage/vsphere/vsphere_statefulsets.go +++ b/test/e2e/storage/vsphere/vsphere_statefulsets.go @@ -21,7 +21,7 @@ import ( "github.com/onsi/ginkgo" "github.com/onsi/gomega" - apierrs "k8s.io/apimachinery/pkg/api/errors" + apierrors "k8s.io/apimachinery/pkg/api/errors" metav1 "k8s.io/apimachinery/pkg/apis/meta/v1" clientset "k8s.io/client-go/kubernetes" "k8s.io/kubernetes/test/e2e/framework" @@ -112,7 +112,7 @@ var _ = utils.SIGDescribe("vsphere statefulset", func() { for _, sspod := range ssPodsBeforeScaleDown.Items { _, err := client.CoreV1().Pods(namespace).Get(sspod.Name, metav1.GetOptions{}) if err != nil { - framework.ExpectEqual(apierrs.IsNotFound(err), true) + framework.ExpectEqual(apierrors.IsNotFound(err), true) for _, volumespec := range sspod.Spec.Volumes { if volumespec.PersistentVolumeClaim != nil { vSpherediskPath := getvSphereVolumePathFromClaim(client, statefulset.Namespace, volumespec.PersistentVolumeClaim.ClaimName) diff --git a/test/utils/create_resources.go b/test/utils/create_resources.go index c40a0672264..21dd9801bfb 100644 --- a/test/utils/create_resources.go +++ b/test/utils/create_resources.go @@ -25,7 +25,7 @@ import ( apps "k8s.io/api/apps/v1" batch "k8s.io/api/batch/v1" "k8s.io/api/core/v1" - apierrs "k8s.io/apimachinery/pkg/api/errors" + apierrors "k8s.io/apimachinery/pkg/api/errors" utilnet "k8s.io/apimachinery/pkg/util/net" "k8s.io/apimachinery/pkg/util/wait" clientset "k8s.io/client-go/kubernetes" @@ -52,12 +52,12 @@ func RetryWithExponentialBackOff(fn wait.ConditionFunc) error { func IsRetryableAPIError(err error) bool { // These errors may indicate a transient error that we can retry in tests. - if apierrs.IsInternalError(err) || apierrs.IsTimeout(err) || apierrs.IsServerTimeout(err) || - apierrs.IsTooManyRequests(err) || utilnet.IsProbableEOF(err) || utilnet.IsConnectionReset(err) { + if apierrors.IsInternalError(err) || apierrors.IsTimeout(err) || apierrors.IsServerTimeout(err) || + apierrors.IsTooManyRequests(err) || utilnet.IsProbableEOF(err) || utilnet.IsConnectionReset(err) { return true } // If the error sends the Retry-After header, we respect it as an explicit confirmation we should retry. - if _, shouldRetry := apierrs.SuggestsClientDelay(err); shouldRetry { + if _, shouldRetry := apierrors.SuggestsClientDelay(err); shouldRetry { return true } return false @@ -69,7 +69,7 @@ func CreatePodWithRetries(c clientset.Interface, namespace string, obj *v1.Pod) } createFunc := func() (bool, error) { _, err := c.CoreV1().Pods(namespace).Create(obj) - if err == nil || apierrs.IsAlreadyExists(err) { + if err == nil || apierrors.IsAlreadyExists(err) { return true, nil } if IsRetryableAPIError(err) { @@ -86,7 +86,7 @@ func CreateRCWithRetries(c clientset.Interface, namespace string, obj *v1.Replic } createFunc := func() (bool, error) { _, err := c.CoreV1().ReplicationControllers(namespace).Create(obj) - if err == nil || apierrs.IsAlreadyExists(err) { + if err == nil || apierrors.IsAlreadyExists(err) { return true, nil } if IsRetryableAPIError(err) { @@ -103,7 +103,7 @@ func CreateReplicaSetWithRetries(c clientset.Interface, namespace string, obj *a } createFunc := func() (bool, error) { _, err := c.AppsV1().ReplicaSets(namespace).Create(obj) - if err == nil || apierrs.IsAlreadyExists(err) { + if err == nil || apierrors.IsAlreadyExists(err) { return true, nil } if IsRetryableAPIError(err) { @@ -120,7 +120,7 @@ func CreateDeploymentWithRetries(c clientset.Interface, namespace string, obj *a } createFunc := func() (bool, error) { _, err := c.AppsV1().Deployments(namespace).Create(obj) - if err == nil || apierrs.IsAlreadyExists(err) { + if err == nil || apierrors.IsAlreadyExists(err) { return true, nil } if IsRetryableAPIError(err) { @@ -137,7 +137,7 @@ func CreateDaemonSetWithRetries(c clientset.Interface, namespace string, obj *ap } createFunc := func() (bool, error) { _, err := c.AppsV1().DaemonSets(namespace).Create(obj) - if err == nil || apierrs.IsAlreadyExists(err) { + if err == nil || apierrors.IsAlreadyExists(err) { return true, nil } if IsRetryableAPIError(err) { @@ -154,7 +154,7 @@ func CreateJobWithRetries(c clientset.Interface, namespace string, obj *batch.Jo } createFunc := func() (bool, error) { _, err := c.BatchV1().Jobs(namespace).Create(obj) - if err == nil || apierrs.IsAlreadyExists(err) { + if err == nil || apierrors.IsAlreadyExists(err) { return true, nil } if IsRetryableAPIError(err) { @@ -171,7 +171,7 @@ func CreateSecretWithRetries(c clientset.Interface, namespace string, obj *v1.Se } createFunc := func() (bool, error) { _, err := c.CoreV1().Secrets(namespace).Create(obj) - if err == nil || apierrs.IsAlreadyExists(err) { + if err == nil || apierrors.IsAlreadyExists(err) { return true, nil } if IsRetryableAPIError(err) { @@ -188,7 +188,7 @@ func CreateConfigMapWithRetries(c clientset.Interface, namespace string, obj *v1 } createFunc := func() (bool, error) { _, err := c.CoreV1().ConfigMaps(namespace).Create(obj) - if err == nil || apierrs.IsAlreadyExists(err) { + if err == nil || apierrors.IsAlreadyExists(err) { return true, nil } if IsRetryableAPIError(err) { @@ -205,7 +205,7 @@ func CreateServiceWithRetries(c clientset.Interface, namespace string, obj *v1.S } createFunc := func() (bool, error) { _, err := c.CoreV1().Services(namespace).Create(obj) - if err == nil || apierrs.IsAlreadyExists(err) { + if err == nil || apierrors.IsAlreadyExists(err) { return true, nil } if IsRetryableAPIError(err) { @@ -222,7 +222,7 @@ func CreateResourceQuotaWithRetries(c clientset.Interface, namespace string, obj } createFunc := func() (bool, error) { _, err := c.CoreV1().ResourceQuotas(namespace).Create(obj) - if err == nil || apierrs.IsAlreadyExists(err) { + if err == nil || apierrors.IsAlreadyExists(err) { return true, nil } if IsRetryableAPIError(err) { @@ -239,7 +239,7 @@ func CreatePersistentVolumeWithRetries(c clientset.Interface, obj *v1.Persistent } createFunc := func() (bool, error) { _, err := c.CoreV1().PersistentVolumes().Create(obj) - if err == nil || apierrs.IsAlreadyExists(err) { + if err == nil || apierrors.IsAlreadyExists(err) { return true, nil } if IsRetryableAPIError(err) { @@ -256,7 +256,7 @@ func CreatePersistentVolumeClaimWithRetries(c clientset.Interface, namespace str } createFunc := func() (bool, error) { _, err := c.CoreV1().PersistentVolumeClaims(namespace).Create(obj) - if err == nil || apierrs.IsAlreadyExists(err) { + if err == nil || apierrors.IsAlreadyExists(err) { return true, nil } if IsRetryableAPIError(err) { diff --git a/test/utils/delete_resources.go b/test/utils/delete_resources.go index 58221f56383..1eb96cddaa6 100644 --- a/test/utils/delete_resources.go +++ b/test/utils/delete_resources.go @@ -21,7 +21,7 @@ package utils import ( "fmt" - apierrs "k8s.io/apimachinery/pkg/api/errors" + apierrors "k8s.io/apimachinery/pkg/api/errors" metav1 "k8s.io/apimachinery/pkg/apis/meta/v1" "k8s.io/apimachinery/pkg/runtime/schema" clientset "k8s.io/client-go/kubernetes" @@ -59,7 +59,7 @@ func deleteResource(c clientset.Interface, kind schema.GroupKind, namespace, nam func DeleteResourceWithRetries(c clientset.Interface, kind schema.GroupKind, namespace, name string, options *metav1.DeleteOptions) error { deleteFunc := func() (bool, error) { err := deleteResource(c, kind, namespace, name, options) - if err == nil || apierrs.IsNotFound(err) { + if err == nil || apierrors.IsNotFound(err) { return true, nil } if IsRetryableAPIError(err) { diff --git a/test/utils/density_utils.go b/test/utils/density_utils.go index 60a32d50c19..d1b583cc623 100644 --- a/test/utils/density_utils.go +++ b/test/utils/density_utils.go @@ -22,7 +22,7 @@ import ( "time" "k8s.io/api/core/v1" - apierrs "k8s.io/apimachinery/pkg/api/errors" + apierrors "k8s.io/apimachinery/pkg/api/errors" metav1 "k8s.io/apimachinery/pkg/apis/meta/v1" "k8s.io/apimachinery/pkg/types" clientset "k8s.io/client-go/kubernetes" @@ -44,7 +44,7 @@ func AddLabelsToNode(c clientset.Interface, nodeName string, labels map[string]s for attempt := 0; attempt < retries; attempt++ { _, err = c.CoreV1().Nodes().Patch(nodeName, types.MergePatchType, []byte(patch)) if err != nil { - if !apierrs.IsConflict(err) { + if !apierrors.IsConflict(err) { return err } } else { @@ -76,7 +76,7 @@ func RemoveLabelOffNode(c clientset.Interface, nodeName string, labelKeys []stri } _, err = c.CoreV1().Nodes().Update(node) if err != nil { - if !apierrs.IsConflict(err) { + if !apierrors.IsConflict(err) { return err } else { klog.V(2).Infof("Conflict when trying to remove a labels %v from %v", labelKeys, nodeName) diff --git a/test/utils/runners.go b/test/utils/runners.go index d5bee2dfdee..9a9d4a2f5f5 100644 --- a/test/utils/runners.go +++ b/test/utils/runners.go @@ -30,7 +30,7 @@ import ( v1 "k8s.io/api/core/v1" storagev1beta1 "k8s.io/api/storage/v1beta1" apiequality "k8s.io/apimachinery/pkg/api/equality" - apierrs "k8s.io/apimachinery/pkg/api/errors" + apierrors "k8s.io/apimachinery/pkg/api/errors" "k8s.io/apimachinery/pkg/api/resource" metav1 "k8s.io/apimachinery/pkg/apis/meta/v1" "k8s.io/apimachinery/pkg/fields" @@ -1085,10 +1085,10 @@ func (s *NodeAllocatableStrategy) createCSINode(nodeName string, client clientse } _, err := client.StorageV1beta1().CSINodes().Create(csiNode) - if apierrs.IsAlreadyExists(err) { + if apierrors.IsAlreadyExists(err) { // Something created CSINode instance after we checked it did not exist. // Make the caller to re-try PrepareDependentObjects by returning Conflict error - err = apierrs.NewConflict(storagev1beta1.Resource("csinodes"), nodeName, err) + err = apierrors.NewConflict(storagev1beta1.Resource("csinodes"), nodeName, err) } return err } @@ -1121,7 +1121,7 @@ func (s *NodeAllocatableStrategy) updateCSINode(csiNode *storagev1beta1.CSINode, func (s *NodeAllocatableStrategy) PrepareDependentObjects(node *v1.Node, client clientset.Interface) error { csiNode, err := client.StorageV1beta1().CSINodes().Get(node.Name, metav1.GetOptions{}) if err != nil { - if apierrs.IsNotFound(err) { + if apierrors.IsNotFound(err) { return s.createCSINode(node.Name, client) } return err @@ -1132,7 +1132,7 @@ func (s *NodeAllocatableStrategy) PrepareDependentObjects(node *v1.Node, client func (s *NodeAllocatableStrategy) CleanupDependentObjects(nodeName string, client clientset.Interface) error { csiNode, err := client.StorageV1beta1().CSINodes().Get(nodeName, metav1.GetOptions{}) if err != nil { - if apierrs.IsNotFound(err) { + if apierrors.IsNotFound(err) { return nil } return err @@ -1158,7 +1158,7 @@ func DoPrepareNode(client clientset.Interface, node *v1.Node, strategy PrepareNo if _, err = client.CoreV1().Nodes().Patch(node.Name, types.MergePatchType, []byte(patch)); err == nil { break } - if !apierrs.IsConflict(err) { + if !apierrors.IsConflict(err) { return fmt.Errorf("Error while applying patch %v to Node %v: %v", string(patch), node.Name, err) } time.Sleep(100 * time.Millisecond) @@ -1171,7 +1171,7 @@ func DoPrepareNode(client clientset.Interface, node *v1.Node, strategy PrepareNo if err = strategy.PrepareDependentObjects(node, client); err == nil { break } - if !apierrs.IsConflict(err) { + if !apierrors.IsConflict(err) { return fmt.Errorf("Error while preparing objects for node %s: %s", node.Name, err) } time.Sleep(100 * time.Millisecond) @@ -1196,7 +1196,7 @@ func DoCleanupNode(client clientset.Interface, nodeName string, strategy Prepare if _, err = client.CoreV1().Nodes().Update(updatedNode); err == nil { break } - if !apierrs.IsConflict(err) { + if !apierrors.IsConflict(err) { return fmt.Errorf("Error when updating Node %v: %v", nodeName, err) } time.Sleep(100 * time.Millisecond) @@ -1210,7 +1210,7 @@ func DoCleanupNode(client clientset.Interface, nodeName string, strategy Prepare if err == nil { break } - if !apierrs.IsConflict(err) { + if !apierrors.IsConflict(err) { return fmt.Errorf("Error when cleaning up Node %v objects: %v", nodeName, err) } time.Sleep(100 * time.Millisecond)