diff --git a/test/e2e/windows/memory_limits.go b/test/e2e/windows/memory_limits.go index 7773e79b679..7d30dc4234d 100644 --- a/test/e2e/windows/memory_limits.go +++ b/test/e2e/windows/memory_limits.go @@ -168,7 +168,7 @@ func overrideAllocatableMemoryTest(ctx context.Context, f *framework.Framework, } } return false - }, 3*time.Minute, 10*time.Second).Should(gomega.Equal(true)) + }, 3*time.Minute, 10*time.Second).Should(gomega.BeTrue()) } diff --git a/test/e2e_node/hugepages_test.go b/test/e2e_node/hugepages_test.go index 29e74b526b4..6974832a876 100644 --- a/test/e2e_node/hugepages_test.go +++ b/test/e2e_node/hugepages_test.go @@ -227,7 +227,7 @@ var _ = SIGDescribe("HugePages [Serial] [Feature:HugePages][NodeSpecialFeature:H framework.ExpectNoError(err, "while getting node status") _, isPresent := node.Status.Capacity["hugepages-3Mi"] return isPresent - }, 30*time.Second, framework.Poll).Should(gomega.Equal(false)) + }, 30*time.Second, framework.Poll).Should(gomega.BeFalse()) }) ginkgo.It("should add resources for new huge page sizes on kubelet restart", func(ctx context.Context) { @@ -247,7 +247,7 @@ var _ = SIGDescribe("HugePages [Serial] [Feature:HugePages][NodeSpecialFeature:H framework.ExpectNoError(err, "while getting node status") _, isPresent := node.Status.Capacity["hugepages-2Mi"] return isPresent - }, 30*time.Second, framework.Poll).Should(gomega.Equal(true)) + }, 30*time.Second, framework.Poll).Should(gomega.BeTrue()) }) ginkgo.When("start the pod", func() { diff --git a/test/e2e_node/memory_manager_test.go b/test/e2e_node/memory_manager_test.go index 0e584d85ecf..464bfce6b0e 100644 --- a/test/e2e_node/memory_manager_test.go +++ b/test/e2e_node/memory_manager_test.go @@ -627,7 +627,7 @@ var _ = SIGDescribe("Memory Manager [Disruptive] [Serial] [Feature:MemoryManager return true }, time.Minute, 5*time.Second).Should( - gomega.Equal(true), + gomega.BeTrue(), "the pod succeeded to start, when it should fail with the admission error", ) })