mirror of
https://github.com/k3s-io/kubernetes.git
synced 2025-08-11 13:02:14 +00:00
Merge pull request #120403 from wlq1212/cheanup/notwork/expectnotequal
e2e_network:stop using deprecated framework.ExpectNotEqual
This commit is contained in:
commit
8eaaf2f9b9
@ -22,7 +22,6 @@ import (
|
||||
"strings"
|
||||
"time"
|
||||
|
||||
"github.com/onsi/ginkgo/v2"
|
||||
v1 "k8s.io/api/core/v1"
|
||||
metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
|
||||
"k8s.io/apimachinery/pkg/util/intstr"
|
||||
@ -36,6 +35,9 @@ import (
|
||||
"k8s.io/kubernetes/test/e2e/network/common"
|
||||
imageutils "k8s.io/kubernetes/test/utils/image"
|
||||
admissionapi "k8s.io/pod-security-admission/api"
|
||||
|
||||
"github.com/onsi/ginkgo/v2"
|
||||
"github.com/onsi/gomega"
|
||||
)
|
||||
|
||||
const (
|
||||
@ -113,7 +115,7 @@ var _ = common.SIGDescribe("Conntrack", func() {
|
||||
}
|
||||
|
||||
ips := e2enode.GetAddressesByTypeAndFamily(&nodes.Items[0], v1.NodeInternalIP, family)
|
||||
framework.ExpectNotEqual(len(ips), 0)
|
||||
gomega.Expect(ips).ToNot(gomega.BeEmpty())
|
||||
|
||||
clientNodeInfo = nodeInfo{
|
||||
name: nodes.Items[0].Name,
|
||||
@ -121,7 +123,7 @@ var _ = common.SIGDescribe("Conntrack", func() {
|
||||
}
|
||||
|
||||
ips = e2enode.GetAddressesByTypeAndFamily(&nodes.Items[1], v1.NodeInternalIP, family)
|
||||
framework.ExpectNotEqual(len(ips), 0)
|
||||
gomega.Expect(ips).ToNot(gomega.BeEmpty())
|
||||
|
||||
serverNodeInfo = nodeInfo{
|
||||
name: nodes.Items[1].Name,
|
||||
|
@ -786,7 +786,7 @@ var _ = common.SIGDescribe("Ingress API", func() {
|
||||
ginkgo.By("deleting")
|
||||
|
||||
expectFinalizer := func(ing *networkingv1.Ingress, msg string) {
|
||||
framework.ExpectNotEqual(ing.DeletionTimestamp, nil, fmt.Sprintf("expected deletionTimestamp, got nil on step: %q, ingress: %+v", msg, ing))
|
||||
gomega.Expect(ing.DeletionTimestamp).ToNot(gomega.BeNil(), "expected deletionTimestamp, got nil on step: %q, ingress: %+v", msg, ing)
|
||||
if len(ing.Finalizers) == 0 {
|
||||
framework.Failf("expected finalizers on ingress, got none on step: %q, ingress: %+v", msg, ing)
|
||||
}
|
||||
|
@ -40,6 +40,7 @@ import (
|
||||
netutils "k8s.io/utils/net"
|
||||
|
||||
"github.com/onsi/ginkgo/v2"
|
||||
"github.com/onsi/gomega"
|
||||
)
|
||||
|
||||
var kubeProxyE2eImage = imageutils.GetE2EImage(imageutils.Agnhost)
|
||||
@ -76,7 +77,7 @@ var _ = common.SIGDescribe("KubeProxy", func() {
|
||||
}
|
||||
|
||||
ips := e2enode.GetAddressesByTypeAndFamily(&nodes.Items[0], v1.NodeInternalIP, family)
|
||||
framework.ExpectNotEqual(len(ips), 0)
|
||||
gomega.Expect(ips).ToNot(gomega.BeEmpty())
|
||||
|
||||
clientNodeInfo := NodeInfo{
|
||||
node: &nodes.Items[0],
|
||||
@ -85,7 +86,7 @@ var _ = common.SIGDescribe("KubeProxy", func() {
|
||||
}
|
||||
|
||||
ips = e2enode.GetAddressesByTypeAndFamily(&nodes.Items[1], v1.NodeInternalIP, family)
|
||||
framework.ExpectNotEqual(len(ips), 0)
|
||||
gomega.Expect(ips).ToNot(gomega.BeEmpty())
|
||||
|
||||
serverNodeInfo := NodeInfo{
|
||||
node: &nodes.Items[1],
|
||||
|
@ -551,7 +551,7 @@ var _ = common.SIGDescribe("LoadBalancers", func() {
|
||||
acceptPod, err = cs.CoreV1().Pods(namespace).Get(ctx, acceptPod.Name, metav1.GetOptions{})
|
||||
framework.ExpectNoError(err, "Unable to get pod %s", acceptPod.Name)
|
||||
gomega.Expect(acceptPod.Status.Phase).To(gomega.Equal(v1.PodRunning))
|
||||
framework.ExpectNotEqual(acceptPod.Status.PodIP, "")
|
||||
gomega.Expect(acceptPod.Status.PodIP).ToNot(gomega.BeEmpty())
|
||||
|
||||
// Create loadbalancer service with source range from node[0] and podAccept
|
||||
svc, err := jig.CreateTCPService(ctx, func(svc *v1.Service) {
|
||||
@ -581,7 +581,7 @@ var _ = common.SIGDescribe("LoadBalancers", func() {
|
||||
dropPod, err = cs.CoreV1().Pods(namespace).Get(ctx, dropPod.Name, metav1.GetOptions{})
|
||||
framework.ExpectNoError(err, "Unable to get pod %s", dropPod.Name)
|
||||
gomega.Expect(acceptPod.Status.Phase).To(gomega.Equal(v1.PodRunning))
|
||||
framework.ExpectNotEqual(acceptPod.Status.PodIP, "")
|
||||
gomega.Expect(acceptPod.Status.PodIP).ToNot(gomega.BeEmpty())
|
||||
|
||||
ginkgo.By("Update service LoadBalancerSourceRange and check reachability")
|
||||
_, err = jig.UpdateService(ctx, func(svc *v1.Service) {
|
||||
@ -1602,8 +1602,8 @@ var _ = common.SIGDescribe("LoadBalancers ESIPP [Slow]", func() {
|
||||
noEndpointNodeMap[n.Name] = ips[0]
|
||||
}
|
||||
}
|
||||
framework.ExpectNotEqual(len(endpointNodeMap), 0)
|
||||
framework.ExpectNotEqual(len(noEndpointNodeMap), 0)
|
||||
gomega.Expect(endpointNodeMap).ToNot(gomega.BeEmpty())
|
||||
gomega.Expect(noEndpointNodeMap).ToNot(gomega.BeEmpty())
|
||||
|
||||
svcTCPPort := int(svc.Spec.Ports[0].Port)
|
||||
svcNodePort := int(svc.Spec.Ports[0].NodePort)
|
||||
|
@ -73,7 +73,7 @@ var _ = common.SIGDescribe("NoSNAT [Feature:NoSNAT] [Slow]", func() {
|
||||
ginkgo.By("creating a test pod on each Node")
|
||||
nodes, err := e2enode.GetReadySchedulableNodes(ctx, cs)
|
||||
framework.ExpectNoError(err)
|
||||
framework.ExpectNotEqual(len(nodes.Items), 0, "no Nodes in the cluster")
|
||||
gomega.Expect(nodes.Items).ToNot(gomega.BeEmpty(), "no Nodes in the cluster")
|
||||
|
||||
for _, node := range nodes.Items {
|
||||
// target Pod at Node
|
||||
|
@ -1401,7 +1401,7 @@ var _ = common.SIGDescribe("Services", func() {
|
||||
gomega.Expect(nodePortCounts).To(gomega.Equal(2), "updated service should have two Ports but found %d Ports", nodePortCounts)
|
||||
|
||||
for _, port := range nodePortService.Spec.Ports {
|
||||
framework.ExpectNotEqual(port.NodePort, 0, "NodePort service failed to allocate NodePort for Port %s", port.Name)
|
||||
gomega.Expect(port.NodePort).ToNot(gomega.BeZero(), "NodePort service failed to allocate NodePort for Port %s", port.Name)
|
||||
framework.Logf("NodePort service allocates NodePort: %d for Port: %s over Protocol: %s", port.NodePort, port.Name, port.Protocol)
|
||||
}
|
||||
})
|
||||
|
Loading…
Reference in New Issue
Block a user