fix: enable expected-actual rule from testifylint in module k8s.io/kubernetes

Signed-off-by: Matthieu MOREL <matthieu.morel35@gmail.com>
This commit is contained in:
Matthieu MOREL 2024-09-27 07:56:31 +02:00
parent a8c955ab42
commit f736cca0e5
20 changed files with 78 additions and 78 deletions

View File

@ -279,7 +279,7 @@ func TestAPIVersionOfDiscoveryEndpoints(t *testing.T) {
} }
apiVersions := metav1.APIVersions{} apiVersions := metav1.APIVersions{}
assert.NoError(decodeResponse(resp, &apiVersions)) assert.NoError(decodeResponse(resp, &apiVersions))
assert.Equal(apiVersions.APIVersion, "") assert.Equal("", apiVersions.APIVersion)
// /api/v1 exists in release-1.1 // /api/v1 exists in release-1.1
resp, err = http.Get(server.URL + "/api/v1") resp, err = http.Get(server.URL + "/api/v1")
@ -288,7 +288,7 @@ func TestAPIVersionOfDiscoveryEndpoints(t *testing.T) {
} }
resourceList := metav1.APIResourceList{} resourceList := metav1.APIResourceList{}
assert.NoError(decodeResponse(resp, &resourceList)) assert.NoError(decodeResponse(resp, &resourceList))
assert.Equal(resourceList.APIVersion, "") assert.Equal("", resourceList.APIVersion)
// /apis exists in release-1.1 // /apis exists in release-1.1
resp, err = http.Get(server.URL + "/apis") resp, err = http.Get(server.URL + "/apis")
@ -297,7 +297,7 @@ func TestAPIVersionOfDiscoveryEndpoints(t *testing.T) {
} }
groupList := metav1.APIGroupList{} groupList := metav1.APIGroupList{}
assert.NoError(decodeResponse(resp, &groupList)) assert.NoError(decodeResponse(resp, &groupList))
assert.Equal(groupList.APIVersion, "") assert.Equal("", groupList.APIVersion)
// /apis/autoscaling doesn't exist in release-1.1, so the APIVersion field // /apis/autoscaling doesn't exist in release-1.1, so the APIVersion field
// should be non-empty in the results returned by the server. // should be non-empty in the results returned by the server.
@ -307,7 +307,7 @@ func TestAPIVersionOfDiscoveryEndpoints(t *testing.T) {
} }
group := metav1.APIGroup{} group := metav1.APIGroup{}
assert.NoError(decodeResponse(resp, &group)) assert.NoError(decodeResponse(resp, &group))
assert.Equal(group.APIVersion, "v1") assert.Equal("v1", group.APIVersion)
// apis/autoscaling/v1 doesn't exist in release-1.1, so the APIVersion field // apis/autoscaling/v1 doesn't exist in release-1.1, so the APIVersion field
// should be non-empty in the results returned by the server. // should be non-empty in the results returned by the server.
@ -318,7 +318,7 @@ func TestAPIVersionOfDiscoveryEndpoints(t *testing.T) {
} }
resourceList = metav1.APIResourceList{} resourceList = metav1.APIResourceList{}
assert.NoError(decodeResponse(resp, &resourceList)) assert.NoError(decodeResponse(resp, &resourceList))
assert.Equal(resourceList.APIVersion, "v1") assert.Equal("v1", resourceList.APIVersion)
} }

View File

@ -111,6 +111,6 @@ func TestAdmissionOptionsAddFlags(t *testing.T) {
} }
// using assert because cannot compare neither pointer nor function of underlying GenericAdmission // using assert because cannot compare neither pointer nor function of underlying GenericAdmission
assert.Equal(t, opts.GenericAdmission.ConfigFile, "admission_control_config.yaml") assert.Equal(t, "admission_control_config.yaml", opts.GenericAdmission.ConfigFile)
assert.Equal(t, opts.GenericAdmission.EnablePlugins, []string{"foo", "bar", "baz"}) assert.Equal(t, []string{"foo", "bar", "baz"}, opts.GenericAdmission.EnablePlugins)
} }

View File

@ -24,7 +24,7 @@ import (
"github.com/stretchr/testify/assert" "github.com/stretchr/testify/assert"
"k8s.io/kubernetes/pkg/kubelet/checkpointmanager/checksum" "k8s.io/kubernetes/pkg/kubelet/checkpointmanager/checksum"
utilstore "k8s.io/kubernetes/pkg/kubelet/checkpointmanager/testing" utilstore "k8s.io/kubernetes/pkg/kubelet/checkpointmanager/testing"
"k8s.io/kubernetes/pkg/kubelet/checkpointmanager/testing/example_checkpoint_formats/v1" v1 "k8s.io/kubernetes/pkg/kubelet/checkpointmanager/testing/example_checkpoint_formats/v1"
) )
var testStore *utilstore.MemStore var testStore *utilstore.MemStore
@ -199,8 +199,8 @@ func TestCheckpointManager(t *testing.T) {
assert.NoError(t, err) assert.NoError(t, err)
actualPortMappings, actualHostNetwork := checkpointOut.GetData() actualPortMappings, actualHostNetwork := checkpointOut.GetData()
expPortMappings, expHostNetwork := tc.checkpoint.GetData() expPortMappings, expHostNetwork := tc.checkpoint.GetData()
assert.Equal(t, actualPortMappings, expPortMappings) assert.Equal(t, expPortMappings, actualPortMappings)
assert.Equal(t, actualHostNetwork, expHostNetwork) assert.Equal(t, expHostNetwork, actualHostNetwork)
} }
// Test it fails if tried to read V1 structure into V2, a different structure from the structure which is checkpointed // Test it fails if tried to read V1 structure into V2, a different structure from the structure which is checkpointed
checkpointV2 := newFakeCheckpointV2("", nil, false) checkpointV2 := newFakeCheckpointV2("", nil, false)
@ -229,7 +229,7 @@ func TestCheckpointManager(t *testing.T) {
keys, err := manager.ListCheckpoints() keys, err := manager.ListCheckpoints()
assert.NoError(t, err) assert.NoError(t, err)
sort.Strings(keys) sort.Strings(keys)
assert.Equal(t, keys, []string{"key1", "key2"}) assert.Equal(t, []string{"key1", "key2"}, keys)
// Test RemoveCheckpoints // Test RemoveCheckpoints
err = manager.RemoveCheckpoint("key1") err = manager.RemoveCheckpoint("key1")
@ -241,7 +241,7 @@ func TestCheckpointManager(t *testing.T) {
// Test ListCheckpoints // Test ListCheckpoints
keys, err = manager.ListCheckpoints() keys, err = manager.ListCheckpoints()
assert.NoError(t, err) assert.NoError(t, err)
assert.Equal(t, keys, []string{"key2"}) assert.Equal(t, []string{"key2"}, keys)
// Test Get NonExisted Checkpoint // Test Get NonExisted Checkpoint
checkpointNE := newFakeCheckpointV1("NE", nil, false) checkpointNE := newFakeCheckpointV1("NE", nil, false)

View File

@ -144,7 +144,7 @@ func TestRun(t *testing.T) {
// Wait for the second callback to be issued. // Wait for the second callback to be issued.
<-callbackChan <-callbackChan
require.Equal(t, callbackCount, 2) require.Equal(t, 2, callbackCount)
} }
func TestAllocate(t *testing.T) { func TestAllocate(t *testing.T) {

View File

@ -905,7 +905,7 @@ func TestHasWindowsHostProcessContainer(t *testing.T) {
pod := &v1.Pod{} pod := &v1.Pod{}
pod.Spec = *testCase.podSpec pod.Spec = *testCase.podSpec
result := HasWindowsHostProcessContainer(pod) result := HasWindowsHostProcessContainer(pod)
assert.Equal(t, result, testCase.expectedResult) assert.Equal(t, testCase.expectedResult, result)
}) })
} }
} }

View File

@ -154,7 +154,7 @@ func TestDetectImagesInitialDetect(t *testing.T) {
_, err := manager.detectImages(ctx, zero) _, err := manager.detectImages(ctx, zero)
assert := assert.New(t) assert := assert.New(t)
require.NoError(t, err) require.NoError(t, err)
assert.Equal(manager.imageRecordsLen(), 3) assert.Equal(3, manager.imageRecordsLen())
noContainer, ok := manager.getImageRecord(imageID(0)) noContainer, ok := manager.getImageRecord(imageID(0))
require.True(t, ok) require.True(t, ok)
assert.Equal(zero, noContainer.firstDetected) assert.Equal(zero, noContainer.firstDetected)
@ -207,7 +207,7 @@ func TestDetectImagesInitialDetectWithRuntimeHandlerInImageCriAPIFeatureGate(t *
_, err := manager.detectImages(ctx, zero) _, err := manager.detectImages(ctx, zero)
assert := assert.New(t) assert := assert.New(t)
require.NoError(t, err) require.NoError(t, err)
assert.Equal(manager.imageRecordsLen(), 3) assert.Equal(3, manager.imageRecordsLen())
noContainer, ok := manager.getImageRecordWithRuntimeHandlerInImageCriAPIFeatureGate(imageID(0), testRuntimeHandler) noContainer, ok := manager.getImageRecordWithRuntimeHandlerInImageCriAPIFeatureGate(imageID(0), testRuntimeHandler)
require.True(t, ok) require.True(t, ok)
assert.Equal(zero, noContainer.firstDetected) assert.Equal(zero, noContainer.firstDetected)
@ -246,7 +246,7 @@ func TestDetectImagesWithNewImage(t *testing.T) {
_, err := manager.detectImages(ctx, zero) _, err := manager.detectImages(ctx, zero)
assert := assert.New(t) assert := assert.New(t)
require.NoError(t, err) require.NoError(t, err)
assert.Equal(manager.imageRecordsLen(), 2) assert.Equal(2, manager.imageRecordsLen())
// Add a new image. // Add a new image.
fakeRuntime.ImageList = []container.Image{ fakeRuntime.ImageList = []container.Image{
@ -259,7 +259,7 @@ func TestDetectImagesWithNewImage(t *testing.T) {
startTime := time.Now().Add(-time.Millisecond) startTime := time.Now().Add(-time.Millisecond)
_, err = manager.detectImages(ctx, detectedTime) _, err = manager.detectImages(ctx, detectedTime)
require.NoError(t, err) require.NoError(t, err)
assert.Equal(manager.imageRecordsLen(), 3) assert.Equal(3, manager.imageRecordsLen())
noContainer, ok := manager.getImageRecord(imageID(0)) noContainer, ok := manager.getImageRecord(imageID(0))
require.True(t, ok) require.True(t, ok)
assert.Equal(zero, noContainer.firstDetected) assert.Equal(zero, noContainer.firstDetected)
@ -403,7 +403,7 @@ func TestDetectImagesContainerStopped(t *testing.T) {
_, err := manager.detectImages(ctx, zero) _, err := manager.detectImages(ctx, zero)
assert := assert.New(t) assert := assert.New(t)
require.NoError(t, err) require.NoError(t, err)
assert.Equal(manager.imageRecordsLen(), 2) assert.Equal(2, manager.imageRecordsLen())
withContainer, ok := manager.getImageRecord(imageID(1)) withContainer, ok := manager.getImageRecord(imageID(1))
require.True(t, ok) require.True(t, ok)
@ -411,7 +411,7 @@ func TestDetectImagesContainerStopped(t *testing.T) {
fakeRuntime.AllPodList = []*containertest.FakePod{} fakeRuntime.AllPodList = []*containertest.FakePod{}
_, err = manager.detectImages(ctx, time.Now()) _, err = manager.detectImages(ctx, time.Now())
require.NoError(t, err) require.NoError(t, err)
assert.Equal(manager.imageRecordsLen(), 2) assert.Equal(2, manager.imageRecordsLen())
container1, ok := manager.getImageRecord(imageID(0)) container1, ok := manager.getImageRecord(imageID(0))
require.True(t, ok) require.True(t, ok)
assert.Equal(zero, container1.firstDetected) assert.Equal(zero, container1.firstDetected)
@ -442,13 +442,13 @@ func TestDetectImagesWithRemovedImages(t *testing.T) {
_, err := manager.detectImages(ctx, zero) _, err := manager.detectImages(ctx, zero)
assert := assert.New(t) assert := assert.New(t)
require.NoError(t, err) require.NoError(t, err)
assert.Equal(manager.imageRecordsLen(), 2) assert.Equal(2, manager.imageRecordsLen())
// Simulate both images being removed. // Simulate both images being removed.
fakeRuntime.ImageList = []container.Image{} fakeRuntime.ImageList = []container.Image{}
_, err = manager.detectImages(ctx, time.Now()) _, err = manager.detectImages(ctx, time.Now())
require.NoError(t, err) require.NoError(t, err)
assert.Equal(manager.imageRecordsLen(), 0) assert.Equal(0, manager.imageRecordsLen())
} }
func TestFreeSpaceImagesInUseContainersAreIgnored(t *testing.T) { func TestFreeSpaceImagesInUseContainersAreIgnored(t *testing.T) {
@ -573,7 +573,7 @@ func TestFreeSpaceRemoveByLeastRecentlyUsed(t *testing.T) {
} }
_, err = manager.detectImages(ctx, time.Now()) _, err = manager.detectImages(ctx, time.Now())
require.NoError(t, err) require.NoError(t, err)
require.Equal(t, manager.imageRecordsLen(), 2) require.Equal(t, 2, manager.imageRecordsLen())
// We're setting the delete time one minute in the future, so the time the image // We're setting the delete time one minute in the future, so the time the image
// was first detected and the delete time are different. // was first detected and the delete time are different.
@ -609,7 +609,7 @@ func TestFreeSpaceTiesBrokenByDetectedTime(t *testing.T) {
fakeRuntime.AllPodList = []*containertest.FakePod{} fakeRuntime.AllPodList = []*containertest.FakePod{}
_, err = manager.detectImages(ctx, time.Now()) _, err = manager.detectImages(ctx, time.Now())
require.NoError(t, err) require.NoError(t, err)
require.Equal(t, manager.imageRecordsLen(), 2) require.Equal(t, 2, manager.imageRecordsLen())
assert := assert.New(t) assert := assert.New(t)
getImagesAndFreeSpace(ctx, t, assert, manager, fakeRuntime, 1024, 2048, 1, time.Now()) getImagesAndFreeSpace(ctx, t, assert, manager, fakeRuntime, 1024, 2048, 1, time.Now())
@ -726,7 +726,7 @@ func TestGarbageCollectImageNotOldEnough(t *testing.T) {
t.Log(fakeClock.Now()) t.Log(fakeClock.Now())
_, err := manager.detectImages(ctx, fakeClock.Now()) _, err := manager.detectImages(ctx, fakeClock.Now())
require.NoError(t, err) require.NoError(t, err)
require.Equal(t, manager.imageRecordsLen(), 2) require.Equal(t, 2, manager.imageRecordsLen())
// no space freed since one image is in used, and another one is not old enough // no space freed since one image is in used, and another one is not old enough
assert := assert.New(t) assert := assert.New(t)
getImagesAndFreeSpace(ctx, t, assert, manager, fakeRuntime, 1024, 0, 2, fakeClock.Now()) getImagesAndFreeSpace(ctx, t, assert, manager, fakeRuntime, 1024, 0, 2, fakeClock.Now())

View File

@ -635,7 +635,7 @@ func TestEvalCRIPullErr(t *testing.T) {
input: crierrors.ErrRegistryUnavailable, input: crierrors.ErrRegistryUnavailable,
assert: func(msg string, err error) { assert: func(msg string, err error) {
assert.ErrorIs(t, err, crierrors.ErrRegistryUnavailable) assert.ErrorIs(t, err, crierrors.ErrRegistryUnavailable)
assert.Equal(t, msg, "image pull failed for test because the registry is unavailable") assert.Equal(t, "image pull failed for test because the registry is unavailable", msg)
}, },
}, },
{ {
@ -643,7 +643,7 @@ func TestEvalCRIPullErr(t *testing.T) {
input: fmt.Errorf("%v: foo", crierrors.ErrRegistryUnavailable), input: fmt.Errorf("%v: foo", crierrors.ErrRegistryUnavailable),
assert: func(msg string, err error) { assert: func(msg string, err error) {
assert.ErrorIs(t, err, crierrors.ErrRegistryUnavailable) assert.ErrorIs(t, err, crierrors.ErrRegistryUnavailable)
assert.Equal(t, msg, "image pull failed for test because the registry is unavailable: foo") assert.Equal(t, "image pull failed for test because the registry is unavailable: foo", msg)
}, },
}, },
{ {
@ -651,7 +651,7 @@ func TestEvalCRIPullErr(t *testing.T) {
input: crierrors.ErrSignatureValidationFailed, input: crierrors.ErrSignatureValidationFailed,
assert: func(msg string, err error) { assert: func(msg string, err error) {
assert.ErrorIs(t, err, crierrors.ErrSignatureValidationFailed) assert.ErrorIs(t, err, crierrors.ErrSignatureValidationFailed)
assert.Equal(t, msg, "image pull failed for test because the signature validation failed") assert.Equal(t, "image pull failed for test because the signature validation failed", msg)
}, },
}, },
{ {
@ -659,7 +659,7 @@ func TestEvalCRIPullErr(t *testing.T) {
input: fmt.Errorf("%w: bar", crierrors.ErrSignatureValidationFailed), input: fmt.Errorf("%w: bar", crierrors.ErrSignatureValidationFailed),
assert: func(msg string, err error) { assert: func(msg string, err error) {
assert.ErrorIs(t, err, crierrors.ErrSignatureValidationFailed) assert.ErrorIs(t, err, crierrors.ErrSignatureValidationFailed)
assert.Equal(t, msg, "image pull failed for test because the signature validation failed: bar") assert.Equal(t, "image pull failed for test because the signature validation failed: bar", msg)
}, },
}, },
} { } {

View File

@ -323,7 +323,7 @@ func TestUpdateNewNodeStatus(t *testing.T) {
actions := kubeClient.Actions() actions := kubeClient.Actions()
require.Len(t, actions, 2) require.Len(t, actions, 2)
require.True(t, actions[1].Matches("patch", "nodes")) require.True(t, actions[1].Matches("patch", "nodes"))
require.Equal(t, actions[1].GetSubresource(), "status") require.Equal(t, "status", actions[1].GetSubresource())
updatedNode, err := applyNodeStatusPatch(&existingNode, actions[1].(core.PatchActionImpl).GetPatch()) updatedNode, err := applyNodeStatusPatch(&existingNode, actions[1].(core.PatchActionImpl).GetPatch())
assert.NoError(t, err) assert.NoError(t, err)
@ -719,7 +719,7 @@ func TestUpdateNodeStatusWithRuntimeStateError(t *testing.T) {
actions := kubeClient.Actions() actions := kubeClient.Actions()
require.Len(t, actions, 2) require.Len(t, actions, 2)
require.True(t, actions[1].Matches("patch", "nodes")) require.True(t, actions[1].Matches("patch", "nodes"))
require.Equal(t, actions[1].GetSubresource(), "status") require.Equal(t, "status", actions[1].GetSubresource())
updatedNode, err := kubeClient.CoreV1().Nodes().Get(ctx, testKubeletHostname, metav1.GetOptions{}) updatedNode, err := kubeClient.CoreV1().Nodes().Get(ctx, testKubeletHostname, metav1.GetOptions{})
require.NoError(t, err, "can't apply node status patch") require.NoError(t, err, "can't apply node status patch")
@ -1642,7 +1642,7 @@ func TestUpdateNewNodeStatusTooLargeReservation(t *testing.T) {
actions := kubeClient.Actions() actions := kubeClient.Actions()
require.Len(t, actions, 1) require.Len(t, actions, 1)
require.True(t, actions[0].Matches("patch", "nodes")) require.True(t, actions[0].Matches("patch", "nodes"))
require.Equal(t, actions[0].GetSubresource(), "status") require.Equal(t, "status", actions[0].GetSubresource())
updatedNode, err := applyNodeStatusPatch(&existingNode, actions[0].(core.PatchActionImpl).GetPatch()) updatedNode, err := applyNodeStatusPatch(&existingNode, actions[0].(core.PatchActionImpl).GetPatch())
assert.NoError(t, err) assert.NoError(t, err)

View File

@ -6115,7 +6115,7 @@ func TestGetNonExistentImagePullSecret(t *testing.T) {
assert.Len(t, fakeRecorder.Events, 1) assert.Len(t, fakeRecorder.Events, 1)
event := <-fakeRecorder.Events event := <-fakeRecorder.Events
assert.Equal(t, event, expectedEvent) assert.Equal(t, expectedEvent, event)
} }
func TestParseGetSubIdsOutput(t *testing.T) { func TestParseGetSubIdsOutput(t *testing.T) {

View File

@ -44,7 +44,7 @@ func TestPullImage(t *testing.T) {
images, err := fakeManager.ListImages(ctx) images, err := fakeManager.ListImages(ctx)
assert.NoError(t, err) assert.NoError(t, err)
assert.Len(t, images, 1) assert.Len(t, images, 1)
assert.Equal(t, images[0].RepoTags, []string{"busybox"}) assert.Equal(t, []string{"busybox"}, images[0].RepoTags)
} }
func TestPullImageWithError(t *testing.T) { func TestPullImageWithError(t *testing.T) {

View File

@ -91,7 +91,7 @@ func TestCreatePodSandbox(t *testing.T) {
assert.NoError(t, err) assert.NoError(t, err)
assert.Len(t, sandboxes, 1) assert.Len(t, sandboxes, 1)
assert.Equal(t, sandboxes[0].Id, fmt.Sprintf("%s_%s_%s_1", pod.Name, pod.Namespace, pod.UID)) assert.Equal(t, sandboxes[0].Id, fmt.Sprintf("%s_%s_%s_1", pod.Name, pod.Namespace, pod.UID))
assert.Equal(t, sandboxes[0].State, runtimeapi.PodSandboxState_SANDBOX_READY) assert.Equal(t, runtimeapi.PodSandboxState_SANDBOX_READY, sandboxes[0].State)
} }
func TestGeneratePodSandboxLinuxConfigSeccomp(t *testing.T) { func TestGeneratePodSandboxLinuxConfigSeccomp(t *testing.T) {

View File

@ -32,8 +32,8 @@ func TestGetStat(t *testing.T) {
fi, err := getStat(event) fi, err := getStat(event)
fiExpected, errExpected := os.Stat(event.Name) fiExpected, errExpected := os.Stat(event.Name)
assert.Equal(t, fi, fiExpected) assert.Equal(t, fiExpected, fi)
assert.Equal(t, err, errExpected) assert.Equal(t, errExpected, err)
} }
func TestGetSocketPath(t *testing.T) { func TestGetSocketPath(t *testing.T) {

View File

@ -918,9 +918,9 @@ func TestCheckpointContainer(t *testing.T) {
t.Errorf("Got error POSTing: %v", err) t.Errorf("Got error POSTing: %v", err)
} }
defer resp.Body.Close() defer resp.Body.Close()
assert.Equal(t, resp.StatusCode, 500) assert.Equal(t, 500, resp.StatusCode)
body, _ := io.ReadAll(resp.Body) body, _ := io.ReadAll(resp.Body)
assert.Equal(t, string(body), "checkpointing of other/foo/checkpointingFailure failed (Returning error for test)") assert.Equal(t, "checkpointing of other/foo/checkpointingFailure failed (Returning error for test)", string(body))
}) })
// Now test a successful checkpoint succeeds // Now test a successful checkpoint succeeds
setPodByNameFunc(fw, podNamespace, podName, expectedContainerName) setPodByNameFunc(fw, podNamespace, podName, expectedContainerName)
@ -929,7 +929,7 @@ func TestCheckpointContainer(t *testing.T) {
if err != nil { if err != nil {
t.Errorf("Got error POSTing: %v", err) t.Errorf("Got error POSTing: %v", err)
} }
assert.Equal(t, resp.StatusCode, 200) assert.Equal(t, 200, resp.StatusCode)
}) })
// Now test for 404 if checkpointing support is explicitly disabled. // Now test for 404 if checkpointing support is explicitly disabled.

View File

@ -93,14 +93,14 @@ func TestSummaryProviderGetStatsNoSplitFileSystem(t *testing.T) {
summary, err := provider.Get(ctx, true) summary, err := provider.Get(ctx, true)
assert.NoError(err) assert.NoError(err)
assert.Equal(summary.Node.NodeName, "test-node") assert.Equal("test-node", summary.Node.NodeName)
assert.Equal(summary.Node.StartTime, systemBootTime) assert.Equal(summary.Node.StartTime, systemBootTime)
assert.Equal(summary.Node.CPU, cgroupStatsMap["/"].cs.CPU) assert.Equal(summary.Node.CPU, cgroupStatsMap["/"].cs.CPU)
assert.Equal(summary.Node.Memory, cgroupStatsMap["/"].cs.Memory) assert.Equal(summary.Node.Memory, cgroupStatsMap["/"].cs.Memory)
assert.Equal(summary.Node.Swap, cgroupStatsMap["/"].cs.Swap) assert.Equal(summary.Node.Swap, cgroupStatsMap["/"].cs.Swap)
assert.Equal(summary.Node.Network, cgroupStatsMap["/"].ns) assert.Equal(summary.Node.Network, cgroupStatsMap["/"].ns)
assert.Equal(summary.Node.Fs, rootFsStats) assert.Equal(summary.Node.Fs, rootFsStats)
assert.Equal(summary.Node.Runtime, &statsapi.RuntimeStats{ContainerFs: imageFsStats, ImageFs: imageFsStats}) assert.Equal(&statsapi.RuntimeStats{ContainerFs: imageFsStats, ImageFs: imageFsStats}, summary.Node.Runtime)
assert.Len(summary.Node.SystemContainers, 4) assert.Len(summary.Node.SystemContainers, 4)
assert.Contains(summary.Node.SystemContainers, statsapi.ContainerStats{ assert.Contains(summary.Node.SystemContainers, statsapi.ContainerStats{
@ -189,7 +189,7 @@ func TestSummaryProviderGetStatsSplitImageFs(t *testing.T) {
summary, err := provider.Get(ctx, true) summary, err := provider.Get(ctx, true)
assert.NoError(err) assert.NoError(err)
assert.Equal(summary.Node.NodeName, "test-node") assert.Equal("test-node", summary.Node.NodeName)
assert.Equal(summary.Node.StartTime, systemBootTime) assert.Equal(summary.Node.StartTime, systemBootTime)
assert.Equal(summary.Node.CPU, cgroupStatsMap["/"].cs.CPU) assert.Equal(summary.Node.CPU, cgroupStatsMap["/"].cs.CPU)
assert.Equal(summary.Node.Memory, cgroupStatsMap["/"].cs.Memory) assert.Equal(summary.Node.Memory, cgroupStatsMap["/"].cs.Memory)
@ -197,7 +197,7 @@ func TestSummaryProviderGetStatsSplitImageFs(t *testing.T) {
assert.Equal(summary.Node.Network, cgroupStatsMap["/"].ns) assert.Equal(summary.Node.Network, cgroupStatsMap["/"].ns)
assert.Equal(summary.Node.Fs, rootFsStats) assert.Equal(summary.Node.Fs, rootFsStats)
// Since we are a split filesystem we want root filesystem to be container fs and image to be image filesystem // Since we are a split filesystem we want root filesystem to be container fs and image to be image filesystem
assert.Equal(summary.Node.Runtime, &statsapi.RuntimeStats{ContainerFs: rootFsStats, ImageFs: imageFsStats}) assert.Equal(&statsapi.RuntimeStats{ContainerFs: rootFsStats, ImageFs: imageFsStats}, summary.Node.Runtime)
assert.Len(summary.Node.SystemContainers, 4) assert.Len(summary.Node.SystemContainers, 4)
assert.Contains(summary.Node.SystemContainers, statsapi.ContainerStats{ assert.Contains(summary.Node.SystemContainers, statsapi.ContainerStats{
@ -276,7 +276,7 @@ func TestSummaryProviderGetCPUAndMemoryStats(t *testing.T) {
summary, err := provider.GetCPUAndMemoryStats(ctx) summary, err := provider.GetCPUAndMemoryStats(ctx)
assert.NoError(err) assert.NoError(err)
assert.Equal(summary.Node.NodeName, "test-node") assert.Equal("test-node", summary.Node.NodeName)
assert.Equal(summary.Node.StartTime, cgroupStatsMap["/"].cs.StartTime) assert.Equal(summary.Node.StartTime, cgroupStatsMap["/"].cs.StartTime)
assert.Equal(summary.Node.CPU, cgroupStatsMap["/"].cs.CPU) assert.Equal(summary.Node.CPU, cgroupStatsMap["/"].cs.CPU)
assert.Equal(summary.Node.Memory, cgroupStatsMap["/"].cs.Memory) assert.Equal(summary.Node.Memory, cgroupStatsMap["/"].cs.Memory)

View File

@ -615,7 +615,7 @@ func TestStaticPod(t *testing.T) {
assert.True(t, isPodStatusByKubeletEqual(&status, &retrievedStatus), "Expected: %+v, Got: %+v", status, retrievedStatus) assert.True(t, isPodStatusByKubeletEqual(&status, &retrievedStatus), "Expected: %+v, Got: %+v", status, retrievedStatus)
t.Logf("Should sync pod because the corresponding mirror pod is created") t.Logf("Should sync pod because the corresponding mirror pod is created")
assert.Equal(t, m.syncBatch(true), 1) assert.Equal(t, 1, m.syncBatch(true))
verifyActions(t, m, []core.Action{getAction(), patchAction()}) verifyActions(t, m, []core.Action{getAction(), patchAction()})
t.Logf("syncBatch should not sync any pods because nothing is changed.") t.Logf("syncBatch should not sync any pods because nothing is changed.")
@ -629,7 +629,7 @@ func TestStaticPod(t *testing.T) {
m.podManager.(mutablePodManager).AddPod(mirrorPod) m.podManager.(mutablePodManager).AddPod(mirrorPod)
t.Logf("Should not update to mirror pod, because UID has changed.") t.Logf("Should not update to mirror pod, because UID has changed.")
assert.Equal(t, m.syncBatch(true), 1) assert.Equal(t, 1, m.syncBatch(true))
verifyActions(t, m, []core.Action{getAction()}) verifyActions(t, m, []core.Action{getAction()})
} }

View File

@ -92,7 +92,7 @@ func testStore(t *testing.T, store Store) {
keys, err := store.List() keys, err := store.List()
assert.NoError(t, err) assert.NoError(t, err)
sort.Strings(keys) sort.Strings(keys)
assert.Equal(t, keys, []string{"id1", "id2"}) assert.Equal(t, []string{"id1", "id2"}, keys)
// Test Delete data // Test Delete data
for _, c := range testCases { for _, c := range testCases {

View File

@ -2397,12 +2397,12 @@ func TestReconcileWithUpdateReconstructedFromAPIServer(t *testing.T) {
if vol.VolumeName == volumeName1 { if vol.VolumeName == volumeName1 {
// devicePath + attachability must have been updated from node.status // devicePath + attachability must have been updated from node.status
assert.True(t, vol.PluginIsAttachable) assert.True(t, vol.PluginIsAttachable)
assert.Equal(t, vol.DevicePath, "fake/path") assert.Equal(t, "fake/path", vol.DevicePath)
} }
if vol.VolumeName == volumeName2 { if vol.VolumeName == volumeName2 {
// only attachability was updated from node.status // only attachability was updated from node.status
assert.False(t, vol.PluginIsAttachable) assert.False(t, vol.PluginIsAttachable)
assert.Equal(t, vol.DevicePath, "/dev/reconstructed") assert.Equal(t, "/dev/reconstructed", vol.DevicePath)
} }
} }
} }

View File

@ -4415,9 +4415,9 @@ func TestEndpointSliceE2E(t *testing.T) {
realServers1, rsErr1 := ipvs.GetRealServers(virtualServers1[0]) realServers1, rsErr1 := ipvs.GetRealServers(virtualServers1[0])
assert.NoError(t, rsErr1, "Expected no error getting real servers") assert.NoError(t, rsErr1, "Expected no error getting real servers")
assert.Len(t, realServers1, 3, "Expected 3 real servers") assert.Len(t, realServers1, 3, "Expected 3 real servers")
assert.Equal(t, realServers1[0].String(), "10.0.1.1:80") assert.Equal(t, "10.0.1.1:80", realServers1[0].String())
assert.Equal(t, realServers1[1].String(), "10.0.1.2:80") assert.Equal(t, "10.0.1.2:80", realServers1[1].String())
assert.Equal(t, realServers1[2].String(), "10.0.1.3:80") assert.Equal(t, "10.0.1.3:80", realServers1[2].String())
fp.OnEndpointSliceDelete(endpointSlice) fp.OnEndpointSliceDelete(endpointSlice)
fp.syncProxyRules() fp.syncProxyRules()
@ -4963,17 +4963,17 @@ func Test_EndpointSliceReadyAndTerminatingCluster(t *testing.T) {
realServers1, rsErr1 := ipvs.GetRealServers(clusterIPServer) realServers1, rsErr1 := ipvs.GetRealServers(clusterIPServer)
assert.NoError(t, rsErr1, "Expected no error getting real servers") assert.NoError(t, rsErr1, "Expected no error getting real servers")
assert.Len(t, realServers1, 3, "Expected 3 real servers") assert.Len(t, realServers1, 3, "Expected 3 real servers")
assert.Equal(t, realServers1[0].String(), "10.0.1.1:80") assert.Equal(t, "10.0.1.1:80", realServers1[0].String())
assert.Equal(t, realServers1[1].String(), "10.0.1.2:80") assert.Equal(t, "10.0.1.2:80", realServers1[1].String())
assert.Equal(t, realServers1[2].String(), "10.0.1.5:80") assert.Equal(t, "10.0.1.5:80", realServers1[2].String())
// externalIP should route to cluster-wide ready endpoints // externalIP should route to cluster-wide ready endpoints
realServers2, rsErr2 := ipvs.GetRealServers(externalIPServer) realServers2, rsErr2 := ipvs.GetRealServers(externalIPServer)
assert.NoError(t, rsErr2, "Expected no error getting real servers") assert.NoError(t, rsErr2, "Expected no error getting real servers")
assert.Len(t, realServers2, 3, "Expected 3 real servers") assert.Len(t, realServers2, 3, "Expected 3 real servers")
assert.Equal(t, realServers2[0].String(), "10.0.1.1:80") assert.Equal(t, "10.0.1.1:80", realServers2[0].String())
assert.Equal(t, realServers2[1].String(), "10.0.1.2:80") assert.Equal(t, "10.0.1.2:80", realServers2[1].String())
assert.Equal(t, realServers1[2].String(), "10.0.1.5:80") assert.Equal(t, "10.0.1.5:80", realServers1[2].String())
fp.OnEndpointSliceDelete(endpointSlice) fp.OnEndpointSliceDelete(endpointSlice)
fp.syncProxyRules() fp.syncProxyRules()
@ -5136,16 +5136,16 @@ func Test_EndpointSliceReadyAndTerminatingLocal(t *testing.T) {
realServers1, rsErr1 := ipvs.GetRealServers(clusterIPServer) realServers1, rsErr1 := ipvs.GetRealServers(clusterIPServer)
assert.NoError(t, rsErr1, "Expected no error getting real servers") assert.NoError(t, rsErr1, "Expected no error getting real servers")
assert.Len(t, realServers1, 3, "Expected 3 real servers") assert.Len(t, realServers1, 3, "Expected 3 real servers")
assert.Equal(t, realServers1[0].String(), "10.0.1.1:80") assert.Equal(t, "10.0.1.1:80", realServers1[0].String())
assert.Equal(t, realServers1[1].String(), "10.0.1.2:80") assert.Equal(t, "10.0.1.2:80", realServers1[1].String())
assert.Equal(t, realServers1[2].String(), "10.0.1.5:80") assert.Equal(t, "10.0.1.5:80", realServers1[2].String())
// externalIP should route to local ready + non-terminating endpoints if they exist // externalIP should route to local ready + non-terminating endpoints if they exist
realServers2, rsErr2 := ipvs.GetRealServers(externalIPServer) realServers2, rsErr2 := ipvs.GetRealServers(externalIPServer)
assert.NoError(t, rsErr2, "Expected no error getting real servers") assert.NoError(t, rsErr2, "Expected no error getting real servers")
assert.Len(t, realServers2, 2, "Expected 2 real servers") assert.Len(t, realServers2, 2, "Expected 2 real servers")
assert.Equal(t, realServers2[0].String(), "10.0.1.1:80") assert.Equal(t, "10.0.1.1:80", realServers2[0].String())
assert.Equal(t, realServers2[1].String(), "10.0.1.2:80") assert.Equal(t, "10.0.1.2:80", realServers2[1].String())
fp.OnEndpointSliceDelete(endpointSlice) fp.OnEndpointSliceDelete(endpointSlice)
fp.syncProxyRules() fp.syncProxyRules()
@ -5307,17 +5307,17 @@ func Test_EndpointSliceOnlyReadyAndTerminatingCluster(t *testing.T) {
realServers1, rsErr1 := ipvs.GetRealServers(clusterIPServer) realServers1, rsErr1 := ipvs.GetRealServers(clusterIPServer)
assert.NoError(t, rsErr1, "Expected no error getting real servers") assert.NoError(t, rsErr1, "Expected no error getting real servers")
assert.Len(t, realServers1, 3, "Expected 1 real servers") assert.Len(t, realServers1, 3, "Expected 1 real servers")
assert.Equal(t, realServers1[0].String(), "10.0.1.1:80") assert.Equal(t, "10.0.1.1:80", realServers1[0].String())
assert.Equal(t, realServers1[1].String(), "10.0.1.2:80") assert.Equal(t, "10.0.1.2:80", realServers1[1].String())
assert.Equal(t, realServers1[2].String(), "10.0.1.4:80") assert.Equal(t, "10.0.1.4:80", realServers1[2].String())
// externalIP should fall back to ready + terminating endpoints // externalIP should fall back to ready + terminating endpoints
realServers2, rsErr2 := ipvs.GetRealServers(externalIPServer) realServers2, rsErr2 := ipvs.GetRealServers(externalIPServer)
assert.NoError(t, rsErr2, "Expected no error getting real servers") assert.NoError(t, rsErr2, "Expected no error getting real servers")
assert.Len(t, realServers2, 3, "Expected 2 real servers") assert.Len(t, realServers2, 3, "Expected 2 real servers")
assert.Equal(t, realServers2[0].String(), "10.0.1.1:80") assert.Equal(t, "10.0.1.1:80", realServers2[0].String())
assert.Equal(t, realServers2[1].String(), "10.0.1.2:80") assert.Equal(t, "10.0.1.2:80", realServers2[1].String())
assert.Equal(t, realServers2[2].String(), "10.0.1.4:80") assert.Equal(t, "10.0.1.4:80", realServers2[2].String())
fp.OnEndpointSliceDelete(endpointSlice) fp.OnEndpointSliceDelete(endpointSlice)
fp.syncProxyRules() fp.syncProxyRules()
@ -5479,14 +5479,14 @@ func Test_EndpointSliceOnlyReadyAndTerminatingLocal(t *testing.T) {
realServers1, rsErr1 := ipvs.GetRealServers(clusterIPServer) realServers1, rsErr1 := ipvs.GetRealServers(clusterIPServer)
assert.NoError(t, rsErr1, "Expected no error getting real servers") assert.NoError(t, rsErr1, "Expected no error getting real servers")
assert.Len(t, realServers1, 1, "Expected 1 real servers") assert.Len(t, realServers1, 1, "Expected 1 real servers")
assert.Equal(t, realServers1[0].String(), "10.0.1.5:80") assert.Equal(t, "10.0.1.5:80", realServers1[0].String())
// externalIP should fall back to local ready + terminating endpoints // externalIP should fall back to local ready + terminating endpoints
realServers2, rsErr2 := ipvs.GetRealServers(externalIPServer) realServers2, rsErr2 := ipvs.GetRealServers(externalIPServer)
assert.NoError(t, rsErr2, "Expected no error getting real servers") assert.NoError(t, rsErr2, "Expected no error getting real servers")
assert.Len(t, realServers2, 2, "Expected 2 real servers") assert.Len(t, realServers2, 2, "Expected 2 real servers")
assert.Equal(t, realServers2[0].String(), "10.0.1.1:80") assert.Equal(t, "10.0.1.1:80", realServers2[0].String())
assert.Equal(t, realServers2[1].String(), "10.0.1.2:80") assert.Equal(t, "10.0.1.2:80", realServers2[1].String())
fp.OnEndpointSliceDelete(endpointSlice) fp.OnEndpointSliceDelete(endpointSlice)
fp.syncProxyRules() fp.syncProxyRules()

View File

@ -4879,17 +4879,17 @@ func TestProxier_OnServiceCIDRsChanged(t *testing.T) {
proxier = &Proxier{ipFamily: v1.IPv4Protocol} proxier = &Proxier{ipFamily: v1.IPv4Protocol}
proxier.OnServiceCIDRsChanged([]string{"172.30.0.0/16", "fd00:10:96::/112"}) proxier.OnServiceCIDRsChanged([]string{"172.30.0.0/16", "fd00:10:96::/112"})
assert.Equal(t, proxier.serviceCIDRs, "172.30.0.0/16") assert.Equal(t, "172.30.0.0/16", proxier.serviceCIDRs)
proxier.OnServiceCIDRsChanged([]string{"172.30.0.0/16", "172.50.0.0/16", "fd00:10:96::/112", "fd00:172:30::/112"}) proxier.OnServiceCIDRsChanged([]string{"172.30.0.0/16", "172.50.0.0/16", "fd00:10:96::/112", "fd00:172:30::/112"})
assert.Equal(t, proxier.serviceCIDRs, "172.30.0.0/16,172.50.0.0/16") assert.Equal(t, "172.30.0.0/16,172.50.0.0/16", proxier.serviceCIDRs)
proxier = &Proxier{ipFamily: v1.IPv6Protocol} proxier = &Proxier{ipFamily: v1.IPv6Protocol}
proxier.OnServiceCIDRsChanged([]string{"172.30.0.0/16", "fd00:10:96::/112"}) proxier.OnServiceCIDRsChanged([]string{"172.30.0.0/16", "fd00:10:96::/112"})
assert.Equal(t, proxier.serviceCIDRs, "fd00:10:96::/112") assert.Equal(t, "fd00:10:96::/112", proxier.serviceCIDRs)
proxier.OnServiceCIDRsChanged([]string{"172.30.0.0/16", "172.50.0.0/16", "fd00:10:96::/112", "fd00:172:30::/112"}) proxier.OnServiceCIDRsChanged([]string{"172.30.0.0/16", "172.50.0.0/16", "fd00:10:96::/112", "fd00:172:30::/112"})
assert.Equal(t, proxier.serviceCIDRs, "fd00:10:96::/112,fd00:172:30::/112") assert.Equal(t, "fd00:10:96::/112,fd00:172:30::/112", proxier.serviceCIDRs)
} }
// TestBadIPs tests that "bad" IPs and CIDRs in Services/Endpoints are rewritten to // TestBadIPs tests that "bad" IPs and CIDRs in Services/Endpoints are rewritten to

View File

@ -176,7 +176,7 @@ func TestOperationGenerator_GenerateExpandAndRecoverVolumeFunc(t *testing.T) {
} }
updatedPVC := expansionResponse.pvc updatedPVC := expansionResponse.pvc
actualResizeStatus := updatedPVC.Status.AllocatedResourceStatuses[v1.ResourceStorage] actualResizeStatus := updatedPVC.Status.AllocatedResourceStatuses[v1.ResourceStorage]
assert.Equal(t, actualResizeStatus, test.expectedResizeStatus) assert.Equal(t, test.expectedResizeStatus, actualResizeStatus)
actualAllocatedSize := updatedPVC.Status.AllocatedResources.Storage() actualAllocatedSize := updatedPVC.Status.AllocatedResources.Storage()
if test.expectedAllocatedSize.Cmp(*actualAllocatedSize) != 0 { if test.expectedAllocatedSize.Cmp(*actualAllocatedSize) != 0 {
t.Fatalf("GenerateExpandAndRecoverVolumeFunc failed: expected allocated size %s, got %s", test.expectedAllocatedSize.String(), actualAllocatedSize.String()) t.Fatalf("GenerateExpandAndRecoverVolumeFunc failed: expected allocated size %s, got %s", test.expectedAllocatedSize.String(), actualAllocatedSize.String())