mirror of
https://github.com/k3s-io/kubernetes.git
synced 2025-07-30 23:15:14 +00:00
fix: enable compares rule from testifylint in module k8s.io/kubernetes
Signed-off-by: Matthieu MOREL <matthieu.morel35@gmail.com>
This commit is contained in:
parent
221bf19ee0
commit
fa0e38981c
@ -2011,7 +2011,7 @@ func TestFeatureGateResourceHealthStatus(t *testing.T) {
|
||||
{ID: "dev2", Health: pluginapi.Unhealthy},
|
||||
})
|
||||
// update chan no data
|
||||
assert.Empty(t, len(testManager.update), 0)
|
||||
assert.Empty(t, testManager.update)
|
||||
|
||||
// update chan receive pod1
|
||||
var wg sync.WaitGroup
|
||||
|
@ -35,7 +35,7 @@ func TestCacheNotInitialized(t *testing.T) {
|
||||
cache := newTestCache()
|
||||
// If the global timestamp is not set, always return nil.
|
||||
d := cache.getIfNewerThan(types.UID("1234"), time.Time{})
|
||||
assert.True(t, d == nil, "should return nil since cache is not initialized")
|
||||
assert.Nil(t, d, "should return nil since cache is not initialized")
|
||||
}
|
||||
|
||||
func getTestPodIDAndStatus(numContainers int) (types.UID, *PodStatus) {
|
||||
@ -178,9 +178,9 @@ func TestDelete(t *testing.T) {
|
||||
|
||||
func verifyNotification(t *testing.T, ch chan *data, expectNotification bool) {
|
||||
if expectNotification {
|
||||
assert.True(t, len(ch) > 0, "Did not receive notification")
|
||||
assert.NotEmpty(t, ch, "Did not receive notification")
|
||||
} else {
|
||||
assert.True(t, len(ch) < 1, "Should not have triggered the notification")
|
||||
assert.Empty(t, ch, "Should not have triggered the notification")
|
||||
}
|
||||
// Drain the channel.
|
||||
for i := 0; i < len(ch); i++ {
|
||||
|
@ -527,7 +527,7 @@ func TestHandlePodCleanupsPerQOS(t *testing.T) {
|
||||
// done within a goroutine and can get called multiple times, so the
|
||||
// Destroy() count in not deterministic on the actual number.
|
||||
// https://github.com/kubernetes/kubernetes/blob/29fdbb065b5e0d195299eb2d260b975cbc554673/pkg/kubelet/kubelet_pods.go#L2006
|
||||
assert.True(t, destroyCount >= 1, "Expect 1 or more destroys")
|
||||
assert.GreaterOrEqual(t, destroyCount, 1, "Expect 1 or more destroys")
|
||||
}
|
||||
|
||||
func TestDispatchWorkOfCompletedPod(t *testing.T) {
|
||||
|
@ -390,7 +390,7 @@ func TestVolumeAttachAndMountControllerDisabled(t *testing.T) {
|
||||
for _, name := range expectedPodVolumes {
|
||||
assert.Contains(t, podVolumes, name, "Volumes for pod %+v", pod)
|
||||
}
|
||||
assert.True(t, testKubelet.volumePlugin.GetNewAttacherCallCount() >= 1, "Expected plugin NewAttacher to be called at least once")
|
||||
assert.GreaterOrEqual(t, testKubelet.volumePlugin.GetNewAttacherCallCount(), 1, "Expected plugin NewAttacher to be called at least once")
|
||||
assert.NoError(t, volumetest.VerifyWaitForAttachCallCount(
|
||||
1 /* expectedWaitForAttachCallCount */, testKubelet.volumePlugin))
|
||||
assert.NoError(t, volumetest.VerifyAttachCallCount(
|
||||
@ -454,7 +454,7 @@ func TestVolumeUnmountAndDetachControllerDisabled(t *testing.T) {
|
||||
assert.Contains(t, podVolumes, name, "Volumes for pod %+v", pod)
|
||||
}
|
||||
|
||||
assert.True(t, testKubelet.volumePlugin.GetNewAttacherCallCount() >= 1, "Expected plugin NewAttacher to be called at least once")
|
||||
assert.GreaterOrEqual(t, testKubelet.volumePlugin.GetNewAttacherCallCount(), 1, "Expected plugin NewAttacher to be called at least once")
|
||||
assert.NoError(t, volumetest.VerifyWaitForAttachCallCount(
|
||||
1 /* expectedWaitForAttachCallCount */, testKubelet.volumePlugin))
|
||||
assert.NoError(t, volumetest.VerifyAttachCallCount(
|
||||
@ -486,7 +486,7 @@ func TestVolumeUnmountAndDetachControllerDisabled(t *testing.T) {
|
||||
|
||||
// Verify volumes detached and no longer reported as in use
|
||||
assert.NoError(t, waitForVolumeDetach(v1.UniqueVolumeName("fake/fake-device"), kubelet.volumeManager))
|
||||
assert.True(t, testKubelet.volumePlugin.GetNewAttacherCallCount() >= 1, "Expected plugin NewAttacher to be called at least once")
|
||||
assert.GreaterOrEqual(t, testKubelet.volumePlugin.GetNewAttacherCallCount(), 1, "Expected plugin NewAttacher to be called at least once")
|
||||
assert.NoError(t, volumetest.VerifyDetachCallCount(
|
||||
1 /* expectedDetachCallCount */, testKubelet.volumePlugin))
|
||||
}
|
||||
@ -566,7 +566,7 @@ func TestVolumeAttachAndMountControllerEnabled(t *testing.T) {
|
||||
for _, name := range expectedPodVolumes {
|
||||
assert.Contains(t, podVolumes, name, "Volumes for pod %+v", pod)
|
||||
}
|
||||
assert.True(t, testKubelet.volumePlugin.GetNewAttacherCallCount() >= 1, "Expected plugin NewAttacher to be called at least once")
|
||||
assert.GreaterOrEqual(t, testKubelet.volumePlugin.GetNewAttacherCallCount(), 1, "Expected plugin NewAttacher to be called at least once")
|
||||
assert.NoError(t, volumetest.VerifyWaitForAttachCallCount(
|
||||
1 /* expectedWaitForAttachCallCount */, testKubelet.volumePlugin))
|
||||
assert.NoError(t, volumetest.VerifyZeroAttachCalls(testKubelet.volumePlugin))
|
||||
@ -654,7 +654,7 @@ func TestVolumeUnmountAndDetachControllerEnabled(t *testing.T) {
|
||||
assert.Contains(t, podVolumes, name, "Volumes for pod %+v", pod)
|
||||
}
|
||||
|
||||
assert.True(t, testKubelet.volumePlugin.GetNewAttacherCallCount() >= 1, "Expected plugin NewAttacher to be called at least once")
|
||||
assert.GreaterOrEqual(t, testKubelet.volumePlugin.GetNewAttacherCallCount(), 1, "Expected plugin NewAttacher to be called at least once")
|
||||
assert.NoError(t, volumetest.VerifyWaitForAttachCallCount(
|
||||
1 /* expectedWaitForAttachCallCount */, testKubelet.volumePlugin))
|
||||
assert.NoError(t, volumetest.VerifyZeroAttachCalls(testKubelet.volumePlugin))
|
||||
@ -684,7 +684,7 @@ func TestVolumeUnmountAndDetachControllerEnabled(t *testing.T) {
|
||||
|
||||
// Verify volumes detached and no longer reported as in use
|
||||
assert.NoError(t, waitForVolumeDetach(v1.UniqueVolumeName("fake/fake-device"), kubelet.volumeManager))
|
||||
assert.True(t, testKubelet.volumePlugin.GetNewAttacherCallCount() >= 1, "Expected plugin NewAttacher to be called at least once")
|
||||
assert.GreaterOrEqual(t, testKubelet.volumePlugin.GetNewAttacherCallCount(), 1, "Expected plugin NewAttacher to be called at least once")
|
||||
assert.NoError(t, volumetest.VerifyZeroDetachCallCount(testKubelet.volumePlugin))
|
||||
}
|
||||
|
||||
|
@ -39,7 +39,7 @@ func TestCacheOperations(t *testing.T) {
|
||||
|
||||
m.Set(setID, Success, &corev1.Pod{})
|
||||
result, found := m.Get(setID)
|
||||
assert.True(t, result == Success, "set result")
|
||||
assert.Equal(t, Success, result, "set result")
|
||||
assert.True(t, found, "set result found")
|
||||
|
||||
m.Remove(setID)
|
||||
|
@ -683,10 +683,10 @@ func TestTerminatePod(t *testing.T) {
|
||||
t.Logf("we expect the container statuses to have changed to terminated")
|
||||
newStatus := expectPodStatus(t, syncer, testPod)
|
||||
for i := range newStatus.ContainerStatuses {
|
||||
assert.False(t, newStatus.ContainerStatuses[i].State.Terminated == nil, "expected containers to be terminated")
|
||||
assert.NotNil(t, newStatus.ContainerStatuses[i].State.Terminated, "expected containers to be terminated")
|
||||
}
|
||||
for i := range newStatus.InitContainerStatuses {
|
||||
assert.False(t, newStatus.InitContainerStatuses[i].State.Terminated == nil, "expected init containers to be terminated")
|
||||
assert.NotNil(t, newStatus.InitContainerStatuses[i].State.Terminated, "expected init containers to be terminated")
|
||||
}
|
||||
|
||||
expectUnknownState := v1.ContainerState{Terminated: &v1.ContainerStateTerminated{Reason: kubecontainer.ContainerReasonStatusUnknown, Message: "The container could not be located when the pod was terminated", ExitCode: 137}}
|
||||
@ -765,13 +765,13 @@ func TestTerminatePodWaiting(t *testing.T) {
|
||||
t.Logf("we expect the container statuses to have changed to terminated")
|
||||
newStatus := expectPodStatus(t, syncer, testPod)
|
||||
for _, container := range newStatus.ContainerStatuses {
|
||||
assert.False(t, container.State.Terminated == nil, "expected containers to be terminated")
|
||||
assert.NotNil(t, container.State.Terminated, "expected containers to be terminated")
|
||||
}
|
||||
for _, container := range newStatus.InitContainerStatuses[:2] {
|
||||
assert.False(t, container.State.Terminated == nil, "expected init containers to be terminated")
|
||||
assert.NotNil(t, container.State.Terminated, "expected init containers to be terminated")
|
||||
}
|
||||
for _, container := range newStatus.InitContainerStatuses[2:] {
|
||||
assert.False(t, container.State.Waiting == nil, "expected init containers to be waiting")
|
||||
assert.NotNil(t, container.State.Waiting, "expected init containers to be waiting")
|
||||
}
|
||||
|
||||
expectUnknownState := v1.ContainerState{Terminated: &v1.ContainerStateTerminated{Reason: kubecontainer.ContainerReasonStatusUnknown, Message: "The container could not be located when the pod was terminated", ExitCode: 137}}
|
||||
|
@ -122,9 +122,9 @@ func TestUserNsManagerAllocate(t *testing.T) {
|
||||
allocated, length, err = m.allocateOne(types.UID(fmt.Sprintf("%d", i)))
|
||||
assert.Equal(t, userNsLength, int(length), "length is not the expected. iter: %v", i)
|
||||
assert.NoError(t, err)
|
||||
assert.True(t, allocated >= minimumMappingUID)
|
||||
assert.GreaterOrEqual(t, allocated, uint32(minimumMappingUID))
|
||||
// The last ID of the userns range (allocated+userNsLength) should be within bounds.
|
||||
assert.True(t, allocated <= minimumMappingUID+mappingLen-userNsLength)
|
||||
assert.LessOrEqual(t, allocated, uint32(minimumMappingUID+mappingLen-userNsLength))
|
||||
allocs = append(allocs, allocated)
|
||||
}
|
||||
for i, v := range allocs {
|
||||
|
Loading…
Reference in New Issue
Block a user