mirror of
https://github.com/k3s-io/kubernetes.git
synced 2025-07-26 05:03:09 +00:00
skip deleted activePods and return nil
This commit is contained in:
parent
35a456b0c6
commit
6157d3cc4a
@ -1014,11 +1014,6 @@ func (m *ManagerImpl) checkPodActive(pod *v1.Pod) bool {
|
|||||||
// for the found one. An empty struct is returned in case no cached state is found.
|
// for the found one. An empty struct is returned in case no cached state is found.
|
||||||
func (m *ManagerImpl) GetDeviceRunContainerOptions(pod *v1.Pod, container *v1.Container) (*DeviceRunContainerOptions, error) {
|
func (m *ManagerImpl) GetDeviceRunContainerOptions(pod *v1.Pod, container *v1.Container) (*DeviceRunContainerOptions, error) {
|
||||||
podUID := string(pod.UID)
|
podUID := string(pod.UID)
|
||||||
if !m.checkPodActive(pod) {
|
|
||||||
klog.Warningf("pod %s has been deleted from activePods, skip getting device run options", podUID)
|
|
||||||
return nil, fmt.Errorf("pod %v is removed from activePods list", podUID)
|
|
||||||
}
|
|
||||||
|
|
||||||
contName := container.Name
|
contName := container.Name
|
||||||
needsReAllocate := false
|
needsReAllocate := false
|
||||||
for k, v := range container.Resources.Limits {
|
for k, v := range container.Resources.Limits {
|
||||||
@ -1030,6 +1025,12 @@ func (m *ManagerImpl) GetDeviceRunContainerOptions(pod *v1.Pod, container *v1.Co
|
|||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
|
|
||||||
|
if !m.checkPodActive(pod) {
|
||||||
|
klog.ErrorS(nil, "pod deleted from activePods, skip to reAllocate", "podUID", podUID)
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
|
||||||
// This is a device plugin resource yet we don't have cached
|
// This is a device plugin resource yet we don't have cached
|
||||||
// resource state. This is likely due to a race during node
|
// resource state. This is likely due to a race during node
|
||||||
// restart. We re-issue allocate request to cover this race.
|
// restart. We re-issue allocate request to cover this race.
|
||||||
|
@ -960,13 +960,23 @@ func TestPodContainerDeviceAllocation(t *testing.T) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
func TestGetDeviceRunContainerOptions(t *testing.T) {
|
func TestGetDeviceRunContainerOptions(t *testing.T) {
|
||||||
res := TestResource{
|
res1 := TestResource{
|
||||||
resourceName: "domain1.com/resource1",
|
resourceName: "domain1.com/resource1",
|
||||||
resourceQuantity: *resource.NewQuantity(int64(2), resource.DecimalSI),
|
resourceQuantity: *resource.NewQuantity(int64(2), resource.DecimalSI),
|
||||||
devs: checkpoint.DevicesPerNUMA{0: []string{"dev1", "dev2"}},
|
devs: checkpoint.DevicesPerNUMA{0: []string{"dev1", "dev2"}},
|
||||||
topology: true,
|
topology: true,
|
||||||
}
|
}
|
||||||
testResources := []TestResource{res}
|
res2 := TestResource{
|
||||||
|
resourceName: "domain2.com/resource2",
|
||||||
|
resourceQuantity: *resource.NewQuantity(int64(1), resource.DecimalSI),
|
||||||
|
devs: checkpoint.DevicesPerNUMA{0: []string{"dev3", "dev4"}},
|
||||||
|
topology: false,
|
||||||
|
}
|
||||||
|
|
||||||
|
testResources := make([]TestResource, 2)
|
||||||
|
testResources = append(testResources, res1)
|
||||||
|
testResources = append(testResources, res2)
|
||||||
|
|
||||||
podsStub := activePodsStub{
|
podsStub := activePodsStub{
|
||||||
activePods: []*v1.Pod{},
|
activePods: []*v1.Pod{},
|
||||||
}
|
}
|
||||||
@ -979,26 +989,37 @@ func TestGetDeviceRunContainerOptions(t *testing.T) {
|
|||||||
testManager, err := getTestManager(tmpDir, podsStub.getActivePods, testResources)
|
testManager, err := getTestManager(tmpDir, podsStub.getActivePods, testResources)
|
||||||
as.Nil(err)
|
as.Nil(err)
|
||||||
|
|
||||||
pod := makePod(v1.ResourceList{v1.ResourceName(res.resourceName): res.resourceQuantity})
|
pod1 := makePod(v1.ResourceList{
|
||||||
activePods := []*v1.Pod{pod}
|
v1.ResourceName(res1.resourceName): res1.resourceQuantity,
|
||||||
|
v1.ResourceName(res2.resourceName): res2.resourceQuantity,
|
||||||
|
})
|
||||||
|
pod2 := makePod(v1.ResourceList{
|
||||||
|
v1.ResourceName(res2.resourceName): res2.resourceQuantity,
|
||||||
|
})
|
||||||
|
|
||||||
|
activePods := []*v1.Pod{pod1, pod2}
|
||||||
podsStub.updateActivePods(activePods)
|
podsStub.updateActivePods(activePods)
|
||||||
|
|
||||||
err = testManager.Allocate(pod, &pod.Spec.Containers[0])
|
err = testManager.Allocate(pod1, &pod1.Spec.Containers[0])
|
||||||
|
as.Nil(err)
|
||||||
|
err = testManager.Allocate(pod2, &pod2.Spec.Containers[0])
|
||||||
as.Nil(err)
|
as.Nil(err)
|
||||||
|
|
||||||
// when pod is in activePods, GetDeviceRunContainerOptions should return
|
// when pod is in activePods, GetDeviceRunContainerOptions should return
|
||||||
_, err = testManager.GetDeviceRunContainerOptions(pod, &pod.Spec.Containers[0])
|
runContainerOpts, err := testManager.GetDeviceRunContainerOptions(pod1, &pod1.Spec.Containers[0])
|
||||||
as.Nil(err)
|
as.Nil(err)
|
||||||
|
as.Equal(len(runContainerOpts.Devices), 3)
|
||||||
|
as.Equal(len(runContainerOpts.Mounts), 2)
|
||||||
|
as.Equal(len(runContainerOpts.Envs), 2)
|
||||||
|
|
||||||
activePods = []*v1.Pod{}
|
activePods = []*v1.Pod{pod2}
|
||||||
podsStub.updateActivePods(activePods)
|
podsStub.updateActivePods(activePods)
|
||||||
|
testManager.UpdateAllocatedDevices()
|
||||||
|
|
||||||
// when pod is removed from activePods,G etDeviceRunContainerOptions should return error
|
// when pod is removed from activePods,G etDeviceRunContainerOptions should return error
|
||||||
_, err = testManager.GetDeviceRunContainerOptions(pod, &pod.Spec.Containers[0])
|
runContainerOpts, err = testManager.GetDeviceRunContainerOptions(pod1, &pod1.Spec.Containers[0])
|
||||||
expectedErr := fmt.Errorf("pod %v is removed from activePods list", pod.UID)
|
as.Nil(err)
|
||||||
as.NotNil(err)
|
as.Nil(runContainerOpts)
|
||||||
if !reflect.DeepEqual(err, expectedErr) {
|
|
||||||
t.Errorf("GetDeviceRunContainerOptions. expected error: %v but got: %v", expectedErr, err)
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestInitContainerDeviceAllocation(t *testing.T) {
|
func TestInitContainerDeviceAllocation(t *testing.T) {
|
||||||
|
Loading…
Reference in New Issue
Block a user