mirror of
https://github.com/k3s-io/kubernetes.git
synced 2025-08-17 23:57:49 +00:00
refactor integ test
This commit is contained in:
parent
6c6be931ee
commit
558945958e
@ -57,11 +57,11 @@ import (
|
|||||||
|
|
||||||
func TestSchedulingGates(t *testing.T) {
|
func TestSchedulingGates(t *testing.T) {
|
||||||
tests := []struct {
|
tests := []struct {
|
||||||
name string
|
name string
|
||||||
pods []*v1.Pod
|
pods []*v1.Pod
|
||||||
want []string
|
schedule []string
|
||||||
rmPodsSchedulingGates []int
|
delete []string
|
||||||
wantPostGatesRemoval []string
|
rmGates []string
|
||||||
}{
|
}{
|
||||||
{
|
{
|
||||||
name: "regular pods",
|
name: "regular pods",
|
||||||
@ -69,7 +69,7 @@ func TestSchedulingGates(t *testing.T) {
|
|||||||
st.MakePod().Name("p1").Container("pause").Obj(),
|
st.MakePod().Name("p1").Container("pause").Obj(),
|
||||||
st.MakePod().Name("p2").Container("pause").Obj(),
|
st.MakePod().Name("p2").Container("pause").Obj(),
|
||||||
},
|
},
|
||||||
want: []string{"p1", "p2"},
|
schedule: []string{"p1", "p2"},
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
name: "one pod carrying scheduling gates",
|
name: "one pod carrying scheduling gates",
|
||||||
@ -77,7 +77,7 @@ func TestSchedulingGates(t *testing.T) {
|
|||||||
st.MakePod().Name("p1").SchedulingGates([]string{"foo"}).Container("pause").Obj(),
|
st.MakePod().Name("p1").SchedulingGates([]string{"foo"}).Container("pause").Obj(),
|
||||||
st.MakePod().Name("p2").Container("pause").Obj(),
|
st.MakePod().Name("p2").Container("pause").Obj(),
|
||||||
},
|
},
|
||||||
want: []string{"p2"},
|
schedule: []string{"p2"},
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
name: "two pod carrying scheduling gates, and remove gates of one pod",
|
name: "two pod carrying scheduling gates, and remove gates of one pod",
|
||||||
@ -86,9 +86,18 @@ func TestSchedulingGates(t *testing.T) {
|
|||||||
st.MakePod().Name("p2").SchedulingGates([]string{"bar"}).Container("pause").Obj(),
|
st.MakePod().Name("p2").SchedulingGates([]string{"bar"}).Container("pause").Obj(),
|
||||||
st.MakePod().Name("p3").Container("pause").Obj(),
|
st.MakePod().Name("p3").Container("pause").Obj(),
|
||||||
},
|
},
|
||||||
want: []string{"p3"},
|
schedule: []string{"p3"},
|
||||||
rmPodsSchedulingGates: []int{1}, // remove gates of 'p2'
|
rmGates: []string{"p2"},
|
||||||
wantPostGatesRemoval: []string{"p2"},
|
},
|
||||||
|
{
|
||||||
|
name: "gated pod schedulable after deleting the scheduled pod and removing gate",
|
||||||
|
pods: []*v1.Pod{
|
||||||
|
st.MakePod().Name("p1").SchedulingGates([]string{"foo"}).Container("pause").Obj(),
|
||||||
|
st.MakePod().Name("p2").Container("pause").Obj(),
|
||||||
|
},
|
||||||
|
schedule: []string{"p2"},
|
||||||
|
delete: []string{"p2"},
|
||||||
|
rmGates: []string{"p1"},
|
||||||
},
|
},
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -107,6 +116,15 @@ func TestSchedulingGates(t *testing.T) {
|
|||||||
testutils.SyncSchedulerInformerFactory(testCtx)
|
testutils.SyncSchedulerInformerFactory(testCtx)
|
||||||
|
|
||||||
cs, ns, ctx := testCtx.ClientSet, testCtx.NS.Name, testCtx.Ctx
|
cs, ns, ctx := testCtx.ClientSet, testCtx.NS.Name, testCtx.Ctx
|
||||||
|
|
||||||
|
// Create node, so we can schedule pods.
|
||||||
|
node := st.MakeNode().Name("node").Obj()
|
||||||
|
if _, err := cs.CoreV1().Nodes().Create(ctx, node, metav1.CreateOptions{}); err != nil {
|
||||||
|
t.Fatal("Failed to create node")
|
||||||
|
|
||||||
|
}
|
||||||
|
|
||||||
|
// Create pods.
|
||||||
for _, p := range tt.pods {
|
for _, p := range tt.pods {
|
||||||
p.Namespace = ns
|
p.Namespace = ns
|
||||||
if _, err := cs.CoreV1().Pods(ns).Create(ctx, p, metav1.CreateOptions{}); err != nil {
|
if _, err := cs.CoreV1().Pods(ns).Create(ctx, p, metav1.CreateOptions{}); err != nil {
|
||||||
@ -122,95 +140,48 @@ func TestSchedulingGates(t *testing.T) {
|
|||||||
t.Fatal(err)
|
t.Fatal(err)
|
||||||
}
|
}
|
||||||
|
|
||||||
// Pop the expected pods out. They should be de-queueable.
|
// Schedule pods.
|
||||||
for _, wantPod := range tt.want {
|
for _, podName := range tt.schedule {
|
||||||
podInfo := testutils.NextPodOrDie(t, testCtx)
|
testCtx.Scheduler.ScheduleOne(testCtx.Ctx)
|
||||||
if got := podInfo.Pod.Name; got != wantPod {
|
if err := wait.PollUntilContextTimeout(ctx, time.Millisecond*200, wait.ForeverTestTimeout, false, testutils.PodScheduled(cs, ns, podName)); err != nil {
|
||||||
t.Errorf("Want %v to be popped out, but got %v", wantPod, got)
|
t.Fatalf("Failed to schedule %s", podName)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
if len(tt.rmPodsSchedulingGates) == 0 {
|
// Delete pods, which triggers AssignedPodDelete event in the scheduling queue.
|
||||||
return
|
for _, podName := range tt.delete {
|
||||||
|
if err := cs.CoreV1().Pods(ns).Delete(ctx, podName, metav1.DeleteOptions{}); err != nil {
|
||||||
|
t.Fatalf("Error calling Delete on %s", podName)
|
||||||
|
}
|
||||||
|
if err := wait.PollUntilContextTimeout(ctx, time.Millisecond*200, wait.ForeverTestTimeout, false, testutils.PodDeleted(ctx, cs, ns, podName)); err != nil {
|
||||||
|
t.Fatalf("Failed to delete %s", podName)
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// Ensure gated pods are not in ActiveQ
|
||||||
|
if len(testCtx.Scheduler.SchedulingQueue.PodsInActiveQ()) > 0 {
|
||||||
|
t.Fatal("Expected no schedulable pods")
|
||||||
|
}
|
||||||
|
|
||||||
// Remove scheduling gates from the pod spec.
|
// Remove scheduling gates from the pod spec.
|
||||||
for _, idx := range tt.rmPodsSchedulingGates {
|
for _, podName := range tt.rmGates {
|
||||||
patch := `{"spec": {"schedulingGates": null}}`
|
patch := `{"spec": {"schedulingGates": null}}`
|
||||||
podName := tt.pods[idx].Name
|
|
||||||
if _, err := cs.CoreV1().Pods(ns).Patch(ctx, podName, types.StrategicMergePatchType, []byte(patch), metav1.PatchOptions{}); err != nil {
|
if _, err := cs.CoreV1().Pods(ns).Patch(ctx, podName, types.StrategicMergePatchType, []byte(patch), metav1.PatchOptions{}); err != nil {
|
||||||
t.Fatalf("Failed to patch pod %v: %v", podName, err)
|
t.Fatalf("Failed to patch pod %v: %v", podName, err)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
// Pop the expected pods out. They should be de-queueable.
|
|
||||||
for _, wantPod := range tt.wantPostGatesRemoval {
|
// Schedule pods which no longer have gates.
|
||||||
podInfo := testutils.NextPodOrDie(t, testCtx)
|
for _, podName := range tt.rmGates {
|
||||||
if got := podInfo.Pod.Name; got != wantPod {
|
testCtx.Scheduler.ScheduleOne(testCtx.Ctx)
|
||||||
t.Errorf("Want %v to be popped out, but got %v", wantPod, got)
|
if err := wait.PollUntilContextTimeout(ctx, time.Millisecond*200, wait.ForeverTestTimeout, false, testutils.PodScheduled(cs, ns, podName)); err != nil {
|
||||||
|
t.Fatalf("Failed to schedule %s", podName)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// We create a gated and non-gated pod. After scheduling and deleting the
|
|
||||||
// non-gated pod, we ensure that the gated pod is schedulable after the removal
|
|
||||||
// of its gate.
|
|
||||||
func TestGatedPodSchedulableAfterEvent(t *testing.T) {
|
|
||||||
testCtx := testutils.InitTestSchedulerWithOptions(
|
|
||||||
t,
|
|
||||||
testutils.InitTestAPIServer(t, "pod-scheduling-gates", nil),
|
|
||||||
0,
|
|
||||||
scheduler.WithPodInitialBackoffSeconds(0),
|
|
||||||
scheduler.WithPodMaxBackoffSeconds(0),
|
|
||||||
)
|
|
||||||
testutils.SyncSchedulerInformerFactory(testCtx)
|
|
||||||
cs, ns, ctx := testCtx.ClientSet, testCtx.NS.Name, testCtx.Ctx
|
|
||||||
|
|
||||||
// create initially gated pod
|
|
||||||
pod1 := st.MakePod().Namespace(ns).Name("p1").Container("pause").SchedulingGates([]string{"foo"}).Obj()
|
|
||||||
if _, err := cs.CoreV1().Pods(ns).Create(ctx, pod1, metav1.CreateOptions{}); err != nil {
|
|
||||||
t.Fatal("Failed to create p1")
|
|
||||||
}
|
|
||||||
|
|
||||||
// create immediately schedulable pod
|
|
||||||
pod2 := st.MakePod().Namespace(ns).Name("p2").Container("pause").Obj()
|
|
||||||
if _, err := cs.CoreV1().Pods(ns).Create(ctx, pod2, metav1.CreateOptions{}); err != nil {
|
|
||||||
t.Fatal("Failed to create p2")
|
|
||||||
}
|
|
||||||
|
|
||||||
// create node on which to schedule pods
|
|
||||||
node := st.MakeNode().Name("node1").Obj()
|
|
||||||
if _, err := cs.CoreV1().Nodes().Create(ctx, node, metav1.CreateOptions{}); err != nil {
|
|
||||||
t.Fatal("Failed to create node1")
|
|
||||||
}
|
|
||||||
|
|
||||||
// schedule p2
|
|
||||||
testCtx.Scheduler.ScheduleOne(testCtx.Ctx)
|
|
||||||
if err := wait.PollUntilContextTimeout(ctx, 200*time.Millisecond, wait.ForeverTestTimeout, false, testutils.PodScheduled(cs, ns, "p2")); err != nil {
|
|
||||||
t.Fatal("Failed to schedule p2")
|
|
||||||
}
|
|
||||||
|
|
||||||
// delete p2, which triggers DeletePodFromCache event
|
|
||||||
if err := cs.CoreV1().Pods(ns).Delete(ctx, "p2", metav1.DeleteOptions{}); err != nil {
|
|
||||||
t.Fatal("Error calling Delete on p2")
|
|
||||||
}
|
|
||||||
if err := wait.PollUntilContextTimeout(ctx, 200*time.Millisecond, wait.ForeverTestTimeout, false, testutils.PodDeleted(ctx, cs, ns, "p2")); err != nil {
|
|
||||||
t.Fatal("Failed to delete p2")
|
|
||||||
}
|
|
||||||
|
|
||||||
// remove gate from p1
|
|
||||||
patch := `{"spec": {"schedulingGates": null}}`
|
|
||||||
if _, err := cs.CoreV1().Pods(ns).Patch(ctx, "p1", types.StrategicMergePatchType, []byte(patch), metav1.PatchOptions{}); err != nil {
|
|
||||||
t.Fatal("Failed to remove schedulingGates from p1")
|
|
||||||
}
|
|
||||||
|
|
||||||
// schedule p1
|
|
||||||
testCtx.Scheduler.ScheduleOne(testCtx.Ctx)
|
|
||||||
if err := wait.PollUntilContextTimeout(ctx, 200*time.Millisecond, wait.ForeverTestTimeout, false, testutils.PodScheduled(cs, ns, "p1")); err != nil {
|
|
||||||
t.Fatal("Failed to schedule p1")
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// TestCoreResourceEnqueue verify Pods failed by in-tree default plugins can be
|
// TestCoreResourceEnqueue verify Pods failed by in-tree default plugins can be
|
||||||
// moved properly upon their registered events.
|
// moved properly upon their registered events.
|
||||||
func TestCoreResourceEnqueue(t *testing.T) {
|
func TestCoreResourceEnqueue(t *testing.T) {
|
||||||
|
Loading…
Reference in New Issue
Block a user