mirror of
https://github.com/k3s-io/kubernetes.git
synced 2025-08-05 10:19:50 +00:00
podtopologyspread: scheduler queueing hints
This commit is contained in:
parent
55f2bc1043
commit
09a80df1d1
@ -164,3 +164,13 @@ func countPodsMatchSelector(podInfos []*framework.PodInfo, selector labels.Selec
|
|||||||
}
|
}
|
||||||
return count
|
return count
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// podLabelsMatchSpreadConstraints returns whether tha labels matches with the selector in any of topologySpreadConstraint
|
||||||
|
func podLabelsMatchSpreadConstraints(constraints []topologySpreadConstraint, labels labels.Set) bool {
|
||||||
|
for _, c := range constraints {
|
||||||
|
if c.Selector.Matches(labels) {
|
||||||
|
return true
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
@ -239,24 +239,9 @@ func getPreFilterState(cycleState *framework.CycleState) (*preFilterState, error
|
|||||||
|
|
||||||
// calPreFilterState computes preFilterState describing how pods are spread on topologies.
|
// calPreFilterState computes preFilterState describing how pods are spread on topologies.
|
||||||
func (pl *PodTopologySpread) calPreFilterState(ctx context.Context, pod *v1.Pod) (*preFilterState, error) {
|
func (pl *PodTopologySpread) calPreFilterState(ctx context.Context, pod *v1.Pod) (*preFilterState, error) {
|
||||||
var constraints []topologySpreadConstraint
|
constraints, err := pl.getConstraints(pod)
|
||||||
var err error
|
if err != nil {
|
||||||
if len(pod.Spec.TopologySpreadConstraints) > 0 {
|
return nil, fmt.Errorf("get constraints from pod: %w", err)
|
||||||
// We have feature gating in APIServer to strip the spec
|
|
||||||
// so don't need to re-check feature gate, just check length of Constraints.
|
|
||||||
constraints, err = pl.filterTopologySpreadConstraints(
|
|
||||||
pod.Spec.TopologySpreadConstraints,
|
|
||||||
pod.Labels,
|
|
||||||
v1.DoNotSchedule,
|
|
||||||
)
|
|
||||||
if err != nil {
|
|
||||||
return nil, fmt.Errorf("obtaining pod's hard topology spread constraints: %w", err)
|
|
||||||
}
|
|
||||||
} else {
|
|
||||||
constraints, err = pl.buildDefaultConstraints(pod, v1.DoNotSchedule)
|
|
||||||
if err != nil {
|
|
||||||
return nil, fmt.Errorf("setting default hard topology spread constraints: %w", err)
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
if len(constraints) == 0 {
|
if len(constraints) == 0 {
|
||||||
return &preFilterState{}, nil
|
return &preFilterState{}, nil
|
||||||
|
@ -19,18 +19,22 @@ package podtopologyspread
|
|||||||
import (
|
import (
|
||||||
"context"
|
"context"
|
||||||
"fmt"
|
"fmt"
|
||||||
|
"reflect"
|
||||||
|
|
||||||
v1 "k8s.io/api/core/v1"
|
v1 "k8s.io/api/core/v1"
|
||||||
|
"k8s.io/apimachinery/pkg/labels"
|
||||||
"k8s.io/apimachinery/pkg/runtime"
|
"k8s.io/apimachinery/pkg/runtime"
|
||||||
"k8s.io/client-go/informers"
|
"k8s.io/client-go/informers"
|
||||||
appslisters "k8s.io/client-go/listers/apps/v1"
|
appslisters "k8s.io/client-go/listers/apps/v1"
|
||||||
corelisters "k8s.io/client-go/listers/core/v1"
|
corelisters "k8s.io/client-go/listers/core/v1"
|
||||||
|
"k8s.io/klog/v2"
|
||||||
"k8s.io/kubernetes/pkg/scheduler/apis/config"
|
"k8s.io/kubernetes/pkg/scheduler/apis/config"
|
||||||
"k8s.io/kubernetes/pkg/scheduler/apis/config/validation"
|
"k8s.io/kubernetes/pkg/scheduler/apis/config/validation"
|
||||||
"k8s.io/kubernetes/pkg/scheduler/framework"
|
"k8s.io/kubernetes/pkg/scheduler/framework"
|
||||||
"k8s.io/kubernetes/pkg/scheduler/framework/parallelize"
|
"k8s.io/kubernetes/pkg/scheduler/framework/parallelize"
|
||||||
"k8s.io/kubernetes/pkg/scheduler/framework/plugins/feature"
|
"k8s.io/kubernetes/pkg/scheduler/framework/plugins/feature"
|
||||||
"k8s.io/kubernetes/pkg/scheduler/framework/plugins/names"
|
"k8s.io/kubernetes/pkg/scheduler/framework/plugins/names"
|
||||||
|
"k8s.io/kubernetes/pkg/scheduler/util"
|
||||||
)
|
)
|
||||||
|
|
||||||
const (
|
const (
|
||||||
@ -141,9 +145,137 @@ func (pl *PodTopologySpread) EventsToRegister() []framework.ClusterEventWithHint
|
|||||||
// an unschedulable Pod schedulable.
|
// an unschedulable Pod schedulable.
|
||||||
// - Delete. An unschedulable Pod may fail due to violating an existing Pod's topology spread constraints,
|
// - Delete. An unschedulable Pod may fail due to violating an existing Pod's topology spread constraints,
|
||||||
// deleting an existing Pod may make it schedulable.
|
// deleting an existing Pod may make it schedulable.
|
||||||
{Event: framework.ClusterEvent{Resource: framework.Pod, ActionType: framework.All}},
|
{Event: framework.ClusterEvent{Resource: framework.Pod, ActionType: framework.All}, QueueingHintFn: pl.isSchedulableAfterPodChange},
|
||||||
// Node add|delete|updateLabel maybe lead an topology key changed,
|
// Node add|delete|update maybe lead an topology key changed,
|
||||||
// and make these pod in scheduling schedulable or unschedulable.
|
// and make these pod in scheduling schedulable or unschedulable.
|
||||||
{Event: framework.ClusterEvent{Resource: framework.Node, ActionType: framework.Add | framework.Delete | framework.UpdateNodeLabel}},
|
{Event: framework.ClusterEvent{Resource: framework.Node, ActionType: framework.Add | framework.Delete | framework.Update}, QueueingHintFn: pl.isSchedulableAfterNodeChange},
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func involvedInTopologySpreading(incomingPod, podWithSpreading *v1.Pod) bool {
|
||||||
|
return incomingPod.Spec.NodeName != "" && incomingPod.Namespace == podWithSpreading.Namespace
|
||||||
|
}
|
||||||
|
|
||||||
|
func (pl *PodTopologySpread) isSchedulableAfterPodChange(logger klog.Logger, pod *v1.Pod, oldObj, newObj interface{}) (framework.QueueingHint, error) {
|
||||||
|
originalPod, modifiedPod, err := util.As[*v1.Pod](oldObj, newObj)
|
||||||
|
if err != nil {
|
||||||
|
return framework.Queue, err
|
||||||
|
}
|
||||||
|
|
||||||
|
if (modifiedPod != nil && !involvedInTopologySpreading(modifiedPod, pod)) || (originalPod != nil && !involvedInTopologySpreading(originalPod, pod)) {
|
||||||
|
logger.V(5).Info("the added/updated/deleted pod is unscheduled or has different namespace with target pod, so it doesn't make the target pod schedulable",
|
||||||
|
"pod", klog.KObj(pod), "originalPod", klog.KObj(originalPod))
|
||||||
|
return framework.QueueSkip, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
constraints, err := pl.getConstraints(pod)
|
||||||
|
if err != nil {
|
||||||
|
return framework.Queue, err
|
||||||
|
}
|
||||||
|
|
||||||
|
// Pod is modified. Return Queue when the label(s) matching topologySpread's selector is added, changed, or deleted.
|
||||||
|
if modifiedPod != nil && originalPod != nil {
|
||||||
|
if reflect.DeepEqual(modifiedPod.Labels, originalPod.Labels) {
|
||||||
|
logger.V(5).Info("the updated pod is unscheduled or has no updated labels or has different namespace with target pod, so it doesn't make the target pod schedulable",
|
||||||
|
"pod", klog.KObj(pod), "modifiedPod", klog.KObj(modifiedPod))
|
||||||
|
return framework.QueueSkip, nil
|
||||||
|
}
|
||||||
|
for _, c := range constraints {
|
||||||
|
if c.Selector.Matches(labels.Set(originalPod.Labels)) != c.Selector.Matches(labels.Set(modifiedPod.Labels)) {
|
||||||
|
// This modification makes this Pod match(or not match) with this constraint.
|
||||||
|
// Maybe now the scheduling result of topology spread gets changed by this change.
|
||||||
|
logger.V(5).Info("a scheduled pod's label was updated and it makes the updated pod match or unmatch the pod's topology spread constraints",
|
||||||
|
"pod", klog.KObj(pod), "modifiedPod", klog.KObj(modifiedPod))
|
||||||
|
return framework.Queue, nil
|
||||||
|
}
|
||||||
|
}
|
||||||
|
// This modification of labels doesn't change whether this Pod would match selector or not in any constraints.
|
||||||
|
logger.V(5).Info("a scheduled pod's label was updated, but it's a change unrelated to the pod's topology spread constraints",
|
||||||
|
"pod", klog.KObj(pod), "modifiedPod", klog.KObj(modifiedPod))
|
||||||
|
return framework.QueueSkip, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// Pod is added. Return Queue when the added Pod has a label that matches with topologySpread's selector.
|
||||||
|
if originalPod != nil {
|
||||||
|
if podLabelsMatchSpreadConstraints(constraints, originalPod.Labels) {
|
||||||
|
logger.V(5).Info("a scheduled pod was created and it matches with the pod's topology spread constraints",
|
||||||
|
"pod", klog.KObj(pod), "createdPod", klog.KObj(originalPod))
|
||||||
|
return framework.Queue, nil
|
||||||
|
}
|
||||||
|
logger.V(5).Info("a scheduled pod was created, but it doesn't matches with the pod's topology spread constraints",
|
||||||
|
"pod", klog.KObj(pod), "createdPod", klog.KObj(originalPod))
|
||||||
|
return framework.QueueSkip, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// Pod is deleted. Return Queue when the deleted Pod has a label that matches with topologySpread's selector.
|
||||||
|
if podLabelsMatchSpreadConstraints(constraints, modifiedPod.Labels) {
|
||||||
|
logger.V(5).Info("a scheduled pod which matches with the pod's topology spread constraints was deleted, and the pod may be schedulable now",
|
||||||
|
"pod", klog.KObj(pod), "deletedPod", klog.KObj(modifiedPod))
|
||||||
|
return framework.Queue, nil
|
||||||
|
}
|
||||||
|
logger.V(5).Info("a scheduled pod was deleted, but it's unrelated to the pod's topology spread constraints",
|
||||||
|
"pod", klog.KObj(pod), "deletedPod", klog.KObj(modifiedPod))
|
||||||
|
|
||||||
|
return framework.QueueSkip, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// getConstraints extracts topologySpreadConstraint(s) from the Pod spec.
|
||||||
|
// If the Pod doesn't have any topologySpreadConstraint, it returns default constraints.
|
||||||
|
func (pl *PodTopologySpread) getConstraints(pod *v1.Pod) ([]topologySpreadConstraint, error) {
|
||||||
|
var constraints []topologySpreadConstraint
|
||||||
|
var err error
|
||||||
|
if len(pod.Spec.TopologySpreadConstraints) > 0 {
|
||||||
|
// We have feature gating in APIServer to strip the spec
|
||||||
|
// so don't need to re-check feature gate, just check length of Constraints.
|
||||||
|
constraints, err = pl.filterTopologySpreadConstraints(
|
||||||
|
pod.Spec.TopologySpreadConstraints,
|
||||||
|
pod.Labels,
|
||||||
|
v1.DoNotSchedule,
|
||||||
|
)
|
||||||
|
if err != nil {
|
||||||
|
return nil, fmt.Errorf("obtaining pod's hard topology spread constraints: %w", err)
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
constraints, err = pl.buildDefaultConstraints(pod, v1.DoNotSchedule)
|
||||||
|
if err != nil {
|
||||||
|
return nil, fmt.Errorf("setting default hard topology spread constraints: %w", err)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return constraints, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (pl *PodTopologySpread) isSchedulableAfterNodeChange(logger klog.Logger, pod *v1.Pod, oldObj, newObj interface{}) (framework.QueueingHint, error) {
|
||||||
|
originalNode, modifiedNode, err := util.As[*v1.Node](oldObj, newObj)
|
||||||
|
if err != nil {
|
||||||
|
return framework.Queue, err
|
||||||
|
}
|
||||||
|
|
||||||
|
constraints, err := pl.getConstraints(pod)
|
||||||
|
if err != nil {
|
||||||
|
return framework.Queue, err
|
||||||
|
}
|
||||||
|
|
||||||
|
// framework.Add/framework.Update: return Queue when node has topologyKey in its labels, else return QueueSkip.
|
||||||
|
//
|
||||||
|
// TODO: we can filter out node update events in a more fine-grained way once preCheck is completely removed.
|
||||||
|
// See: https://github.com/kubernetes/kubernetes/issues/110175
|
||||||
|
if modifiedNode != nil {
|
||||||
|
if !nodeLabelsMatchSpreadConstraints(modifiedNode.Labels, constraints) {
|
||||||
|
logger.V(5).Info("the created/updated node doesn't match pod topology spread constraints",
|
||||||
|
"pod", klog.KObj(pod), "node", klog.KObj(modifiedNode))
|
||||||
|
return framework.QueueSkip, nil
|
||||||
|
}
|
||||||
|
logger.V(5).Info("node that match topology spread constraints was created/updated, and the pod may be schedulable now",
|
||||||
|
"pod", klog.KObj(pod), "node", klog.KObj(modifiedNode))
|
||||||
|
return framework.Queue, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// framework.Delete: return Queue when node has topologyKey in its labels, else return QueueSkip.
|
||||||
|
if !nodeLabelsMatchSpreadConstraints(originalNode.Labels, constraints) {
|
||||||
|
logger.V(5).Info("the deleted node doesn't match pod topology spread constraints", "pod", klog.KObj(pod), "node", klog.KObj(originalNode))
|
||||||
|
return framework.QueueSkip, nil
|
||||||
|
}
|
||||||
|
logger.V(5).Info("node that match topology spread constraints was deleted, and the pod may be schedulable now",
|
||||||
|
"pod", klog.KObj(pod), "node", klog.KObj(originalNode))
|
||||||
|
return framework.Queue, nil
|
||||||
|
}
|
||||||
|
315
pkg/scheduler/framework/plugins/podtopologyspread/plugin_test.go
Normal file
315
pkg/scheduler/framework/plugins/podtopologyspread/plugin_test.go
Normal file
@ -0,0 +1,315 @@
|
|||||||
|
/*
|
||||||
|
Copyright 2023 The Kubernetes Authors.
|
||||||
|
|
||||||
|
Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
you may not use this file except in compliance with the License.
|
||||||
|
You may obtain a copy of the License at
|
||||||
|
|
||||||
|
http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
|
||||||
|
Unless required by applicable law or agreed to in writing, software
|
||||||
|
distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
See the License for the specific language governing permissions and
|
||||||
|
limitations under the License.
|
||||||
|
*/
|
||||||
|
|
||||||
|
package podtopologyspread
|
||||||
|
|
||||||
|
import (
|
||||||
|
"testing"
|
||||||
|
|
||||||
|
"github.com/stretchr/testify/require"
|
||||||
|
v1 "k8s.io/api/core/v1"
|
||||||
|
"k8s.io/klog/v2/ktesting"
|
||||||
|
"k8s.io/kubernetes/pkg/scheduler/apis/config"
|
||||||
|
"k8s.io/kubernetes/pkg/scheduler/framework"
|
||||||
|
plugintesting "k8s.io/kubernetes/pkg/scheduler/framework/plugins/testing"
|
||||||
|
"k8s.io/kubernetes/pkg/scheduler/internal/cache"
|
||||||
|
st "k8s.io/kubernetes/pkg/scheduler/testing"
|
||||||
|
)
|
||||||
|
|
||||||
|
func Test_isSchedulableAfterNodeChange(t *testing.T) {
|
||||||
|
testcases := []struct {
|
||||||
|
name string
|
||||||
|
pod *v1.Pod
|
||||||
|
oldNode, newNode *v1.Node
|
||||||
|
expectedHint framework.QueueingHint
|
||||||
|
expectedErr bool
|
||||||
|
}{
|
||||||
|
{
|
||||||
|
name: "node updates label which matches topologyKey",
|
||||||
|
pod: st.MakePod().Name("p").SpreadConstraint(1, "zone", v1.DoNotSchedule, fooSelector, nil, nil, nil, nil).
|
||||||
|
Obj(),
|
||||||
|
oldNode: st.MakeNode().Name("node-a").Label("zone", "zone1").Obj(),
|
||||||
|
newNode: st.MakeNode().Name("node-a").Label("zone", "zone2").Obj(),
|
||||||
|
expectedHint: framework.Queue,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "node that doesn't match topologySpreadConstraints updates non-related label",
|
||||||
|
pod: st.MakePod().Name("p").SpreadConstraint(1, "zone", v1.DoNotSchedule, fooSelector, nil, nil, nil, nil).
|
||||||
|
Obj(),
|
||||||
|
oldNode: st.MakeNode().Name("node-a").Label("foo", "bar1").Obj(),
|
||||||
|
newNode: st.MakeNode().Name("node-a").Label("foo", "bar2").Obj(),
|
||||||
|
expectedHint: framework.QueueSkip,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "node that match topologySpreadConstraints adds non-related label",
|
||||||
|
pod: st.MakePod().Name("p").
|
||||||
|
SpreadConstraint(1, "zone", v1.DoNotSchedule, fooSelector, nil, nil, nil, nil).
|
||||||
|
SpreadConstraint(1, "node", v1.DoNotSchedule, fooSelector, nil, nil, nil, nil).
|
||||||
|
Obj(),
|
||||||
|
oldNode: st.MakeNode().Name("node-a").Label("zone", "zone1").Label("node", "node1").Label("foo", "bar").Obj(),
|
||||||
|
newNode: st.MakeNode().Name("node-a").Label("zone", "zone1").Label("node", "node1").Obj(),
|
||||||
|
expectedHint: framework.Queue,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "create node with non-related labels",
|
||||||
|
pod: st.MakePod().Name("p").SpreadConstraint(1, "zone", v1.DoNotSchedule, fooSelector, nil, nil, nil, nil).
|
||||||
|
Obj(),
|
||||||
|
newNode: st.MakeNode().Name("node-a").Label("foo", "bar").Obj(),
|
||||||
|
expectedHint: framework.QueueSkip,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "create node with related labels",
|
||||||
|
pod: st.MakePod().Name("p").SpreadConstraint(1, "zone", v1.DoNotSchedule, fooSelector, nil, nil, nil, nil).
|
||||||
|
Obj(),
|
||||||
|
newNode: st.MakeNode().Name("node-a").Label("zone", "zone1").Obj(),
|
||||||
|
expectedHint: framework.Queue,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "delete node with non-related labels",
|
||||||
|
pod: st.MakePod().Name("p").SpreadConstraint(1, "zone", v1.DoNotSchedule, fooSelector, nil, nil, nil, nil).
|
||||||
|
Obj(),
|
||||||
|
oldNode: st.MakeNode().Name("node-a").Label("foo", "bar").Obj(),
|
||||||
|
expectedHint: framework.QueueSkip,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "delete node with related labels",
|
||||||
|
pod: st.MakePod().Name("p").SpreadConstraint(1, "zone", v1.DoNotSchedule, fooSelector, nil, nil, nil, nil).
|
||||||
|
Obj(),
|
||||||
|
oldNode: st.MakeNode().Name("node-a").Label("zone", "zone1").Obj(),
|
||||||
|
expectedHint: framework.Queue,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "add node with related labels that only match one of topologySpreadConstraints",
|
||||||
|
pod: st.MakePod().Name("p").
|
||||||
|
SpreadConstraint(1, "zone", v1.DoNotSchedule, fooSelector, nil, nil, nil, nil).
|
||||||
|
SpreadConstraint(1, "node", v1.DoNotSchedule, fooSelector, nil, nil, nil, nil).
|
||||||
|
Obj(),
|
||||||
|
newNode: st.MakeNode().Name("node-a").Label("zone", "zone1").Obj(),
|
||||||
|
expectedHint: framework.QueueSkip,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "add node with related labels that match all topologySpreadConstraints",
|
||||||
|
pod: st.MakePod().Name("p").
|
||||||
|
SpreadConstraint(1, "zone", v1.DoNotSchedule, fooSelector, nil, nil, nil, nil).
|
||||||
|
SpreadConstraint(1, "node", v1.DoNotSchedule, fooSelector, nil, nil, nil, nil).
|
||||||
|
Obj(),
|
||||||
|
newNode: st.MakeNode().Name("node-a").Label("zone", "zone1").Label("node", "node1").Obj(),
|
||||||
|
expectedHint: framework.Queue,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "update node with related labels that only match one of topologySpreadConstraints",
|
||||||
|
pod: st.MakePod().Name("p").
|
||||||
|
SpreadConstraint(1, "zone", v1.DoNotSchedule, fooSelector, nil, nil, nil, nil).
|
||||||
|
SpreadConstraint(1, "node", v1.DoNotSchedule, fooSelector, nil, nil, nil, nil).
|
||||||
|
Obj(),
|
||||||
|
oldNode: st.MakeNode().Name("node-a").Label("zone", "zone1").Obj(),
|
||||||
|
newNode: st.MakeNode().Name("node-a").Label("zone", "zone1").Obj(),
|
||||||
|
expectedHint: framework.QueueSkip,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "update node with related labels that match all topologySpreadConstraints",
|
||||||
|
pod: st.MakePod().Name("p").
|
||||||
|
SpreadConstraint(1, "zone", v1.DoNotSchedule, fooSelector, nil, nil, nil, nil).
|
||||||
|
SpreadConstraint(1, "node", v1.DoNotSchedule, fooSelector, nil, nil, nil, nil).
|
||||||
|
Obj(),
|
||||||
|
oldNode: st.MakeNode().Name("node-a").Label("zone", "zone1").Label("node", "node1").Obj(),
|
||||||
|
newNode: st.MakeNode().Name("node-a").Label("zone", "zone1").Label("node", "node2").Obj(),
|
||||||
|
expectedHint: framework.Queue,
|
||||||
|
},
|
||||||
|
}
|
||||||
|
|
||||||
|
for _, tc := range testcases {
|
||||||
|
t.Run(tc.name, func(t *testing.T) {
|
||||||
|
logger, ctx := ktesting.NewTestContext(t)
|
||||||
|
snapshot := cache.NewSnapshot(nil, nil)
|
||||||
|
pl := plugintesting.SetupPlugin(ctx, t, topologySpreadFunc, &config.PodTopologySpreadArgs{DefaultingType: config.ListDefaulting}, snapshot)
|
||||||
|
p := pl.(*PodTopologySpread)
|
||||||
|
actualHint, err := p.isSchedulableAfterNodeChange(logger, tc.pod, tc.oldNode, tc.newNode)
|
||||||
|
if tc.expectedErr {
|
||||||
|
require.Error(t, err)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
require.NoError(t, err)
|
||||||
|
require.Equal(t, tc.expectedHint, actualHint)
|
||||||
|
})
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func Test_isSchedulableAfterPodChange(t *testing.T) {
|
||||||
|
testcases := []struct {
|
||||||
|
name string
|
||||||
|
pod *v1.Pod
|
||||||
|
oldPod, newPod *v1.Pod
|
||||||
|
expectedHint framework.QueueingHint
|
||||||
|
expectedErr bool
|
||||||
|
}{
|
||||||
|
{
|
||||||
|
name: "add pod with labels match topologySpreadConstraints selector",
|
||||||
|
pod: st.MakePod().Name("p").Label("foo", "").
|
||||||
|
SpreadConstraint(1, "zone", v1.DoNotSchedule, fooSelector, nil, nil, nil, nil).
|
||||||
|
Obj(),
|
||||||
|
newPod: st.MakePod().Node("fake-node").Label("foo", "").Obj(),
|
||||||
|
expectedHint: framework.Queue,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "add un-scheduled pod",
|
||||||
|
pod: st.MakePod().Name("p").Label("foo", "").
|
||||||
|
SpreadConstraint(1, "zone", v1.DoNotSchedule, fooSelector, nil, nil, nil, nil).
|
||||||
|
Obj(),
|
||||||
|
newPod: st.MakePod().Label("foo", "").Obj(),
|
||||||
|
expectedHint: framework.QueueSkip,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "update un-scheduled pod",
|
||||||
|
pod: st.MakePod().Name("p").Label("foo", "").
|
||||||
|
SpreadConstraint(1, "zone", v1.DoNotSchedule, fooSelector, nil, nil, nil, nil).
|
||||||
|
Obj(),
|
||||||
|
newPod: st.MakePod().Label("foo", "").Obj(),
|
||||||
|
oldPod: st.MakePod().Label("bar", "").Obj(),
|
||||||
|
expectedHint: framework.QueueSkip,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "delete un-scheduled pod",
|
||||||
|
pod: st.MakePod().Name("p").Label("foo", "").
|
||||||
|
SpreadConstraint(1, "zone", v1.DoNotSchedule, fooSelector, nil, nil, nil, nil).
|
||||||
|
Obj(),
|
||||||
|
oldPod: st.MakePod().Label("foo", "").Obj(),
|
||||||
|
expectedHint: framework.QueueSkip,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "add pod with different namespace",
|
||||||
|
pod: st.MakePod().Name("p").Label("foo", "").
|
||||||
|
SpreadConstraint(1, "zone", v1.DoNotSchedule, fooSelector, nil, nil, nil, nil).
|
||||||
|
Obj(),
|
||||||
|
newPod: st.MakePod().Node("fake-node").Namespace("fake-namespace").Label("foo", "").Obj(),
|
||||||
|
expectedHint: framework.QueueSkip,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "add pod with labels don't match topologySpreadConstraints selector",
|
||||||
|
pod: st.MakePod().Name("p").Label("foo", "").
|
||||||
|
SpreadConstraint(1, "zone", v1.DoNotSchedule, fooSelector, nil, nil, nil, nil).
|
||||||
|
Obj(),
|
||||||
|
newPod: st.MakePod().Node("fake-node").Label("bar", "").Obj(),
|
||||||
|
expectedHint: framework.QueueSkip,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "delete pod with labels that match topologySpreadConstraints selector",
|
||||||
|
pod: st.MakePod().Name("p").Label("foo", "").
|
||||||
|
SpreadConstraint(1, "zone", v1.DoNotSchedule, fooSelector, nil, nil, nil, nil).
|
||||||
|
Obj(),
|
||||||
|
oldPod: st.MakePod().Node("fake-node").Label("foo", "").Obj(),
|
||||||
|
expectedHint: framework.Queue,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "delete pod with labels that don't match topologySpreadConstraints selector",
|
||||||
|
pod: st.MakePod().Name("p").Label("foo", "").
|
||||||
|
SpreadConstraint(1, "zone", v1.DoNotSchedule, fooSelector, nil, nil, nil, nil).
|
||||||
|
Obj(),
|
||||||
|
oldPod: st.MakePod().Node("fake-node").Label("bar", "").Obj(),
|
||||||
|
expectedHint: framework.QueueSkip,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "update pod's non-related label",
|
||||||
|
pod: st.MakePod().Name("p").Label("foo", "").
|
||||||
|
SpreadConstraint(1, "zone", v1.DoNotSchedule, fooSelector, nil, nil, nil, nil).
|
||||||
|
Obj(),
|
||||||
|
oldPod: st.MakePod().Node("fake-node").Label("foo", "").Label("bar", "bar1").Obj(),
|
||||||
|
newPod: st.MakePod().Node("fake-node").Label("foo", "").Label("bar", "bar2").Obj(),
|
||||||
|
expectedHint: framework.QueueSkip,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "add pod's label that matches topologySpreadConstraints selector",
|
||||||
|
pod: st.MakePod().Name("p").Label("foo", "").
|
||||||
|
SpreadConstraint(1, "zone", v1.DoNotSchedule, fooSelector, nil, nil, nil, nil).
|
||||||
|
Obj(),
|
||||||
|
oldPod: st.MakePod().Node("fake-node").Obj(),
|
||||||
|
newPod: st.MakePod().Node("fake-node").Label("foo", "").Obj(),
|
||||||
|
expectedHint: framework.Queue,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "delete pod label that matches topologySpreadConstraints selector",
|
||||||
|
pod: st.MakePod().Name("p").Label("foo", "").
|
||||||
|
SpreadConstraint(1, "zone", v1.DoNotSchedule, fooSelector, nil, nil, nil, nil).
|
||||||
|
Obj(),
|
||||||
|
oldPod: st.MakePod().Node("fake-node").Label("foo", "").Obj(),
|
||||||
|
newPod: st.MakePod().Node("fake-node").Obj(),
|
||||||
|
expectedHint: framework.Queue,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "change pod's label that matches topologySpreadConstraints selector",
|
||||||
|
pod: st.MakePod().Name("p").Label("foo", "").
|
||||||
|
SpreadConstraint(1, "zone", v1.DoNotSchedule, fooSelector, nil, nil, nil, nil).
|
||||||
|
Obj(),
|
||||||
|
oldPod: st.MakePod().Node("fake-node").Label("foo", "foo1").Obj(),
|
||||||
|
newPod: st.MakePod().Node("fake-node").Label("foo", "foo2").Obj(),
|
||||||
|
expectedHint: framework.QueueSkip,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "change pod's label that doesn't match topologySpreadConstraints selector",
|
||||||
|
pod: st.MakePod().Name("p").Label("foo", "").
|
||||||
|
SpreadConstraint(1, "zone", v1.DoNotSchedule, fooSelector, nil, nil, nil, nil).
|
||||||
|
Obj(),
|
||||||
|
oldPod: st.MakePod().Node("fake-node").Label("foo", "").Label("bar", "bar1").Obj(),
|
||||||
|
newPod: st.MakePod().Node("fake-node").Label("foo", "").Label("bar", "bar2").Obj(),
|
||||||
|
expectedHint: framework.QueueSkip,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "add pod's label that matches topologySpreadConstraints selector with multi topologySpreadConstraints",
|
||||||
|
pod: st.MakePod().Name("p").Label("foo", "").
|
||||||
|
SpreadConstraint(1, "zone", v1.DoNotSchedule, fooSelector, nil, nil, nil, nil).
|
||||||
|
SpreadConstraint(1, "node", v1.DoNotSchedule, barSelector, nil, nil, nil, nil).
|
||||||
|
Obj(),
|
||||||
|
oldPod: st.MakePod().Node("fake-node").Label("foo", "").Obj(),
|
||||||
|
newPod: st.MakePod().Node("fake-node").Label("foo", "").Label("bar", "bar2").Obj(),
|
||||||
|
expectedHint: framework.Queue,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "change pod's label that doesn't match topologySpreadConstraints selector with multi topologySpreadConstraints",
|
||||||
|
pod: st.MakePod().Name("p").Label("foo", "").
|
||||||
|
SpreadConstraint(1, "zone", v1.DoNotSchedule, fooSelector, nil, nil, nil, nil).
|
||||||
|
SpreadConstraint(1, "node", v1.DoNotSchedule, barSelector, nil, nil, nil, nil).
|
||||||
|
Obj(),
|
||||||
|
oldPod: st.MakePod().Node("fake-node").Label("foo", "").Obj(),
|
||||||
|
newPod: st.MakePod().Node("fake-node").Label("foo", "").Label("baz", "").Obj(),
|
||||||
|
expectedHint: framework.QueueSkip,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "change pod's label that match topologySpreadConstraints selector with multi topologySpreadConstraints",
|
||||||
|
pod: st.MakePod().Name("p").Label("foo", "").
|
||||||
|
SpreadConstraint(1, "zone", v1.DoNotSchedule, fooSelector, nil, nil, nil, nil).
|
||||||
|
SpreadConstraint(1, "node", v1.DoNotSchedule, barSelector, nil, nil, nil, nil).
|
||||||
|
Obj(),
|
||||||
|
oldPod: st.MakePod().Node("fake-node").Label("foo", "").Label("bar", "").Obj(),
|
||||||
|
newPod: st.MakePod().Node("fake-node").Label("foo", "").Label("bar", "bar2").Obj(),
|
||||||
|
expectedHint: framework.QueueSkip,
|
||||||
|
},
|
||||||
|
}
|
||||||
|
for _, tc := range testcases {
|
||||||
|
t.Run(tc.name, func(t *testing.T) {
|
||||||
|
logger, ctx := ktesting.NewTestContext(t)
|
||||||
|
snapshot := cache.NewSnapshot(nil, nil)
|
||||||
|
pl := plugintesting.SetupPlugin(ctx, t, topologySpreadFunc, &config.PodTopologySpreadArgs{DefaultingType: config.ListDefaulting}, snapshot)
|
||||||
|
p := pl.(*PodTopologySpread)
|
||||||
|
actualHint, err := p.isSchedulableAfterPodChange(logger, tc.pod, tc.oldPod, tc.newPod)
|
||||||
|
if tc.expectedErr {
|
||||||
|
require.Error(t, err)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
require.NoError(t, err)
|
||||||
|
require.Equal(t, tc.expectedHint, actualHint)
|
||||||
|
})
|
||||||
|
}
|
||||||
|
}
|
Loading…
Reference in New Issue
Block a user