mirror of
https://github.com/k3s-io/kubernetes.git
synced 2025-07-29 06:27:05 +00:00
fix: make updatePodOther private
This commit is contained in:
parent
0b133c7fa9
commit
0dee497876
@ -48,7 +48,7 @@ import (
|
|||||||
|
|
||||||
func (sched *Scheduler) onStorageClassAdd(obj interface{}) {
|
func (sched *Scheduler) onStorageClassAdd(obj interface{}) {
|
||||||
start := time.Now()
|
start := time.Now()
|
||||||
defer metrics.EventHandlingLatency.WithLabelValues(queue.StorageClassAdd.Label).Observe(metrics.SinceInSeconds(start))
|
defer metrics.EventHandlingLatency.WithLabelValues(framework.StorageClassAdd.Label).Observe(metrics.SinceInSeconds(start))
|
||||||
logger := sched.logger
|
logger := sched.logger
|
||||||
sc, ok := obj.(*storagev1.StorageClass)
|
sc, ok := obj.(*storagev1.StorageClass)
|
||||||
if !ok {
|
if !ok {
|
||||||
@ -63,13 +63,13 @@ func (sched *Scheduler) onStorageClassAdd(obj interface{}) {
|
|||||||
// We don't need to invalidate cached results because results will not be
|
// We don't need to invalidate cached results because results will not be
|
||||||
// cached for pod that has unbound immediate PVCs.
|
// cached for pod that has unbound immediate PVCs.
|
||||||
if sc.VolumeBindingMode != nil && *sc.VolumeBindingMode == storagev1.VolumeBindingWaitForFirstConsumer {
|
if sc.VolumeBindingMode != nil && *sc.VolumeBindingMode == storagev1.VolumeBindingWaitForFirstConsumer {
|
||||||
sched.SchedulingQueue.MoveAllToActiveOrBackoffQueue(logger, queue.StorageClassAdd, nil, sc, nil)
|
sched.SchedulingQueue.MoveAllToActiveOrBackoffQueue(logger, framework.StorageClassAdd, nil, sc, nil)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
func (sched *Scheduler) addNodeToCache(obj interface{}) {
|
func (sched *Scheduler) addNodeToCache(obj interface{}) {
|
||||||
start := time.Now()
|
start := time.Now()
|
||||||
defer metrics.EventHandlingLatency.WithLabelValues(queue.NodeAdd.Label).Observe(metrics.SinceInSeconds(start))
|
defer metrics.EventHandlingLatency.WithLabelValues(framework.NodeAdd.Label).Observe(metrics.SinceInSeconds(start))
|
||||||
logger := sched.logger
|
logger := sched.logger
|
||||||
node, ok := obj.(*v1.Node)
|
node, ok := obj.(*v1.Node)
|
||||||
if !ok {
|
if !ok {
|
||||||
@ -79,7 +79,7 @@ func (sched *Scheduler) addNodeToCache(obj interface{}) {
|
|||||||
|
|
||||||
logger.V(3).Info("Add event for node", "node", klog.KObj(node))
|
logger.V(3).Info("Add event for node", "node", klog.KObj(node))
|
||||||
nodeInfo := sched.Cache.AddNode(logger, node)
|
nodeInfo := sched.Cache.AddNode(logger, node)
|
||||||
sched.SchedulingQueue.MoveAllToActiveOrBackoffQueue(logger, queue.NodeAdd, nil, node, preCheckForNode(nodeInfo))
|
sched.SchedulingQueue.MoveAllToActiveOrBackoffQueue(logger, framework.NodeAdd, nil, node, preCheckForNode(nodeInfo))
|
||||||
}
|
}
|
||||||
|
|
||||||
func (sched *Scheduler) updateNodeInCache(oldObj, newObj interface{}) {
|
func (sched *Scheduler) updateNodeInCache(oldObj, newObj interface{}) {
|
||||||
@ -98,7 +98,7 @@ func (sched *Scheduler) updateNodeInCache(oldObj, newObj interface{}) {
|
|||||||
|
|
||||||
logger.V(4).Info("Update event for node", "node", klog.KObj(newNode))
|
logger.V(4).Info("Update event for node", "node", klog.KObj(newNode))
|
||||||
nodeInfo := sched.Cache.UpdateNode(logger, oldNode, newNode)
|
nodeInfo := sched.Cache.UpdateNode(logger, oldNode, newNode)
|
||||||
events := queue.NodeSchedulingPropertiesChange(newNode, oldNode)
|
events := framework.NodeSchedulingPropertiesChange(newNode, oldNode)
|
||||||
|
|
||||||
// Save the time it takes to update the node in the cache.
|
// Save the time it takes to update the node in the cache.
|
||||||
updatingDuration := metrics.SinceInSeconds(start)
|
updatingDuration := metrics.SinceInSeconds(start)
|
||||||
@ -115,7 +115,7 @@ func (sched *Scheduler) updateNodeInCache(oldObj, newObj interface{}) {
|
|||||||
|
|
||||||
func (sched *Scheduler) deleteNodeFromCache(obj interface{}) {
|
func (sched *Scheduler) deleteNodeFromCache(obj interface{}) {
|
||||||
start := time.Now()
|
start := time.Now()
|
||||||
defer metrics.EventHandlingLatency.WithLabelValues(queue.NodeDelete.Label).Observe(metrics.SinceInSeconds(start))
|
defer metrics.EventHandlingLatency.WithLabelValues(framework.NodeDelete.Label).Observe(metrics.SinceInSeconds(start))
|
||||||
|
|
||||||
logger := sched.logger
|
logger := sched.logger
|
||||||
var node *v1.Node
|
var node *v1.Node
|
||||||
@ -142,7 +142,7 @@ func (sched *Scheduler) deleteNodeFromCache(obj interface{}) {
|
|||||||
|
|
||||||
func (sched *Scheduler) addPodToSchedulingQueue(obj interface{}) {
|
func (sched *Scheduler) addPodToSchedulingQueue(obj interface{}) {
|
||||||
start := time.Now()
|
start := time.Now()
|
||||||
defer metrics.EventHandlingLatency.WithLabelValues(queue.UnscheduledPodAdd.Label).Observe(metrics.SinceInSeconds(start))
|
defer metrics.EventHandlingLatency.WithLabelValues(framework.UnscheduledPodAdd.Label).Observe(metrics.SinceInSeconds(start))
|
||||||
|
|
||||||
logger := sched.logger
|
logger := sched.logger
|
||||||
pod := obj.(*v1.Pod)
|
pod := obj.(*v1.Pod)
|
||||||
@ -154,7 +154,7 @@ func (sched *Scheduler) addPodToSchedulingQueue(obj interface{}) {
|
|||||||
|
|
||||||
func (sched *Scheduler) updatePodInSchedulingQueue(oldObj, newObj interface{}) {
|
func (sched *Scheduler) updatePodInSchedulingQueue(oldObj, newObj interface{}) {
|
||||||
start := time.Now()
|
start := time.Now()
|
||||||
defer metrics.EventHandlingLatency.WithLabelValues(queue.UnscheduledPodUpdate.Label).Observe(metrics.SinceInSeconds(start))
|
defer metrics.EventHandlingLatency.WithLabelValues(framework.UnscheduledPodUpdate.Label).Observe(metrics.SinceInSeconds(start))
|
||||||
logger := sched.logger
|
logger := sched.logger
|
||||||
oldPod, newPod := oldObj.(*v1.Pod), newObj.(*v1.Pod)
|
oldPod, newPod := oldObj.(*v1.Pod), newObj.(*v1.Pod)
|
||||||
// Bypass update event that carries identical objects; otherwise, a duplicated
|
// Bypass update event that carries identical objects; otherwise, a duplicated
|
||||||
@ -179,7 +179,7 @@ func (sched *Scheduler) updatePodInSchedulingQueue(oldObj, newObj interface{}) {
|
|||||||
|
|
||||||
func (sched *Scheduler) deletePodFromSchedulingQueue(obj interface{}) {
|
func (sched *Scheduler) deletePodFromSchedulingQueue(obj interface{}) {
|
||||||
start := time.Now()
|
start := time.Now()
|
||||||
defer metrics.EventHandlingLatency.WithLabelValues(queue.UnscheduledPodDelete.Label).Observe(metrics.SinceInSeconds(start))
|
defer metrics.EventHandlingLatency.WithLabelValues(framework.UnscheduledPodDelete.Label).Observe(metrics.SinceInSeconds(start))
|
||||||
|
|
||||||
logger := sched.logger
|
logger := sched.logger
|
||||||
var pod *v1.Pod
|
var pod *v1.Pod
|
||||||
@ -213,13 +213,13 @@ func (sched *Scheduler) deletePodFromSchedulingQueue(obj interface{}) {
|
|||||||
// removing it from the scheduler cache. In this case, signal a AssignedPodDelete
|
// removing it from the scheduler cache. In this case, signal a AssignedPodDelete
|
||||||
// event to immediately retry some unscheduled Pods.
|
// event to immediately retry some unscheduled Pods.
|
||||||
if fwk.RejectWaitingPod(pod.UID) {
|
if fwk.RejectWaitingPod(pod.UID) {
|
||||||
sched.SchedulingQueue.MoveAllToActiveOrBackoffQueue(logger, queue.AssignedPodDelete, pod, nil, nil)
|
sched.SchedulingQueue.MoveAllToActiveOrBackoffQueue(logger, framework.AssignedPodDelete, pod, nil, nil)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
func (sched *Scheduler) addPodToCache(obj interface{}) {
|
func (sched *Scheduler) addPodToCache(obj interface{}) {
|
||||||
start := time.Now()
|
start := time.Now()
|
||||||
defer metrics.EventHandlingLatency.WithLabelValues(queue.AssignedPodAdd.Label).Observe(metrics.SinceInSeconds(start))
|
defer metrics.EventHandlingLatency.WithLabelValues(framework.AssignedPodAdd.Label).Observe(metrics.SinceInSeconds(start))
|
||||||
|
|
||||||
logger := sched.logger
|
logger := sched.logger
|
||||||
pod, ok := obj.(*v1.Pod)
|
pod, ok := obj.(*v1.Pod)
|
||||||
@ -243,7 +243,7 @@ func (sched *Scheduler) addPodToCache(obj interface{}) {
|
|||||||
// Here we use MoveAllToActiveOrBackoffQueue only when QueueingHint is enabled.
|
// Here we use MoveAllToActiveOrBackoffQueue only when QueueingHint is enabled.
|
||||||
// (We cannot switch to MoveAllToActiveOrBackoffQueue right away because of throughput concern.)
|
// (We cannot switch to MoveAllToActiveOrBackoffQueue right away because of throughput concern.)
|
||||||
if utilfeature.DefaultFeatureGate.Enabled(features.SchedulerQueueingHints) {
|
if utilfeature.DefaultFeatureGate.Enabled(features.SchedulerQueueingHints) {
|
||||||
sched.SchedulingQueue.MoveAllToActiveOrBackoffQueue(logger, queue.AssignedPodAdd, nil, pod, nil)
|
sched.SchedulingQueue.MoveAllToActiveOrBackoffQueue(logger, framework.AssignedPodAdd, nil, pod, nil)
|
||||||
} else {
|
} else {
|
||||||
sched.SchedulingQueue.AssignedPodAdded(logger, pod)
|
sched.SchedulingQueue.AssignedPodAdded(logger, pod)
|
||||||
}
|
}
|
||||||
@ -251,7 +251,7 @@ func (sched *Scheduler) addPodToCache(obj interface{}) {
|
|||||||
|
|
||||||
func (sched *Scheduler) updatePodInCache(oldObj, newObj interface{}) {
|
func (sched *Scheduler) updatePodInCache(oldObj, newObj interface{}) {
|
||||||
start := time.Now()
|
start := time.Now()
|
||||||
defer metrics.EventHandlingLatency.WithLabelValues(queue.AssignedPodUpdate.Label).Observe(metrics.SinceInSeconds(start))
|
defer metrics.EventHandlingLatency.WithLabelValues(framework.AssignedPodUpdate.Label).Observe(metrics.SinceInSeconds(start))
|
||||||
|
|
||||||
logger := sched.logger
|
logger := sched.logger
|
||||||
oldPod, ok := oldObj.(*v1.Pod)
|
oldPod, ok := oldObj.(*v1.Pod)
|
||||||
@ -270,7 +270,7 @@ func (sched *Scheduler) updatePodInCache(oldObj, newObj interface{}) {
|
|||||||
logger.Error(err, "Scheduler cache UpdatePod failed", "pod", klog.KObj(oldPod))
|
logger.Error(err, "Scheduler cache UpdatePod failed", "pod", klog.KObj(oldPod))
|
||||||
}
|
}
|
||||||
|
|
||||||
events := queue.PodSchedulingPropertiesChange(newPod, oldPod)
|
events := framework.PodSchedulingPropertiesChange(newPod, oldPod)
|
||||||
for _, evt := range events {
|
for _, evt := range events {
|
||||||
// SchedulingQueue.AssignedPodUpdated has a problem:
|
// SchedulingQueue.AssignedPodUpdated has a problem:
|
||||||
// It internally pre-filters Pods to move to activeQ,
|
// It internally pre-filters Pods to move to activeQ,
|
||||||
@ -282,7 +282,7 @@ func (sched *Scheduler) updatePodInCache(oldObj, newObj interface{}) {
|
|||||||
// Here we use MoveAllToActiveOrBackoffQueue only when QueueingHint is enabled.
|
// Here we use MoveAllToActiveOrBackoffQueue only when QueueingHint is enabled.
|
||||||
// (We cannot switch to MoveAllToActiveOrBackoffQueue right away because of throughput concern.)
|
// (We cannot switch to MoveAllToActiveOrBackoffQueue right away because of throughput concern.)
|
||||||
if utilfeature.DefaultFeatureGate.Enabled(features.SchedulerQueueingHints) {
|
if utilfeature.DefaultFeatureGate.Enabled(features.SchedulerQueueingHints) {
|
||||||
sched.SchedulingQueue.MoveAllToActiveOrBackoffQueue(logger, queue.AssignedPodUpdate, oldPod, newPod, nil)
|
sched.SchedulingQueue.MoveAllToActiveOrBackoffQueue(logger, framework.AssignedPodUpdate, oldPod, newPod, nil)
|
||||||
} else {
|
} else {
|
||||||
sched.SchedulingQueue.AssignedPodUpdated(logger, oldPod, newPod, evt)
|
sched.SchedulingQueue.AssignedPodUpdated(logger, oldPod, newPod, evt)
|
||||||
}
|
}
|
||||||
@ -291,7 +291,7 @@ func (sched *Scheduler) updatePodInCache(oldObj, newObj interface{}) {
|
|||||||
|
|
||||||
func (sched *Scheduler) deletePodFromCache(obj interface{}) {
|
func (sched *Scheduler) deletePodFromCache(obj interface{}) {
|
||||||
start := time.Now()
|
start := time.Now()
|
||||||
defer metrics.EventHandlingLatency.WithLabelValues(queue.AssignedPodDelete.Label).Observe(metrics.SinceInSeconds(start))
|
defer metrics.EventHandlingLatency.WithLabelValues(framework.AssignedPodDelete.Label).Observe(metrics.SinceInSeconds(start))
|
||||||
|
|
||||||
logger := sched.logger
|
logger := sched.logger
|
||||||
var pod *v1.Pod
|
var pod *v1.Pod
|
||||||
@ -315,7 +315,7 @@ func (sched *Scheduler) deletePodFromCache(obj interface{}) {
|
|||||||
logger.Error(err, "Scheduler cache RemovePod failed", "pod", klog.KObj(pod))
|
logger.Error(err, "Scheduler cache RemovePod failed", "pod", klog.KObj(pod))
|
||||||
}
|
}
|
||||||
|
|
||||||
sched.SchedulingQueue.MoveAllToActiveOrBackoffQueue(logger, queue.AssignedPodDelete, pod, nil, nil)
|
sched.SchedulingQueue.MoveAllToActiveOrBackoffQueue(logger, framework.AssignedPodDelete, pod, nil, nil)
|
||||||
}
|
}
|
||||||
|
|
||||||
// assignedPod selects pods that are assigned (scheduled and running).
|
// assignedPod selects pods that are assigned (scheduled and running).
|
||||||
@ -578,8 +578,8 @@ func addAllEventHandlers(
|
|||||||
cache.ResourceEventHandlerFuncs{
|
cache.ResourceEventHandlerFuncs{
|
||||||
UpdateFunc: func(old, obj interface{}) {
|
UpdateFunc: func(old, obj interface{}) {
|
||||||
start := time.Now()
|
start := time.Now()
|
||||||
defer metrics.EventHandlingLatency.WithLabelValues(queue.StorageClassUpdate.Label).Observe(metrics.SinceInSeconds(start))
|
defer metrics.EventHandlingLatency.WithLabelValues(framework.StorageClassUpdate.Label).Observe(metrics.SinceInSeconds(start))
|
||||||
sched.SchedulingQueue.MoveAllToActiveOrBackoffQueue(logger, queue.StorageClassUpdate, old, obj, nil)
|
sched.SchedulingQueue.MoveAllToActiveOrBackoffQueue(logger, framework.StorageClassUpdate, old, obj, nil)
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
); err != nil {
|
); err != nil {
|
||||||
|
@ -14,7 +14,7 @@ See the License for the specific language governing permissions and
|
|||||||
limitations under the License.
|
limitations under the License.
|
||||||
*/
|
*/
|
||||||
|
|
||||||
package queue
|
package framework
|
||||||
|
|
||||||
import (
|
import (
|
||||||
v1 "k8s.io/api/core/v1"
|
v1 "k8s.io/api/core/v1"
|
||||||
@ -22,7 +22,6 @@ import (
|
|||||||
utilfeature "k8s.io/apiserver/pkg/util/feature"
|
utilfeature "k8s.io/apiserver/pkg/util/feature"
|
||||||
"k8s.io/kubernetes/pkg/api/v1/resource"
|
"k8s.io/kubernetes/pkg/api/v1/resource"
|
||||||
"k8s.io/kubernetes/pkg/features"
|
"k8s.io/kubernetes/pkg/features"
|
||||||
"k8s.io/kubernetes/pkg/scheduler/framework"
|
|
||||||
)
|
)
|
||||||
|
|
||||||
const (
|
const (
|
||||||
@ -41,72 +40,72 @@ const (
|
|||||||
|
|
||||||
var (
|
var (
|
||||||
// AssignedPodAdd is the event when an assigned pod is added.
|
// AssignedPodAdd is the event when an assigned pod is added.
|
||||||
AssignedPodAdd = framework.ClusterEvent{Resource: framework.Pod, ActionType: framework.Add, Label: "AssignedPodAdd"}
|
AssignedPodAdd = ClusterEvent{Resource: Pod, ActionType: Add, Label: "AssignedPodAdd"}
|
||||||
// NodeAdd is the event when a new node is added to the cluster.
|
// NodeAdd is the event when a new node is added to the cluster.
|
||||||
NodeAdd = framework.ClusterEvent{Resource: framework.Node, ActionType: framework.Add, Label: "NodeAdd"}
|
NodeAdd = ClusterEvent{Resource: Node, ActionType: Add, Label: "NodeAdd"}
|
||||||
// NodeDelete is the event when a node is deleted from the cluster.
|
// NodeDelete is the event when a node is deleted from the cluster.
|
||||||
NodeDelete = framework.ClusterEvent{Resource: framework.Node, ActionType: framework.Delete, Label: "NodeDelete"}
|
NodeDelete = ClusterEvent{Resource: Node, ActionType: Delete, Label: "NodeDelete"}
|
||||||
// AssignedPodUpdate is the event when an assigned pod is updated.
|
// AssignedPodUpdate is the event when an assigned pod is updated.
|
||||||
AssignedPodUpdate = framework.ClusterEvent{Resource: framework.Pod, ActionType: framework.Update, Label: "AssignedPodUpdate"}
|
AssignedPodUpdate = ClusterEvent{Resource: Pod, ActionType: Update, Label: "AssignedPodUpdate"}
|
||||||
// UnscheduledPodAdd is the event when an unscheduled pod is added.
|
// UnscheduledPodAdd is the event when an unscheduled pod is added.
|
||||||
UnscheduledPodAdd = framework.ClusterEvent{Resource: framework.Pod, ActionType: framework.Update, Label: "UnschedulablePodAdd"}
|
UnscheduledPodAdd = ClusterEvent{Resource: Pod, ActionType: Update, Label: "UnschedulablePodAdd"}
|
||||||
// AssignedPodOtherUpdate is the event when an assigned pod got updated in fields that are not covered by framework.UpdatePodXXX.
|
|
||||||
AssignedPodOtherUpdate = framework.ClusterEvent{Resource: framework.Pod, ActionType: framework.UpdatePodOther, Label: "AssignedPodUpdate"}
|
|
||||||
// UnscheduledPodUpdate is the event when an unscheduled pod is updated.
|
// UnscheduledPodUpdate is the event when an unscheduled pod is updated.
|
||||||
UnscheduledPodUpdate = framework.ClusterEvent{Resource: framework.Pod, ActionType: framework.Update, Label: "UnschedulablePodUpdate"}
|
UnscheduledPodUpdate = ClusterEvent{Resource: Pod, ActionType: Update, Label: "UnschedulablePodUpdate"}
|
||||||
// UnscheduledPodDelete is the event when an unscheduled pod is deleted.
|
// UnscheduledPodDelete is the event when an unscheduled pod is deleted.
|
||||||
UnscheduledPodDelete = framework.ClusterEvent{Resource: framework.Pod, ActionType: framework.Update, Label: "UnschedulablePodDelete"}
|
UnscheduledPodDelete = ClusterEvent{Resource: Pod, ActionType: Update, Label: "UnschedulablePodDelete"}
|
||||||
|
// assignedPodOtherUpdate is the event when an assigned pod got updated in fields that are not covered by UpdatePodXXX.
|
||||||
|
assignedPodOtherUpdate = ClusterEvent{Resource: Pod, ActionType: updatePodOther, Label: "AssignedPodUpdate"}
|
||||||
// AssignedPodDelete is the event when an assigned pod is deleted.
|
// AssignedPodDelete is the event when an assigned pod is deleted.
|
||||||
AssignedPodDelete = framework.ClusterEvent{Resource: framework.Pod, ActionType: framework.Delete, Label: "AssignedPodDelete"}
|
AssignedPodDelete = ClusterEvent{Resource: Pod, ActionType: Delete, Label: "AssignedPodDelete"}
|
||||||
// PodRequestChange is the event when a pod's resource request is changed.
|
// PodRequestChange is the event when a pod's resource request is changed.
|
||||||
PodRequestChange = framework.ClusterEvent{Resource: framework.Pod, ActionType: framework.UpdatePodRequest, Label: "PodRequestChange"}
|
PodRequestChange = ClusterEvent{Resource: Pod, ActionType: UpdatePodRequest, Label: "PodRequestChange"}
|
||||||
// PodLabelChange is the event when a pod's label is changed.
|
// PodLabelChange is the event when a pod's label is changed.
|
||||||
PodLabelChange = framework.ClusterEvent{Resource: framework.Pod, ActionType: framework.UpdatePodLabel, Label: "PodLabelChange"}
|
PodLabelChange = ClusterEvent{Resource: Pod, ActionType: UpdatePodLabel, Label: "PodLabelChange"}
|
||||||
// NodeSpecUnschedulableChange is the event when unschedulable node spec is changed.
|
// NodeSpecUnschedulableChange is the event when unschedulable node spec is changed.
|
||||||
NodeSpecUnschedulableChange = framework.ClusterEvent{Resource: framework.Node, ActionType: framework.UpdateNodeTaint, Label: "NodeSpecUnschedulableChange"}
|
NodeSpecUnschedulableChange = ClusterEvent{Resource: Node, ActionType: UpdateNodeTaint, Label: "NodeSpecUnschedulableChange"}
|
||||||
// NodeAllocatableChange is the event when node allocatable is changed.
|
// NodeAllocatableChange is the event when node allocatable is changed.
|
||||||
NodeAllocatableChange = framework.ClusterEvent{Resource: framework.Node, ActionType: framework.UpdateNodeAllocatable, Label: "NodeAllocatableChange"}
|
NodeAllocatableChange = ClusterEvent{Resource: Node, ActionType: UpdateNodeAllocatable, Label: "NodeAllocatableChange"}
|
||||||
// NodeLabelChange is the event when node label is changed.
|
// NodeLabelChange is the event when node label is changed.
|
||||||
NodeLabelChange = framework.ClusterEvent{Resource: framework.Node, ActionType: framework.UpdateNodeLabel, Label: "NodeLabelChange"}
|
NodeLabelChange = ClusterEvent{Resource: Node, ActionType: UpdateNodeLabel, Label: "NodeLabelChange"}
|
||||||
// NodeAnnotationChange is the event when node annotation is changed.
|
// NodeAnnotationChange is the event when node annotation is changed.
|
||||||
NodeAnnotationChange = framework.ClusterEvent{Resource: framework.Node, ActionType: framework.UpdateNodeAnnotation, Label: "NodeAnnotationChange"}
|
NodeAnnotationChange = ClusterEvent{Resource: Node, ActionType: UpdateNodeAnnotation, Label: "NodeAnnotationChange"}
|
||||||
// NodeTaintChange is the event when node taint is changed.
|
// NodeTaintChange is the event when node taint is changed.
|
||||||
NodeTaintChange = framework.ClusterEvent{Resource: framework.Node, ActionType: framework.UpdateNodeTaint, Label: "NodeTaintChange"}
|
NodeTaintChange = ClusterEvent{Resource: Node, ActionType: UpdateNodeTaint, Label: "NodeTaintChange"}
|
||||||
// NodeConditionChange is the event when node condition is changed.
|
// NodeConditionChange is the event when node condition is changed.
|
||||||
NodeConditionChange = framework.ClusterEvent{Resource: framework.Node, ActionType: framework.UpdateNodeCondition, Label: "NodeConditionChange"}
|
NodeConditionChange = ClusterEvent{Resource: Node, ActionType: UpdateNodeCondition, Label: "NodeConditionChange"}
|
||||||
// PvAdd is the event when a persistent volume is added in the cluster.
|
// PvAdd is the event when a persistent volume is added in the cluster.
|
||||||
PvAdd = framework.ClusterEvent{Resource: framework.PersistentVolume, ActionType: framework.Add, Label: "PvAdd"}
|
PvAdd = ClusterEvent{Resource: PersistentVolume, ActionType: Add, Label: "PvAdd"}
|
||||||
// PvUpdate is the event when a persistent volume is updated in the cluster.
|
// PvUpdate is the event when a persistent volume is updated in the cluster.
|
||||||
PvUpdate = framework.ClusterEvent{Resource: framework.PersistentVolume, ActionType: framework.Update, Label: "PvUpdate"}
|
PvUpdate = ClusterEvent{Resource: PersistentVolume, ActionType: Update, Label: "PvUpdate"}
|
||||||
// PvcAdd is the event when a persistent volume claim is added in the cluster.
|
// PvcAdd is the event when a persistent volume claim is added in the cluster.
|
||||||
PvcAdd = framework.ClusterEvent{Resource: framework.PersistentVolumeClaim, ActionType: framework.Add, Label: "PvcAdd"}
|
PvcAdd = ClusterEvent{Resource: PersistentVolumeClaim, ActionType: Add, Label: "PvcAdd"}
|
||||||
// PvcUpdate is the event when a persistent volume claim is updated in the cluster.
|
// PvcUpdate is the event when a persistent volume claim is updated in the cluster.
|
||||||
PvcUpdate = framework.ClusterEvent{Resource: framework.PersistentVolumeClaim, ActionType: framework.Update, Label: "PvcUpdate"}
|
PvcUpdate = ClusterEvent{Resource: PersistentVolumeClaim, ActionType: Update, Label: "PvcUpdate"}
|
||||||
// StorageClassAdd is the event when a StorageClass is added in the cluster.
|
// StorageClassAdd is the event when a StorageClass is added in the cluster.
|
||||||
StorageClassAdd = framework.ClusterEvent{Resource: framework.StorageClass, ActionType: framework.Add, Label: "StorageClassAdd"}
|
StorageClassAdd = ClusterEvent{Resource: StorageClass, ActionType: Add, Label: "StorageClassAdd"}
|
||||||
// StorageClassUpdate is the event when a StorageClass is updated in the cluster.
|
// StorageClassUpdate is the event when a StorageClass is updated in the cluster.
|
||||||
StorageClassUpdate = framework.ClusterEvent{Resource: framework.StorageClass, ActionType: framework.Update, Label: "StorageClassUpdate"}
|
StorageClassUpdate = ClusterEvent{Resource: StorageClass, ActionType: Update, Label: "StorageClassUpdate"}
|
||||||
// CSINodeAdd is the event when a CSI node is added in the cluster.
|
// CSINodeAdd is the event when a CSI node is added in the cluster.
|
||||||
CSINodeAdd = framework.ClusterEvent{Resource: framework.CSINode, ActionType: framework.Add, Label: "CSINodeAdd"}
|
CSINodeAdd = ClusterEvent{Resource: CSINode, ActionType: Add, Label: "CSINodeAdd"}
|
||||||
// CSINodeUpdate is the event when a CSI node is updated in the cluster.
|
// CSINodeUpdate is the event when a CSI node is updated in the cluster.
|
||||||
CSINodeUpdate = framework.ClusterEvent{Resource: framework.CSINode, ActionType: framework.Update, Label: "CSINodeUpdate"}
|
CSINodeUpdate = ClusterEvent{Resource: CSINode, ActionType: Update, Label: "CSINodeUpdate"}
|
||||||
// CSIDriverAdd is the event when a CSI driver is added in the cluster.
|
// CSIDriverAdd is the event when a CSI driver is added in the cluster.
|
||||||
CSIDriverAdd = framework.ClusterEvent{Resource: framework.CSIDriver, ActionType: framework.Add, Label: "CSIDriverAdd"}
|
CSIDriverAdd = ClusterEvent{Resource: CSIDriver, ActionType: Add, Label: "CSIDriverAdd"}
|
||||||
// CSIDriverUpdate is the event when a CSI driver is updated in the cluster.
|
// CSIDriverUpdate is the event when a CSI driver is updated in the cluster.
|
||||||
CSIDriverUpdate = framework.ClusterEvent{Resource: framework.CSIDriver, ActionType: framework.Update, Label: "CSIDriverUpdate"}
|
CSIDriverUpdate = ClusterEvent{Resource: CSIDriver, ActionType: Update, Label: "CSIDriverUpdate"}
|
||||||
// CSIStorageCapacityAdd is the event when a CSI storage capacity is added in the cluster.
|
// CSIStorageCapacityAdd is the event when a CSI storage capacity is added in the cluster.
|
||||||
CSIStorageCapacityAdd = framework.ClusterEvent{Resource: framework.CSIStorageCapacity, ActionType: framework.Add, Label: "CSIStorageCapacityAdd"}
|
CSIStorageCapacityAdd = ClusterEvent{Resource: CSIStorageCapacity, ActionType: Add, Label: "CSIStorageCapacityAdd"}
|
||||||
// CSIStorageCapacityUpdate is the event when a CSI storage capacity is updated in the cluster.
|
// CSIStorageCapacityUpdate is the event when a CSI storage capacity is updated in the cluster.
|
||||||
CSIStorageCapacityUpdate = framework.ClusterEvent{Resource: framework.CSIStorageCapacity, ActionType: framework.Update, Label: "CSIStorageCapacityUpdate"}
|
CSIStorageCapacityUpdate = ClusterEvent{Resource: CSIStorageCapacity, ActionType: Update, Label: "CSIStorageCapacityUpdate"}
|
||||||
// WildCardEvent semantically matches all resources on all actions.
|
// WildCardEvent semantically matches all resources on all actions.
|
||||||
WildCardEvent = framework.ClusterEvent{Resource: framework.WildCard, ActionType: framework.All, Label: "WildCardEvent"}
|
WildCardEvent = ClusterEvent{Resource: WildCard, ActionType: All, Label: "WildCardEvent"}
|
||||||
// UnschedulableTimeout is the event when a pod stays in unschedulable for longer than timeout.
|
// UnschedulableTimeout is the event when a pod stays in unschedulable for longer than timeout.
|
||||||
UnschedulableTimeout = framework.ClusterEvent{Resource: framework.WildCard, ActionType: framework.All, Label: "UnschedulableTimeout"}
|
UnschedulableTimeout = ClusterEvent{Resource: WildCard, ActionType: All, Label: "UnschedulableTimeout"}
|
||||||
)
|
)
|
||||||
|
|
||||||
// PodSchedulingPropertiesChange interprets the update of a pod and returns corresponding UpdatePodXYZ event(s).
|
// PodSchedulingPropertiesChange interprets the update of a pod and returns corresponding UpdatePodXYZ event(s).
|
||||||
// Once we have other pod update events, we should update here as well.
|
// Once we have other pod update events, we should update here as well.
|
||||||
func PodSchedulingPropertiesChange(newPod *v1.Pod, oldPod *v1.Pod) (events []framework.ClusterEvent) {
|
func PodSchedulingPropertiesChange(newPod *v1.Pod, oldPod *v1.Pod) (events []ClusterEvent) {
|
||||||
podChangeExtracters := []podChangeExtractor{
|
podChangeExtracters := []podChangeExtractor{
|
||||||
extractPodLabelsChange,
|
extractPodLabelsChange,
|
||||||
extractPodResourceRequestChange,
|
extractPodResourceRequestChange,
|
||||||
@ -121,15 +120,15 @@ func PodSchedulingPropertiesChange(newPod *v1.Pod, oldPod *v1.Pod) (events []fra
|
|||||||
if len(events) == 0 {
|
if len(events) == 0 {
|
||||||
// When no specific event is found, we use AssignedPodOtherUpdate,
|
// When no specific event is found, we use AssignedPodOtherUpdate,
|
||||||
// which should only trigger plugins registering a general Pod/Update event.
|
// which should only trigger plugins registering a general Pod/Update event.
|
||||||
events = append(events, AssignedPodOtherUpdate)
|
events = append(events, assignedPodOtherUpdate)
|
||||||
}
|
}
|
||||||
|
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
type podChangeExtractor func(newNode *v1.Pod, oldNode *v1.Pod) *framework.ClusterEvent
|
type podChangeExtractor func(newNode *v1.Pod, oldNode *v1.Pod) *ClusterEvent
|
||||||
|
|
||||||
func extractPodResourceRequestChange(newPod, oldPod *v1.Pod) *framework.ClusterEvent {
|
func extractPodResourceRequestChange(newPod, oldPod *v1.Pod) *ClusterEvent {
|
||||||
opt := resource.PodResourcesOptions{
|
opt := resource.PodResourcesOptions{
|
||||||
InPlacePodVerticalScalingEnabled: utilfeature.DefaultFeatureGate.Enabled(features.InPlacePodVerticalScaling),
|
InPlacePodVerticalScalingEnabled: utilfeature.DefaultFeatureGate.Enabled(features.InPlacePodVerticalScaling),
|
||||||
}
|
}
|
||||||
@ -139,7 +138,7 @@ func extractPodResourceRequestChange(newPod, oldPod *v1.Pod) *framework.ClusterE
|
|||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func extractPodLabelsChange(newPod *v1.Pod, oldPod *v1.Pod) *framework.ClusterEvent {
|
func extractPodLabelsChange(newPod *v1.Pod, oldPod *v1.Pod) *ClusterEvent {
|
||||||
if isLabelChanged(newPod.GetLabels(), oldPod.GetLabels()) {
|
if isLabelChanged(newPod.GetLabels(), oldPod.GetLabels()) {
|
||||||
return &PodLabelChange
|
return &PodLabelChange
|
||||||
}
|
}
|
||||||
@ -147,7 +146,7 @@ func extractPodLabelsChange(newPod *v1.Pod, oldPod *v1.Pod) *framework.ClusterEv
|
|||||||
}
|
}
|
||||||
|
|
||||||
// NodeSchedulingPropertiesChange interprets the update of a node and returns corresponding UpdateNodeXYZ event(s).
|
// NodeSchedulingPropertiesChange interprets the update of a node and returns corresponding UpdateNodeXYZ event(s).
|
||||||
func NodeSchedulingPropertiesChange(newNode *v1.Node, oldNode *v1.Node) (events []framework.ClusterEvent) {
|
func NodeSchedulingPropertiesChange(newNode *v1.Node, oldNode *v1.Node) (events []ClusterEvent) {
|
||||||
nodeChangeExtracters := []nodeChangeExtractor{
|
nodeChangeExtracters := []nodeChangeExtractor{
|
||||||
extractNodeSpecUnschedulableChange,
|
extractNodeSpecUnschedulableChange,
|
||||||
extractNodeAllocatableChange,
|
extractNodeAllocatableChange,
|
||||||
@ -165,16 +164,16 @@ func NodeSchedulingPropertiesChange(newNode *v1.Node, oldNode *v1.Node) (events
|
|||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
type nodeChangeExtractor func(newNode *v1.Node, oldNode *v1.Node) *framework.ClusterEvent
|
type nodeChangeExtractor func(newNode *v1.Node, oldNode *v1.Node) *ClusterEvent
|
||||||
|
|
||||||
func extractNodeAllocatableChange(newNode *v1.Node, oldNode *v1.Node) *framework.ClusterEvent {
|
func extractNodeAllocatableChange(newNode *v1.Node, oldNode *v1.Node) *ClusterEvent {
|
||||||
if !equality.Semantic.DeepEqual(oldNode.Status.Allocatable, newNode.Status.Allocatable) {
|
if !equality.Semantic.DeepEqual(oldNode.Status.Allocatable, newNode.Status.Allocatable) {
|
||||||
return &NodeAllocatableChange
|
return &NodeAllocatableChange
|
||||||
}
|
}
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func extractNodeLabelsChange(newNode *v1.Node, oldNode *v1.Node) *framework.ClusterEvent {
|
func extractNodeLabelsChange(newNode *v1.Node, oldNode *v1.Node) *ClusterEvent {
|
||||||
if isLabelChanged(newNode.GetLabels(), oldNode.GetLabels()) {
|
if isLabelChanged(newNode.GetLabels(), oldNode.GetLabels()) {
|
||||||
return &NodeLabelChange
|
return &NodeLabelChange
|
||||||
}
|
}
|
||||||
@ -185,14 +184,14 @@ func isLabelChanged(newLabels map[string]string, oldLabels map[string]string) bo
|
|||||||
return !equality.Semantic.DeepEqual(newLabels, oldLabels)
|
return !equality.Semantic.DeepEqual(newLabels, oldLabels)
|
||||||
}
|
}
|
||||||
|
|
||||||
func extractNodeTaintsChange(newNode *v1.Node, oldNode *v1.Node) *framework.ClusterEvent {
|
func extractNodeTaintsChange(newNode *v1.Node, oldNode *v1.Node) *ClusterEvent {
|
||||||
if !equality.Semantic.DeepEqual(newNode.Spec.Taints, oldNode.Spec.Taints) {
|
if !equality.Semantic.DeepEqual(newNode.Spec.Taints, oldNode.Spec.Taints) {
|
||||||
return &NodeTaintChange
|
return &NodeTaintChange
|
||||||
}
|
}
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func extractNodeConditionsChange(newNode *v1.Node, oldNode *v1.Node) *framework.ClusterEvent {
|
func extractNodeConditionsChange(newNode *v1.Node, oldNode *v1.Node) *ClusterEvent {
|
||||||
strip := func(conditions []v1.NodeCondition) map[v1.NodeConditionType]v1.ConditionStatus {
|
strip := func(conditions []v1.NodeCondition) map[v1.NodeConditionType]v1.ConditionStatus {
|
||||||
conditionStatuses := make(map[v1.NodeConditionType]v1.ConditionStatus, len(conditions))
|
conditionStatuses := make(map[v1.NodeConditionType]v1.ConditionStatus, len(conditions))
|
||||||
for i := range conditions {
|
for i := range conditions {
|
||||||
@ -206,14 +205,14 @@ func extractNodeConditionsChange(newNode *v1.Node, oldNode *v1.Node) *framework.
|
|||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func extractNodeSpecUnschedulableChange(newNode *v1.Node, oldNode *v1.Node) *framework.ClusterEvent {
|
func extractNodeSpecUnschedulableChange(newNode *v1.Node, oldNode *v1.Node) *ClusterEvent {
|
||||||
if newNode.Spec.Unschedulable != oldNode.Spec.Unschedulable && !newNode.Spec.Unschedulable {
|
if newNode.Spec.Unschedulable != oldNode.Spec.Unschedulable && !newNode.Spec.Unschedulable {
|
||||||
return &NodeSpecUnschedulableChange
|
return &NodeSpecUnschedulableChange
|
||||||
}
|
}
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func extractNodeAnnotationsChange(newNode *v1.Node, oldNode *v1.Node) *framework.ClusterEvent {
|
func extractNodeAnnotationsChange(newNode *v1.Node, oldNode *v1.Node) *ClusterEvent {
|
||||||
if !equality.Semantic.DeepEqual(oldNode.GetAnnotations(), newNode.GetAnnotations()) {
|
if !equality.Semantic.DeepEqual(oldNode.GetAnnotations(), newNode.GetAnnotations()) {
|
||||||
return &NodeAnnotationChange
|
return &NodeAnnotationChange
|
||||||
}
|
}
|
@ -14,7 +14,7 @@ See the License for the specific language governing permissions and
|
|||||||
limitations under the License.
|
limitations under the License.
|
||||||
*/
|
*/
|
||||||
|
|
||||||
package queue
|
package framework
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"reflect"
|
"reflect"
|
||||||
@ -24,7 +24,6 @@ import (
|
|||||||
v1 "k8s.io/api/core/v1"
|
v1 "k8s.io/api/core/v1"
|
||||||
"k8s.io/apimachinery/pkg/api/resource"
|
"k8s.io/apimachinery/pkg/api/resource"
|
||||||
metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
|
metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
|
||||||
"k8s.io/kubernetes/pkg/scheduler/framework"
|
|
||||||
st "k8s.io/kubernetes/pkg/scheduler/testing"
|
st "k8s.io/kubernetes/pkg/scheduler/testing"
|
||||||
)
|
)
|
||||||
|
|
||||||
@ -188,7 +187,7 @@ func TestNodeSchedulingPropertiesChange(t *testing.T) {
|
|||||||
name string
|
name string
|
||||||
newNode *v1.Node
|
newNode *v1.Node
|
||||||
oldNode *v1.Node
|
oldNode *v1.Node
|
||||||
wantEvents []framework.ClusterEvent
|
wantEvents []ClusterEvent
|
||||||
}{
|
}{
|
||||||
{
|
{
|
||||||
name: "no specific changed applied",
|
name: "no specific changed applied",
|
||||||
@ -200,7 +199,7 @@ func TestNodeSchedulingPropertiesChange(t *testing.T) {
|
|||||||
name: "only node spec unavailable changed",
|
name: "only node spec unavailable changed",
|
||||||
newNode: st.MakeNode().Unschedulable(false).Obj(),
|
newNode: st.MakeNode().Unschedulable(false).Obj(),
|
||||||
oldNode: st.MakeNode().Unschedulable(true).Obj(),
|
oldNode: st.MakeNode().Unschedulable(true).Obj(),
|
||||||
wantEvents: []framework.ClusterEvent{NodeSpecUnschedulableChange},
|
wantEvents: []ClusterEvent{NodeSpecUnschedulableChange},
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
name: "only node allocatable changed",
|
name: "only node allocatable changed",
|
||||||
@ -214,13 +213,13 @@ func TestNodeSchedulingPropertiesChange(t *testing.T) {
|
|||||||
v1.ResourceMemory: "100m",
|
v1.ResourceMemory: "100m",
|
||||||
v1.ResourceName("example.com/foo"): "2"},
|
v1.ResourceName("example.com/foo"): "2"},
|
||||||
).Obj(),
|
).Obj(),
|
||||||
wantEvents: []framework.ClusterEvent{NodeAllocatableChange},
|
wantEvents: []ClusterEvent{NodeAllocatableChange},
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
name: "only node label changed",
|
name: "only node label changed",
|
||||||
newNode: st.MakeNode().Label("foo", "bar").Obj(),
|
newNode: st.MakeNode().Label("foo", "bar").Obj(),
|
||||||
oldNode: st.MakeNode().Label("foo", "fuz").Obj(),
|
oldNode: st.MakeNode().Label("foo", "fuz").Obj(),
|
||||||
wantEvents: []framework.ClusterEvent{NodeLabelChange},
|
wantEvents: []ClusterEvent{NodeLabelChange},
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
name: "only node taint changed",
|
name: "only node taint changed",
|
||||||
@ -230,13 +229,13 @@ func TestNodeSchedulingPropertiesChange(t *testing.T) {
|
|||||||
oldNode: st.MakeNode().Taints([]v1.Taint{
|
oldNode: st.MakeNode().Taints([]v1.Taint{
|
||||||
{Key: v1.TaintNodeUnschedulable, Value: "foo", Effect: v1.TaintEffectNoSchedule},
|
{Key: v1.TaintNodeUnschedulable, Value: "foo", Effect: v1.TaintEffectNoSchedule},
|
||||||
}).Obj(),
|
}).Obj(),
|
||||||
wantEvents: []framework.ClusterEvent{NodeTaintChange},
|
wantEvents: []ClusterEvent{NodeTaintChange},
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
name: "only node annotation changed",
|
name: "only node annotation changed",
|
||||||
newNode: st.MakeNode().Annotation("foo", "bar").Obj(),
|
newNode: st.MakeNode().Annotation("foo", "bar").Obj(),
|
||||||
oldNode: st.MakeNode().Annotation("foo", "fuz").Obj(),
|
oldNode: st.MakeNode().Annotation("foo", "fuz").Obj(),
|
||||||
wantEvents: []framework.ClusterEvent{NodeAnnotationChange},
|
wantEvents: []ClusterEvent{NodeAnnotationChange},
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
name: "only node condition changed",
|
name: "only node condition changed",
|
||||||
@ -247,7 +246,7 @@ func TestNodeSchedulingPropertiesChange(t *testing.T) {
|
|||||||
"Ready",
|
"Ready",
|
||||||
"Ready",
|
"Ready",
|
||||||
).Obj(),
|
).Obj(),
|
||||||
wantEvents: []framework.ClusterEvent{NodeConditionChange},
|
wantEvents: []ClusterEvent{NodeConditionChange},
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
name: "both node label and node taint changed",
|
name: "both node label and node taint changed",
|
||||||
@ -259,7 +258,7 @@ func TestNodeSchedulingPropertiesChange(t *testing.T) {
|
|||||||
oldNode: st.MakeNode().Taints([]v1.Taint{
|
oldNode: st.MakeNode().Taints([]v1.Taint{
|
||||||
{Key: v1.TaintNodeUnschedulable, Value: "foo", Effect: v1.TaintEffectNoSchedule},
|
{Key: v1.TaintNodeUnschedulable, Value: "foo", Effect: v1.TaintEffectNoSchedule},
|
||||||
}).Obj(),
|
}).Obj(),
|
||||||
wantEvents: []framework.ClusterEvent{NodeLabelChange, NodeTaintChange},
|
wantEvents: []ClusterEvent{NodeLabelChange, NodeTaintChange},
|
||||||
},
|
},
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -339,31 +338,31 @@ func Test_podSchedulingPropertiesChange(t *testing.T) {
|
|||||||
name string
|
name string
|
||||||
newPod *v1.Pod
|
newPod *v1.Pod
|
||||||
oldPod *v1.Pod
|
oldPod *v1.Pod
|
||||||
want []framework.ClusterEvent
|
want []ClusterEvent
|
||||||
}{
|
}{
|
||||||
{
|
{
|
||||||
name: "only label is updated",
|
name: "only label is updated",
|
||||||
newPod: st.MakePod().Label("foo", "bar").Obj(),
|
newPod: st.MakePod().Label("foo", "bar").Obj(),
|
||||||
oldPod: st.MakePod().Label("foo", "bar2").Obj(),
|
oldPod: st.MakePod().Label("foo", "bar2").Obj(),
|
||||||
want: []framework.ClusterEvent{PodLabelChange},
|
want: []ClusterEvent{PodLabelChange},
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
name: "only pod's resource request is updated",
|
name: "only pod's resource request is updated",
|
||||||
oldPod: podWithSmallRequest,
|
oldPod: podWithSmallRequest,
|
||||||
newPod: podWithBigRequest,
|
newPod: podWithBigRequest,
|
||||||
want: []framework.ClusterEvent{PodRequestChange},
|
want: []ClusterEvent{PodRequestChange},
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
name: "both pod's resource request and label are updated",
|
name: "both pod's resource request and label are updated",
|
||||||
oldPod: podWithSmallRequest,
|
oldPod: podWithSmallRequest,
|
||||||
newPod: podWithBigRequestAndLabel,
|
newPod: podWithBigRequestAndLabel,
|
||||||
want: []framework.ClusterEvent{PodLabelChange, PodRequestChange},
|
want: []ClusterEvent{PodLabelChange, PodRequestChange},
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
name: "untracked properties of pod is updated",
|
name: "untracked properties of pod is updated",
|
||||||
newPod: st.MakePod().Annotation("foo", "bar").Obj(),
|
newPod: st.MakePod().Annotation("foo", "bar").Obj(),
|
||||||
oldPod: st.MakePod().Annotation("foo", "bar2").Obj(),
|
oldPod: st.MakePod().Annotation("foo", "bar2").Obj(),
|
||||||
want: []framework.ClusterEvent{AssignedPodOtherUpdate},
|
want: []ClusterEvent{assignedPodOtherUpdate},
|
||||||
},
|
},
|
||||||
}
|
}
|
||||||
for _, tt := range tests {
|
for _, tt := range tests {
|
@ -68,17 +68,15 @@ const (
|
|||||||
UpdatePodLabel
|
UpdatePodLabel
|
||||||
// UpdatePodRequest is a update for pod's resource request calculated by resource.PodRequests() function.
|
// UpdatePodRequest is a update for pod's resource request calculated by resource.PodRequests() function.
|
||||||
UpdatePodRequest
|
UpdatePodRequest
|
||||||
// UpdatePodOther is a update for pod's other fields.
|
|
||||||
// NOT RECOMMENDED using it in your plugin's EventsToRegister,
|
// updatePodOther is a update for pod's other fields.
|
||||||
// use Pod/Update instead when you have to subscribe Pod updates which are not covered by other UpdatePodXYZ events.
|
// It's used only for the internal event handling, and thus unexported.
|
||||||
// Otherwise, your plugin might be broken when the upstream supports a new Pod specific Update event.
|
updatePodOther
|
||||||
// It's used only for the internal event handling.
|
|
||||||
UpdatePodOther
|
|
||||||
|
|
||||||
All ActionType = 1<<iota - 1
|
All ActionType = 1<<iota - 1
|
||||||
|
|
||||||
// Use the general Update type if you don't either know or care the specific sub-Update type to use.
|
// Use the general Update type if you don't either know or care the specific sub-Update type to use.
|
||||||
Update = UpdateNodeAllocatable | UpdateNodeLabel | UpdateNodeTaint | UpdateNodeCondition | UpdateNodeAnnotation | UpdatePodLabel | UpdatePodRequest | UpdatePodOther
|
Update = UpdateNodeAllocatable | UpdateNodeLabel | UpdateNodeTaint | UpdateNodeCondition | UpdateNodeAnnotation | UpdatePodLabel | UpdatePodRequest | updatePodOther
|
||||||
)
|
)
|
||||||
|
|
||||||
// GVK is short for group/version/kind, which can uniquely represent a particular API resource.
|
// GVK is short for group/version/kind, which can uniquely represent a particular API resource.
|
||||||
|
@ -614,7 +614,7 @@ func (p *PriorityQueue) moveToActiveQ(logger klog.Logger, pInfo *framework.Queue
|
|||||||
_ = p.podBackoffQ.Delete(pInfo) // Don't need to react when pInfo is not found.
|
_ = p.podBackoffQ.Delete(pInfo) // Don't need to react when pInfo is not found.
|
||||||
logger.V(5).Info("Pod moved to an internal scheduling queue", "pod", klog.KObj(pInfo.Pod), "event", event, "queue", activeQ)
|
logger.V(5).Info("Pod moved to an internal scheduling queue", "pod", klog.KObj(pInfo.Pod), "event", event, "queue", activeQ)
|
||||||
metrics.SchedulerQueueIncomingPods.WithLabelValues("active", event).Inc()
|
metrics.SchedulerQueueIncomingPods.WithLabelValues("active", event).Inc()
|
||||||
if event == PodAdd || event == PodUpdate {
|
if event == framework.PodAdd || event == framework.PodUpdate {
|
||||||
p.AddNominatedPod(logger, pInfo.PodInfo, nil)
|
p.AddNominatedPod(logger, pInfo.PodInfo, nil)
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -628,7 +628,7 @@ func (p *PriorityQueue) Add(logger klog.Logger, pod *v1.Pod) error {
|
|||||||
defer p.lock.Unlock()
|
defer p.lock.Unlock()
|
||||||
|
|
||||||
pInfo := p.newQueuedPodInfo(pod)
|
pInfo := p.newQueuedPodInfo(pod)
|
||||||
if added, err := p.moveToActiveQ(logger, pInfo, PodAdd); !added {
|
if added, err := p.moveToActiveQ(logger, pInfo, framework.PodAdd); !added {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
p.cond.Broadcast()
|
p.cond.Broadcast()
|
||||||
@ -679,7 +679,7 @@ func (p *PriorityQueue) activate(logger klog.Logger, pod *v1.Pod) bool {
|
|||||||
return false
|
return false
|
||||||
}
|
}
|
||||||
|
|
||||||
added, _ := p.moveToActiveQ(logger, pInfo, ForceActivate)
|
added, _ := p.moveToActiveQ(logger, pInfo, framework.ForceActivate)
|
||||||
return added
|
return added
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -803,12 +803,12 @@ func (p *PriorityQueue) addUnschedulableWithoutQueueingHint(logger klog.Logger,
|
|||||||
if err := p.podBackoffQ.Add(pInfo); err != nil {
|
if err := p.podBackoffQ.Add(pInfo); err != nil {
|
||||||
return fmt.Errorf("error adding pod %v to the backoff queue: %v", klog.KObj(pod), err)
|
return fmt.Errorf("error adding pod %v to the backoff queue: %v", klog.KObj(pod), err)
|
||||||
}
|
}
|
||||||
logger.V(5).Info("Pod moved to an internal scheduling queue", "pod", klog.KObj(pod), "event", ScheduleAttemptFailure, "queue", backoffQ)
|
logger.V(5).Info("Pod moved to an internal scheduling queue", "pod", klog.KObj(pod), "event", framework.ScheduleAttemptFailure, "queue", backoffQ)
|
||||||
metrics.SchedulerQueueIncomingPods.WithLabelValues("backoff", ScheduleAttemptFailure).Inc()
|
metrics.SchedulerQueueIncomingPods.WithLabelValues("backoff", framework.ScheduleAttemptFailure).Inc()
|
||||||
} else {
|
} else {
|
||||||
p.unschedulablePods.addOrUpdate(pInfo)
|
p.unschedulablePods.addOrUpdate(pInfo)
|
||||||
logger.V(5).Info("Pod moved to an internal scheduling queue", "pod", klog.KObj(pod), "event", ScheduleAttemptFailure, "queue", unschedulablePods)
|
logger.V(5).Info("Pod moved to an internal scheduling queue", "pod", klog.KObj(pod), "event", framework.ScheduleAttemptFailure, "queue", unschedulablePods)
|
||||||
metrics.SchedulerQueueIncomingPods.WithLabelValues("unschedulable", ScheduleAttemptFailure).Inc()
|
metrics.SchedulerQueueIncomingPods.WithLabelValues("unschedulable", framework.ScheduleAttemptFailure).Inc()
|
||||||
}
|
}
|
||||||
|
|
||||||
p.AddNominatedPod(logger, pInfo.PodInfo, nil)
|
p.AddNominatedPod(logger, pInfo.PodInfo, nil)
|
||||||
@ -857,8 +857,8 @@ func (p *PriorityQueue) AddUnschedulableIfNotPresent(logger klog.Logger, pInfo *
|
|||||||
schedulingHint := p.determineSchedulingHintForInFlightPod(logger, pInfo)
|
schedulingHint := p.determineSchedulingHintForInFlightPod(logger, pInfo)
|
||||||
|
|
||||||
// In this case, we try to requeue this Pod to activeQ/backoffQ.
|
// In this case, we try to requeue this Pod to activeQ/backoffQ.
|
||||||
queue := p.requeuePodViaQueueingHint(logger, pInfo, schedulingHint, ScheduleAttemptFailure)
|
queue := p.requeuePodViaQueueingHint(logger, pInfo, schedulingHint, framework.ScheduleAttemptFailure)
|
||||||
logger.V(3).Info("Pod moved to an internal scheduling queue", "pod", klog.KObj(pod), "event", ScheduleAttemptFailure, "queue", queue, "schedulingCycle", podSchedulingCycle, "hint", schedulingHint, "unschedulable plugins", rejectorPlugins)
|
logger.V(3).Info("Pod moved to an internal scheduling queue", "pod", klog.KObj(pod), "event", framework.ScheduleAttemptFailure, "queue", queue, "schedulingCycle", podSchedulingCycle, "hint", schedulingHint, "unschedulable plugins", rejectorPlugins)
|
||||||
if queue == activeQ {
|
if queue == activeQ {
|
||||||
// When the Pod is moved to activeQ, need to let p.cond know so that the Pod will be pop()ed out.
|
// When the Pod is moved to activeQ, need to let p.cond know so that the Pod will be pop()ed out.
|
||||||
p.cond.Broadcast()
|
p.cond.Broadcast()
|
||||||
@ -888,7 +888,7 @@ func (p *PriorityQueue) flushBackoffQCompleted(logger klog.Logger) {
|
|||||||
logger.Error(err, "Unable to pop pod from backoff queue despite backoff completion", "pod", klog.KObj(pod))
|
logger.Error(err, "Unable to pop pod from backoff queue despite backoff completion", "pod", klog.KObj(pod))
|
||||||
break
|
break
|
||||||
}
|
}
|
||||||
if added, _ := p.moveToActiveQ(logger, pInfo, BackoffComplete); added {
|
if added, _ := p.moveToActiveQ(logger, pInfo, framework.BackoffComplete); added {
|
||||||
activated = true
|
activated = true
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@ -914,7 +914,7 @@ func (p *PriorityQueue) flushUnschedulablePodsLeftover(logger klog.Logger) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
if len(podsToMove) > 0 {
|
if len(podsToMove) > 0 {
|
||||||
p.movePodsToActiveOrBackoffQueue(logger, podsToMove, UnschedulableTimeout, nil, nil)
|
p.movePodsToActiveOrBackoffQueue(logger, podsToMove, framework.UnschedulableTimeout, nil, nil)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -1051,7 +1051,7 @@ func (p *PriorityQueue) Update(logger klog.Logger, oldPod, newPod *v1.Pod) error
|
|||||||
// We can clean it up once we change updatePodInSchedulingQueue to call MoveAllToActiveOrBackoffQueue.
|
// We can clean it up once we change updatePodInSchedulingQueue to call MoveAllToActiveOrBackoffQueue.
|
||||||
// See https://github.com/kubernetes/kubernetes/pull/125578#discussion_r1648338033 for more context.
|
// See https://github.com/kubernetes/kubernetes/pull/125578#discussion_r1648338033 for more context.
|
||||||
p.inFlightEvents.PushBack(&clusterEvent{
|
p.inFlightEvents.PushBack(&clusterEvent{
|
||||||
event: UnscheduledPodUpdate,
|
event: framework.UnscheduledPodUpdate,
|
||||||
oldObj: oldPod,
|
oldObj: oldPod,
|
||||||
newObj: newPod,
|
newObj: newPod,
|
||||||
})
|
})
|
||||||
@ -1088,12 +1088,12 @@ func (p *PriorityQueue) Update(logger klog.Logger, oldPod, newPod *v1.Pod) error
|
|||||||
// Plugins have to implement a QueueingHint for Pod/Update event
|
// Plugins have to implement a QueueingHint for Pod/Update event
|
||||||
// if the rejection from them could be resolved by updating unscheduled Pods itself.
|
// if the rejection from them could be resolved by updating unscheduled Pods itself.
|
||||||
|
|
||||||
events := PodSchedulingPropertiesChange(newPod, oldPod)
|
events := framework.PodSchedulingPropertiesChange(newPod, oldPod)
|
||||||
for _, evt := range events {
|
for _, evt := range events {
|
||||||
hint := p.isPodWorthRequeuing(logger, pInfo, evt, oldPod, newPod)
|
hint := p.isPodWorthRequeuing(logger, pInfo, evt, oldPod, newPod)
|
||||||
queue := p.requeuePodViaQueueingHint(logger, pInfo, hint, UnscheduledPodUpdate.Label)
|
queue := p.requeuePodViaQueueingHint(logger, pInfo, hint, framework.UnscheduledPodUpdate.Label)
|
||||||
if queue != unschedulablePods {
|
if queue != unschedulablePods {
|
||||||
logger.V(5).Info("Pod moved to an internal scheduling queue because the Pod is updated", "pod", klog.KObj(newPod), "event", PodUpdate, "queue", queue)
|
logger.V(5).Info("Pod moved to an internal scheduling queue because the Pod is updated", "pod", klog.KObj(newPod), "event", framework.PodUpdate, "queue", queue)
|
||||||
p.unschedulablePods.delete(usPodInfo.Pod, gated)
|
p.unschedulablePods.delete(usPodInfo.Pod, gated)
|
||||||
}
|
}
|
||||||
if queue == activeQ {
|
if queue == activeQ {
|
||||||
@ -1111,11 +1111,11 @@ func (p *PriorityQueue) Update(logger klog.Logger, oldPod, newPod *v1.Pod) error
|
|||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
p.unschedulablePods.delete(usPodInfo.Pod, gated)
|
p.unschedulablePods.delete(usPodInfo.Pod, gated)
|
||||||
logger.V(5).Info("Pod moved to an internal scheduling queue", "pod", klog.KObj(pInfo.Pod), "event", PodUpdate, "queue", backoffQ)
|
logger.V(5).Info("Pod moved to an internal scheduling queue", "pod", klog.KObj(pInfo.Pod), "event", framework.PodUpdate, "queue", backoffQ)
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
if added, err := p.moveToActiveQ(logger, pInfo, BackoffComplete); !added {
|
if added, err := p.moveToActiveQ(logger, pInfo, framework.BackoffComplete); !added {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
p.cond.Broadcast()
|
p.cond.Broadcast()
|
||||||
@ -1128,7 +1128,7 @@ func (p *PriorityQueue) Update(logger klog.Logger, oldPod, newPod *v1.Pod) error
|
|||||||
}
|
}
|
||||||
// If pod is not in any of the queues, we put it in the active queue.
|
// If pod is not in any of the queues, we put it in the active queue.
|
||||||
pInfo := p.newQueuedPodInfo(newPod)
|
pInfo := p.newQueuedPodInfo(newPod)
|
||||||
if added, err := p.moveToActiveQ(logger, pInfo, PodUpdate); !added {
|
if added, err := p.moveToActiveQ(logger, pInfo, framework.PodUpdate); !added {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
p.cond.Broadcast()
|
p.cond.Broadcast()
|
||||||
@ -1161,7 +1161,7 @@ func (p *PriorityQueue) AssignedPodAdded(logger klog.Logger, pod *v1.Pod) {
|
|||||||
|
|
||||||
// Pre-filter Pods to move by getUnschedulablePodsWithCrossTopologyTerm
|
// Pre-filter Pods to move by getUnschedulablePodsWithCrossTopologyTerm
|
||||||
// because Pod related events shouldn't make Pods that rejected by single-node scheduling requirement schedulable.
|
// because Pod related events shouldn't make Pods that rejected by single-node scheduling requirement schedulable.
|
||||||
p.movePodsToActiveOrBackoffQueue(logger, p.getUnschedulablePodsWithCrossTopologyTerm(logger, pod), AssignedPodAdd, nil, pod)
|
p.movePodsToActiveOrBackoffQueue(logger, p.getUnschedulablePodsWithCrossTopologyTerm(logger, pod), framework.AssignedPodAdd, nil, pod)
|
||||||
p.lock.Unlock()
|
p.lock.Unlock()
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -1172,7 +1172,7 @@ func (p *PriorityQueue) AssignedPodUpdated(logger klog.Logger, oldPod, newPod *v
|
|||||||
if event.Resource == framework.Pod && event.ActionType&framework.UpdatePodRequest != 0 {
|
if event.Resource == framework.Pod && event.ActionType&framework.UpdatePodRequest != 0 {
|
||||||
// In this case, we don't want to pre-filter Pods by getUnschedulablePodsWithCrossTopologyTerm
|
// In this case, we don't want to pre-filter Pods by getUnschedulablePodsWithCrossTopologyTerm
|
||||||
// because Pod related events may make Pods that were rejected by NodeResourceFit schedulable.
|
// because Pod related events may make Pods that were rejected by NodeResourceFit schedulable.
|
||||||
p.moveAllToActiveOrBackoffQueue(logger, AssignedPodUpdate, oldPod, newPod, nil)
|
p.moveAllToActiveOrBackoffQueue(logger, framework.AssignedPodUpdate, oldPod, newPod, nil)
|
||||||
} else {
|
} else {
|
||||||
// Pre-filter Pods to move by getUnschedulablePodsWithCrossTopologyTerm
|
// Pre-filter Pods to move by getUnschedulablePodsWithCrossTopologyTerm
|
||||||
// because Pod related events only make Pods rejected by cross topology term schedulable.
|
// because Pod related events only make Pods rejected by cross topology term schedulable.
|
||||||
@ -1250,7 +1250,7 @@ func (p *PriorityQueue) requeuePodViaQueueingHint(logger klog.Logger, pInfo *fra
|
|||||||
}
|
}
|
||||||
|
|
||||||
p.unschedulablePods.addOrUpdate(pInfo)
|
p.unschedulablePods.addOrUpdate(pInfo)
|
||||||
metrics.SchedulerQueueIncomingPods.WithLabelValues("unschedulable", ScheduleAttemptFailure).Inc()
|
metrics.SchedulerQueueIncomingPods.WithLabelValues("unschedulable", framework.ScheduleAttemptFailure).Inc()
|
||||||
return unschedulablePods
|
return unschedulablePods
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -225,13 +225,13 @@ func Test_InFlightPods(t *testing.T) {
|
|||||||
// This Pod shouldn't be added to inFlightPods because SchedulingQueueHint is disabled.
|
// This Pod shouldn't be added to inFlightPods because SchedulingQueueHint is disabled.
|
||||||
{podPopped: pod},
|
{podPopped: pod},
|
||||||
// This event shouldn't be added to inFlightEvents because SchedulingQueueHint is disabled.
|
// This event shouldn't be added to inFlightEvents because SchedulingQueueHint is disabled.
|
||||||
{eventHappens: &PvAdd},
|
{eventHappens: &framework.PvAdd},
|
||||||
},
|
},
|
||||||
wantInFlightPods: nil,
|
wantInFlightPods: nil,
|
||||||
wantInFlightEvents: nil,
|
wantInFlightEvents: nil,
|
||||||
queueingHintMap: QueueingHintMapPerProfile{
|
queueingHintMap: QueueingHintMapPerProfile{
|
||||||
"": {
|
"": {
|
||||||
PvAdd: {
|
framework.PvAdd: {
|
||||||
{
|
{
|
||||||
PluginName: "fooPlugin1",
|
PluginName: "fooPlugin1",
|
||||||
QueueingHintFn: queueHintReturnQueue,
|
QueueingHintFn: queueHintReturnQueue,
|
||||||
@ -246,18 +246,18 @@ func Test_InFlightPods(t *testing.T) {
|
|||||||
initialPods: []*v1.Pod{pod},
|
initialPods: []*v1.Pod{pod},
|
||||||
actions: []action{
|
actions: []action{
|
||||||
// This won't be added to inFlightEvents because no inFlightPods at this point.
|
// This won't be added to inFlightEvents because no inFlightPods at this point.
|
||||||
{eventHappens: &PvcAdd},
|
{eventHappens: &framework.PvcAdd},
|
||||||
{podPopped: pod},
|
{podPopped: pod},
|
||||||
// This gets added for the pod.
|
// This gets added for the pod.
|
||||||
{eventHappens: &PvAdd},
|
{eventHappens: &framework.PvAdd},
|
||||||
// This doesn't get added because no plugin is interested in PvUpdate.
|
// This doesn't get added because no plugin is interested in framework.PvUpdate.
|
||||||
{eventHappens: &PvUpdate},
|
{eventHappens: &framework.PvUpdate},
|
||||||
},
|
},
|
||||||
wantInFlightPods: []*v1.Pod{pod},
|
wantInFlightPods: []*v1.Pod{pod},
|
||||||
wantInFlightEvents: []interface{}{pod, PvAdd},
|
wantInFlightEvents: []interface{}{pod, framework.PvAdd},
|
||||||
queueingHintMap: QueueingHintMapPerProfile{
|
queueingHintMap: QueueingHintMapPerProfile{
|
||||||
"": {
|
"": {
|
||||||
PvAdd: {
|
framework.PvAdd: {
|
||||||
{
|
{
|
||||||
PluginName: "fooPlugin1",
|
PluginName: "fooPlugin1",
|
||||||
QueueingHintFn: queueHintReturnQueue,
|
QueueingHintFn: queueHintReturnQueue,
|
||||||
@ -272,32 +272,32 @@ func Test_InFlightPods(t *testing.T) {
|
|||||||
initialPods: []*v1.Pod{pod, pod2},
|
initialPods: []*v1.Pod{pod, pod2},
|
||||||
actions: []action{
|
actions: []action{
|
||||||
// This won't be added to inFlightEvents because no inFlightPods at this point.
|
// This won't be added to inFlightEvents because no inFlightPods at this point.
|
||||||
{eventHappens: &PvcAdd},
|
{eventHappens: &framework.PvcAdd},
|
||||||
{podPopped: pod},
|
{podPopped: pod},
|
||||||
{eventHappens: &PvAdd},
|
{eventHappens: &framework.PvAdd},
|
||||||
{podPopped: pod2},
|
{podPopped: pod2},
|
||||||
{eventHappens: &NodeAdd},
|
{eventHappens: &framework.NodeAdd},
|
||||||
// This pod will be requeued to backoffQ because no plugin is registered as unschedulable plugin.
|
// This pod will be requeued to backoffQ because no plugin is registered as unschedulable plugin.
|
||||||
{podEnqueued: newQueuedPodInfoForLookup(pod)},
|
{podEnqueued: newQueuedPodInfoForLookup(pod)},
|
||||||
},
|
},
|
||||||
wantBackoffQPodNames: []string{"targetpod"},
|
wantBackoffQPodNames: []string{"targetpod"},
|
||||||
wantInFlightPods: []*v1.Pod{pod2}, // only pod2 is registered because pod is already enqueued back.
|
wantInFlightPods: []*v1.Pod{pod2}, // only pod2 is registered because pod is already enqueued back.
|
||||||
wantInFlightEvents: []interface{}{pod2, NodeAdd},
|
wantInFlightEvents: []interface{}{pod2, framework.NodeAdd},
|
||||||
queueingHintMap: QueueingHintMapPerProfile{
|
queueingHintMap: QueueingHintMapPerProfile{
|
||||||
"": {
|
"": {
|
||||||
PvAdd: {
|
framework.PvAdd: {
|
||||||
{
|
{
|
||||||
PluginName: "fooPlugin1",
|
PluginName: "fooPlugin1",
|
||||||
QueueingHintFn: queueHintReturnQueue,
|
QueueingHintFn: queueHintReturnQueue,
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
NodeAdd: {
|
framework.NodeAdd: {
|
||||||
{
|
{
|
||||||
PluginName: "fooPlugin1",
|
PluginName: "fooPlugin1",
|
||||||
QueueingHintFn: queueHintReturnQueue,
|
QueueingHintFn: queueHintReturnQueue,
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
PvcAdd: {
|
framework.PvcAdd: {
|
||||||
{
|
{
|
||||||
PluginName: "fooPlugin1",
|
PluginName: "fooPlugin1",
|
||||||
QueueingHintFn: queueHintReturnQueue,
|
QueueingHintFn: queueHintReturnQueue,
|
||||||
@ -312,14 +312,14 @@ func Test_InFlightPods(t *testing.T) {
|
|||||||
initialPods: []*v1.Pod{pod, pod2},
|
initialPods: []*v1.Pod{pod, pod2},
|
||||||
actions: []action{
|
actions: []action{
|
||||||
// This won't be added to inFlightEvents because no inFlightPods at this point.
|
// This won't be added to inFlightEvents because no inFlightPods at this point.
|
||||||
{eventHappens: &PvcAdd},
|
{eventHappens: &framework.PvcAdd},
|
||||||
{podPopped: pod},
|
{podPopped: pod},
|
||||||
{eventHappens: &PvAdd},
|
{eventHappens: &framework.PvAdd},
|
||||||
{podPopped: pod2},
|
{podPopped: pod2},
|
||||||
{eventHappens: &NodeAdd},
|
{eventHappens: &framework.NodeAdd},
|
||||||
// This pod will be requeued to backoffQ because no plugin is registered as unschedulable plugin.
|
// This pod will be requeued to backoffQ because no plugin is registered as unschedulable plugin.
|
||||||
{podEnqueued: newQueuedPodInfoForLookup(pod)},
|
{podEnqueued: newQueuedPodInfoForLookup(pod)},
|
||||||
{eventHappens: &CSINodeUpdate},
|
{eventHappens: &framework.CSINodeUpdate},
|
||||||
// This pod will be requeued to backoffQ because no plugin is registered as unschedulable plugin.
|
// This pod will be requeued to backoffQ because no plugin is registered as unschedulable plugin.
|
||||||
{podEnqueued: newQueuedPodInfoForLookup(pod2)},
|
{podEnqueued: newQueuedPodInfoForLookup(pod2)},
|
||||||
},
|
},
|
||||||
@ -327,25 +327,25 @@ func Test_InFlightPods(t *testing.T) {
|
|||||||
wantInFlightPods: nil, // empty
|
wantInFlightPods: nil, // empty
|
||||||
queueingHintMap: QueueingHintMapPerProfile{
|
queueingHintMap: QueueingHintMapPerProfile{
|
||||||
"": {
|
"": {
|
||||||
PvAdd: {
|
framework.PvAdd: {
|
||||||
{
|
{
|
||||||
PluginName: "fooPlugin1",
|
PluginName: "fooPlugin1",
|
||||||
QueueingHintFn: queueHintReturnQueue,
|
QueueingHintFn: queueHintReturnQueue,
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
NodeAdd: {
|
framework.NodeAdd: {
|
||||||
{
|
{
|
||||||
PluginName: "fooPlugin1",
|
PluginName: "fooPlugin1",
|
||||||
QueueingHintFn: queueHintReturnQueue,
|
QueueingHintFn: queueHintReturnQueue,
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
PvcAdd: {
|
framework.PvcAdd: {
|
||||||
{
|
{
|
||||||
PluginName: "fooPlugin1",
|
PluginName: "fooPlugin1",
|
||||||
QueueingHintFn: queueHintReturnQueue,
|
QueueingHintFn: queueHintReturnQueue,
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
CSINodeUpdate: {
|
framework.CSINodeUpdate: {
|
||||||
{
|
{
|
||||||
PluginName: "fooPlugin1",
|
PluginName: "fooPlugin1",
|
||||||
QueueingHintFn: queueHintReturnQueue,
|
QueueingHintFn: queueHintReturnQueue,
|
||||||
@ -360,34 +360,34 @@ func Test_InFlightPods(t *testing.T) {
|
|||||||
initialPods: []*v1.Pod{pod, pod2, pod3},
|
initialPods: []*v1.Pod{pod, pod2, pod3},
|
||||||
actions: []action{
|
actions: []action{
|
||||||
// This won't be added to inFlightEvents because no inFlightPods at this point.
|
// This won't be added to inFlightEvents because no inFlightPods at this point.
|
||||||
{eventHappens: &PvcAdd},
|
{eventHappens: &framework.PvcAdd},
|
||||||
{podPopped: pod},
|
{podPopped: pod},
|
||||||
{eventHappens: &PvAdd},
|
{eventHappens: &framework.PvAdd},
|
||||||
{podPopped: pod2},
|
{podPopped: pod2},
|
||||||
{eventHappens: &NodeAdd},
|
{eventHappens: &framework.NodeAdd},
|
||||||
// This Pod won't be requeued again.
|
// This Pod won't be requeued again.
|
||||||
{podPopped: pod3},
|
{podPopped: pod3},
|
||||||
{eventHappens: &AssignedPodAdd},
|
{eventHappens: &framework.AssignedPodAdd},
|
||||||
{podEnqueued: newQueuedPodInfoForLookup(pod2)},
|
{podEnqueued: newQueuedPodInfoForLookup(pod2)},
|
||||||
},
|
},
|
||||||
wantBackoffQPodNames: []string{"targetpod2"},
|
wantBackoffQPodNames: []string{"targetpod2"},
|
||||||
wantInFlightPods: []*v1.Pod{pod, pod3},
|
wantInFlightPods: []*v1.Pod{pod, pod3},
|
||||||
wantInFlightEvents: []interface{}{pod, PvAdd, NodeAdd, pod3, AssignedPodAdd},
|
wantInFlightEvents: []interface{}{pod, framework.PvAdd, framework.NodeAdd, pod3, framework.AssignedPodAdd},
|
||||||
queueingHintMap: QueueingHintMapPerProfile{
|
queueingHintMap: QueueingHintMapPerProfile{
|
||||||
"": {
|
"": {
|
||||||
PvAdd: {
|
framework.PvAdd: {
|
||||||
{
|
{
|
||||||
PluginName: "fooPlugin1",
|
PluginName: "fooPlugin1",
|
||||||
QueueingHintFn: queueHintReturnQueue,
|
QueueingHintFn: queueHintReturnQueue,
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
NodeAdd: {
|
framework.NodeAdd: {
|
||||||
{
|
{
|
||||||
PluginName: "fooPlugin1",
|
PluginName: "fooPlugin1",
|
||||||
QueueingHintFn: queueHintReturnQueue,
|
QueueingHintFn: queueHintReturnQueue,
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
AssignedPodAdd: {
|
framework.AssignedPodAdd: {
|
||||||
{
|
{
|
||||||
PluginName: "fooPlugin1",
|
PluginName: "fooPlugin1",
|
||||||
QueueingHintFn: queueHintReturnQueue,
|
QueueingHintFn: queueHintReturnQueue,
|
||||||
@ -401,7 +401,7 @@ func Test_InFlightPods(t *testing.T) {
|
|||||||
initialPods: []*v1.Pod{pod},
|
initialPods: []*v1.Pod{pod},
|
||||||
actions: []action{
|
actions: []action{
|
||||||
{podPopped: pod},
|
{podPopped: pod},
|
||||||
{eventHappens: &AssignedPodAdd},
|
{eventHappens: &framework.AssignedPodAdd},
|
||||||
{podEnqueued: newQueuedPodInfoForLookup(pod, "fooPlugin1")},
|
{podEnqueued: newQueuedPodInfoForLookup(pod, "fooPlugin1")},
|
||||||
},
|
},
|
||||||
wantBackoffQPodNames: []string{"targetpod"},
|
wantBackoffQPodNames: []string{"targetpod"},
|
||||||
@ -411,7 +411,7 @@ func Test_InFlightPods(t *testing.T) {
|
|||||||
"": {
|
"": {
|
||||||
// This hint fn tells that this event doesn't make a Pod schedulable.
|
// This hint fn tells that this event doesn't make a Pod schedulable.
|
||||||
// However, this QueueingHintFn will be ignored actually because SchedulingQueueHint is disabled.
|
// However, this QueueingHintFn will be ignored actually because SchedulingQueueHint is disabled.
|
||||||
AssignedPodAdd: {
|
framework.AssignedPodAdd: {
|
||||||
{
|
{
|
||||||
PluginName: "fooPlugin1",
|
PluginName: "fooPlugin1",
|
||||||
QueueingHintFn: queueHintReturnSkip,
|
QueueingHintFn: queueHintReturnSkip,
|
||||||
@ -425,9 +425,9 @@ func Test_InFlightPods(t *testing.T) {
|
|||||||
isSchedulingQueueHintEnabled: true,
|
isSchedulingQueueHintEnabled: true,
|
||||||
initialPods: []*v1.Pod{pod},
|
initialPods: []*v1.Pod{pod},
|
||||||
actions: []action{
|
actions: []action{
|
||||||
{eventHappens: &WildCardEvent},
|
{eventHappens: &framework.WildCardEvent},
|
||||||
{podPopped: pod},
|
{podPopped: pod},
|
||||||
{eventHappens: &AssignedPodAdd},
|
{eventHappens: &framework.AssignedPodAdd},
|
||||||
// This Pod won't be requeued to activeQ/backoffQ because fooPlugin1 returns QueueSkip.
|
// This Pod won't be requeued to activeQ/backoffQ because fooPlugin1 returns QueueSkip.
|
||||||
{podEnqueued: newQueuedPodInfoForLookup(pod, "fooPlugin1")},
|
{podEnqueued: newQueuedPodInfoForLookup(pod, "fooPlugin1")},
|
||||||
},
|
},
|
||||||
@ -436,9 +436,9 @@ func Test_InFlightPods(t *testing.T) {
|
|||||||
wantInFlightEvents: nil,
|
wantInFlightEvents: nil,
|
||||||
queueingHintMap: QueueingHintMapPerProfile{
|
queueingHintMap: QueueingHintMapPerProfile{
|
||||||
"": {
|
"": {
|
||||||
// fooPlugin1 has a queueing hint function for AssignedPodAdd,
|
// fooPlugin1 has a queueing hint function for framework.AssignedPodAdd,
|
||||||
// but hint fn tells that this event doesn't make a Pod scheudlable.
|
// but hint fn tells that this event doesn't make a Pod scheudlable.
|
||||||
AssignedPodAdd: {
|
framework.AssignedPodAdd: {
|
||||||
{
|
{
|
||||||
PluginName: "fooPlugin1",
|
PluginName: "fooPlugin1",
|
||||||
QueueingHintFn: queueHintReturnSkip,
|
QueueingHintFn: queueHintReturnSkip,
|
||||||
@ -453,7 +453,7 @@ func Test_InFlightPods(t *testing.T) {
|
|||||||
initialPods: []*v1.Pod{pod},
|
initialPods: []*v1.Pod{pod},
|
||||||
actions: []action{
|
actions: []action{
|
||||||
{podPopped: pod},
|
{podPopped: pod},
|
||||||
{eventHappens: &AssignedPodAdd},
|
{eventHappens: &framework.AssignedPodAdd},
|
||||||
{podEnqueued: newQueuedPodInfoForLookup(pod)},
|
{podEnqueued: newQueuedPodInfoForLookup(pod)},
|
||||||
},
|
},
|
||||||
wantBackoffQPodNames: []string{"targetpod"},
|
wantBackoffQPodNames: []string{"targetpod"},
|
||||||
@ -462,7 +462,7 @@ func Test_InFlightPods(t *testing.T) {
|
|||||||
queueingHintMap: QueueingHintMapPerProfile{
|
queueingHintMap: QueueingHintMapPerProfile{
|
||||||
"": {
|
"": {
|
||||||
// It will be ignored because no failed plugin.
|
// It will be ignored because no failed plugin.
|
||||||
AssignedPodAdd: {
|
framework.AssignedPodAdd: {
|
||||||
{
|
{
|
||||||
PluginName: "fooPlugin1",
|
PluginName: "fooPlugin1",
|
||||||
QueueingHintFn: queueHintReturnQueue,
|
QueueingHintFn: queueHintReturnQueue,
|
||||||
@ -477,7 +477,7 @@ func Test_InFlightPods(t *testing.T) {
|
|||||||
initialPods: []*v1.Pod{pod},
|
initialPods: []*v1.Pod{pod},
|
||||||
actions: []action{
|
actions: []action{
|
||||||
{podPopped: pod},
|
{podPopped: pod},
|
||||||
{eventHappens: &NodeAdd},
|
{eventHappens: &framework.NodeAdd},
|
||||||
{podEnqueued: newQueuedPodInfoForLookup(pod, "fooPlugin1")},
|
{podEnqueued: newQueuedPodInfoForLookup(pod, "fooPlugin1")},
|
||||||
},
|
},
|
||||||
wantUnschedPodPoolPodNames: []string{"targetpod"},
|
wantUnschedPodPoolPodNames: []string{"targetpod"},
|
||||||
@ -485,10 +485,10 @@ func Test_InFlightPods(t *testing.T) {
|
|||||||
wantInFlightEvents: nil,
|
wantInFlightEvents: nil,
|
||||||
queueingHintMap: QueueingHintMapPerProfile{
|
queueingHintMap: QueueingHintMapPerProfile{
|
||||||
"": {
|
"": {
|
||||||
// fooPlugin1 has no queueing hint function for NodeAdd.
|
// fooPlugin1 has no queueing hint function for framework.NodeAdd.
|
||||||
AssignedPodAdd: {
|
framework.AssignedPodAdd: {
|
||||||
{
|
{
|
||||||
// It will be ignored because the event is not NodeAdd.
|
// It will be ignored because the event is not framework.NodeAdd.
|
||||||
PluginName: "fooPlugin1",
|
PluginName: "fooPlugin1",
|
||||||
QueueingHintFn: queueHintReturnQueue,
|
QueueingHintFn: queueHintReturnQueue,
|
||||||
},
|
},
|
||||||
@ -502,7 +502,7 @@ func Test_InFlightPods(t *testing.T) {
|
|||||||
initialPods: []*v1.Pod{pod},
|
initialPods: []*v1.Pod{pod},
|
||||||
actions: []action{
|
actions: []action{
|
||||||
{podPopped: pod},
|
{podPopped: pod},
|
||||||
{eventHappens: &AssignedPodAdd},
|
{eventHappens: &framework.AssignedPodAdd},
|
||||||
{podEnqueued: newQueuedPodInfoForLookup(pod, "fooPlugin1")},
|
{podEnqueued: newQueuedPodInfoForLookup(pod, "fooPlugin1")},
|
||||||
},
|
},
|
||||||
wantUnschedPodPoolPodNames: []string{"targetpod"},
|
wantUnschedPodPoolPodNames: []string{"targetpod"},
|
||||||
@ -510,9 +510,9 @@ func Test_InFlightPods(t *testing.T) {
|
|||||||
wantInFlightEvents: nil,
|
wantInFlightEvents: nil,
|
||||||
queueingHintMap: QueueingHintMapPerProfile{
|
queueingHintMap: QueueingHintMapPerProfile{
|
||||||
"": {
|
"": {
|
||||||
// fooPlugin1 has a queueing hint function for AssignedPodAdd,
|
// fooPlugin1 has a queueing hint function for framework.AssignedPodAdd,
|
||||||
// but hint fn tells that this event doesn't make a Pod scheudlable.
|
// but hint fn tells that this event doesn't make a Pod scheudlable.
|
||||||
AssignedPodAdd: {
|
framework.AssignedPodAdd: {
|
||||||
{
|
{
|
||||||
PluginName: "fooPlugin1",
|
PluginName: "fooPlugin1",
|
||||||
QueueingHintFn: queueHintReturnSkip,
|
QueueingHintFn: queueHintReturnSkip,
|
||||||
@ -527,7 +527,7 @@ func Test_InFlightPods(t *testing.T) {
|
|||||||
initialPods: []*v1.Pod{pod},
|
initialPods: []*v1.Pod{pod},
|
||||||
actions: []action{
|
actions: []action{
|
||||||
{podPopped: pod},
|
{podPopped: pod},
|
||||||
{eventHappens: &AssignedPodAdd},
|
{eventHappens: &framework.AssignedPodAdd},
|
||||||
{podEnqueued: &framework.QueuedPodInfo{
|
{podEnqueued: &framework.QueuedPodInfo{
|
||||||
PodInfo: mustNewPodInfo(pod),
|
PodInfo: mustNewPodInfo(pod),
|
||||||
UnschedulablePlugins: sets.New("fooPlugin2", "fooPlugin3"),
|
UnschedulablePlugins: sets.New("fooPlugin2", "fooPlugin3"),
|
||||||
@ -539,7 +539,7 @@ func Test_InFlightPods(t *testing.T) {
|
|||||||
wantInFlightEvents: nil,
|
wantInFlightEvents: nil,
|
||||||
queueingHintMap: QueueingHintMapPerProfile{
|
queueingHintMap: QueueingHintMapPerProfile{
|
||||||
"": {
|
"": {
|
||||||
AssignedPodAdd: {
|
framework.AssignedPodAdd: {
|
||||||
{
|
{
|
||||||
PluginName: "fooPlugin3",
|
PluginName: "fooPlugin3",
|
||||||
QueueingHintFn: queueHintReturnSkip,
|
QueueingHintFn: queueHintReturnSkip,
|
||||||
@ -563,7 +563,7 @@ func Test_InFlightPods(t *testing.T) {
|
|||||||
initialPods: []*v1.Pod{pod},
|
initialPods: []*v1.Pod{pod},
|
||||||
actions: []action{
|
actions: []action{
|
||||||
{podPopped: pod},
|
{podPopped: pod},
|
||||||
{eventHappens: &AssignedPodAdd},
|
{eventHappens: &framework.AssignedPodAdd},
|
||||||
{podEnqueued: newQueuedPodInfoForLookup(pod, "fooPlugin1", "fooPlugin2")},
|
{podEnqueued: newQueuedPodInfoForLookup(pod, "fooPlugin1", "fooPlugin2")},
|
||||||
},
|
},
|
||||||
wantBackoffQPodNames: []string{"targetpod"},
|
wantBackoffQPodNames: []string{"targetpod"},
|
||||||
@ -571,7 +571,7 @@ func Test_InFlightPods(t *testing.T) {
|
|||||||
wantInFlightEvents: nil,
|
wantInFlightEvents: nil,
|
||||||
queueingHintMap: QueueingHintMapPerProfile{
|
queueingHintMap: QueueingHintMapPerProfile{
|
||||||
"": {
|
"": {
|
||||||
AssignedPodAdd: {
|
framework.AssignedPodAdd: {
|
||||||
{
|
{
|
||||||
// it will be ignored because the hint fn returns Skip that is weaker than queueHintReturnQueue from fooPlugin1.
|
// it will be ignored because the hint fn returns Skip that is weaker than queueHintReturnQueue from fooPlugin1.
|
||||||
PluginName: "fooPlugin2",
|
PluginName: "fooPlugin2",
|
||||||
@ -592,9 +592,9 @@ func Test_InFlightPods(t *testing.T) {
|
|||||||
initialPods: []*v1.Pod{pod, pod2},
|
initialPods: []*v1.Pod{pod, pod2},
|
||||||
actions: []action{
|
actions: []action{
|
||||||
{callback: func(t *testing.T, q *PriorityQueue) { poppedPod = popPod(t, logger, q, pod) }},
|
{callback: func(t *testing.T, q *PriorityQueue) { poppedPod = popPod(t, logger, q, pod) }},
|
||||||
{eventHappens: &NodeAdd},
|
{eventHappens: &framework.NodeAdd},
|
||||||
{callback: func(t *testing.T, q *PriorityQueue) { poppedPod2 = popPod(t, logger, q, pod2) }},
|
{callback: func(t *testing.T, q *PriorityQueue) { poppedPod2 = popPod(t, logger, q, pod2) }},
|
||||||
{eventHappens: &AssignedPodAdd},
|
{eventHappens: &framework.AssignedPodAdd},
|
||||||
{callback: func(t *testing.T, q *PriorityQueue) {
|
{callback: func(t *testing.T, q *PriorityQueue) {
|
||||||
logger, _ := ktesting.NewTestContext(t)
|
logger, _ := ktesting.NewTestContext(t)
|
||||||
err := q.AddUnschedulableIfNotPresent(logger, poppedPod, q.SchedulingCycle())
|
err := q.AddUnschedulableIfNotPresent(logger, poppedPod, q.SchedulingCycle())
|
||||||
@ -617,7 +617,7 @@ func Test_InFlightPods(t *testing.T) {
|
|||||||
wantInFlightEvents: nil,
|
wantInFlightEvents: nil,
|
||||||
queueingHintMap: QueueingHintMapPerProfile{
|
queueingHintMap: QueueingHintMapPerProfile{
|
||||||
"": {
|
"": {
|
||||||
AssignedPodAdd: {
|
framework.AssignedPodAdd: {
|
||||||
{
|
{
|
||||||
// it will be ignored because the hint fn returns QueueSkip that is weaker than queueHintReturnQueueImmediately from fooPlugin1.
|
// it will be ignored because the hint fn returns QueueSkip that is weaker than queueHintReturnQueueImmediately from fooPlugin1.
|
||||||
PluginName: "fooPlugin3",
|
PluginName: "fooPlugin3",
|
||||||
@ -653,7 +653,7 @@ func Test_InFlightPods(t *testing.T) {
|
|||||||
t.Errorf("Unexpected error from AddUnschedulableIfNotPresent: %v", err)
|
t.Errorf("Unexpected error from AddUnschedulableIfNotPresent: %v", err)
|
||||||
}
|
}
|
||||||
}},
|
}},
|
||||||
{eventHappens: &PvAdd}, // Active again.
|
{eventHappens: &framework.PvAdd}, // Active again.
|
||||||
{callback: func(t *testing.T, q *PriorityQueue) {
|
{callback: func(t *testing.T, q *PriorityQueue) {
|
||||||
poppedPod = popPod(t, logger, q, pod)
|
poppedPod = popPod(t, logger, q, pod)
|
||||||
if len(poppedPod.UnschedulablePlugins) > 0 {
|
if len(poppedPod.UnschedulablePlugins) > 0 {
|
||||||
@ -670,7 +670,7 @@ func Test_InFlightPods(t *testing.T) {
|
|||||||
},
|
},
|
||||||
queueingHintMap: QueueingHintMapPerProfile{
|
queueingHintMap: QueueingHintMapPerProfile{
|
||||||
"": {
|
"": {
|
||||||
PvAdd: {
|
framework.PvAdd: {
|
||||||
{
|
{
|
||||||
// The hint fn tells that this event makes a Pod scheudlable immediately.
|
// The hint fn tells that this event makes a Pod scheudlable immediately.
|
||||||
PluginName: "fooPlugin1",
|
PluginName: "fooPlugin1",
|
||||||
@ -808,7 +808,7 @@ func TestPop(t *testing.T) {
|
|||||||
pod := st.MakePod().Name("targetpod").UID("pod1").Obj()
|
pod := st.MakePod().Name("targetpod").UID("pod1").Obj()
|
||||||
queueingHintMap := QueueingHintMapPerProfile{
|
queueingHintMap := QueueingHintMapPerProfile{
|
||||||
"": {
|
"": {
|
||||||
PvAdd: {
|
framework.PvAdd: {
|
||||||
{
|
{
|
||||||
// The hint fn tells that this event makes a Pod scheudlable.
|
// The hint fn tells that this event makes a Pod scheudlable.
|
||||||
PluginName: "fooPlugin1",
|
PluginName: "fooPlugin1",
|
||||||
@ -836,7 +836,7 @@ func TestPop(t *testing.T) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// Activate it again.
|
// Activate it again.
|
||||||
q.MoveAllToActiveOrBackoffQueue(logger, PvAdd, nil, nil, nil)
|
q.MoveAllToActiveOrBackoffQueue(logger, framework.PvAdd, nil, nil, nil)
|
||||||
|
|
||||||
// Now check result of Pop.
|
// Now check result of Pop.
|
||||||
poppedPod = popPod(t, logger, q, pod)
|
poppedPod = popPod(t, logger, q, pod)
|
||||||
@ -917,7 +917,7 @@ func TestPriorityQueue_AddUnschedulableIfNotPresent_Backoff(t *testing.T) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// move all pods to active queue when we were trying to schedule them
|
// move all pods to active queue when we were trying to schedule them
|
||||||
q.MoveAllToActiveOrBackoffQueue(logger, WildCardEvent, nil, nil, nil)
|
q.MoveAllToActiveOrBackoffQueue(logger, framework.WildCardEvent, nil, nil, nil)
|
||||||
oldCycle := q.SchedulingCycle()
|
oldCycle := q.SchedulingCycle()
|
||||||
|
|
||||||
firstPod, _ := q.Pop(logger)
|
firstPod, _ := q.Pop(logger)
|
||||||
@ -981,7 +981,7 @@ func TestPriorityQueue_Update(t *testing.T) {
|
|||||||
skipPlugin := "skipPlugin"
|
skipPlugin := "skipPlugin"
|
||||||
queueingHintMap := QueueingHintMapPerProfile{
|
queueingHintMap := QueueingHintMapPerProfile{
|
||||||
"": {
|
"": {
|
||||||
UnscheduledPodUpdate: {
|
framework.UnscheduledPodUpdate: {
|
||||||
{
|
{
|
||||||
PluginName: queuePlugin,
|
PluginName: queuePlugin,
|
||||||
QueueingHintFn: queueHintReturnQueue,
|
QueueingHintFn: queueHintReturnQueue,
|
||||||
@ -1184,7 +1184,7 @@ func TestPriorityQueue_UpdateWhenInflight(t *testing.T) {
|
|||||||
featuregatetesting.SetFeatureGateDuringTest(t, utilfeature.DefaultFeatureGate, features.SchedulerQueueingHints, true)
|
featuregatetesting.SetFeatureGateDuringTest(t, utilfeature.DefaultFeatureGate, features.SchedulerQueueingHints, true)
|
||||||
m := makeEmptyQueueingHintMapPerProfile()
|
m := makeEmptyQueueingHintMapPerProfile()
|
||||||
// fakePlugin could change its scheduling result by any updates in Pods.
|
// fakePlugin could change its scheduling result by any updates in Pods.
|
||||||
m[""][UnscheduledPodUpdate] = []*QueueingHintFunction{
|
m[""][framework.UnscheduledPodUpdate] = []*QueueingHintFunction{
|
||||||
{
|
{
|
||||||
PluginName: "fakePlugin",
|
PluginName: "fakePlugin",
|
||||||
QueueingHintFn: queueHintReturnQueue,
|
QueueingHintFn: queueHintReturnQueue,
|
||||||
@ -1405,7 +1405,7 @@ func TestPriorityQueue_addToActiveQ(t *testing.T) {
|
|||||||
m := map[string][]framework.PreEnqueuePlugin{"": tt.plugins}
|
m := map[string][]framework.PreEnqueuePlugin{"": tt.plugins}
|
||||||
q := NewTestQueueWithObjects(ctx, newDefaultQueueSort(), []runtime.Object{tt.pod}, WithPreEnqueuePluginMap(m),
|
q := NewTestQueueWithObjects(ctx, newDefaultQueueSort(), []runtime.Object{tt.pod}, WithPreEnqueuePluginMap(m),
|
||||||
WithPodInitialBackoffDuration(time.Second*30), WithPodMaxBackoffDuration(time.Second*60))
|
WithPodInitialBackoffDuration(time.Second*30), WithPodMaxBackoffDuration(time.Second*60))
|
||||||
got, _ := q.moveToActiveQ(logger, q.newQueuedPodInfo(tt.pod), PodAdd)
|
got, _ := q.moveToActiveQ(logger, q.newQueuedPodInfo(tt.pod), framework.PodAdd)
|
||||||
if got != tt.wantSuccess {
|
if got != tt.wantSuccess {
|
||||||
t.Errorf("Unexpected result: want %v, but got %v", tt.wantSuccess, got)
|
t.Errorf("Unexpected result: want %v, but got %v", tt.wantSuccess, got)
|
||||||
}
|
}
|
||||||
@ -1432,11 +1432,11 @@ func BenchmarkMoveAllToActiveOrBackoffQueue(b *testing.B) {
|
|||||||
}{
|
}{
|
||||||
{
|
{
|
||||||
name: "baseline",
|
name: "baseline",
|
||||||
moveEvent: UnschedulableTimeout,
|
moveEvent: framework.UnschedulableTimeout,
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
name: "worst",
|
name: "worst",
|
||||||
moveEvent: NodeAdd,
|
moveEvent: framework.NodeAdd,
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
name: "random",
|
name: "random",
|
||||||
@ -1450,24 +1450,24 @@ func BenchmarkMoveAllToActiveOrBackoffQueue(b *testing.B) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
events := []framework.ClusterEvent{
|
events := []framework.ClusterEvent{
|
||||||
NodeAdd,
|
framework.NodeAdd,
|
||||||
NodeTaintChange,
|
framework.NodeTaintChange,
|
||||||
NodeAllocatableChange,
|
framework.NodeAllocatableChange,
|
||||||
NodeConditionChange,
|
framework.NodeConditionChange,
|
||||||
NodeLabelChange,
|
framework.NodeLabelChange,
|
||||||
NodeAnnotationChange,
|
framework.NodeAnnotationChange,
|
||||||
PvcAdd,
|
framework.PvcAdd,
|
||||||
PvcUpdate,
|
framework.PvcUpdate,
|
||||||
PvAdd,
|
framework.PvAdd,
|
||||||
PvUpdate,
|
framework.PvUpdate,
|
||||||
StorageClassAdd,
|
framework.StorageClassAdd,
|
||||||
StorageClassUpdate,
|
framework.StorageClassUpdate,
|
||||||
CSINodeAdd,
|
framework.CSINodeAdd,
|
||||||
CSINodeUpdate,
|
framework.CSINodeUpdate,
|
||||||
CSIDriverAdd,
|
framework.CSIDriverAdd,
|
||||||
CSIDriverUpdate,
|
framework.CSIDriverUpdate,
|
||||||
CSIStorageCapacityAdd,
|
framework.CSIStorageCapacityAdd,
|
||||||
CSIStorageCapacityUpdate,
|
framework.CSIStorageCapacityUpdate,
|
||||||
}
|
}
|
||||||
|
|
||||||
pluginNum := 20
|
pluginNum := 20
|
||||||
@ -1486,7 +1486,7 @@ func BenchmarkMoveAllToActiveOrBackoffQueue(b *testing.B) {
|
|||||||
c := testingclock.NewFakeClock(time.Now())
|
c := testingclock.NewFakeClock(time.Now())
|
||||||
|
|
||||||
m := makeEmptyQueueingHintMapPerProfile()
|
m := makeEmptyQueueingHintMapPerProfile()
|
||||||
// - All plugins registered for events[0], which is NodeAdd.
|
// - All plugins registered for events[0], which is framework.NodeAdd.
|
||||||
// - 1/2 of plugins registered for events[1]
|
// - 1/2 of plugins registered for events[1]
|
||||||
// - 1/3 of plugins registered for events[2]
|
// - 1/3 of plugins registered for events[2]
|
||||||
// - ...
|
// - ...
|
||||||
@ -1604,7 +1604,7 @@ func TestPriorityQueue_MoveAllToActiveOrBackoffQueueWithQueueingHint(t *testing.
|
|||||||
t.Run(test.name, func(t *testing.T) {
|
t.Run(test.name, func(t *testing.T) {
|
||||||
logger, ctx := ktesting.NewTestContext(t)
|
logger, ctx := ktesting.NewTestContext(t)
|
||||||
m := makeEmptyQueueingHintMapPerProfile()
|
m := makeEmptyQueueingHintMapPerProfile()
|
||||||
m[""][NodeAdd] = []*QueueingHintFunction{
|
m[""][framework.NodeAdd] = []*QueueingHintFunction{
|
||||||
{
|
{
|
||||||
PluginName: "foo",
|
PluginName: "foo",
|
||||||
QueueingHintFn: test.hint,
|
QueueingHintFn: test.hint,
|
||||||
@ -1623,7 +1623,7 @@ func TestPriorityQueue_MoveAllToActiveOrBackoffQueueWithQueueingHint(t *testing.
|
|||||||
}
|
}
|
||||||
cl.Step(test.duration)
|
cl.Step(test.duration)
|
||||||
|
|
||||||
q.MoveAllToActiveOrBackoffQueue(logger, NodeAdd, nil, nil, nil)
|
q.MoveAllToActiveOrBackoffQueue(logger, framework.NodeAdd, nil, nil, nil)
|
||||||
|
|
||||||
if q.podBackoffQ.Len() == 0 && test.expectedQ == backoffQ {
|
if q.podBackoffQ.Len() == 0 && test.expectedQ == backoffQ {
|
||||||
t.Fatalf("expected pod to be queued to backoffQ, but it was not")
|
t.Fatalf("expected pod to be queued to backoffQ, but it was not")
|
||||||
@ -1648,7 +1648,7 @@ func TestPriorityQueue_MoveAllToActiveOrBackoffQueue(t *testing.T) {
|
|||||||
m := makeEmptyQueueingHintMapPerProfile()
|
m := makeEmptyQueueingHintMapPerProfile()
|
||||||
featuregatetesting.SetFeatureGateDuringTest(t, utilfeature.DefaultFeatureGate, features.SchedulerQueueingHints, true)
|
featuregatetesting.SetFeatureGateDuringTest(t, utilfeature.DefaultFeatureGate, features.SchedulerQueueingHints, true)
|
||||||
|
|
||||||
m[""][NodeAdd] = []*QueueingHintFunction{
|
m[""][framework.NodeAdd] = []*QueueingHintFunction{
|
||||||
{
|
{
|
||||||
PluginName: "fooPlugin",
|
PluginName: "fooPlugin",
|
||||||
QueueingHintFn: queueHintReturnQueue,
|
QueueingHintFn: queueHintReturnQueue,
|
||||||
@ -1703,7 +1703,7 @@ func TestPriorityQueue_MoveAllToActiveOrBackoffQueue(t *testing.T) {
|
|||||||
expectInFlightPods(t, q)
|
expectInFlightPods(t, q)
|
||||||
// This NodeAdd event moves unschedulablePodInfo and highPriorityPodInfo to the backoffQ,
|
// This NodeAdd event moves unschedulablePodInfo and highPriorityPodInfo to the backoffQ,
|
||||||
// because of the queueing hint function registered for NodeAdd/fooPlugin.
|
// because of the queueing hint function registered for NodeAdd/fooPlugin.
|
||||||
q.MoveAllToActiveOrBackoffQueue(logger, NodeAdd, nil, nil, nil)
|
q.MoveAllToActiveOrBackoffQueue(logger, framework.NodeAdd, nil, nil, nil)
|
||||||
q.Add(logger, medPriorityPodInfo.Pod)
|
q.Add(logger, medPriorityPodInfo.Pod)
|
||||||
if q.activeQ.Len() != 1 {
|
if q.activeQ.Len() != 1 {
|
||||||
t.Errorf("Expected 1 item to be in activeQ, but got: %v", q.activeQ.Len())
|
t.Errorf("Expected 1 item to be in activeQ, but got: %v", q.activeQ.Len())
|
||||||
@ -1775,7 +1775,7 @@ func TestPriorityQueue_MoveAllToActiveOrBackoffQueue(t *testing.T) {
|
|||||||
// and the pods will be moved into activeQ.
|
// and the pods will be moved into activeQ.
|
||||||
c.Step(q.podInitialBackoffDuration)
|
c.Step(q.podInitialBackoffDuration)
|
||||||
q.flushBackoffQCompleted(logger) // flush the completed backoffQ to move hpp1 to activeQ.
|
q.flushBackoffQCompleted(logger) // flush the completed backoffQ to move hpp1 to activeQ.
|
||||||
q.MoveAllToActiveOrBackoffQueue(logger, NodeAdd, nil, nil, nil)
|
q.MoveAllToActiveOrBackoffQueue(logger, framework.NodeAdd, nil, nil, nil)
|
||||||
if q.activeQ.Len() != 4 {
|
if q.activeQ.Len() != 4 {
|
||||||
t.Errorf("Expected 4 items to be in activeQ, but got: %v", q.activeQ.Len())
|
t.Errorf("Expected 4 items to be in activeQ, but got: %v", q.activeQ.Len())
|
||||||
}
|
}
|
||||||
@ -1796,7 +1796,7 @@ func TestPriorityQueue_MoveAllToActiveOrBackoffQueueWithOutQueueingHint(t *testi
|
|||||||
defer cancel()
|
defer cancel()
|
||||||
m := makeEmptyQueueingHintMapPerProfile()
|
m := makeEmptyQueueingHintMapPerProfile()
|
||||||
featuregatetesting.SetFeatureGateDuringTest(t, utilfeature.DefaultFeatureGate, features.SchedulerQueueingHints, false)
|
featuregatetesting.SetFeatureGateDuringTest(t, utilfeature.DefaultFeatureGate, features.SchedulerQueueingHints, false)
|
||||||
m[""][NodeAdd] = []*QueueingHintFunction{
|
m[""][framework.NodeAdd] = []*QueueingHintFunction{
|
||||||
{
|
{
|
||||||
PluginName: "fooPlugin",
|
PluginName: "fooPlugin",
|
||||||
QueueingHintFn: queueHintReturnQueue,
|
QueueingHintFn: queueHintReturnQueue,
|
||||||
@ -1832,7 +1832,7 @@ func TestPriorityQueue_MoveAllToActiveOrBackoffQueueWithOutQueueingHint(t *testi
|
|||||||
}
|
}
|
||||||
// This NodeAdd event moves unschedulablePodInfo and highPriorityPodInfo to the backoffQ,
|
// This NodeAdd event moves unschedulablePodInfo and highPriorityPodInfo to the backoffQ,
|
||||||
// because of the queueing hint function registered for NodeAdd/fooPlugin.
|
// because of the queueing hint function registered for NodeAdd/fooPlugin.
|
||||||
q.MoveAllToActiveOrBackoffQueue(logger, NodeAdd, nil, nil, nil)
|
q.MoveAllToActiveOrBackoffQueue(logger, framework.NodeAdd, nil, nil, nil)
|
||||||
if q.activeQ.Len() != 1 {
|
if q.activeQ.Len() != 1 {
|
||||||
t.Errorf("Expected 1 item to be in activeQ, but got: %v", q.activeQ.Len())
|
t.Errorf("Expected 1 item to be in activeQ, but got: %v", q.activeQ.Len())
|
||||||
}
|
}
|
||||||
@ -1888,7 +1888,7 @@ func TestPriorityQueue_MoveAllToActiveOrBackoffQueueWithOutQueueingHint(t *testi
|
|||||||
// and the pods will be moved into activeQ.
|
// and the pods will be moved into activeQ.
|
||||||
c.Step(q.podInitialBackoffDuration)
|
c.Step(q.podInitialBackoffDuration)
|
||||||
q.flushBackoffQCompleted(logger) // flush the completed backoffQ to move hpp1 to activeQ.
|
q.flushBackoffQCompleted(logger) // flush the completed backoffQ to move hpp1 to activeQ.
|
||||||
q.MoveAllToActiveOrBackoffQueue(logger, NodeAdd, nil, nil, nil)
|
q.MoveAllToActiveOrBackoffQueue(logger, framework.NodeAdd, nil, nil, nil)
|
||||||
if q.activeQ.Len() != 4 {
|
if q.activeQ.Len() != 4 {
|
||||||
t.Errorf("Expected 4 items to be in activeQ, but got: %v", q.activeQ.Len())
|
t.Errorf("Expected 4 items to be in activeQ, but got: %v", q.activeQ.Len())
|
||||||
}
|
}
|
||||||
@ -1985,7 +1985,7 @@ func TestPriorityQueue_AssignedPodAdded_(t *testing.T) {
|
|||||||
|
|
||||||
c := testingclock.NewFakeClock(time.Now())
|
c := testingclock.NewFakeClock(time.Now())
|
||||||
m := makeEmptyQueueingHintMapPerProfile()
|
m := makeEmptyQueueingHintMapPerProfile()
|
||||||
m[""][AssignedPodAdd] = []*QueueingHintFunction{
|
m[""][framework.AssignedPodAdd] = []*QueueingHintFunction{
|
||||||
{
|
{
|
||||||
PluginName: "fakePlugin",
|
PluginName: "fakePlugin",
|
||||||
QueueingHintFn: queueHintReturnQueue,
|
QueueingHintFn: queueHintReturnQueue,
|
||||||
@ -2146,7 +2146,7 @@ func TestPriorityQueue_PendingPods(t *testing.T) {
|
|||||||
t.Errorf("Unexpected pending pods summary: want %v, but got %v.", wantSummary, gotSummary)
|
t.Errorf("Unexpected pending pods summary: want %v, but got %v.", wantSummary, gotSummary)
|
||||||
}
|
}
|
||||||
// Move all to active queue. We should still see the same set of pods.
|
// Move all to active queue. We should still see the same set of pods.
|
||||||
q.MoveAllToActiveOrBackoffQueue(logger, WildCardEvent, nil, nil, nil)
|
q.MoveAllToActiveOrBackoffQueue(logger, framework.WildCardEvent, nil, nil, nil)
|
||||||
gotPods, gotSummary = q.PendingPods()
|
gotPods, gotSummary = q.PendingPods()
|
||||||
if diff := cmp.Diff(expectedSet, makeSet(gotPods)); diff != "" {
|
if diff := cmp.Diff(expectedSet, makeSet(gotPods)); diff != "" {
|
||||||
t.Errorf("Unexpected list of pending Pods (-want, +got):\n%s", diff)
|
t.Errorf("Unexpected list of pending Pods (-want, +got):\n%s", diff)
|
||||||
@ -2432,7 +2432,7 @@ func TestRecentlyTriedPodsGoBack(t *testing.T) {
|
|||||||
}
|
}
|
||||||
c.Step(DefaultPodInitialBackoffDuration)
|
c.Step(DefaultPodInitialBackoffDuration)
|
||||||
// Move all unschedulable pods to the active queue.
|
// Move all unschedulable pods to the active queue.
|
||||||
q.MoveAllToActiveOrBackoffQueue(logger, UnschedulableTimeout, nil, nil, nil)
|
q.MoveAllToActiveOrBackoffQueue(logger, framework.UnschedulableTimeout, nil, nil, nil)
|
||||||
// Simulation is over. Now let's pop all pods. The pod popped first should be
|
// Simulation is over. Now let's pop all pods. The pod popped first should be
|
||||||
// the last one we pop here.
|
// the last one we pop here.
|
||||||
for i := 0; i < 5; i++ {
|
for i := 0; i < 5; i++ {
|
||||||
@ -2483,7 +2483,7 @@ func TestPodFailedSchedulingMultipleTimesDoesNotBlockNewerPod(t *testing.T) {
|
|||||||
// Move clock to make the unschedulable pods complete backoff.
|
// Move clock to make the unschedulable pods complete backoff.
|
||||||
c.Step(DefaultPodInitialBackoffDuration + time.Second)
|
c.Step(DefaultPodInitialBackoffDuration + time.Second)
|
||||||
// Move all unschedulable pods to the active queue.
|
// Move all unschedulable pods to the active queue.
|
||||||
q.MoveAllToActiveOrBackoffQueue(logger, UnschedulableTimeout, nil, nil, nil)
|
q.MoveAllToActiveOrBackoffQueue(logger, framework.UnschedulableTimeout, nil, nil, nil)
|
||||||
|
|
||||||
// Simulate a pod being popped by the scheduler,
|
// Simulate a pod being popped by the scheduler,
|
||||||
// At this time, unschedulable pod should be popped.
|
// At this time, unschedulable pod should be popped.
|
||||||
@ -2516,7 +2516,7 @@ func TestPodFailedSchedulingMultipleTimesDoesNotBlockNewerPod(t *testing.T) {
|
|||||||
// Move clock to make the unschedulable pods complete backoff.
|
// Move clock to make the unschedulable pods complete backoff.
|
||||||
c.Step(DefaultPodInitialBackoffDuration + time.Second)
|
c.Step(DefaultPodInitialBackoffDuration + time.Second)
|
||||||
// Move all unschedulable pods to the active queue.
|
// Move all unschedulable pods to the active queue.
|
||||||
q.MoveAllToActiveOrBackoffQueue(logger, UnschedulableTimeout, nil, nil, nil)
|
q.MoveAllToActiveOrBackoffQueue(logger, framework.UnschedulableTimeout, nil, nil, nil)
|
||||||
|
|
||||||
// At this time, newerPod should be popped
|
// At this time, newerPod should be popped
|
||||||
// because it is the oldest tried pod.
|
// because it is the oldest tried pod.
|
||||||
@ -2563,7 +2563,7 @@ func TestHighPriorityBackoff(t *testing.T) {
|
|||||||
t.Fatalf("unexpected error from AddUnschedulableIfNotPresent: %v", err)
|
t.Fatalf("unexpected error from AddUnschedulableIfNotPresent: %v", err)
|
||||||
}
|
}
|
||||||
// Move all unschedulable pods to the active queue.
|
// Move all unschedulable pods to the active queue.
|
||||||
q.MoveAllToActiveOrBackoffQueue(logger, WildCardEvent, nil, nil, nil)
|
q.MoveAllToActiveOrBackoffQueue(logger, framework.WildCardEvent, nil, nil, nil)
|
||||||
|
|
||||||
p, err = q.Pop(logger)
|
p, err = q.Pop(logger)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
@ -2579,7 +2579,7 @@ func TestHighPriorityBackoff(t *testing.T) {
|
|||||||
func TestHighPriorityFlushUnschedulablePodsLeftover(t *testing.T) {
|
func TestHighPriorityFlushUnschedulablePodsLeftover(t *testing.T) {
|
||||||
c := testingclock.NewFakeClock(time.Now())
|
c := testingclock.NewFakeClock(time.Now())
|
||||||
m := makeEmptyQueueingHintMapPerProfile()
|
m := makeEmptyQueueingHintMapPerProfile()
|
||||||
m[""][NodeAdd] = []*QueueingHintFunction{
|
m[""][framework.NodeAdd] = []*QueueingHintFunction{
|
||||||
{
|
{
|
||||||
PluginName: "fakePlugin",
|
PluginName: "fakePlugin",
|
||||||
QueueingHintFn: queueHintReturnQueue,
|
QueueingHintFn: queueHintReturnQueue,
|
||||||
@ -2796,7 +2796,7 @@ var (
|
|||||||
queue.podBackoffQ.Add(pInfo)
|
queue.podBackoffQ.Add(pInfo)
|
||||||
}
|
}
|
||||||
moveAllToActiveOrBackoffQ = func(t *testing.T, logger klog.Logger, queue *PriorityQueue, _ *framework.QueuedPodInfo) {
|
moveAllToActiveOrBackoffQ = func(t *testing.T, logger klog.Logger, queue *PriorityQueue, _ *framework.QueuedPodInfo) {
|
||||||
queue.MoveAllToActiveOrBackoffQueue(logger, UnschedulableTimeout, nil, nil, nil)
|
queue.MoveAllToActiveOrBackoffQueue(logger, framework.UnschedulableTimeout, nil, nil, nil)
|
||||||
}
|
}
|
||||||
flushBackoffQ = func(t *testing.T, logger klog.Logger, queue *PriorityQueue, _ *framework.QueuedPodInfo) {
|
flushBackoffQ = func(t *testing.T, logger klog.Logger, queue *PriorityQueue, _ *framework.QueuedPodInfo) {
|
||||||
queue.clock.(*testingclock.FakeClock).Step(2 * time.Second)
|
queue.clock.(*testingclock.FakeClock).Step(2 * time.Second)
|
||||||
@ -3431,7 +3431,7 @@ func TestBackOffFlow(t *testing.T) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// An event happens.
|
// An event happens.
|
||||||
q.MoveAllToActiveOrBackoffQueue(logger, UnschedulableTimeout, nil, nil, nil)
|
q.MoveAllToActiveOrBackoffQueue(logger, framework.UnschedulableTimeout, nil, nil, nil)
|
||||||
|
|
||||||
if _, ok, _ := q.podBackoffQ.Get(podInfo); !ok {
|
if _, ok, _ := q.podBackoffQ.Get(podInfo); !ok {
|
||||||
t.Errorf("pod %v is not in the backoff queue", podID)
|
t.Errorf("pod %v is not in the backoff queue", podID)
|
||||||
@ -3480,20 +3480,20 @@ func TestMoveAllToActiveOrBackoffQueue_PreEnqueueChecks(t *testing.T) {
|
|||||||
{
|
{
|
||||||
name: "nil PreEnqueueCheck",
|
name: "nil PreEnqueueCheck",
|
||||||
podInfos: podInfos,
|
podInfos: podInfos,
|
||||||
event: WildCardEvent,
|
event: framework.WildCardEvent,
|
||||||
want: []string{"p0", "p1", "p2", "p3", "p4"},
|
want: []string{"p0", "p1", "p2", "p3", "p4"},
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
name: "move Pods with priority greater than 2",
|
name: "move Pods with priority greater than 2",
|
||||||
podInfos: podInfos,
|
podInfos: podInfos,
|
||||||
event: WildCardEvent,
|
event: framework.WildCardEvent,
|
||||||
preEnqueueCheck: func(pod *v1.Pod) bool { return *pod.Spec.Priority >= 2 },
|
preEnqueueCheck: func(pod *v1.Pod) bool { return *pod.Spec.Priority >= 2 },
|
||||||
want: []string{"p2", "p3", "p4"},
|
want: []string{"p2", "p3", "p4"},
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
name: "move Pods with even priority and greater than 2",
|
name: "move Pods with even priority and greater than 2",
|
||||||
podInfos: podInfos,
|
podInfos: podInfos,
|
||||||
event: WildCardEvent,
|
event: framework.WildCardEvent,
|
||||||
preEnqueueCheck: func(pod *v1.Pod) bool {
|
preEnqueueCheck: func(pod *v1.Pod) bool {
|
||||||
return *pod.Spec.Priority%2 == 0 && *pod.Spec.Priority >= 2
|
return *pod.Spec.Priority%2 == 0 && *pod.Spec.Priority >= 2
|
||||||
},
|
},
|
||||||
@ -3502,7 +3502,7 @@ func TestMoveAllToActiveOrBackoffQueue_PreEnqueueChecks(t *testing.T) {
|
|||||||
{
|
{
|
||||||
name: "move Pods with even and negative priority",
|
name: "move Pods with even and negative priority",
|
||||||
podInfos: podInfos,
|
podInfos: podInfos,
|
||||||
event: WildCardEvent,
|
event: framework.WildCardEvent,
|
||||||
preEnqueueCheck: func(pod *v1.Pod) bool {
|
preEnqueueCheck: func(pod *v1.Pod) bool {
|
||||||
return *pod.Spec.Priority%2 == 0 && *pod.Spec.Priority < 0
|
return *pod.Spec.Priority%2 == 0 && *pod.Spec.Priority < 0
|
||||||
},
|
},
|
||||||
@ -3510,7 +3510,7 @@ func TestMoveAllToActiveOrBackoffQueue_PreEnqueueChecks(t *testing.T) {
|
|||||||
{
|
{
|
||||||
name: "preCheck isn't called if the event is not interested by any plugins",
|
name: "preCheck isn't called if the event is not interested by any plugins",
|
||||||
podInfos: podInfos,
|
podInfos: podInfos,
|
||||||
event: PvAdd, // No plugin is interested in this event.
|
event: framework.PvAdd, // No plugin is interested in this event.
|
||||||
preEnqueueCheck: func(pod *v1.Pod) bool {
|
preEnqueueCheck: func(pod *v1.Pod) bool {
|
||||||
panic("preCheck shouldn't be called")
|
panic("preCheck shouldn't be called")
|
||||||
},
|
},
|
||||||
@ -3665,17 +3665,17 @@ func Test_isPodWorthRequeuing(t *testing.T) {
|
|||||||
UnschedulablePlugins: sets.New("fooPlugin1"),
|
UnschedulablePlugins: sets.New("fooPlugin1"),
|
||||||
PodInfo: mustNewPodInfo(st.MakePod().Name("pod1").Namespace("ns1").UID("1").Obj()),
|
PodInfo: mustNewPodInfo(st.MakePod().Name("pod1").Namespace("ns1").UID("1").Obj()),
|
||||||
},
|
},
|
||||||
event: NodeAdd,
|
event: framework.NodeAdd,
|
||||||
oldObj: nil,
|
oldObj: nil,
|
||||||
newObj: st.MakeNode().Obj(),
|
newObj: st.MakeNode().Obj(),
|
||||||
expected: queueSkip,
|
expected: queueSkip,
|
||||||
expectedExecutionCount: 0,
|
expectedExecutionCount: 0,
|
||||||
queueingHintMap: QueueingHintMapPerProfile{
|
queueingHintMap: QueueingHintMapPerProfile{
|
||||||
"": {
|
"": {
|
||||||
// no queueing hint function for NodeAdd.
|
// no queueing hint function for framework.NodeAdd.
|
||||||
AssignedPodAdd: {
|
framework.AssignedPodAdd: {
|
||||||
{
|
{
|
||||||
// It will be ignored because the event is not NodeAdd.
|
// It will be ignored because the event is not framework.NodeAdd.
|
||||||
PluginName: "fooPlugin1",
|
PluginName: "fooPlugin1",
|
||||||
QueueingHintFn: queueHintReturnQueue,
|
QueueingHintFn: queueHintReturnQueue,
|
||||||
},
|
},
|
||||||
@ -3689,14 +3689,14 @@ func Test_isPodWorthRequeuing(t *testing.T) {
|
|||||||
UnschedulablePlugins: sets.New("fooPlugin1"),
|
UnschedulablePlugins: sets.New("fooPlugin1"),
|
||||||
PodInfo: mustNewPodInfo(st.MakePod().Name("pod1").Namespace("ns1").UID("1").Obj()),
|
PodInfo: mustNewPodInfo(st.MakePod().Name("pod1").Namespace("ns1").UID("1").Obj()),
|
||||||
},
|
},
|
||||||
event: NodeAdd,
|
event: framework.NodeAdd,
|
||||||
oldObj: nil,
|
oldObj: nil,
|
||||||
newObj: st.MakeNode().Obj(),
|
newObj: st.MakeNode().Obj(),
|
||||||
expected: queueAfterBackoff,
|
expected: queueAfterBackoff,
|
||||||
expectedExecutionCount: 1,
|
expectedExecutionCount: 1,
|
||||||
queueingHintMap: QueueingHintMapPerProfile{
|
queueingHintMap: QueueingHintMapPerProfile{
|
||||||
"": {
|
"": {
|
||||||
NodeAdd: {
|
framework.NodeAdd: {
|
||||||
{
|
{
|
||||||
PluginName: "fooPlugin1",
|
PluginName: "fooPlugin1",
|
||||||
QueueingHintFn: queueHintReturnErr,
|
QueueingHintFn: queueHintReturnErr,
|
||||||
@ -3711,7 +3711,7 @@ func Test_isPodWorthRequeuing(t *testing.T) {
|
|||||||
UnschedulablePlugins: sets.New("fooPlugin1"),
|
UnschedulablePlugins: sets.New("fooPlugin1"),
|
||||||
PodInfo: mustNewPodInfo(st.MakePod().Name("pod1").Namespace("ns1").UID("1").Obj()),
|
PodInfo: mustNewPodInfo(st.MakePod().Name("pod1").Namespace("ns1").UID("1").Obj()),
|
||||||
},
|
},
|
||||||
event: WildCardEvent,
|
event: framework.WildCardEvent,
|
||||||
oldObj: nil,
|
oldObj: nil,
|
||||||
newObj: st.MakeNode().Obj(),
|
newObj: st.MakeNode().Obj(),
|
||||||
expected: queueAfterBackoff,
|
expected: queueAfterBackoff,
|
||||||
@ -3725,14 +3725,14 @@ func Test_isPodWorthRequeuing(t *testing.T) {
|
|||||||
PendingPlugins: sets.New("fooPlugin2"),
|
PendingPlugins: sets.New("fooPlugin2"),
|
||||||
PodInfo: mustNewPodInfo(st.MakePod().Name("pod1").Namespace("ns1").UID("1").Obj()),
|
PodInfo: mustNewPodInfo(st.MakePod().Name("pod1").Namespace("ns1").UID("1").Obj()),
|
||||||
},
|
},
|
||||||
event: NodeAdd,
|
event: framework.NodeAdd,
|
||||||
oldObj: nil,
|
oldObj: nil,
|
||||||
newObj: st.MakeNode().Node,
|
newObj: st.MakeNode().Node,
|
||||||
expected: queueImmediately,
|
expected: queueImmediately,
|
||||||
expectedExecutionCount: 2,
|
expectedExecutionCount: 2,
|
||||||
queueingHintMap: QueueingHintMapPerProfile{
|
queueingHintMap: QueueingHintMapPerProfile{
|
||||||
"": {
|
"": {
|
||||||
NodeAdd: {
|
framework.NodeAdd: {
|
||||||
{
|
{
|
||||||
PluginName: "fooPlugin1",
|
PluginName: "fooPlugin1",
|
||||||
// It returns Queue and it's interpreted as queueAfterBackoff.
|
// It returns Queue and it's interpreted as queueAfterBackoff.
|
||||||
@ -3763,14 +3763,14 @@ func Test_isPodWorthRequeuing(t *testing.T) {
|
|||||||
UnschedulablePlugins: sets.New("fooPlugin1", "fooPlugin2"),
|
UnschedulablePlugins: sets.New("fooPlugin1", "fooPlugin2"),
|
||||||
PodInfo: mustNewPodInfo(st.MakePod().Name("pod1").Namespace("ns1").UID("1").Obj()),
|
PodInfo: mustNewPodInfo(st.MakePod().Name("pod1").Namespace("ns1").UID("1").Obj()),
|
||||||
},
|
},
|
||||||
event: NodeAdd,
|
event: framework.NodeAdd,
|
||||||
oldObj: nil,
|
oldObj: nil,
|
||||||
newObj: st.MakeNode().Obj(),
|
newObj: st.MakeNode().Obj(),
|
||||||
expected: queueAfterBackoff,
|
expected: queueAfterBackoff,
|
||||||
expectedExecutionCount: 2,
|
expectedExecutionCount: 2,
|
||||||
queueingHintMap: QueueingHintMapPerProfile{
|
queueingHintMap: QueueingHintMapPerProfile{
|
||||||
"": {
|
"": {
|
||||||
NodeAdd: {
|
framework.NodeAdd: {
|
||||||
{
|
{
|
||||||
// Skip will be ignored
|
// Skip will be ignored
|
||||||
PluginName: "fooPlugin1",
|
PluginName: "fooPlugin1",
|
||||||
@ -3791,14 +3791,14 @@ func Test_isPodWorthRequeuing(t *testing.T) {
|
|||||||
UnschedulablePlugins: sets.New("fooPlugin1", "fooPlugin2"),
|
UnschedulablePlugins: sets.New("fooPlugin1", "fooPlugin2"),
|
||||||
PodInfo: mustNewPodInfo(st.MakePod().Name("pod1").Namespace("ns1").UID("1").Obj()),
|
PodInfo: mustNewPodInfo(st.MakePod().Name("pod1").Namespace("ns1").UID("1").Obj()),
|
||||||
},
|
},
|
||||||
event: NodeAdd,
|
event: framework.NodeAdd,
|
||||||
oldObj: nil,
|
oldObj: nil,
|
||||||
newObj: st.MakeNode().Node,
|
newObj: st.MakeNode().Node,
|
||||||
expected: queueSkip,
|
expected: queueSkip,
|
||||||
expectedExecutionCount: 2,
|
expectedExecutionCount: 2,
|
||||||
queueingHintMap: QueueingHintMapPerProfile{
|
queueingHintMap: QueueingHintMapPerProfile{
|
||||||
"": {
|
"": {
|
||||||
NodeAdd: {
|
framework.NodeAdd: {
|
||||||
{
|
{
|
||||||
PluginName: "fooPlugin1",
|
PluginName: "fooPlugin1",
|
||||||
QueueingHintFn: queueHintReturnSkip,
|
QueueingHintFn: queueHintReturnSkip,
|
||||||
@ -3846,7 +3846,7 @@ func Test_isPodWorthRequeuing(t *testing.T) {
|
|||||||
QueueingHintFn: queueHintReturnQueue,
|
QueueingHintFn: queueHintReturnQueue,
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
NodeAdd: { // not executed because NodeAdd is unrelated.
|
framework.NodeAdd: { // not executed because NodeAdd is unrelated.
|
||||||
{
|
{
|
||||||
PluginName: "fooPlugin1",
|
PluginName: "fooPlugin1",
|
||||||
QueueingHintFn: queueHintReturnQueue,
|
QueueingHintFn: queueHintReturnQueue,
|
||||||
|
@ -38,7 +38,6 @@ import (
|
|||||||
"k8s.io/kubernetes/pkg/apis/core/validation"
|
"k8s.io/kubernetes/pkg/apis/core/validation"
|
||||||
"k8s.io/kubernetes/pkg/scheduler/framework"
|
"k8s.io/kubernetes/pkg/scheduler/framework"
|
||||||
"k8s.io/kubernetes/pkg/scheduler/framework/parallelize"
|
"k8s.io/kubernetes/pkg/scheduler/framework/parallelize"
|
||||||
internalqueue "k8s.io/kubernetes/pkg/scheduler/internal/queue"
|
|
||||||
"k8s.io/kubernetes/pkg/scheduler/metrics"
|
"k8s.io/kubernetes/pkg/scheduler/metrics"
|
||||||
"k8s.io/kubernetes/pkg/scheduler/util"
|
"k8s.io/kubernetes/pkg/scheduler/util"
|
||||||
utiltrace "k8s.io/utils/trace"
|
utiltrace "k8s.io/utils/trace"
|
||||||
@ -355,11 +354,11 @@ func (sched *Scheduler) handleBindingCycleError(
|
|||||||
// It's intentional to "defer" this operation; otherwise MoveAllToActiveOrBackoffQueue() would
|
// It's intentional to "defer" this operation; otherwise MoveAllToActiveOrBackoffQueue() would
|
||||||
// add this event to in-flight events and thus move the assumed pod to backoffQ anyways if the plugins don't have appropriate QueueingHint.
|
// add this event to in-flight events and thus move the assumed pod to backoffQ anyways if the plugins don't have appropriate QueueingHint.
|
||||||
if status.IsRejected() {
|
if status.IsRejected() {
|
||||||
defer sched.SchedulingQueue.MoveAllToActiveOrBackoffQueue(logger, internalqueue.AssignedPodDelete, assumedPod, nil, func(pod *v1.Pod) bool {
|
defer sched.SchedulingQueue.MoveAllToActiveOrBackoffQueue(logger, framework.AssignedPodDelete, assumedPod, nil, func(pod *v1.Pod) bool {
|
||||||
return assumedPod.UID != pod.UID
|
return assumedPod.UID != pod.UID
|
||||||
})
|
})
|
||||||
} else {
|
} else {
|
||||||
sched.SchedulingQueue.MoveAllToActiveOrBackoffQueue(logger, internalqueue.AssignedPodDelete, assumedPod, nil, nil)
|
sched.SchedulingQueue.MoveAllToActiveOrBackoffQueue(logger, framework.AssignedPodDelete, assumedPod, nil, nil)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
Loading…
Reference in New Issue
Block a user