mirror of
https://github.com/k3s-io/kubernetes.git
synced 2025-07-31 23:37:01 +00:00
MarkPodsNotReady retry fix
This commit is contained in:
parent
8577711b61
commit
9759a982c5
@ -236,6 +236,8 @@ type Controller struct {
|
||||
// workers that are responsible for tainting nodes.
|
||||
zoneNoExecuteTainter map[string]*scheduler.RateLimitedTimedQueue
|
||||
|
||||
nodesToRetry sync.Map
|
||||
|
||||
zoneStates map[string]ZoneState
|
||||
|
||||
daemonSetStore appsv1listers.DaemonSetLister
|
||||
@ -351,6 +353,7 @@ func NewNodeLifecycleController(
|
||||
nodeMonitorGracePeriod: nodeMonitorGracePeriod,
|
||||
zonePodEvictor: make(map[string]*scheduler.RateLimitedTimedQueue),
|
||||
zoneNoExecuteTainter: make(map[string]*scheduler.RateLimitedTimedQueue),
|
||||
nodesToRetry: sync.Map{},
|
||||
zoneStates: make(map[string]ZoneState),
|
||||
podEvictionTimeout: podEvictionTimeout,
|
||||
evictionLimiterQPS: evictionLimiterQPS,
|
||||
@ -467,6 +470,10 @@ func NewNodeLifecycleController(
|
||||
nc.nodeUpdateQueue.Add(newNode.Name)
|
||||
return nil
|
||||
}),
|
||||
DeleteFunc: nodeutil.CreateDeleteNodeHandler(func(node *v1.Node) error {
|
||||
nc.nodesToRetry.Delete(node.Name)
|
||||
return nil
|
||||
}),
|
||||
})
|
||||
|
||||
if nc.taintNodeByCondition {
|
||||
@ -787,25 +794,38 @@ func (nc *Controller) monitorNodeHealth() error {
|
||||
nc.processNoTaintBaseEviction(node, &observedReadyCondition, gracePeriod)
|
||||
}
|
||||
|
||||
// Report node event.
|
||||
if currentReadyCondition.Status != v1.ConditionTrue && observedReadyCondition.Status == v1.ConditionTrue {
|
||||
_, needsRetry := nc.nodesToRetry.Load(node.Name)
|
||||
switch {
|
||||
case currentReadyCondition.Status != v1.ConditionTrue && observedReadyCondition.Status == v1.ConditionTrue:
|
||||
// Report node event only once when status changed.
|
||||
nodeutil.RecordNodeStatusChange(nc.recorder, node, "NodeNotReady")
|
||||
pods, err := listPodsFromNode(nc.kubeClient, node.Name)
|
||||
if err != nil {
|
||||
utilruntime.HandleError(fmt.Errorf("Unable to list pods from node %v: %v", node.Name, err))
|
||||
fallthrough
|
||||
case needsRetry && observedReadyCondition.Status != v1.ConditionTrue:
|
||||
if err := nc.markPodsNotReady(node.Name); err != nil {
|
||||
utilruntime.HandleError(err)
|
||||
nc.nodesToRetry.Store(node.Name, struct{}{})
|
||||
continue
|
||||
}
|
||||
if err = nodeutil.MarkPodsNotReady(nc.kubeClient, pods, node.Name); err != nil {
|
||||
utilruntime.HandleError(fmt.Errorf("Unable to mark all pods NotReady on node %v: %v", node.Name, err))
|
||||
}
|
||||
}
|
||||
}
|
||||
nc.nodesToRetry.Delete(node.Name)
|
||||
}
|
||||
nc.handleDisruption(zoneToNodeConditions, nodes)
|
||||
|
||||
return nil
|
||||
}
|
||||
|
||||
func (nc *Controller) markPodsNotReady(nodeName string) error {
|
||||
pods, err := listPodsFromNode(nc.kubeClient, nodeName)
|
||||
if err != nil {
|
||||
return fmt.Errorf("unable to list pods from node %v: %v", nodeName, err)
|
||||
}
|
||||
if err = nodeutil.MarkPodsNotReady(nc.kubeClient, pods, nodeName); err != nil {
|
||||
return fmt.Errorf("unable to mark all pods NotReady on node %v: %v", nodeName, err)
|
||||
}
|
||||
return nil
|
||||
}
|
||||
|
||||
func (nc *Controller) processTaintBaseEviction(node *v1.Node, observedReadyCondition *v1.NodeCondition) {
|
||||
decisionTimestamp := nc.now()
|
||||
// Check eviction timeout against decisionTimestamp
|
||||
|
Loading…
Reference in New Issue
Block a user