Merge pull request #130356 from aojea/automated-cherry-pick-of-#130348-upstream-release-1.32

Automated cherry pick of #130348: Revert "Add random interval to nodeStatusReport interval every time after an actual node status change"
This commit is contained in:
Kubernetes Prow Robot 2025-02-24 16:50:28 -08:00 committed by GitHub
commit 85ff7928f4
No known key found for this signature in database
GPG Key ID: B5690EEEBB952194
3 changed files with 4 additions and 100 deletions

View File

@ -1269,12 +1269,6 @@ type Kubelet struct {
// status to master. It is only used when node lease feature is enabled.
nodeStatusReportFrequency time.Duration
// delayAfterNodeStatusChange is the one-time random duration that we add to the next node status report interval
// every time when there's an actual node status change. But all future node status update that is not caused by
// real status change will stick with nodeStatusReportFrequency. The random duration is a uniform distribution over
// [-0.5*nodeStatusReportFrequency, 0.5*nodeStatusReportFrequency]
delayAfterNodeStatusChange time.Duration
// lastStatusReportTime is the time when node status was last reported.
lastStatusReportTime time.Time

View File

@ -19,7 +19,6 @@ package kubelet
import (
"context"
"fmt"
"math/rand"
"net"
goruntime "runtime"
"sort"
@ -580,21 +579,13 @@ func (kl *Kubelet) tryUpdateNodeStatus(ctx context.Context, tryNumber int) error
}
node, changed := kl.updateNode(ctx, originalNode)
// no need to update the status yet
if !changed && !kl.isUpdateStatusPeriodExperid() {
shouldPatchNodeStatus := changed || kl.clock.Since(kl.lastStatusReportTime) >= kl.nodeStatusReportFrequency
if !shouldPatchNodeStatus {
kl.markVolumesFromNode(node)
return nil
}
// We need to update the node status, if this is caused by a node change we want to calculate a new
// random delay so we avoid all the nodes to reach the apiserver at the same time. If the update is not related
// to a node change, because we run over the period, we reset the random delay so the node keeps updating
// its status at the same cadence
if changed {
kl.delayAfterNodeStatusChange = kl.calculateDelay()
} else {
kl.delayAfterNodeStatusChange = 0
}
updatedNode, err := kl.patchNodeStatus(originalNode, node)
if err == nil {
kl.markVolumesFromNode(updatedNode)
@ -602,20 +593,6 @@ func (kl *Kubelet) tryUpdateNodeStatus(ctx context.Context, tryNumber int) error
return err
}
func (kl *Kubelet) isUpdateStatusPeriodExperid() bool {
if kl.lastStatusReportTime.IsZero() {
return false
}
if kl.clock.Since(kl.lastStatusReportTime) >= kl.nodeStatusReportFrequency+kl.delayAfterNodeStatusChange {
return true
}
return false
}
func (kl *Kubelet) calculateDelay() time.Duration {
return time.Duration(float64(kl.nodeStatusReportFrequency) * (-0.5 + rand.Float64()))
}
// updateNode creates a copy of originalNode and runs update logic on it.
// It returns the updated node object and a bool indicating if anything has been changed.
func (kl *Kubelet) updateNode(ctx context.Context, originalNode *v1.Node) (*v1.Node, bool) {

View File

@ -849,10 +849,7 @@ func TestUpdateNodeStatusWithLease(t *testing.T) {
// Since this test retroactively overrides the stub container manager,
// we have to regenerate default status setters.
kubelet.setNodeStatusFuncs = kubelet.defaultNodeStatusFuncs()
// You will add up to 50% of nodeStatusReportFrequency of additional random latency for
// kubelet to determine if update node status is needed due to time passage. We need to
// take that into consideration to ensure this test pass all time.
kubelet.nodeStatusReportFrequency = 30 * time.Second
kubelet.nodeStatusReportFrequency = time.Minute
kubeClient := testKubelet.fakeKubeClient
existingNode := &v1.Node{ObjectMeta: metav1.ObjectMeta{Name: testKubeletHostname}}
@ -3091,67 +3088,3 @@ func TestUpdateNodeAddresses(t *testing.T) {
})
}
}
func TestIsUpdateStatusPeriodExperid(t *testing.T) {
testcases := []struct {
name string
lastStatusReportTime time.Time
delayAfterNodeStatusChange time.Duration
expectExpired bool
}{
{
name: "no status update before and no delay",
lastStatusReportTime: time.Time{},
delayAfterNodeStatusChange: 0,
expectExpired: false,
},
{
name: "no status update before and existing delay",
lastStatusReportTime: time.Time{},
delayAfterNodeStatusChange: 30 * time.Second,
expectExpired: false,
},
{
name: "not expired and no delay",
lastStatusReportTime: time.Now().Add(-4 * time.Minute),
delayAfterNodeStatusChange: 0,
expectExpired: false,
},
{
name: "not expired",
lastStatusReportTime: time.Now().Add(-5 * time.Minute),
delayAfterNodeStatusChange: time.Minute,
expectExpired: false,
},
{
name: "expired",
lastStatusReportTime: time.Now().Add(-4 * time.Minute),
delayAfterNodeStatusChange: -2 * time.Minute,
expectExpired: true,
},
}
testKubelet := newTestKubelet(t, false /* controllerAttachDetachEnabled */)
defer testKubelet.Cleanup()
kubelet := testKubelet.kubelet
kubelet.nodeStatusReportFrequency = 5 * time.Minute
for _, tc := range testcases {
kubelet.lastStatusReportTime = tc.lastStatusReportTime
kubelet.delayAfterNodeStatusChange = tc.delayAfterNodeStatusChange
expired := kubelet.isUpdateStatusPeriodExperid()
assert.Equal(t, tc.expectExpired, expired, tc.name)
}
}
func TestCalculateDelay(t *testing.T) {
testKubelet := newTestKubelet(t, false /* controllerAttachDetachEnabled */)
defer testKubelet.Cleanup()
kubelet := testKubelet.kubelet
kubelet.nodeStatusReportFrequency = 5 * time.Minute
for i := 0; i < 100; i++ {
randomDelay := kubelet.calculateDelay()
assert.LessOrEqual(t, randomDelay.Abs(), kubelet.nodeStatusReportFrequency/2)
}
}