mirror of
https://github.com/k3s-io/kubernetes.git
synced 2025-08-05 18:24:07 +00:00
Merge pull request #80173 from gaorong/node-lease-renew-interval
make node lease renew interval more heuristic
This commit is contained in:
commit
ab266f725f
@ -871,7 +871,7 @@ func NewMainKubelet(kubeCfg *kubeletconfiginternal.KubeletConfiguration,
|
|||||||
klet.softAdmitHandlers.AddPodAdmitHandler(lifecycle.NewNoNewPrivsAdmitHandler(klet.containerRuntime))
|
klet.softAdmitHandlers.AddPodAdmitHandler(lifecycle.NewNoNewPrivsAdmitHandler(klet.containerRuntime))
|
||||||
|
|
||||||
if utilfeature.DefaultFeatureGate.Enabled(features.NodeLease) {
|
if utilfeature.DefaultFeatureGate.Enabled(features.NodeLease) {
|
||||||
klet.nodeLeaseController = nodelease.NewController(klet.clock, klet.heartbeatClient, string(klet.nodeName), kubeCfg.NodeLeaseDurationSeconds, klet.onRepeatedHeartbeatFailure)
|
klet.nodeLeaseController = nodelease.NewController(klet.clock, klet.heartbeatClient, string(klet.nodeName), kubeCfg.NodeLeaseDurationSeconds, kubeCfg.NodeStatusUpdateFrequency.Duration, klet.onRepeatedHeartbeatFailure)
|
||||||
}
|
}
|
||||||
|
|
||||||
klet.softAdmitHandlers.AddPodAdmitHandler(lifecycle.NewProcMountAdmitHandler(klet.containerRuntime))
|
klet.softAdmitHandlers.AddPodAdmitHandler(lifecycle.NewProcMountAdmitHandler(klet.containerRuntime))
|
||||||
|
@ -34,8 +34,8 @@ import (
|
|||||||
)
|
)
|
||||||
|
|
||||||
const (
|
const (
|
||||||
// renewInterval is the interval at which the lease is renewed
|
// defaultRenewInterval is the default interval at which the lease is renewed
|
||||||
renewInterval = 10 * time.Second
|
defaultRenewInterval = 10 * time.Second
|
||||||
// maxUpdateRetries is the number of immediate, successive retries the Kubelet will attempt
|
// maxUpdateRetries is the number of immediate, successive retries the Kubelet will attempt
|
||||||
// when renewing the lease before it waits for the renewal interval before trying again,
|
// when renewing the lease before it waits for the renewal interval before trying again,
|
||||||
// similar to what we do for node status retries
|
// similar to what we do for node status retries
|
||||||
@ -60,11 +60,21 @@ type controller struct {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// NewController constructs and returns a controller
|
// NewController constructs and returns a controller
|
||||||
func NewController(clock clock.Clock, client clientset.Interface, holderIdentity string, leaseDurationSeconds int32, onRepeatedHeartbeatFailure func()) Controller {
|
func NewController(clock clock.Clock, client clientset.Interface, holderIdentity string, leaseDurationSeconds int32, nodeStatusUpdateFrequency time.Duration, onRepeatedHeartbeatFailure func()) Controller {
|
||||||
var leaseClient coordclientset.LeaseInterface
|
var leaseClient coordclientset.LeaseInterface
|
||||||
if client != nil {
|
if client != nil {
|
||||||
leaseClient = client.CoordinationV1().Leases(corev1.NamespaceNodeLease)
|
leaseClient = client.CoordinationV1().Leases(corev1.NamespaceNodeLease)
|
||||||
}
|
}
|
||||||
|
renewInterval := defaultRenewInterval
|
||||||
|
// Users are able to decrease the timeout after which nodes are being
|
||||||
|
// marked as "Ready: Unknown" by NodeLifecycleController to values
|
||||||
|
// smaller than defaultRenewInterval. Until the knob to configure
|
||||||
|
// lease renew interval is exposed to user, we temporarily decrease
|
||||||
|
// renewInterval based on the NodeStatusUpdateFrequency.
|
||||||
|
if renewInterval > nodeStatusUpdateFrequency {
|
||||||
|
renewInterval = nodeStatusUpdateFrequency
|
||||||
|
}
|
||||||
|
|
||||||
return &controller{
|
return &controller{
|
||||||
client: client,
|
client: client,
|
||||||
leaseClient: leaseClient,
|
leaseClient: leaseClient,
|
||||||
|
Loading…
Reference in New Issue
Block a user