mirror of
https://github.com/k3s-io/kubernetes.git
synced 2025-08-06 18:54:06 +00:00
sidecars: terminate sidecars after main containers
Sidecars should terminate: - after all main containers have exited - serialized and in reverse order
This commit is contained in:
parent
6d70013af5
commit
7bcc98c46b
@ -289,7 +289,7 @@ func (m *kubeGenericRuntimeManager) startContainer(ctx context.Context, podSandb
|
|||||||
"podUID", pod.UID, "containerName", container.Name, "containerID", kubeContainerID.String())
|
"podUID", pod.UID, "containerName", container.Name, "containerID", kubeContainerID.String())
|
||||||
// do not record the message in the event so that secrets won't leak from the server.
|
// do not record the message in the event so that secrets won't leak from the server.
|
||||||
m.recordContainerEvent(pod, container, kubeContainerID.ID, v1.EventTypeWarning, events.FailedPostStartHook, "PostStartHook failed")
|
m.recordContainerEvent(pod, container, kubeContainerID.ID, v1.EventTypeWarning, events.FailedPostStartHook, "PostStartHook failed")
|
||||||
if err := m.killContainer(ctx, pod, kubeContainerID, container.Name, "FailedPostStartHook", reasonFailedPostStartHook, nil); err != nil {
|
if err := m.killContainer(ctx, pod, kubeContainerID, container.Name, "FailedPostStartHook", reasonFailedPostStartHook, nil, nil); err != nil {
|
||||||
klog.ErrorS(err, "Failed to kill container", "pod", klog.KObj(pod),
|
klog.ErrorS(err, "Failed to kill container", "pod", klog.KObj(pod),
|
||||||
"podUID", pod.UID, "containerName", container.Name, "containerID", kubeContainerID.String())
|
"podUID", pod.UID, "containerName", container.Name, "containerID", kubeContainerID.String())
|
||||||
}
|
}
|
||||||
@ -703,7 +703,7 @@ func (m *kubeGenericRuntimeManager) restoreSpecsFromContainerLabels(ctx context.
|
|||||||
// killContainer kills a container through the following steps:
|
// killContainer kills a container through the following steps:
|
||||||
// * Run the pre-stop lifecycle hooks (if applicable).
|
// * Run the pre-stop lifecycle hooks (if applicable).
|
||||||
// * Stop the container.
|
// * Stop the container.
|
||||||
func (m *kubeGenericRuntimeManager) killContainer(ctx context.Context, pod *v1.Pod, containerID kubecontainer.ContainerID, containerName string, message string, reason containerKillReason, gracePeriodOverride *int64) error {
|
func (m *kubeGenericRuntimeManager) killContainer(ctx context.Context, pod *v1.Pod, containerID kubecontainer.ContainerID, containerName string, message string, reason containerKillReason, gracePeriodOverride *int64, ordering *terminationOrdering) error {
|
||||||
var containerSpec *v1.Container
|
var containerSpec *v1.Container
|
||||||
if pod != nil {
|
if pod != nil {
|
||||||
if containerSpec = kubecontainer.GetContainerSpec(pod, containerName); containerSpec == nil {
|
if containerSpec = kubecontainer.GetContainerSpec(pod, containerName); containerSpec == nil {
|
||||||
@ -738,6 +738,13 @@ func (m *kubeGenericRuntimeManager) killContainer(ctx context.Context, pod *v1.P
|
|||||||
gracePeriod = gracePeriod - m.executePreStopHook(ctx, pod, containerID, containerSpec, gracePeriod)
|
gracePeriod = gracePeriod - m.executePreStopHook(ctx, pod, containerID, containerSpec, gracePeriod)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// if we care about termination ordering, then wait for this container's turn to exit if there is
|
||||||
|
// time remaining
|
||||||
|
if ordering != nil && gracePeriod > 0 {
|
||||||
|
// grace period is only in seconds, so the time we've waited gets truncated downward
|
||||||
|
gracePeriod -= int64(ordering.waitForTurn(containerName, gracePeriod))
|
||||||
|
}
|
||||||
|
|
||||||
// always give containers a minimal shutdown window to avoid unnecessary SIGKILLs
|
// always give containers a minimal shutdown window to avoid unnecessary SIGKILLs
|
||||||
if gracePeriod < minimumGracePeriodInSeconds {
|
if gracePeriod < minimumGracePeriodInSeconds {
|
||||||
gracePeriod = minimumGracePeriodInSeconds
|
gracePeriod = minimumGracePeriodInSeconds
|
||||||
@ -755,6 +762,10 @@ func (m *kubeGenericRuntimeManager) killContainer(ctx context.Context, pod *v1.P
|
|||||||
klog.V(3).InfoS("Container exited normally", "pod", klog.KObj(pod), "podUID", pod.UID,
|
klog.V(3).InfoS("Container exited normally", "pod", klog.KObj(pod), "podUID", pod.UID,
|
||||||
"containerName", containerName, "containerID", containerID.String())
|
"containerName", containerName, "containerID", containerID.String())
|
||||||
|
|
||||||
|
if ordering != nil {
|
||||||
|
ordering.containerTerminated(containerName)
|
||||||
|
}
|
||||||
|
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -764,13 +775,22 @@ func (m *kubeGenericRuntimeManager) killContainersWithSyncResult(ctx context.Con
|
|||||||
wg := sync.WaitGroup{}
|
wg := sync.WaitGroup{}
|
||||||
|
|
||||||
wg.Add(len(runningPod.Containers))
|
wg.Add(len(runningPod.Containers))
|
||||||
|
var termOrdering *terminationOrdering
|
||||||
|
// we only care about container termination ordering if the sidecars feature is enabled
|
||||||
|
if utilfeature.DefaultFeatureGate.Enabled(features.SidecarContainers) {
|
||||||
|
var runningContainerNames []string
|
||||||
|
for _, container := range runningPod.Containers {
|
||||||
|
runningContainerNames = append(runningContainerNames, container.Name)
|
||||||
|
}
|
||||||
|
termOrdering = newTerminationOrdering(pod, runningContainerNames)
|
||||||
|
}
|
||||||
for _, container := range runningPod.Containers {
|
for _, container := range runningPod.Containers {
|
||||||
go func(container *kubecontainer.Container) {
|
go func(container *kubecontainer.Container) {
|
||||||
defer utilruntime.HandleCrash()
|
defer utilruntime.HandleCrash()
|
||||||
defer wg.Done()
|
defer wg.Done()
|
||||||
|
|
||||||
killContainerResult := kubecontainer.NewSyncResult(kubecontainer.KillContainer, container.Name)
|
killContainerResult := kubecontainer.NewSyncResult(kubecontainer.KillContainer, container.Name)
|
||||||
if err := m.killContainer(ctx, pod, container.ID, container.Name, "", reasonUnknown, gracePeriodOverride); err != nil {
|
if err := m.killContainer(ctx, pod, container.ID, container.Name, "", reasonUnknown, gracePeriodOverride, termOrdering); err != nil {
|
||||||
killContainerResult.Fail(kubecontainer.ErrKillContainer, err.Error())
|
killContainerResult.Fail(kubecontainer.ErrKillContainer, err.Error())
|
||||||
// Use runningPod for logging as the pod passed in could be *nil*.
|
// Use runningPod for logging as the pod passed in could be *nil*.
|
||||||
klog.ErrorS(err, "Kill container failed", "pod", klog.KRef(runningPod.Namespace, runningPod.Name), "podUID", runningPod.ID,
|
klog.ErrorS(err, "Kill container failed", "pod", klog.KRef(runningPod.Namespace, runningPod.Name), "podUID", runningPod.ID,
|
||||||
|
@ -38,6 +38,7 @@ import (
|
|||||||
|
|
||||||
v1 "k8s.io/api/core/v1"
|
v1 "k8s.io/api/core/v1"
|
||||||
runtimeapi "k8s.io/cri-api/pkg/apis/runtime/v1"
|
runtimeapi "k8s.io/cri-api/pkg/apis/runtime/v1"
|
||||||
|
|
||||||
"k8s.io/kubernetes/pkg/features"
|
"k8s.io/kubernetes/pkg/features"
|
||||||
kubecontainer "k8s.io/kubernetes/pkg/kubelet/container"
|
kubecontainer "k8s.io/kubernetes/pkg/kubelet/container"
|
||||||
containertest "k8s.io/kubernetes/pkg/kubelet/container/testing"
|
containertest "k8s.io/kubernetes/pkg/kubelet/container/testing"
|
||||||
@ -128,7 +129,7 @@ func TestKillContainer(t *testing.T) {
|
|||||||
|
|
||||||
for _, test := range tests {
|
for _, test := range tests {
|
||||||
ctx := context.Background()
|
ctx := context.Background()
|
||||||
err := m.killContainer(ctx, test.pod, test.containerID, test.containerName, test.reason, "", &test.gracePeriodOverride)
|
err := m.killContainer(ctx, test.pod, test.containerID, test.containerName, test.reason, "", &test.gracePeriodOverride, nil)
|
||||||
if test.succeed != (err == nil) {
|
if test.succeed != (err == nil) {
|
||||||
t.Errorf("%s: expected %v, got %v (%v)", test.caseName, test.succeed, (err == nil), err)
|
t.Errorf("%s: expected %v, got %v (%v)", test.caseName, test.succeed, (err == nil), err)
|
||||||
}
|
}
|
||||||
@ -402,7 +403,7 @@ func testLifeCycleHook(t *testing.T, testPod *v1.Pod, testContainer *v1.Containe
|
|||||||
t.Run("PreStop-CMDExec", func(t *testing.T) {
|
t.Run("PreStop-CMDExec", func(t *testing.T) {
|
||||||
ctx := context.Background()
|
ctx := context.Background()
|
||||||
testContainer.Lifecycle = cmdLifeCycle
|
testContainer.Lifecycle = cmdLifeCycle
|
||||||
m.killContainer(ctx, testPod, cID, "foo", "testKill", "", &gracePeriod)
|
_ = m.killContainer(ctx, testPod, cID, "foo", "testKill", "", &gracePeriod, nil)
|
||||||
if fakeRunner.Cmd[0] != cmdLifeCycle.PreStop.Exec.Command[0] {
|
if fakeRunner.Cmd[0] != cmdLifeCycle.PreStop.Exec.Command[0] {
|
||||||
t.Errorf("CMD Prestop hook was not invoked")
|
t.Errorf("CMD Prestop hook was not invoked")
|
||||||
}
|
}
|
||||||
@ -416,8 +417,7 @@ func testLifeCycleHook(t *testing.T, testPod *v1.Pod, testContainer *v1.Containe
|
|||||||
defer featuregatetesting.SetFeatureGateDuringTest(t, utilfeature.DefaultFeatureGate, features.ConsistentHTTPGetHandlers, false)()
|
defer featuregatetesting.SetFeatureGateDuringTest(t, utilfeature.DefaultFeatureGate, features.ConsistentHTTPGetHandlers, false)()
|
||||||
httpLifeCycle.PreStop.HTTPGet.Port = intstr.IntOrString{}
|
httpLifeCycle.PreStop.HTTPGet.Port = intstr.IntOrString{}
|
||||||
testContainer.Lifecycle = httpLifeCycle
|
testContainer.Lifecycle = httpLifeCycle
|
||||||
m.killContainer(ctx, testPod, cID, "foo", "testKill", "", &gracePeriod)
|
_ = m.killContainer(ctx, testPod, cID, "foo", "testKill", "", &gracePeriod, nil)
|
||||||
|
|
||||||
if fakeHTTP.req == nil || !strings.Contains(fakeHTTP.req.URL.String(), httpLifeCycle.PreStop.HTTPGet.Host) {
|
if fakeHTTP.req == nil || !strings.Contains(fakeHTTP.req.URL.String(), httpLifeCycle.PreStop.HTTPGet.Host) {
|
||||||
t.Errorf("HTTP Prestop hook was not invoked")
|
t.Errorf("HTTP Prestop hook was not invoked")
|
||||||
}
|
}
|
||||||
@ -427,8 +427,7 @@ func testLifeCycleHook(t *testing.T, testPod *v1.Pod, testContainer *v1.Containe
|
|||||||
defer func() { fakeHTTP.req = nil }()
|
defer func() { fakeHTTP.req = nil }()
|
||||||
httpLifeCycle.PreStop.HTTPGet.Port = intstr.FromInt32(80)
|
httpLifeCycle.PreStop.HTTPGet.Port = intstr.FromInt32(80)
|
||||||
testContainer.Lifecycle = httpLifeCycle
|
testContainer.Lifecycle = httpLifeCycle
|
||||||
m.killContainer(ctx, testPod, cID, "foo", "testKill", "", &gracePeriod)
|
_ = m.killContainer(ctx, testPod, cID, "foo", "testKill", "", &gracePeriod, nil)
|
||||||
|
|
||||||
if fakeHTTP.req == nil || !strings.Contains(fakeHTTP.req.URL.String(), httpLifeCycle.PreStop.HTTPGet.Host) {
|
if fakeHTTP.req == nil || !strings.Contains(fakeHTTP.req.URL.String(), httpLifeCycle.PreStop.HTTPGet.Host) {
|
||||||
t.Errorf("HTTP Prestop hook was not invoked")
|
t.Errorf("HTTP Prestop hook was not invoked")
|
||||||
}
|
}
|
||||||
@ -443,8 +442,7 @@ func testLifeCycleHook(t *testing.T, testPod *v1.Pod, testContainer *v1.Containe
|
|||||||
testPod.DeletionGracePeriodSeconds = &gracePeriodLocal
|
testPod.DeletionGracePeriodSeconds = &gracePeriodLocal
|
||||||
testPod.Spec.TerminationGracePeriodSeconds = &gracePeriodLocal
|
testPod.Spec.TerminationGracePeriodSeconds = &gracePeriodLocal
|
||||||
|
|
||||||
m.killContainer(ctx, testPod, cID, "foo", "testKill", "", &gracePeriodLocal)
|
_ = m.killContainer(ctx, testPod, cID, "foo", "testKill", "", &gracePeriodLocal, nil)
|
||||||
|
|
||||||
if fakeHTTP.req != nil {
|
if fakeHTTP.req != nil {
|
||||||
t.Errorf("HTTP Prestop hook Should not execute when gracePeriod is 0")
|
t.Errorf("HTTP Prestop hook Should not execute when gracePeriod is 0")
|
||||||
}
|
}
|
||||||
|
@ -141,7 +141,7 @@ func (cgc *containerGC) removeOldestN(ctx context.Context, containers []containe
|
|||||||
ID: containers[i].id,
|
ID: containers[i].id,
|
||||||
}
|
}
|
||||||
message := "Container is in unknown state, try killing it before removal"
|
message := "Container is in unknown state, try killing it before removal"
|
||||||
if err := cgc.manager.killContainer(ctx, nil, id, containers[i].name, message, reasonUnknown, nil); err != nil {
|
if err := cgc.manager.killContainer(ctx, nil, id, containers[i].name, message, reasonUnknown, nil, nil); err != nil {
|
||||||
klog.ErrorS(err, "Failed to stop container", "containerID", containers[i].id)
|
klog.ErrorS(err, "Failed to stop container", "containerID", containers[i].id)
|
||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
|
@ -44,6 +44,7 @@ import (
|
|||||||
"k8s.io/component-base/logs/logreduction"
|
"k8s.io/component-base/logs/logreduction"
|
||||||
internalapi "k8s.io/cri-api/pkg/apis"
|
internalapi "k8s.io/cri-api/pkg/apis"
|
||||||
runtimeapi "k8s.io/cri-api/pkg/apis/runtime/v1"
|
runtimeapi "k8s.io/cri-api/pkg/apis/runtime/v1"
|
||||||
|
|
||||||
"k8s.io/kubernetes/pkg/api/legacyscheme"
|
"k8s.io/kubernetes/pkg/api/legacyscheme"
|
||||||
podutil "k8s.io/kubernetes/pkg/api/v1/pod"
|
podutil "k8s.io/kubernetes/pkg/api/v1/pod"
|
||||||
"k8s.io/kubernetes/pkg/credentialprovider"
|
"k8s.io/kubernetes/pkg/credentialprovider"
|
||||||
@ -1090,7 +1091,7 @@ func (m *kubeGenericRuntimeManager) SyncPod(ctx context.Context, pod *v1.Pod, po
|
|||||||
klog.V(3).InfoS("Killing unwanted container for pod", "containerName", containerInfo.name, "containerID", containerID, "pod", klog.KObj(pod))
|
klog.V(3).InfoS("Killing unwanted container for pod", "containerName", containerInfo.name, "containerID", containerID, "pod", klog.KObj(pod))
|
||||||
killContainerResult := kubecontainer.NewSyncResult(kubecontainer.KillContainer, containerInfo.name)
|
killContainerResult := kubecontainer.NewSyncResult(kubecontainer.KillContainer, containerInfo.name)
|
||||||
result.AddSyncResult(killContainerResult)
|
result.AddSyncResult(killContainerResult)
|
||||||
if err := m.killContainer(ctx, pod, containerID, containerInfo.name, containerInfo.message, containerInfo.reason, nil); err != nil {
|
if err := m.killContainer(ctx, pod, containerID, containerInfo.name, containerInfo.message, containerInfo.reason, nil, nil); err != nil {
|
||||||
killContainerResult.Fail(kubecontainer.ErrKillContainer, err.Error())
|
killContainerResult.Fail(kubecontainer.ErrKillContainer, err.Error())
|
||||||
klog.ErrorS(err, "killContainer for pod failed", "containerName", containerInfo.name, "containerID", containerID, "pod", klog.KObj(pod))
|
klog.ErrorS(err, "killContainer for pod failed", "containerName", containerInfo.name, "containerID", containerID, "pod", klog.KObj(pod))
|
||||||
return
|
return
|
||||||
|
114
pkg/kubelet/kuberuntime/kuberuntime_termination_order.go
Normal file
114
pkg/kubelet/kuberuntime/kuberuntime_termination_order.go
Normal file
@ -0,0 +1,114 @@
|
|||||||
|
/*
|
||||||
|
Copyright 2023 The Kubernetes Authors.
|
||||||
|
|
||||||
|
Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
you may not use this file except in compliance with the License.
|
||||||
|
You may obtain a copy of the License at
|
||||||
|
|
||||||
|
http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
|
||||||
|
Unless required by applicable law or agreed to in writing, software
|
||||||
|
distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
See the License for the specific language governing permissions and
|
||||||
|
limitations under the License.
|
||||||
|
*/
|
||||||
|
|
||||||
|
package kuberuntime
|
||||||
|
|
||||||
|
import (
|
||||||
|
"time"
|
||||||
|
|
||||||
|
v1 "k8s.io/api/core/v1"
|
||||||
|
|
||||||
|
"k8s.io/kubernetes/pkg/kubelet/types"
|
||||||
|
)
|
||||||
|
|
||||||
|
// terminationOrdering is used to enforce a termination ordering for sidecar containers. It sets up
|
||||||
|
// dependencies between sidecars and allows the pod termination process to wait until the grace period
|
||||||
|
// expires, or all dependent containers have finished terminating.
|
||||||
|
type terminationOrdering struct {
|
||||||
|
// terminated is a map from container name to a channel, that if closed
|
||||||
|
// indicates that the container with that name was terminated
|
||||||
|
terminated map[string]chan struct{}
|
||||||
|
// prereqs is a map from container name to a list of channel that the container
|
||||||
|
// must wait on to ensure termination ordering
|
||||||
|
prereqs map[string][]chan struct{}
|
||||||
|
}
|
||||||
|
|
||||||
|
// newTerminationOrdering constructs a terminationOrdering based on the pod spec and the currently running containers.
|
||||||
|
func newTerminationOrdering(pod *v1.Pod, runningContainerNames []string) *terminationOrdering {
|
||||||
|
to := &terminationOrdering{
|
||||||
|
prereqs: map[string][]chan struct{}{},
|
||||||
|
terminated: map[string]chan struct{}{},
|
||||||
|
}
|
||||||
|
|
||||||
|
runningContainers := map[string]struct{}{}
|
||||||
|
for _, name := range runningContainerNames {
|
||||||
|
runningContainers[name] = struct{}{}
|
||||||
|
}
|
||||||
|
|
||||||
|
var mainContainerChannels []chan struct{}
|
||||||
|
// sidecar containers need to wait on main containers, so we create a channel per main container
|
||||||
|
// for them to wait on
|
||||||
|
for _, c := range pod.Spec.Containers {
|
||||||
|
channel := make(chan struct{})
|
||||||
|
to.terminated[c.Name] = channel
|
||||||
|
mainContainerChannels = append(mainContainerChannels, channel)
|
||||||
|
|
||||||
|
// if its not a running container, pre-close the channel so nothing waits on it
|
||||||
|
if _, isRunning := runningContainers[c.Name]; !isRunning {
|
||||||
|
close(channel)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
var previousSidecarName string
|
||||||
|
for i := range pod.Spec.InitContainers {
|
||||||
|
// get the init containers in reverse order
|
||||||
|
ic := pod.Spec.InitContainers[len(pod.Spec.InitContainers)-i-1]
|
||||||
|
|
||||||
|
to.terminated[ic.Name] = make(chan struct{})
|
||||||
|
|
||||||
|
if types.IsRestartableInitContainer(&ic) {
|
||||||
|
// sidecars need to wait for all main containers to exit
|
||||||
|
to.prereqs[ic.Name] = append(to.prereqs[ic.Name], mainContainerChannels...)
|
||||||
|
|
||||||
|
// if there is a later sidecar, this container needs to wait for it to finish
|
||||||
|
if previousSidecarName != "" {
|
||||||
|
to.prereqs[ic.Name] = append(to.prereqs[ic.Name], to.terminated[previousSidecarName])
|
||||||
|
}
|
||||||
|
previousSidecarName = ic.Name
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return to
|
||||||
|
}
|
||||||
|
|
||||||
|
// waitForTurn waits until it is time for the container with the specified name to begin terminating, up until
|
||||||
|
// the specified grace period. If gracePeriod = 0, there is no wait.
|
||||||
|
func (o *terminationOrdering) waitForTurn(name string, gracePeriod int64) float64 {
|
||||||
|
// if there is no grace period, we don't wait
|
||||||
|
if gracePeriod <= 0 {
|
||||||
|
return 0
|
||||||
|
}
|
||||||
|
|
||||||
|
start := time.Now()
|
||||||
|
remainingGrace := time.NewTimer(time.Duration(gracePeriod) * time.Second)
|
||||||
|
|
||||||
|
for _, c := range o.prereqs[name] {
|
||||||
|
select {
|
||||||
|
case <-c:
|
||||||
|
case <-remainingGrace.C:
|
||||||
|
// grace period expired, so immediately exit
|
||||||
|
return time.Since(start).Seconds()
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return time.Since(start).Seconds()
|
||||||
|
}
|
||||||
|
|
||||||
|
// containerTerminated should be called once the container with the speecified name has exited.
|
||||||
|
func (o *terminationOrdering) containerTerminated(name string) {
|
||||||
|
if ch, ok := o.terminated[name]; ok {
|
||||||
|
close(ch)
|
||||||
|
}
|
||||||
|
}
|
292
pkg/kubelet/kuberuntime/kuberuntime_termination_order_test.go
Normal file
292
pkg/kubelet/kuberuntime/kuberuntime_termination_order_test.go
Normal file
@ -0,0 +1,292 @@
|
|||||||
|
/*
|
||||||
|
Copyright 2023 The Kubernetes Authors.
|
||||||
|
|
||||||
|
Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
you may not use this file except in compliance with the License.
|
||||||
|
You may obtain a copy of the License at
|
||||||
|
|
||||||
|
http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
|
||||||
|
Unless required by applicable law or agreed to in writing, software
|
||||||
|
distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
See the License for the specific language governing permissions and
|
||||||
|
limitations under the License.
|
||||||
|
*/
|
||||||
|
|
||||||
|
package kuberuntime
|
||||||
|
|
||||||
|
import (
|
||||||
|
"sync"
|
||||||
|
"testing"
|
||||||
|
"time"
|
||||||
|
|
||||||
|
v1 "k8s.io/api/core/v1"
|
||||||
|
metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
|
||||||
|
)
|
||||||
|
|
||||||
|
func TestTerminationOrderingSidecarStopAfterMain(t *testing.T) {
|
||||||
|
restartPolicy := v1.ContainerRestartPolicyAlways
|
||||||
|
pod := &v1.Pod{
|
||||||
|
ObjectMeta: metav1.ObjectMeta{
|
||||||
|
UID: "12345678",
|
||||||
|
Name: "bar",
|
||||||
|
Namespace: "new",
|
||||||
|
},
|
||||||
|
Spec: v1.PodSpec{
|
||||||
|
InitContainers: []v1.Container{
|
||||||
|
{
|
||||||
|
Name: "init",
|
||||||
|
Image: "busybox",
|
||||||
|
ImagePullPolicy: v1.PullIfNotPresent,
|
||||||
|
RestartPolicy: &restartPolicy,
|
||||||
|
},
|
||||||
|
},
|
||||||
|
Containers: []v1.Container{
|
||||||
|
{
|
||||||
|
Name: "main",
|
||||||
|
Image: "busybox",
|
||||||
|
ImagePullPolicy: v1.PullIfNotPresent,
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
}
|
||||||
|
to := newTerminationOrdering(pod, getContainerNames(pod))
|
||||||
|
|
||||||
|
var wg sync.WaitGroup
|
||||||
|
wg.Add(1)
|
||||||
|
var sidecarWaitDelay int64
|
||||||
|
var mainWaitDelay int64
|
||||||
|
go func() {
|
||||||
|
sidecarWaitDelay = int64(to.waitForTurn("init", 30))
|
||||||
|
to.containerTerminated("init")
|
||||||
|
wg.Done()
|
||||||
|
}()
|
||||||
|
|
||||||
|
wg.Add(1)
|
||||||
|
go func() {
|
||||||
|
mainWaitDelay = int64(to.waitForTurn("main", 0))
|
||||||
|
time.Sleep(1 * time.Second)
|
||||||
|
to.containerTerminated("main")
|
||||||
|
wg.Done()
|
||||||
|
}()
|
||||||
|
wg.Wait()
|
||||||
|
if sidecarWaitDelay != 1 {
|
||||||
|
t.Errorf("expected sidecar to wait for main container to exit, got delay of %d", sidecarWaitDelay)
|
||||||
|
}
|
||||||
|
if mainWaitDelay != 0 {
|
||||||
|
t.Errorf("expected main container to not wait to exit, got delay of %d", mainWaitDelay)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestTerminationOrderingSidecarsInReverseOrder(t *testing.T) {
|
||||||
|
restartPolicy := v1.ContainerRestartPolicyAlways
|
||||||
|
pod := &v1.Pod{
|
||||||
|
ObjectMeta: metav1.ObjectMeta{
|
||||||
|
UID: "12345678",
|
||||||
|
Name: "bar",
|
||||||
|
Namespace: "new",
|
||||||
|
},
|
||||||
|
Spec: v1.PodSpec{
|
||||||
|
InitContainers: []v1.Container{
|
||||||
|
{
|
||||||
|
Name: "sc1",
|
||||||
|
Image: "busybox",
|
||||||
|
ImagePullPolicy: v1.PullIfNotPresent,
|
||||||
|
RestartPolicy: &restartPolicy,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Name: "sc2",
|
||||||
|
Image: "busybox",
|
||||||
|
ImagePullPolicy: v1.PullIfNotPresent,
|
||||||
|
RestartPolicy: &restartPolicy,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Name: "sc3",
|
||||||
|
Image: "busybox",
|
||||||
|
ImagePullPolicy: v1.PullIfNotPresent,
|
||||||
|
RestartPolicy: &restartPolicy,
|
||||||
|
},
|
||||||
|
},
|
||||||
|
Containers: []v1.Container{
|
||||||
|
{
|
||||||
|
Name: "main",
|
||||||
|
Image: "busybox",
|
||||||
|
ImagePullPolicy: v1.PullIfNotPresent,
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
}
|
||||||
|
to := newTerminationOrdering(pod, getContainerNames(pod))
|
||||||
|
|
||||||
|
var wg sync.WaitGroup
|
||||||
|
var delays sync.Map
|
||||||
|
|
||||||
|
waitAndExit := func(name string) {
|
||||||
|
delay := int64(to.waitForTurn(name, 30))
|
||||||
|
delays.Store(name, delay)
|
||||||
|
time.Sleep(1 * time.Second)
|
||||||
|
to.containerTerminated(name)
|
||||||
|
wg.Done()
|
||||||
|
}
|
||||||
|
for _, ic := range pod.Spec.InitContainers {
|
||||||
|
wg.Add(1)
|
||||||
|
go waitAndExit(ic.Name)
|
||||||
|
}
|
||||||
|
for _, c := range pod.Spec.Containers {
|
||||||
|
wg.Add(1)
|
||||||
|
go waitAndExit(c.Name)
|
||||||
|
}
|
||||||
|
|
||||||
|
// wait for our simulated containers to exit
|
||||||
|
wg.Wait()
|
||||||
|
|
||||||
|
getDelay := func(name string) int64 {
|
||||||
|
delay, ok := delays.Load(name)
|
||||||
|
if !ok {
|
||||||
|
t.Errorf("unable to find delay for container %s", name)
|
||||||
|
}
|
||||||
|
return delay.(int64)
|
||||||
|
}
|
||||||
|
|
||||||
|
for _, tc := range []struct {
|
||||||
|
containerName string
|
||||||
|
expectedDelay int64
|
||||||
|
}{
|
||||||
|
// sidecars should exit in reverse order, so
|
||||||
|
// sc1 = 3 (main container + sc3 + sc2)
|
||||||
|
{
|
||||||
|
containerName: "sc1",
|
||||||
|
expectedDelay: 3,
|
||||||
|
},
|
||||||
|
// sc2 = 2 (main container + sc3)
|
||||||
|
{
|
||||||
|
containerName: "sc2",
|
||||||
|
expectedDelay: 2,
|
||||||
|
},
|
||||||
|
// sc3 = 1 (main container)
|
||||||
|
{
|
||||||
|
containerName: "sc3",
|
||||||
|
expectedDelay: 1,
|
||||||
|
},
|
||||||
|
// main container = 0 delay, nothing to wait on
|
||||||
|
{
|
||||||
|
containerName: "main",
|
||||||
|
expectedDelay: 0,
|
||||||
|
},
|
||||||
|
} {
|
||||||
|
if got := getDelay(tc.containerName); got != tc.expectedDelay {
|
||||||
|
t.Errorf("expected delay for container %s = %d, got %d", tc.containerName, tc.expectedDelay, got)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestTerminationOrderingObeysGrace(t *testing.T) {
|
||||||
|
restartPolicy := v1.ContainerRestartPolicyAlways
|
||||||
|
pod := &v1.Pod{
|
||||||
|
ObjectMeta: metav1.ObjectMeta{
|
||||||
|
UID: "12345678",
|
||||||
|
Name: "bar",
|
||||||
|
Namespace: "new",
|
||||||
|
},
|
||||||
|
Spec: v1.PodSpec{
|
||||||
|
InitContainers: []v1.Container{
|
||||||
|
{
|
||||||
|
Name: "sc1",
|
||||||
|
Image: "busybox",
|
||||||
|
ImagePullPolicy: v1.PullIfNotPresent,
|
||||||
|
RestartPolicy: &restartPolicy,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Name: "sc2",
|
||||||
|
Image: "busybox",
|
||||||
|
ImagePullPolicy: v1.PullIfNotPresent,
|
||||||
|
RestartPolicy: &restartPolicy,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Name: "sc3",
|
||||||
|
Image: "busybox",
|
||||||
|
ImagePullPolicy: v1.PullIfNotPresent,
|
||||||
|
RestartPolicy: &restartPolicy,
|
||||||
|
},
|
||||||
|
},
|
||||||
|
Containers: []v1.Container{
|
||||||
|
{
|
||||||
|
Name: "main",
|
||||||
|
Image: "busybox",
|
||||||
|
ImagePullPolicy: v1.PullIfNotPresent,
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
}
|
||||||
|
to := newTerminationOrdering(pod, getContainerNames(pod))
|
||||||
|
|
||||||
|
var wg sync.WaitGroup
|
||||||
|
var delays sync.Map
|
||||||
|
|
||||||
|
waitAndExit := func(name string) {
|
||||||
|
// just a two second grace period which is not long enough for all of the waits to finish
|
||||||
|
delay := int64(to.waitForTurn(name, 2))
|
||||||
|
delays.Store(name, delay)
|
||||||
|
time.Sleep(1 * time.Second)
|
||||||
|
to.containerTerminated(name)
|
||||||
|
wg.Done()
|
||||||
|
}
|
||||||
|
for _, ic := range pod.Spec.InitContainers {
|
||||||
|
wg.Add(1)
|
||||||
|
go waitAndExit(ic.Name)
|
||||||
|
}
|
||||||
|
for _, c := range pod.Spec.Containers {
|
||||||
|
wg.Add(1)
|
||||||
|
go waitAndExit(c.Name)
|
||||||
|
}
|
||||||
|
|
||||||
|
// wait for our simulated containers to exit
|
||||||
|
wg.Wait()
|
||||||
|
|
||||||
|
getDelay := func(name string) int64 {
|
||||||
|
delay, ok := delays.Load(name)
|
||||||
|
if !ok {
|
||||||
|
t.Errorf("unable to find delay for container %s", name)
|
||||||
|
}
|
||||||
|
return delay.(int64)
|
||||||
|
}
|
||||||
|
|
||||||
|
for _, tc := range []struct {
|
||||||
|
containerName string
|
||||||
|
expectedDelay int64
|
||||||
|
}{
|
||||||
|
{
|
||||||
|
containerName: "sc1",
|
||||||
|
// overall grace period limits the amount of time waited here
|
||||||
|
expectedDelay: 2,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
containerName: "sc2",
|
||||||
|
expectedDelay: 2,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
containerName: "sc3",
|
||||||
|
expectedDelay: 1,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
containerName: "main",
|
||||||
|
expectedDelay: 0,
|
||||||
|
},
|
||||||
|
} {
|
||||||
|
if got := getDelay(tc.containerName); got != tc.expectedDelay {
|
||||||
|
t.Errorf("expected delay for container %s = %d, got %d", tc.containerName, tc.expectedDelay, got)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func getContainerNames(p *v1.Pod) []string {
|
||||||
|
var running []string
|
||||||
|
for _, ic := range p.Spec.InitContainers {
|
||||||
|
running = append(running, ic.Name)
|
||||||
|
}
|
||||||
|
for _, c := range p.Spec.Containers {
|
||||||
|
running = append(running, c.Name)
|
||||||
|
}
|
||||||
|
return running
|
||||||
|
}
|
@ -40,6 +40,9 @@ type execCommand struct {
|
|||||||
StartDelay int
|
StartDelay int
|
||||||
// Delay is how long the container should delay before exiting
|
// Delay is how long the container should delay before exiting
|
||||||
Delay int
|
Delay int
|
||||||
|
// LoopForever if set will cause the command to log once per second in a loop until
|
||||||
|
// terminated
|
||||||
|
LoopForever bool
|
||||||
// TerminationSeconds is the time it takes for the container before
|
// TerminationSeconds is the time it takes for the container before
|
||||||
// terminating if it catches SIGTERM.
|
// terminating if it catches SIGTERM.
|
||||||
TerminationSeconds int
|
TerminationSeconds int
|
||||||
@ -86,6 +89,9 @@ func ExecCommand(name string, c execCommand) []string {
|
|||||||
if c.Delay != 0 {
|
if c.Delay != 0 {
|
||||||
fmt.Fprint(&cmd, sleepCommand(c.Delay))
|
fmt.Fprint(&cmd, sleepCommand(c.Delay))
|
||||||
}
|
}
|
||||||
|
if c.LoopForever {
|
||||||
|
fmt.Fprintf(&cmd, "while true; do echo %s '%s Looping' | tee -a %s >> /proc/1/fd/1 ; sleep 1 ; done; ", timeCmd, name, containerLog)
|
||||||
|
}
|
||||||
fmt.Fprintf(&cmd, "echo %s '%s Exiting' | tee -a %s >> /proc/1/fd/1; ", timeCmd, name, containerLog)
|
fmt.Fprintf(&cmd, "echo %s '%s Exiting' | tee -a %s >> /proc/1/fd/1; ", timeCmd, name, containerLog)
|
||||||
fmt.Fprintf(&cmd, "exit %d", c.ExitCode)
|
fmt.Fprintf(&cmd, "exit %d", c.ExitCode)
|
||||||
return []string{"sh", "-c", cmd.String()}
|
return []string{"sh", "-c", cmd.String()}
|
||||||
@ -110,8 +116,8 @@ type containerOutputList []containerOutput
|
|||||||
|
|
||||||
func (o containerOutputList) String() string {
|
func (o containerOutputList) String() string {
|
||||||
var b bytes.Buffer
|
var b bytes.Buffer
|
||||||
for _, v := range o {
|
for i, v := range o {
|
||||||
fmt.Fprintf(&b, "%s %f %s %s\n", v.timestamp, v.timeSinceBoot, v.containerName, v.command)
|
fmt.Fprintf(&b, "%d) %s %f %s %s\n", i, v.timestamp, v.timeSinceBoot, v.containerName, v.command)
|
||||||
}
|
}
|
||||||
return b.String()
|
return b.String()
|
||||||
}
|
}
|
||||||
@ -186,10 +192,10 @@ func (o containerOutputList) ExitsBefore(lhs, rhs string) error {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// this works even for the same names (restart case)
|
// this works even for the same names (restart case)
|
||||||
rhsExit := o.findIndex(rhs, "Starting", lhsExit+1)
|
rhsExit := o.findIndex(rhs, "Exiting", lhsExit+1)
|
||||||
|
|
||||||
if rhsExit == -1 {
|
if rhsExit == -1 {
|
||||||
return fmt.Errorf("couldn't find that %s starting before %s exited, got\n%v", rhs, lhs, o)
|
return fmt.Errorf("couldn't find that %s starting before %s exited (starting at idx %d), got\n%v", rhs, lhs, lhsExit+1, o)
|
||||||
}
|
}
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
@ -278,6 +284,32 @@ func (o containerOutputList) findIndex(name string, command string, startIdx int
|
|||||||
}
|
}
|
||||||
return -1
|
return -1
|
||||||
}
|
}
|
||||||
|
func (o containerOutputList) findLastIndex(name string, command string) int {
|
||||||
|
found := -1
|
||||||
|
for i, v := range o {
|
||||||
|
if v.containerName == name && v.command == command {
|
||||||
|
found = i
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return found
|
||||||
|
}
|
||||||
|
|
||||||
|
// TimeOfStart returns the time since the node boot in floating point seconds that the specified container started.
|
||||||
|
func (o containerOutputList) TimeOfStart(name string) (float64, error) {
|
||||||
|
idx := o.findIndex(name, "Starting", 0)
|
||||||
|
if idx == -1 {
|
||||||
|
return 0.0, fmt.Errorf("couldn't find that %s ever started, got\n%v", name, o)
|
||||||
|
}
|
||||||
|
return o[idx].timeSinceBoot, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (o containerOutputList) TimeOfLastLoop(name string) (float64, error) {
|
||||||
|
idx := o.findLastIndex(name, "Looping")
|
||||||
|
if idx == -1 {
|
||||||
|
return 0.0, fmt.Errorf("couldn't find that %s ever looped, got\n%v", name, o)
|
||||||
|
}
|
||||||
|
return o[idx].timeSinceBoot, nil
|
||||||
|
}
|
||||||
|
|
||||||
// parseOutput combines the container log from all of the init and regular
|
// parseOutput combines the container log from all of the init and regular
|
||||||
// containers and parses/sorts the outputs to produce an execution log
|
// containers and parses/sorts the outputs to produce an execution log
|
||||||
|
@ -2505,4 +2505,458 @@ var _ = SIGDescribe("[NodeAlphaFeature:SidecarContainers] Containers Lifecycle",
|
|||||||
framework.ExpectNoError(results.Exits(restartableInit1))
|
framework.ExpectNoError(results.Exits(restartableInit1))
|
||||||
framework.ExpectNoError(results.Starts(regular1))
|
framework.ExpectNoError(results.Starts(regular1))
|
||||||
})
|
})
|
||||||
|
|
||||||
|
ginkgo.It("should terminate sidecars in reverse order after all main containers have exited", func() {
|
||||||
|
restartableInit1 := "restartable-init-1"
|
||||||
|
restartableInit2 := "restartable-init-2"
|
||||||
|
restartableInit3 := "restartable-init-3"
|
||||||
|
regular1 := "regular-1"
|
||||||
|
|
||||||
|
pod := &v1.Pod{
|
||||||
|
ObjectMeta: metav1.ObjectMeta{
|
||||||
|
Name: "serialize-termination",
|
||||||
|
},
|
||||||
|
Spec: v1.PodSpec{
|
||||||
|
RestartPolicy: v1.RestartPolicyNever,
|
||||||
|
InitContainers: []v1.Container{
|
||||||
|
{
|
||||||
|
Name: restartableInit1,
|
||||||
|
Image: busyboxImage,
|
||||||
|
RestartPolicy: &containerRestartPolicyAlways,
|
||||||
|
Command: ExecCommand(restartableInit1, execCommand{
|
||||||
|
Delay: 60,
|
||||||
|
TerminationSeconds: 5,
|
||||||
|
ExitCode: 0,
|
||||||
|
}),
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Name: restartableInit2,
|
||||||
|
Image: busyboxImage,
|
||||||
|
RestartPolicy: &containerRestartPolicyAlways,
|
||||||
|
Command: ExecCommand(restartableInit2, execCommand{
|
||||||
|
Delay: 60,
|
||||||
|
TerminationSeconds: 5,
|
||||||
|
ExitCode: 0,
|
||||||
|
}),
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Name: restartableInit3,
|
||||||
|
Image: busyboxImage,
|
||||||
|
RestartPolicy: &containerRestartPolicyAlways,
|
||||||
|
Command: ExecCommand(restartableInit3, execCommand{
|
||||||
|
Delay: 60,
|
||||||
|
TerminationSeconds: 5,
|
||||||
|
ExitCode: 0,
|
||||||
|
}),
|
||||||
|
},
|
||||||
|
},
|
||||||
|
Containers: []v1.Container{
|
||||||
|
{
|
||||||
|
Name: regular1,
|
||||||
|
Image: busyboxImage,
|
||||||
|
Command: ExecCommand(regular1, execCommand{
|
||||||
|
Delay: 5,
|
||||||
|
ExitCode: 0,
|
||||||
|
}),
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
}
|
||||||
|
|
||||||
|
preparePod(pod)
|
||||||
|
|
||||||
|
client := e2epod.NewPodClient(f)
|
||||||
|
pod = client.Create(context.TODO(), pod)
|
||||||
|
|
||||||
|
err := e2epod.WaitTimeoutForPodNoLongerRunningInNamespace(context.TODO(), f.ClientSet, pod.Name, pod.Namespace, 5*time.Minute)
|
||||||
|
framework.ExpectNoError(err)
|
||||||
|
|
||||||
|
pod, err = client.Get(context.TODO(), pod.Name, metav1.GetOptions{})
|
||||||
|
framework.ExpectNoError(err)
|
||||||
|
|
||||||
|
results := parseOutput(context.TODO(), f, pod)
|
||||||
|
|
||||||
|
ginkgo.By("Analyzing results")
|
||||||
|
framework.ExpectNoError(results.StartsBefore(restartableInit1, restartableInit2))
|
||||||
|
framework.ExpectNoError(results.StartsBefore(restartableInit1, restartableInit3))
|
||||||
|
framework.ExpectNoError(results.StartsBefore(restartableInit2, restartableInit3))
|
||||||
|
framework.ExpectNoError(results.StartsBefore(restartableInit1, regular1))
|
||||||
|
framework.ExpectNoError(results.StartsBefore(restartableInit2, regular1))
|
||||||
|
framework.ExpectNoError(results.StartsBefore(restartableInit3, regular1))
|
||||||
|
|
||||||
|
// main containers exit first
|
||||||
|
framework.ExpectNoError(results.ExitsBefore(regular1, restartableInit1))
|
||||||
|
framework.ExpectNoError(results.ExitsBefore(regular1, restartableInit2))
|
||||||
|
framework.ExpectNoError(results.ExitsBefore(regular1, restartableInit3))
|
||||||
|
// followed by sidecars in reverse order
|
||||||
|
framework.ExpectNoError(results.ExitsBefore(restartableInit3, restartableInit2))
|
||||||
|
framework.ExpectNoError(results.ExitsBefore(restartableInit2, restartableInit1))
|
||||||
|
})
|
||||||
|
|
||||||
|
ginkgo.It("should terminate sidecars simultaneously if prestop doesn't exit", func() {
|
||||||
|
restartableInit1 := "restartable-init-1"
|
||||||
|
restartableInit2 := "restartable-init-2"
|
||||||
|
restartableInit3 := "restartable-init-3"
|
||||||
|
regular1 := "regular-1"
|
||||||
|
|
||||||
|
makePrestop := func(containerName string) *v1.Lifecycle {
|
||||||
|
return &v1.Lifecycle{
|
||||||
|
PreStop: &v1.LifecycleHandler{
|
||||||
|
Exec: &v1.ExecAction{
|
||||||
|
Command: ExecCommand(prefixedName(PreStopPrefix, containerName), execCommand{
|
||||||
|
ExitCode: 0,
|
||||||
|
ContainerName: containerName,
|
||||||
|
LoopForever: true,
|
||||||
|
}),
|
||||||
|
},
|
||||||
|
},
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
pod := &v1.Pod{
|
||||||
|
ObjectMeta: metav1.ObjectMeta{
|
||||||
|
Name: "serialize-termination",
|
||||||
|
},
|
||||||
|
Spec: v1.PodSpec{
|
||||||
|
RestartPolicy: v1.RestartPolicyNever,
|
||||||
|
InitContainers: []v1.Container{
|
||||||
|
{
|
||||||
|
Name: restartableInit1,
|
||||||
|
Image: busyboxImage,
|
||||||
|
RestartPolicy: &containerRestartPolicyAlways,
|
||||||
|
Command: ExecCommand(restartableInit1, execCommand{
|
||||||
|
Delay: 60,
|
||||||
|
TerminationSeconds: 5,
|
||||||
|
ExitCode: 0,
|
||||||
|
}),
|
||||||
|
Lifecycle: makePrestop(restartableInit1),
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Name: restartableInit2,
|
||||||
|
Image: busyboxImage,
|
||||||
|
RestartPolicy: &containerRestartPolicyAlways,
|
||||||
|
Command: ExecCommand(restartableInit2, execCommand{
|
||||||
|
Delay: 60,
|
||||||
|
TerminationSeconds: 5,
|
||||||
|
ExitCode: 0,
|
||||||
|
}),
|
||||||
|
Lifecycle: makePrestop(restartableInit2),
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Name: restartableInit3,
|
||||||
|
Image: busyboxImage,
|
||||||
|
RestartPolicy: &containerRestartPolicyAlways,
|
||||||
|
Command: ExecCommand(restartableInit3, execCommand{
|
||||||
|
Delay: 60,
|
||||||
|
TerminationSeconds: 5,
|
||||||
|
ExitCode: 0,
|
||||||
|
}),
|
||||||
|
Lifecycle: makePrestop(restartableInit3),
|
||||||
|
},
|
||||||
|
},
|
||||||
|
Containers: []v1.Container{
|
||||||
|
{
|
||||||
|
Name: regular1,
|
||||||
|
Image: busyboxImage,
|
||||||
|
Command: ExecCommand(regular1, execCommand{
|
||||||
|
Delay: 5,
|
||||||
|
ExitCode: 0,
|
||||||
|
}),
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
}
|
||||||
|
|
||||||
|
preparePod(pod)
|
||||||
|
|
||||||
|
client := e2epod.NewPodClient(f)
|
||||||
|
pod = client.Create(context.TODO(), pod)
|
||||||
|
|
||||||
|
err := e2epod.WaitTimeoutForPodNoLongerRunningInNamespace(context.TODO(), f.ClientSet, pod.Name, pod.Namespace, 5*time.Minute)
|
||||||
|
framework.ExpectNoError(err)
|
||||||
|
|
||||||
|
pod, err = client.Get(context.TODO(), pod.Name, metav1.GetOptions{})
|
||||||
|
framework.ExpectNoError(err)
|
||||||
|
|
||||||
|
results := parseOutput(context.TODO(), f, pod)
|
||||||
|
|
||||||
|
ginkgo.By("Analyzing results")
|
||||||
|
framework.ExpectNoError(results.StartsBefore(restartableInit1, restartableInit2))
|
||||||
|
framework.ExpectNoError(results.StartsBefore(restartableInit1, restartableInit3))
|
||||||
|
framework.ExpectNoError(results.StartsBefore(restartableInit2, restartableInit3))
|
||||||
|
framework.ExpectNoError(results.StartsBefore(restartableInit1, regular1))
|
||||||
|
framework.ExpectNoError(results.StartsBefore(restartableInit2, regular1))
|
||||||
|
framework.ExpectNoError(results.StartsBefore(restartableInit3, regular1))
|
||||||
|
|
||||||
|
ps1, err := results.TimeOfStart(prefixedName(PreStopPrefix, restartableInit1))
|
||||||
|
framework.ExpectNoError(err)
|
||||||
|
ps2, err := results.TimeOfStart(prefixedName(PreStopPrefix, restartableInit2))
|
||||||
|
framework.ExpectNoError(err)
|
||||||
|
ps3, err := results.TimeOfStart(prefixedName(PreStopPrefix, restartableInit3))
|
||||||
|
framework.ExpectNoError(err)
|
||||||
|
|
||||||
|
ps1Last, err := results.TimeOfLastLoop(prefixedName(PreStopPrefix, restartableInit1))
|
||||||
|
framework.ExpectNoError(err)
|
||||||
|
ps2Last, err := results.TimeOfLastLoop(prefixedName(PreStopPrefix, restartableInit2))
|
||||||
|
framework.ExpectNoError(err)
|
||||||
|
ps3Last, err := results.TimeOfLastLoop(prefixedName(PreStopPrefix, restartableInit3))
|
||||||
|
framework.ExpectNoError(err)
|
||||||
|
|
||||||
|
const simulToleration = 0.5
|
||||||
|
// should all end together since they loop infinitely and exceed their grace period
|
||||||
|
gomega.Expect(ps1Last-ps2Last).To(gomega.BeNumerically("~", 0, simulToleration),
|
||||||
|
fmt.Sprintf("expected PostStart 1 & PostStart 2 to be killed at the same time, got %s", results))
|
||||||
|
gomega.Expect(ps1Last-ps3Last).To(gomega.BeNumerically("~", 0, simulToleration),
|
||||||
|
fmt.Sprintf("expected PostStart 1 & PostStart 3 to be killed at the same time, got %s", results))
|
||||||
|
gomega.Expect(ps2Last-ps3Last).To(gomega.BeNumerically("~", 0, simulToleration),
|
||||||
|
fmt.Sprintf("expected PostStart 2 & PostStart 3 to be killed at the same time, got %s", results))
|
||||||
|
|
||||||
|
// 30 seconds + 2 second minimum grace for the SIGKILL
|
||||||
|
const lifetimeToleration = 1
|
||||||
|
gomega.Expect(ps1Last-ps1).To(gomega.BeNumerically("~", 32, lifetimeToleration),
|
||||||
|
fmt.Sprintf("expected PostStart 1 to live for ~32 seconds, got %s", results))
|
||||||
|
gomega.Expect(ps2Last-ps2).To(gomega.BeNumerically("~", 32, lifetimeToleration),
|
||||||
|
fmt.Sprintf("expected PostStart 2 to live for ~32 seconds, got %s", results))
|
||||||
|
gomega.Expect(ps3Last-ps3).To(gomega.BeNumerically("~", 32, lifetimeToleration),
|
||||||
|
fmt.Sprintf("expected PostStart 3 to live for ~32 seconds, got %s", results))
|
||||||
|
|
||||||
|
})
|
||||||
|
|
||||||
|
ginkgo.It("should call sidecar container PreStop hook simultaneously", func() {
|
||||||
|
restartableInit1 := "restartable-init-1"
|
||||||
|
restartableInit2 := "restartable-init-2"
|
||||||
|
restartableInit3 := "restartable-init-3"
|
||||||
|
regular1 := "regular-1"
|
||||||
|
|
||||||
|
makePrestop := func(containerName string) *v1.Lifecycle {
|
||||||
|
return &v1.Lifecycle{
|
||||||
|
PreStop: &v1.LifecycleHandler{
|
||||||
|
Exec: &v1.ExecAction{
|
||||||
|
Command: ExecCommand(prefixedName(PreStopPrefix, containerName), execCommand{
|
||||||
|
Delay: 1,
|
||||||
|
ExitCode: 0,
|
||||||
|
ContainerName: containerName,
|
||||||
|
}),
|
||||||
|
},
|
||||||
|
},
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
pod := &v1.Pod{
|
||||||
|
ObjectMeta: metav1.ObjectMeta{
|
||||||
|
Name: "serialize-termination-simul-prestop",
|
||||||
|
},
|
||||||
|
Spec: v1.PodSpec{
|
||||||
|
RestartPolicy: v1.RestartPolicyNever,
|
||||||
|
InitContainers: []v1.Container{
|
||||||
|
{
|
||||||
|
Name: restartableInit1,
|
||||||
|
Image: busyboxImage,
|
||||||
|
RestartPolicy: &containerRestartPolicyAlways,
|
||||||
|
Command: ExecCommand(restartableInit1, execCommand{
|
||||||
|
Delay: 60,
|
||||||
|
TerminationSeconds: 5,
|
||||||
|
ExitCode: 0,
|
||||||
|
}),
|
||||||
|
Lifecycle: makePrestop(restartableInit1),
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Name: restartableInit2,
|
||||||
|
Image: busyboxImage,
|
||||||
|
RestartPolicy: &containerRestartPolicyAlways,
|
||||||
|
Command: ExecCommand(restartableInit2, execCommand{
|
||||||
|
Delay: 60,
|
||||||
|
TerminationSeconds: 5,
|
||||||
|
ExitCode: 0,
|
||||||
|
}),
|
||||||
|
Lifecycle: makePrestop(restartableInit2),
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Name: restartableInit3,
|
||||||
|
Image: busyboxImage,
|
||||||
|
RestartPolicy: &containerRestartPolicyAlways,
|
||||||
|
Command: ExecCommand(restartableInit3, execCommand{
|
||||||
|
Delay: 60,
|
||||||
|
TerminationSeconds: 5,
|
||||||
|
ExitCode: 0,
|
||||||
|
}),
|
||||||
|
Lifecycle: makePrestop(restartableInit3),
|
||||||
|
},
|
||||||
|
},
|
||||||
|
Containers: []v1.Container{
|
||||||
|
{
|
||||||
|
Name: regular1,
|
||||||
|
Image: busyboxImage,
|
||||||
|
Command: ExecCommand(regular1, execCommand{
|
||||||
|
Delay: 5,
|
||||||
|
ExitCode: 0,
|
||||||
|
}),
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
}
|
||||||
|
|
||||||
|
preparePod(pod)
|
||||||
|
|
||||||
|
client := e2epod.NewPodClient(f)
|
||||||
|
pod = client.Create(context.TODO(), pod)
|
||||||
|
|
||||||
|
err := e2epod.WaitTimeoutForPodNoLongerRunningInNamespace(context.TODO(), f.ClientSet, pod.Name, pod.Namespace, 5*time.Minute)
|
||||||
|
framework.ExpectNoError(err)
|
||||||
|
|
||||||
|
pod, err = client.Get(context.TODO(), pod.Name, metav1.GetOptions{})
|
||||||
|
framework.ExpectNoError(err)
|
||||||
|
|
||||||
|
results := parseOutput(context.TODO(), f, pod)
|
||||||
|
|
||||||
|
ginkgo.By("Analyzing results")
|
||||||
|
framework.ExpectNoError(results.StartsBefore(restartableInit1, restartableInit2))
|
||||||
|
framework.ExpectNoError(results.StartsBefore(restartableInit1, restartableInit3))
|
||||||
|
framework.ExpectNoError(results.StartsBefore(restartableInit2, restartableInit3))
|
||||||
|
framework.ExpectNoError(results.StartsBefore(restartableInit1, regular1))
|
||||||
|
framework.ExpectNoError(results.StartsBefore(restartableInit2, regular1))
|
||||||
|
framework.ExpectNoError(results.StartsBefore(restartableInit3, regular1))
|
||||||
|
|
||||||
|
// main containers exit first
|
||||||
|
framework.ExpectNoError(results.ExitsBefore(regular1, restartableInit1))
|
||||||
|
framework.ExpectNoError(results.ExitsBefore(regular1, restartableInit2))
|
||||||
|
framework.ExpectNoError(results.ExitsBefore(regular1, restartableInit3))
|
||||||
|
|
||||||
|
// followed by sidecars in reverse order
|
||||||
|
framework.ExpectNoError(results.ExitsBefore(restartableInit3, restartableInit2))
|
||||||
|
framework.ExpectNoError(results.ExitsBefore(restartableInit2, restartableInit1))
|
||||||
|
|
||||||
|
// and the pre-stop hooks should have been called simultaneously
|
||||||
|
ps1, err := results.TimeOfStart(prefixedName(PreStopPrefix, restartableInit1))
|
||||||
|
framework.ExpectNoError(err)
|
||||||
|
ps2, err := results.TimeOfStart(prefixedName(PreStopPrefix, restartableInit2))
|
||||||
|
framework.ExpectNoError(err)
|
||||||
|
ps3, err := results.TimeOfStart(prefixedName(PreStopPrefix, restartableInit3))
|
||||||
|
framework.ExpectNoError(err)
|
||||||
|
|
||||||
|
const toleration = 0.5
|
||||||
|
gomega.Expect(ps1-ps2).To(gomega.BeNumerically("~", 0, toleration),
|
||||||
|
fmt.Sprintf("expected PostStart 1 & PostStart 2 to start at the same time, got %s", results))
|
||||||
|
gomega.Expect(ps1-ps3).To(gomega.BeNumerically("~", 0, toleration),
|
||||||
|
fmt.Sprintf("expected PostStart 1 & PostStart 3 to start at the same time, got %s", results))
|
||||||
|
gomega.Expect(ps2-ps3).To(gomega.BeNumerically("~", 0, toleration),
|
||||||
|
fmt.Sprintf("expected PostStart 2 & PostStart 3 to start at the same time, got %s", results))
|
||||||
|
})
|
||||||
|
|
||||||
|
ginkgo.It("should not hang in termination if terminated during initialization", func() {
|
||||||
|
startInit := "start-init"
|
||||||
|
restartableInit1 := "restartable-init-1"
|
||||||
|
restartableInit2 := "restartable-init-2"
|
||||||
|
restartableInit3 := "restartable-init-3"
|
||||||
|
regular1 := "regular-1"
|
||||||
|
|
||||||
|
makePrestop := func(containerName string) *v1.Lifecycle {
|
||||||
|
return &v1.Lifecycle{
|
||||||
|
PreStop: &v1.LifecycleHandler{
|
||||||
|
Exec: &v1.ExecAction{
|
||||||
|
Command: ExecCommand(prefixedName(PreStopPrefix, containerName), execCommand{
|
||||||
|
Delay: 1,
|
||||||
|
ExitCode: 0,
|
||||||
|
ContainerName: containerName,
|
||||||
|
}),
|
||||||
|
},
|
||||||
|
},
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
pod := &v1.Pod{
|
||||||
|
ObjectMeta: metav1.ObjectMeta{
|
||||||
|
Name: "dont-hang-if-terminated-in-init",
|
||||||
|
},
|
||||||
|
Spec: v1.PodSpec{
|
||||||
|
RestartPolicy: v1.RestartPolicyNever,
|
||||||
|
InitContainers: []v1.Container{
|
||||||
|
{
|
||||||
|
Name: startInit,
|
||||||
|
Image: busyboxImage,
|
||||||
|
Command: ExecCommand(startInit, execCommand{
|
||||||
|
Delay: 300,
|
||||||
|
TerminationSeconds: 0,
|
||||||
|
ExitCode: 0,
|
||||||
|
}),
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Name: restartableInit1,
|
||||||
|
Image: busyboxImage,
|
||||||
|
RestartPolicy: &containerRestartPolicyAlways,
|
||||||
|
Command: ExecCommand(restartableInit1, execCommand{
|
||||||
|
Delay: 60,
|
||||||
|
TerminationSeconds: 5,
|
||||||
|
ExitCode: 0,
|
||||||
|
}),
|
||||||
|
Lifecycle: makePrestop(restartableInit1),
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Name: restartableInit2,
|
||||||
|
Image: busyboxImage,
|
||||||
|
RestartPolicy: &containerRestartPolicyAlways,
|
||||||
|
Command: ExecCommand(restartableInit2, execCommand{
|
||||||
|
Delay: 60,
|
||||||
|
TerminationSeconds: 5,
|
||||||
|
ExitCode: 0,
|
||||||
|
}),
|
||||||
|
Lifecycle: makePrestop(restartableInit2),
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Name: restartableInit3,
|
||||||
|
Image: busyboxImage,
|
||||||
|
RestartPolicy: &containerRestartPolicyAlways,
|
||||||
|
Command: ExecCommand(restartableInit3, execCommand{
|
||||||
|
Delay: 60,
|
||||||
|
TerminationSeconds: 5,
|
||||||
|
ExitCode: 0,
|
||||||
|
}),
|
||||||
|
Lifecycle: makePrestop(restartableInit3),
|
||||||
|
},
|
||||||
|
},
|
||||||
|
Containers: []v1.Container{
|
||||||
|
{
|
||||||
|
Name: regular1,
|
||||||
|
Image: busyboxImage,
|
||||||
|
Command: ExecCommand(regular1, execCommand{
|
||||||
|
Delay: 5,
|
||||||
|
ExitCode: 0,
|
||||||
|
}),
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
}
|
||||||
|
|
||||||
|
preparePod(pod)
|
||||||
|
|
||||||
|
client := e2epod.NewPodClient(f)
|
||||||
|
pod = client.Create(context.TODO(), pod)
|
||||||
|
|
||||||
|
err := e2epod.WaitForPodCondition(context.TODO(), f.ClientSet, pod.Namespace, pod.Name, "pod pending and init running", 2*time.Minute, func(pod *v1.Pod) (bool, error) {
|
||||||
|
if pod.Status.Phase != v1.PodPending {
|
||||||
|
return false, fmt.Errorf("pod should be in pending phase")
|
||||||
|
}
|
||||||
|
if len(pod.Status.InitContainerStatuses) < 1 {
|
||||||
|
return false, nil
|
||||||
|
}
|
||||||
|
containerStatus := pod.Status.InitContainerStatuses[0]
|
||||||
|
return *containerStatus.Started && containerStatus.State.Running != nil, nil
|
||||||
|
})
|
||||||
|
framework.ExpectNoError(err)
|
||||||
|
|
||||||
|
// the init container is running, so we stop the pod before the sidecars even start
|
||||||
|
start := time.Now()
|
||||||
|
grace := int64(3)
|
||||||
|
ginkgo.By("deleting the pod")
|
||||||
|
err = client.Delete(context.TODO(), pod.Name, metav1.DeleteOptions{GracePeriodSeconds: &grace})
|
||||||
|
framework.ExpectNoError(err)
|
||||||
|
ginkgo.By("waiting for the pod to disappear")
|
||||||
|
err = e2epod.WaitForPodNotFoundInNamespace(context.TODO(), f.ClientSet, pod.Name, pod.Namespace, 120*time.Second)
|
||||||
|
framework.ExpectNoError(err)
|
||||||
|
|
||||||
|
buffer := int64(2)
|
||||||
|
deleteTime := time.Since(start).Seconds()
|
||||||
|
// should delete quickly and not try to start/wait on any sidecars since they never started
|
||||||
|
gomega.Expect(deleteTime).To(gomega.BeNumerically("<", grace+buffer), fmt.Sprintf("should delete in < %d seconds, took %f", grace+buffer, deleteTime))
|
||||||
|
})
|
||||||
})
|
})
|
||||||
|
Loading…
Reference in New Issue
Block a user