mirror of
https://github.com/k3s-io/kubernetes.git
synced 2025-07-21 02:41:25 +00:00
Merge pull request #128029 from bouaouda-achraf/oom-adjust-score-pod-aware
kubelet(OOM-score-adj): change OOM score adjustment calculation for sidecar container
This commit is contained in:
commit
2fabf5fa63
43
pkg/kubelet/qos/helpers.go
Normal file
43
pkg/kubelet/qos/helpers.go
Normal file
@ -0,0 +1,43 @@
|
||||
/*
|
||||
Copyright 2024 The Kubernetes Authors.
|
||||
|
||||
Licensed under the Apache License, Version 2.0 (the "License");
|
||||
you may not use this file except in compliance with the License.
|
||||
You may obtain a copy of the License at
|
||||
|
||||
http://www.apache.org/licenses/LICENSE-2.0
|
||||
|
||||
Unless required by applicable law or agreed to in writing, software
|
||||
distributed under the License is distributed on an "AS IS" BASIS,
|
||||
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
See the License for the specific language governing permissions and
|
||||
limitations under the License.
|
||||
*/
|
||||
|
||||
// Package qos contains helper functions for quality of service.
|
||||
// For each resource (memory, CPU) Kubelet supports three classes of containers.
|
||||
// Memory guaranteed containers will receive the highest priority and will get all the resources
|
||||
// they need.
|
||||
// Burstable containers will be guaranteed their request and can "burst" and use more resources
|
||||
// when available.
|
||||
// Best-Effort containers, which don't specify a request, can use resources only if not being used
|
||||
// by other pods.
|
||||
|
||||
package qos // import "k8s.io/kubernetes/pkg/kubelet/qos"
|
||||
|
||||
import (
|
||||
v1 "k8s.io/api/core/v1"
|
||||
)
|
||||
|
||||
// minRegularContainerMemory returns the minimum memory resource quantity
|
||||
// across all regular containers in pod.Spec.Containers.
|
||||
// It does not include initContainers (both restartable and non-restartable).
|
||||
func minRegularContainerMemory(pod v1.Pod) int64 {
|
||||
memoryValue := pod.Spec.Containers[0].Resources.Requests.Memory().Value()
|
||||
for _, container := range pod.Spec.Containers[1:] {
|
||||
if container.Resources.Requests.Memory().Value() < memoryValue {
|
||||
memoryValue = container.Resources.Requests.Memory().Value()
|
||||
}
|
||||
}
|
||||
return memoryValue
|
||||
}
|
@ -18,7 +18,9 @@ package qos
|
||||
|
||||
import (
|
||||
v1 "k8s.io/api/core/v1"
|
||||
utilfeature "k8s.io/apiserver/pkg/util/feature"
|
||||
v1qos "k8s.io/kubernetes/pkg/apis/core/v1/helper/qos"
|
||||
"k8s.io/kubernetes/pkg/features"
|
||||
"k8s.io/kubernetes/pkg/kubelet/types"
|
||||
)
|
||||
|
||||
@ -63,6 +65,20 @@ func GetContainerOOMScoreAdjust(pod *v1.Pod, container *v1.Container, memoryCapa
|
||||
// Note that this is a heuristic, it won't work if a container has many small processes.
|
||||
memoryRequest := container.Resources.Requests.Memory().Value()
|
||||
oomScoreAdjust := 1000 - (1000*memoryRequest)/memoryCapacity
|
||||
|
||||
// adapt the sidecarContainer memoryRequest for OOM ADJ calculation
|
||||
// calculate the oom score adjustment based on: max-memory( currentSideCarContainer , min-memory(regular containers) ) .
|
||||
if utilfeature.DefaultFeatureGate.Enabled(features.SidecarContainers) && isSidecarContainer(pod, container) {
|
||||
// check min memory quantity in regular containers
|
||||
minMemoryRequest := minRegularContainerMemory(*pod)
|
||||
minMemoryOomScoreAdjust := 1000 - (1000*minMemoryRequest)/memoryCapacity
|
||||
// the OOM adjustment for sidecar container will match
|
||||
// or fall below the OOM score adjustment of regular containers in the Pod.
|
||||
if oomScoreAdjust > minMemoryOomScoreAdjust {
|
||||
oomScoreAdjust = minMemoryOomScoreAdjust
|
||||
}
|
||||
}
|
||||
|
||||
// A guaranteed pod using 100% of memory can have an OOM score of 10. Ensure
|
||||
// that burstable pods have a higher OOM score adjustment.
|
||||
if int(oomScoreAdjust) < (1000 + guaranteedOOMScoreAdj) {
|
||||
@ -74,3 +90,18 @@ func GetContainerOOMScoreAdjust(pod *v1.Pod, container *v1.Container, memoryCapa
|
||||
}
|
||||
return int(oomScoreAdjust)
|
||||
}
|
||||
|
||||
// isSidecarContainer returns a boolean indicating whether a container is a sidecar or not.
|
||||
// Since v1.Container does not directly specify whether a container is a sidecar,
|
||||
// this function uses available indicators (container.RestartPolicy == v1.ContainerRestartPolicyAlways)
|
||||
// to make that determination.
|
||||
func isSidecarContainer(pod *v1.Pod, container *v1.Container) bool {
|
||||
if container.RestartPolicy != nil && *container.RestartPolicy == v1.ContainerRestartPolicyAlways {
|
||||
for _, initContainer := range pod.Spec.InitContainers {
|
||||
if initContainer.Name == container.Name {
|
||||
return true
|
||||
}
|
||||
}
|
||||
}
|
||||
return false
|
||||
}
|
||||
|
@ -23,6 +23,10 @@ import (
|
||||
v1 "k8s.io/api/core/v1"
|
||||
"k8s.io/apimachinery/pkg/api/resource"
|
||||
"k8s.io/kubernetes/pkg/apis/scheduling"
|
||||
|
||||
utilfeature "k8s.io/apiserver/pkg/util/feature"
|
||||
featuregatetesting "k8s.io/component-base/featuregate/testing"
|
||||
"k8s.io/kubernetes/pkg/features"
|
||||
)
|
||||
|
||||
const (
|
||||
@ -34,6 +38,7 @@ var (
|
||||
Spec: v1.PodSpec{
|
||||
Containers: []v1.Container{
|
||||
{
|
||||
Name: "cpu-limit",
|
||||
Resources: v1.ResourceRequirements{
|
||||
Limits: v1.ResourceList{
|
||||
v1.ResourceName(v1.ResourceCPU): resource.MustParse("10"),
|
||||
@ -48,6 +53,7 @@ var (
|
||||
Spec: v1.PodSpec{
|
||||
Containers: []v1.Container{
|
||||
{
|
||||
Name: "memory-limit-cpu-request",
|
||||
Resources: v1.ResourceRequirements{
|
||||
Requests: v1.ResourceList{
|
||||
v1.ResourceName(v1.ResourceCPU): resource.MustParse("0"),
|
||||
@ -65,6 +71,7 @@ var (
|
||||
Spec: v1.PodSpec{
|
||||
Containers: []v1.Container{
|
||||
{
|
||||
Name: "zero-memory-limit",
|
||||
Resources: v1.ResourceRequirements{
|
||||
Limits: v1.ResourceList{
|
||||
v1.ResourceName(v1.ResourceMemory): resource.MustParse("0"),
|
||||
@ -79,6 +86,7 @@ var (
|
||||
Spec: v1.PodSpec{
|
||||
Containers: []v1.Container{
|
||||
{
|
||||
Name: "no-request-limit",
|
||||
Resources: v1.ResourceRequirements{},
|
||||
},
|
||||
},
|
||||
@ -89,6 +97,7 @@ var (
|
||||
Spec: v1.PodSpec{
|
||||
Containers: []v1.Container{
|
||||
{
|
||||
Name: "equal-request-limit-cpu-memory",
|
||||
Resources: v1.ResourceRequirements{
|
||||
Requests: v1.ResourceList{
|
||||
v1.ResourceName(v1.ResourceMemory): resource.MustParse("10G"),
|
||||
@ -108,6 +117,7 @@ var (
|
||||
Spec: v1.PodSpec{
|
||||
Containers: []v1.Container{
|
||||
{
|
||||
Name: "cpu-unlimited-memory-limited-with-requests",
|
||||
Resources: v1.ResourceRequirements{
|
||||
Requests: v1.ResourceList{
|
||||
v1.ResourceName(v1.ResourceMemory): resource.MustParse(strconv.FormatInt(standardMemoryAmount/2, 10)),
|
||||
@ -126,6 +136,7 @@ var (
|
||||
Spec: v1.PodSpec{
|
||||
Containers: []v1.Container{
|
||||
{
|
||||
Name: "request-no-limit",
|
||||
Resources: v1.ResourceRequirements{
|
||||
Requests: v1.ResourceList{
|
||||
v1.ResourceName(v1.ResourceMemory): resource.MustParse(strconv.FormatInt(standardMemoryAmount-1, 10)),
|
||||
@ -145,6 +156,7 @@ var (
|
||||
Priority: &systemCritical,
|
||||
Containers: []v1.Container{
|
||||
{
|
||||
Name: "cluster-critical",
|
||||
Resources: v1.ResourceRequirements{},
|
||||
},
|
||||
},
|
||||
@ -159,81 +171,273 @@ var (
|
||||
Priority: &systemNodeCritical,
|
||||
Containers: []v1.Container{
|
||||
{
|
||||
Name: "node-critical",
|
||||
Resources: v1.ResourceRequirements{},
|
||||
},
|
||||
},
|
||||
},
|
||||
}
|
||||
sampleDefaultMemRequest = resource.MustParse(strconv.FormatInt(standardMemoryAmount/8, 10))
|
||||
sampleDefaultMemLimit = resource.MustParse(strconv.FormatInt(1000+(standardMemoryAmount/8), 10))
|
||||
|
||||
sampleContainer = v1.Container{
|
||||
Name: "main-1",
|
||||
Resources: v1.ResourceRequirements{
|
||||
Requests: v1.ResourceList{
|
||||
v1.ResourceName(v1.ResourceMemory): sampleDefaultMemRequest,
|
||||
},
|
||||
Limits: v1.ResourceList{
|
||||
v1.ResourceName(v1.ResourceMemory): sampleDefaultMemLimit,
|
||||
},
|
||||
},
|
||||
}
|
||||
|
||||
burstableUniqueContainerPod = v1.Pod{
|
||||
Spec: v1.PodSpec{
|
||||
Containers: []v1.Container{
|
||||
{
|
||||
Name: "burstable-unique-container",
|
||||
Resources: v1.ResourceRequirements{
|
||||
Requests: v1.ResourceList{
|
||||
v1.ResourceName(v1.ResourceMemory): sampleDefaultMemRequest,
|
||||
},
|
||||
Limits: v1.ResourceList{
|
||||
v1.ResourceName(v1.ResourceMemory): sampleDefaultMemLimit,
|
||||
},
|
||||
},
|
||||
},
|
||||
},
|
||||
},
|
||||
}
|
||||
|
||||
sampleInitContainer = v1.Container{
|
||||
Name: "init-container",
|
||||
Resources: v1.ResourceRequirements{
|
||||
Requests: v1.ResourceList{
|
||||
v1.ResourceName(v1.ResourceMemory): sampleDefaultMemRequest,
|
||||
},
|
||||
Limits: v1.ResourceList{
|
||||
v1.ResourceName(v1.ResourceMemory): sampleDefaultMemLimit,
|
||||
},
|
||||
},
|
||||
}
|
||||
restartPolicyAlways = v1.ContainerRestartPolicyAlways
|
||||
sampleSidecarContainer = v1.Container{
|
||||
Name: "sidecar-container",
|
||||
RestartPolicy: &restartPolicyAlways,
|
||||
Resources: v1.ResourceRequirements{
|
||||
Requests: v1.ResourceList{
|
||||
v1.ResourceName(v1.ResourceMemory): sampleDefaultMemRequest,
|
||||
},
|
||||
Limits: v1.ResourceList{
|
||||
v1.ResourceName(v1.ResourceMemory): sampleDefaultMemLimit,
|
||||
},
|
||||
},
|
||||
}
|
||||
|
||||
sampleSmallSidecarContainer = v1.Container{
|
||||
Name: "sidecar-small-container",
|
||||
RestartPolicy: &restartPolicyAlways,
|
||||
Resources: v1.ResourceRequirements{
|
||||
Requests: v1.ResourceList{
|
||||
v1.ResourceName(v1.ResourceMemory): resource.MustParse(strconv.FormatInt(standardMemoryAmount/20, 10)),
|
||||
},
|
||||
Limits: v1.ResourceList{
|
||||
v1.ResourceName(v1.ResourceMemory): sampleDefaultMemLimit,
|
||||
},
|
||||
},
|
||||
}
|
||||
|
||||
sampleBigSidecarContainer = v1.Container{
|
||||
Name: "sidecar-big-container",
|
||||
RestartPolicy: &restartPolicyAlways,
|
||||
Resources: v1.ResourceRequirements{
|
||||
Requests: v1.ResourceList{
|
||||
v1.ResourceName(v1.ResourceMemory): resource.MustParse(strconv.FormatInt(standardMemoryAmount/2, 10)),
|
||||
},
|
||||
Limits: v1.ResourceList{
|
||||
v1.ResourceName(v1.ResourceMemory): sampleDefaultMemLimit,
|
||||
},
|
||||
},
|
||||
}
|
||||
|
||||
burstableMixedUniqueMainContainerPod = v1.Pod{
|
||||
Spec: v1.PodSpec{
|
||||
InitContainers: []v1.Container{
|
||||
sampleInitContainer,
|
||||
}, Containers: []v1.Container{
|
||||
sampleContainer,
|
||||
},
|
||||
},
|
||||
}
|
||||
|
||||
burstableMixedMultiContainerSameRequestPod = v1.Pod{
|
||||
Spec: v1.PodSpec{
|
||||
InitContainers: []v1.Container{
|
||||
sampleInitContainer, sampleSidecarContainer,
|
||||
}, Containers: []v1.Container{
|
||||
sampleContainer,
|
||||
},
|
||||
},
|
||||
}
|
||||
|
||||
burstableMixedMultiContainerSmallSidecarPod = v1.Pod{
|
||||
Spec: v1.PodSpec{
|
||||
InitContainers: []v1.Container{
|
||||
sampleInitContainer, sampleSmallSidecarContainer,
|
||||
}, Containers: []v1.Container{
|
||||
sampleContainer,
|
||||
},
|
||||
},
|
||||
}
|
||||
|
||||
burstableMixedMultiContainerBigSidecarContainerPod = v1.Pod{
|
||||
Spec: v1.PodSpec{
|
||||
InitContainers: []v1.Container{
|
||||
sampleInitContainer, sampleBigSidecarContainer,
|
||||
}, Containers: []v1.Container{
|
||||
sampleContainer,
|
||||
},
|
||||
},
|
||||
}
|
||||
)
|
||||
|
||||
type lowHighOOMScoreAdjTest struct {
|
||||
lowOOMScoreAdj int
|
||||
highOOMScoreAdj int
|
||||
}
|
||||
type oomTest struct {
|
||||
pod *v1.Pod
|
||||
memoryCapacity int64
|
||||
lowOOMScoreAdj int // The max oom_score_adj score the container should be assigned.
|
||||
highOOMScoreAdj int // The min oom_score_adj score the container should be assigned.
|
||||
pod *v1.Pod
|
||||
memoryCapacity int64
|
||||
lowHighOOMScoreAdj map[string]lowHighOOMScoreAdjTest // [container-name] : min and max oom_score_adj score the container should be assigned.
|
||||
sidecarContainersFeatureEnabled bool
|
||||
}
|
||||
|
||||
func TestGetContainerOOMScoreAdjust(t *testing.T) {
|
||||
oomTests := []oomTest{
|
||||
{
|
||||
pod: &cpuLimit,
|
||||
memoryCapacity: 4000000000,
|
||||
lowOOMScoreAdj: 999,
|
||||
highOOMScoreAdj: 999,
|
||||
oomTests := map[string]oomTest{
|
||||
"cpu-limit": {
|
||||
pod: &cpuLimit,
|
||||
memoryCapacity: 4000000000,
|
||||
lowHighOOMScoreAdj: map[string]lowHighOOMScoreAdjTest{
|
||||
"cpu-limit": {lowOOMScoreAdj: 999, highOOMScoreAdj: 999},
|
||||
},
|
||||
},
|
||||
{
|
||||
pod: &memoryLimitCPURequest,
|
||||
memoryCapacity: 8000000000,
|
||||
lowOOMScoreAdj: 999,
|
||||
highOOMScoreAdj: 999,
|
||||
"memory-limit-cpu-request": {
|
||||
pod: &memoryLimitCPURequest,
|
||||
memoryCapacity: 8000000000,
|
||||
lowHighOOMScoreAdj: map[string]lowHighOOMScoreAdjTest{
|
||||
"memory-limit-cpu-request": {lowOOMScoreAdj: 999, highOOMScoreAdj: 999},
|
||||
},
|
||||
},
|
||||
{
|
||||
pod: &zeroMemoryLimit,
|
||||
memoryCapacity: 7230457451,
|
||||
lowOOMScoreAdj: 1000,
|
||||
highOOMScoreAdj: 1000,
|
||||
"zero-memory-limit": {
|
||||
pod: &zeroMemoryLimit,
|
||||
memoryCapacity: 7230457451,
|
||||
lowHighOOMScoreAdj: map[string]lowHighOOMScoreAdjTest{
|
||||
"zero-memory-limit": {lowOOMScoreAdj: 1000, highOOMScoreAdj: 1000},
|
||||
},
|
||||
},
|
||||
{
|
||||
pod: &noRequestLimit,
|
||||
memoryCapacity: 4000000000,
|
||||
lowOOMScoreAdj: 1000,
|
||||
highOOMScoreAdj: 1000,
|
||||
"no-request-limit": {
|
||||
pod: &noRequestLimit,
|
||||
memoryCapacity: 4000000000,
|
||||
lowHighOOMScoreAdj: map[string]lowHighOOMScoreAdjTest{
|
||||
"no-request-limit": {lowOOMScoreAdj: 1000, highOOMScoreAdj: 1000},
|
||||
},
|
||||
},
|
||||
{
|
||||
pod: &equalRequestLimitCPUMemory,
|
||||
memoryCapacity: 123456789,
|
||||
lowOOMScoreAdj: -997,
|
||||
highOOMScoreAdj: -997,
|
||||
"equal-request-limit-cpu-memory": {
|
||||
pod: &equalRequestLimitCPUMemory,
|
||||
memoryCapacity: 123456789,
|
||||
lowHighOOMScoreAdj: map[string]lowHighOOMScoreAdjTest{
|
||||
"equal-request-limit-cpu-memory": {lowOOMScoreAdj: -997, highOOMScoreAdj: -997},
|
||||
},
|
||||
},
|
||||
{
|
||||
pod: &cpuUnlimitedMemoryLimitedWithRequests,
|
||||
memoryCapacity: standardMemoryAmount,
|
||||
lowOOMScoreAdj: 495,
|
||||
highOOMScoreAdj: 505,
|
||||
"cpu-unlimited-memory-limited-with-requests": {
|
||||
pod: &cpuUnlimitedMemoryLimitedWithRequests,
|
||||
memoryCapacity: standardMemoryAmount,
|
||||
lowHighOOMScoreAdj: map[string]lowHighOOMScoreAdjTest{
|
||||
"cpu-unlimited-memory-limited-with-requests": {lowOOMScoreAdj: 495, highOOMScoreAdj: 505},
|
||||
},
|
||||
},
|
||||
{
|
||||
pod: &requestNoLimit,
|
||||
memoryCapacity: standardMemoryAmount,
|
||||
lowOOMScoreAdj: 3,
|
||||
highOOMScoreAdj: 3,
|
||||
"request-no-limit": {
|
||||
pod: &requestNoLimit,
|
||||
memoryCapacity: standardMemoryAmount,
|
||||
lowHighOOMScoreAdj: map[string]lowHighOOMScoreAdjTest{
|
||||
"request-no-limit": {lowOOMScoreAdj: 3, highOOMScoreAdj: 3},
|
||||
},
|
||||
},
|
||||
{
|
||||
pod: &clusterCritical,
|
||||
memoryCapacity: 4000000000,
|
||||
lowOOMScoreAdj: 1000,
|
||||
highOOMScoreAdj: 1000,
|
||||
"cluster-critical": {
|
||||
pod: &clusterCritical,
|
||||
memoryCapacity: 4000000000,
|
||||
lowHighOOMScoreAdj: map[string]lowHighOOMScoreAdjTest{
|
||||
"cluster-critical": {lowOOMScoreAdj: 1000, highOOMScoreAdj: 1000},
|
||||
},
|
||||
},
|
||||
{
|
||||
pod: &nodeCritical,
|
||||
memoryCapacity: 4000000000,
|
||||
lowOOMScoreAdj: -997,
|
||||
highOOMScoreAdj: -997,
|
||||
"node-critical": {
|
||||
pod: &nodeCritical,
|
||||
memoryCapacity: 4000000000,
|
||||
lowHighOOMScoreAdj: map[string]lowHighOOMScoreAdjTest{
|
||||
"node-critical": {lowOOMScoreAdj: -997, highOOMScoreAdj: -997},
|
||||
},
|
||||
},
|
||||
"burstable-unique-container-pod": {
|
||||
pod: &burstableUniqueContainerPod,
|
||||
memoryCapacity: standardMemoryAmount,
|
||||
lowHighOOMScoreAdj: map[string]lowHighOOMScoreAdjTest{
|
||||
"burstable-unique-container": {lowOOMScoreAdj: 875, highOOMScoreAdj: 880},
|
||||
},
|
||||
sidecarContainersFeatureEnabled: true,
|
||||
},
|
||||
"burstable-mixed-unique-main-container-pod": {
|
||||
pod: &burstableMixedUniqueMainContainerPod,
|
||||
memoryCapacity: standardMemoryAmount,
|
||||
lowHighOOMScoreAdj: map[string]lowHighOOMScoreAdjTest{
|
||||
"init-container": {lowOOMScoreAdj: 875, highOOMScoreAdj: 880},
|
||||
"main-1": {lowOOMScoreAdj: 875, highOOMScoreAdj: 880},
|
||||
},
|
||||
sidecarContainersFeatureEnabled: true,
|
||||
},
|
||||
"burstable-mixed-multi-container-small-sidecar-pod": {
|
||||
pod: &burstableMixedMultiContainerSmallSidecarPod,
|
||||
memoryCapacity: standardMemoryAmount,
|
||||
lowHighOOMScoreAdj: map[string]lowHighOOMScoreAdjTest{
|
||||
"init-container": {lowOOMScoreAdj: 875, highOOMScoreAdj: 880},
|
||||
"sidecar-small-container": {lowOOMScoreAdj: 875, highOOMScoreAdj: 875},
|
||||
"main-1": {lowOOMScoreAdj: 875, highOOMScoreAdj: 875},
|
||||
},
|
||||
sidecarContainersFeatureEnabled: true,
|
||||
},
|
||||
"burstable-mixed-multi-container-sample-request-pod": {
|
||||
pod: &burstableMixedMultiContainerSameRequestPod,
|
||||
memoryCapacity: standardMemoryAmount,
|
||||
lowHighOOMScoreAdj: map[string]lowHighOOMScoreAdjTest{
|
||||
"init-container": {lowOOMScoreAdj: 875, highOOMScoreAdj: 880},
|
||||
"sidecar-container": {lowOOMScoreAdj: 875, highOOMScoreAdj: 875},
|
||||
"main-1": {lowOOMScoreAdj: 875, highOOMScoreAdj: 875},
|
||||
},
|
||||
sidecarContainersFeatureEnabled: true,
|
||||
},
|
||||
"burstable-mixed-multi-container-big-sidecar-container-pod": {
|
||||
pod: &burstableMixedMultiContainerBigSidecarContainerPod,
|
||||
memoryCapacity: standardMemoryAmount,
|
||||
lowHighOOMScoreAdj: map[string]lowHighOOMScoreAdjTest{
|
||||
"init-container": {lowOOMScoreAdj: 875, highOOMScoreAdj: 880},
|
||||
"sidecar-big-container": {lowOOMScoreAdj: 500, highOOMScoreAdj: 500},
|
||||
"main-1": {lowOOMScoreAdj: 875, highOOMScoreAdj: 875},
|
||||
},
|
||||
sidecarContainersFeatureEnabled: true,
|
||||
},
|
||||
}
|
||||
for _, test := range oomTests {
|
||||
oomScoreAdj := GetContainerOOMScoreAdjust(test.pod, &test.pod.Spec.Containers[0], test.memoryCapacity)
|
||||
if oomScoreAdj < test.lowOOMScoreAdj || oomScoreAdj > test.highOOMScoreAdj {
|
||||
t.Errorf("oom_score_adj should be between %d and %d, but was %d", test.lowOOMScoreAdj, test.highOOMScoreAdj, oomScoreAdj)
|
||||
}
|
||||
for name, test := range oomTests {
|
||||
t.Run(name, func(t *testing.T) {
|
||||
featuregatetesting.SetFeatureGateDuringTest(t, utilfeature.DefaultFeatureGate, features.SidecarContainers, test.sidecarContainersFeatureEnabled)
|
||||
listContainers := test.pod.Spec.InitContainers
|
||||
listContainers = append(listContainers, test.pod.Spec.Containers...)
|
||||
for _, container := range listContainers {
|
||||
oomScoreAdj := GetContainerOOMScoreAdjust(test.pod, &container, test.memoryCapacity)
|
||||
if oomScoreAdj < test.lowHighOOMScoreAdj[container.Name].lowOOMScoreAdj || oomScoreAdj > test.lowHighOOMScoreAdj[container.Name].highOOMScoreAdj {
|
||||
t.Errorf("oom_score_adj %s should be between %d and %d, but was %d", container.Name, test.lowHighOOMScoreAdj[container.Name].lowOOMScoreAdj, test.lowHighOOMScoreAdj[container.Name].highOOMScoreAdj, oomScoreAdj)
|
||||
}
|
||||
}
|
||||
})
|
||||
|
||||
}
|
||||
}
|
||||
|
Loading…
Reference in New Issue
Block a user