mirror of
https://github.com/k3s-io/kubernetes.git
synced 2025-07-22 11:21:47 +00:00
Refactor CPUMananger-e2e-tests so that it be reused by topology-manager-e2e-testsuite.
Signed-off-by: Deepthi Dharwar <ddharwar@redhat.com>
This commit is contained in:
parent
1dd25a9efc
commit
4abbce4549
@ -258,16 +258,295 @@ func enableCPUManagerInKubelet(f *framework.Framework, cleanStateFile bool) (old
|
|||||||
return oldCfg
|
return oldCfg
|
||||||
}
|
}
|
||||||
|
|
||||||
func runCPUManagerTests(f *framework.Framework) {
|
func runGuPodTest(f *framework.Framework) {
|
||||||
var cpuCap, cpuAlloc int64
|
var ctnAttrs []ctnAttribute
|
||||||
var oldCfg *kubeletconfig.KubeletConfiguration
|
var cpu1 int
|
||||||
|
var err error
|
||||||
|
var cpuList []int
|
||||||
|
var pod *v1.Pod
|
||||||
|
var expAllowedCPUsListRegex string
|
||||||
|
|
||||||
|
ctnAttrs = []ctnAttribute{
|
||||||
|
{
|
||||||
|
ctnName: "gu-container",
|
||||||
|
cpuRequest: "1000m",
|
||||||
|
cpuLimit: "1000m",
|
||||||
|
},
|
||||||
|
}
|
||||||
|
pod = makeCPUManagerPod("gu-pod", ctnAttrs)
|
||||||
|
pod = f.PodClient().CreateSync(pod)
|
||||||
|
|
||||||
|
ginkgo.By("checking if the expected cpuset was assigned")
|
||||||
|
cpu1 = 1
|
||||||
|
if isHTEnabled() {
|
||||||
|
cpuList = cpuset.MustParse(getCPUSiblingList(0)).ToSlice()
|
||||||
|
cpu1 = cpuList[1]
|
||||||
|
} else if isMultiNUMA() {
|
||||||
|
cpuList = cpuset.MustParse(getCoreSiblingList(0)).ToSlice()
|
||||||
|
cpu1 = cpuList[1]
|
||||||
|
}
|
||||||
|
expAllowedCPUsListRegex = fmt.Sprintf("^%d\n$", cpu1)
|
||||||
|
err = f.PodClient().MatchContainerOutput(pod.Name, pod.Spec.Containers[0].Name, expAllowedCPUsListRegex)
|
||||||
|
framework.ExpectNoError(err, "expected log not found in container [%s] of pod [%s]",
|
||||||
|
pod.Spec.Containers[0].Name, pod.Name)
|
||||||
|
|
||||||
|
ginkgo.By("by deleting the pods and waiting for container removal")
|
||||||
|
deletePods(f, []string{pod.Name})
|
||||||
|
waitForContainerRemoval(pod.Spec.Containers[0].Name, pod.Name, pod.Namespace)
|
||||||
|
}
|
||||||
|
|
||||||
|
func runNonGuPodTest(f *framework.Framework, cpuCap int64) {
|
||||||
|
var ctnAttrs []ctnAttribute
|
||||||
|
var err error
|
||||||
|
var pod *v1.Pod
|
||||||
|
var expAllowedCPUsListRegex string
|
||||||
|
|
||||||
|
ctnAttrs = []ctnAttribute{
|
||||||
|
{
|
||||||
|
ctnName: "non-gu-container",
|
||||||
|
cpuRequest: "100m",
|
||||||
|
cpuLimit: "200m",
|
||||||
|
},
|
||||||
|
}
|
||||||
|
pod = makeCPUManagerPod("non-gu-pod", ctnAttrs)
|
||||||
|
pod = f.PodClient().CreateSync(pod)
|
||||||
|
|
||||||
|
ginkgo.By("checking if the expected cpuset was assigned")
|
||||||
|
expAllowedCPUsListRegex = fmt.Sprintf("^0-%d\n$", cpuCap-1)
|
||||||
|
err = f.PodClient().MatchContainerOutput(pod.Name, pod.Spec.Containers[0].Name, expAllowedCPUsListRegex)
|
||||||
|
framework.ExpectNoError(err, "expected log not found in container [%s] of pod [%s]",
|
||||||
|
pod.Spec.Containers[0].Name, pod.Name)
|
||||||
|
|
||||||
|
ginkgo.By("by deleting the pods and waiting for container removal")
|
||||||
|
deletePods(f, []string{pod.Name})
|
||||||
|
waitForContainerRemoval(pod.Spec.Containers[0].Name, pod.Name, pod.Namespace)
|
||||||
|
}
|
||||||
|
|
||||||
|
func runMultipleGuNonGuPods(f *framework.Framework, cpuCap int64, cpuAlloc int64) {
|
||||||
|
|
||||||
var cpuListString, expAllowedCPUsListRegex string
|
var cpuListString, expAllowedCPUsListRegex string
|
||||||
var cpuList []int
|
var cpuList []int
|
||||||
var cpu1, cpu2 int
|
var cpu1 int
|
||||||
var cset cpuset.CPUSet
|
var cset cpuset.CPUSet
|
||||||
var err error
|
var err error
|
||||||
var ctnAttrs []ctnAttribute
|
var ctnAttrs []ctnAttribute
|
||||||
var pod, pod1, pod2 *v1.Pod
|
var pod1, pod2 *v1.Pod
|
||||||
|
|
||||||
|
ctnAttrs = []ctnAttribute{
|
||||||
|
{
|
||||||
|
ctnName: "gu-container",
|
||||||
|
cpuRequest: "1000m",
|
||||||
|
cpuLimit: "1000m",
|
||||||
|
},
|
||||||
|
}
|
||||||
|
pod1 = makeCPUManagerPod("gu-pod", ctnAttrs)
|
||||||
|
pod1 = f.PodClient().CreateSync(pod1)
|
||||||
|
|
||||||
|
ctnAttrs = []ctnAttribute{
|
||||||
|
{
|
||||||
|
ctnName: "non-gu-container",
|
||||||
|
cpuRequest: "200m",
|
||||||
|
cpuLimit: "300m",
|
||||||
|
},
|
||||||
|
}
|
||||||
|
pod2 = makeCPUManagerPod("non-gu-pod", ctnAttrs)
|
||||||
|
pod2 = f.PodClient().CreateSync(pod2)
|
||||||
|
|
||||||
|
ginkgo.By("checking if the expected cpuset was assigned")
|
||||||
|
cpu1 = 1
|
||||||
|
if isHTEnabled() {
|
||||||
|
cpuList = cpuset.MustParse(getCPUSiblingList(0)).ToSlice()
|
||||||
|
cpu1 = cpuList[1]
|
||||||
|
} else if isMultiNUMA() {
|
||||||
|
cpuList = cpuset.MustParse(getCoreSiblingList(0)).ToSlice()
|
||||||
|
cpu1 = cpuList[1]
|
||||||
|
}
|
||||||
|
expAllowedCPUsListRegex = fmt.Sprintf("^%d\n$", cpu1)
|
||||||
|
err = f.PodClient().MatchContainerOutput(pod1.Name, pod1.Spec.Containers[0].Name, expAllowedCPUsListRegex)
|
||||||
|
framework.ExpectNoError(err, "expected log not found in container [%s] of pod [%s]",
|
||||||
|
pod1.Spec.Containers[0].Name, pod1.Name)
|
||||||
|
|
||||||
|
cpuListString = "0"
|
||||||
|
if cpuAlloc > 2 {
|
||||||
|
cset = cpuset.MustParse(fmt.Sprintf("0-%d", cpuCap-1))
|
||||||
|
cpuListString = fmt.Sprintf("%s", cset.Difference(cpuset.NewCPUSet(cpu1)))
|
||||||
|
}
|
||||||
|
expAllowedCPUsListRegex = fmt.Sprintf("^%s\n$", cpuListString)
|
||||||
|
err = f.PodClient().MatchContainerOutput(pod2.Name, pod2.Spec.Containers[0].Name, expAllowedCPUsListRegex)
|
||||||
|
framework.ExpectNoError(err, "expected log not found in container [%s] of pod [%s]",
|
||||||
|
pod2.Spec.Containers[0].Name, pod2.Name)
|
||||||
|
ginkgo.By("by deleting the pods and waiting for container removal")
|
||||||
|
deletePods(f, []string{pod1.Name, pod2.Name})
|
||||||
|
waitForContainerRemoval(pod1.Spec.Containers[0].Name, pod1.Name, pod1.Namespace)
|
||||||
|
waitForContainerRemoval(pod2.Spec.Containers[0].Name, pod2.Name, pod2.Namespace)
|
||||||
|
}
|
||||||
|
|
||||||
|
func runMultipleCPUGuPod(f *framework.Framework) {
|
||||||
|
var cpuListString, expAllowedCPUsListRegex string
|
||||||
|
var cpuList []int
|
||||||
|
var cset cpuset.CPUSet
|
||||||
|
var err error
|
||||||
|
var ctnAttrs []ctnAttribute
|
||||||
|
var pod *v1.Pod
|
||||||
|
|
||||||
|
ctnAttrs = []ctnAttribute{
|
||||||
|
{
|
||||||
|
ctnName: "gu-container",
|
||||||
|
cpuRequest: "2000m",
|
||||||
|
cpuLimit: "2000m",
|
||||||
|
},
|
||||||
|
}
|
||||||
|
pod = makeCPUManagerPod("gu-pod", ctnAttrs)
|
||||||
|
pod = f.PodClient().CreateSync(pod)
|
||||||
|
|
||||||
|
ginkgo.By("checking if the expected cpuset was assigned")
|
||||||
|
cpuListString = "1-2"
|
||||||
|
if isMultiNUMA() {
|
||||||
|
cpuList = cpuset.MustParse(getCoreSiblingList(0)).ToSlice()
|
||||||
|
if !isHTEnabled() {
|
||||||
|
cset = cpuset.MustParse(fmt.Sprintf("%d-%d", cpuList[1], cpuList[2]))
|
||||||
|
} else {
|
||||||
|
cset = cpuset.MustParse(getCPUSiblingList(int64(cpuList[1])))
|
||||||
|
}
|
||||||
|
cpuListString = fmt.Sprintf("%s", cset)
|
||||||
|
} else if isHTEnabled() {
|
||||||
|
cpuListString = "2-3"
|
||||||
|
cpuList = cpuset.MustParse(getCPUSiblingList(0)).ToSlice()
|
||||||
|
if cpuList[1] != 1 {
|
||||||
|
cset = cpuset.MustParse(getCPUSiblingList(1))
|
||||||
|
cpuListString = fmt.Sprintf("%s", cset)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
expAllowedCPUsListRegex = fmt.Sprintf("^%s\n$", cpuListString)
|
||||||
|
err = f.PodClient().MatchContainerOutput(pod.Name, pod.Spec.Containers[0].Name, expAllowedCPUsListRegex)
|
||||||
|
framework.ExpectNoError(err, "expected log not found in container [%s] of pod [%s]",
|
||||||
|
pod.Spec.Containers[0].Name, pod.Name)
|
||||||
|
|
||||||
|
ginkgo.By("by deleting the pods and waiting for container removal")
|
||||||
|
deletePods(f, []string{pod.Name})
|
||||||
|
waitForContainerRemoval(pod.Spec.Containers[0].Name, pod.Name, pod.Namespace)
|
||||||
|
}
|
||||||
|
|
||||||
|
func runMultipleCPUContainersGuPod(f *framework.Framework) {
|
||||||
|
|
||||||
|
var expAllowedCPUsListRegex string
|
||||||
|
var cpuList []int
|
||||||
|
var cpu1, cpu2 int
|
||||||
|
var err error
|
||||||
|
var ctnAttrs []ctnAttribute
|
||||||
|
var pod *v1.Pod
|
||||||
|
ctnAttrs = []ctnAttribute{
|
||||||
|
{
|
||||||
|
ctnName: "gu-container1",
|
||||||
|
cpuRequest: "1000m",
|
||||||
|
cpuLimit: "1000m",
|
||||||
|
},
|
||||||
|
{
|
||||||
|
ctnName: "gu-container2",
|
||||||
|
cpuRequest: "1000m",
|
||||||
|
cpuLimit: "1000m",
|
||||||
|
},
|
||||||
|
}
|
||||||
|
pod = makeCPUManagerPod("gu-pod", ctnAttrs)
|
||||||
|
pod = f.PodClient().CreateSync(pod)
|
||||||
|
|
||||||
|
ginkgo.By("checking if the expected cpuset was assigned")
|
||||||
|
cpu1, cpu2 = 1, 2
|
||||||
|
if isHTEnabled() {
|
||||||
|
cpuList = cpuset.MustParse(getCPUSiblingList(0)).ToSlice()
|
||||||
|
if cpuList[1] != 1 {
|
||||||
|
cpu1, cpu2 = cpuList[1], 1
|
||||||
|
}
|
||||||
|
if isMultiNUMA() {
|
||||||
|
cpuList = cpuset.MustParse(getCoreSiblingList(0)).ToSlice()
|
||||||
|
cpu2 = cpuList[1]
|
||||||
|
}
|
||||||
|
} else if isMultiNUMA() {
|
||||||
|
cpuList = cpuset.MustParse(getCoreSiblingList(0)).ToSlice()
|
||||||
|
cpu1, cpu2 = cpuList[1], cpuList[2]
|
||||||
|
}
|
||||||
|
expAllowedCPUsListRegex = fmt.Sprintf("^%d|%d\n$", cpu1, cpu2)
|
||||||
|
err = f.PodClient().MatchContainerOutput(pod.Name, pod.Spec.Containers[0].Name, expAllowedCPUsListRegex)
|
||||||
|
framework.ExpectNoError(err, "expected log not found in container [%s] of pod [%s]",
|
||||||
|
pod.Spec.Containers[0].Name, pod.Name)
|
||||||
|
|
||||||
|
err = f.PodClient().MatchContainerOutput(pod.Name, pod.Spec.Containers[1].Name, expAllowedCPUsListRegex)
|
||||||
|
framework.ExpectNoError(err, "expected log not found in container [%s] of pod [%s]",
|
||||||
|
pod.Spec.Containers[1].Name, pod.Name)
|
||||||
|
|
||||||
|
ginkgo.By("by deleting the pods and waiting for container removal")
|
||||||
|
deletePods(f, []string{pod.Name})
|
||||||
|
waitForContainerRemoval(pod.Spec.Containers[0].Name, pod.Name, pod.Namespace)
|
||||||
|
waitForContainerRemoval(pod.Spec.Containers[1].Name, pod.Name, pod.Namespace)
|
||||||
|
}
|
||||||
|
|
||||||
|
func runMultipleGuPods(f *framework.Framework) {
|
||||||
|
var expAllowedCPUsListRegex string
|
||||||
|
var cpuList []int
|
||||||
|
var cpu1, cpu2 int
|
||||||
|
var err error
|
||||||
|
var ctnAttrs []ctnAttribute
|
||||||
|
var pod1, pod2 *v1.Pod
|
||||||
|
|
||||||
|
ctnAttrs = []ctnAttribute{
|
||||||
|
{
|
||||||
|
ctnName: "gu-container1",
|
||||||
|
cpuRequest: "1000m",
|
||||||
|
cpuLimit: "1000m",
|
||||||
|
},
|
||||||
|
}
|
||||||
|
pod1 = makeCPUManagerPod("gu-pod1", ctnAttrs)
|
||||||
|
pod1 = f.PodClient().CreateSync(pod1)
|
||||||
|
|
||||||
|
ctnAttrs = []ctnAttribute{
|
||||||
|
{
|
||||||
|
ctnName: "gu-container2",
|
||||||
|
cpuRequest: "1000m",
|
||||||
|
cpuLimit: "1000m",
|
||||||
|
},
|
||||||
|
}
|
||||||
|
pod2 = makeCPUManagerPod("gu-pod2", ctnAttrs)
|
||||||
|
pod2 = f.PodClient().CreateSync(pod2)
|
||||||
|
|
||||||
|
ginkgo.By("checking if the expected cpuset was assigned")
|
||||||
|
cpu1, cpu2 = 1, 2
|
||||||
|
if isHTEnabled() {
|
||||||
|
cpuList = cpuset.MustParse(getCPUSiblingList(0)).ToSlice()
|
||||||
|
if cpuList[1] != 1 {
|
||||||
|
cpu1, cpu2 = cpuList[1], 1
|
||||||
|
}
|
||||||
|
if isMultiNUMA() {
|
||||||
|
cpuList = cpuset.MustParse(getCoreSiblingList(0)).ToSlice()
|
||||||
|
cpu2 = cpuList[1]
|
||||||
|
}
|
||||||
|
} else if isMultiNUMA() {
|
||||||
|
cpuList = cpuset.MustParse(getCoreSiblingList(0)).ToSlice()
|
||||||
|
cpu1, cpu2 = cpuList[1], cpuList[2]
|
||||||
|
}
|
||||||
|
expAllowedCPUsListRegex = fmt.Sprintf("^%d\n$", cpu1)
|
||||||
|
err = f.PodClient().MatchContainerOutput(pod1.Name, pod1.Spec.Containers[0].Name, expAllowedCPUsListRegex)
|
||||||
|
framework.ExpectNoError(err, "expected log not found in container [%s] of pod [%s]",
|
||||||
|
pod1.Spec.Containers[0].Name, pod1.Name)
|
||||||
|
|
||||||
|
expAllowedCPUsListRegex = fmt.Sprintf("^%d\n$", cpu2)
|
||||||
|
err = f.PodClient().MatchContainerOutput(pod2.Name, pod2.Spec.Containers[0].Name, expAllowedCPUsListRegex)
|
||||||
|
framework.ExpectNoError(err, "expected log not found in container [%s] of pod [%s]",
|
||||||
|
pod2.Spec.Containers[0].Name, pod2.Name)
|
||||||
|
ginkgo.By("by deleting the pods and waiting for container removal")
|
||||||
|
deletePods(f, []string{pod1.Name, pod2.Name})
|
||||||
|
waitForContainerRemoval(pod1.Spec.Containers[0].Name, pod1.Name, pod1.Namespace)
|
||||||
|
waitForContainerRemoval(pod2.Spec.Containers[0].Name, pod2.Name, pod2.Namespace)
|
||||||
|
}
|
||||||
|
|
||||||
|
func runCPUManagerTests(f *framework.Framework) {
|
||||||
|
var cpuCap, cpuAlloc int64
|
||||||
|
var oldCfg *kubeletconfig.KubeletConfiguration
|
||||||
|
var expAllowedCPUsListRegex string
|
||||||
|
var cpuList []int
|
||||||
|
var cpu1 int
|
||||||
|
var err error
|
||||||
|
var ctnAttrs []ctnAttribute
|
||||||
|
var pod *v1.Pod
|
||||||
|
|
||||||
ginkgo.It("should assign CPUs as expected based on the Pod spec", func() {
|
ginkgo.It("should assign CPUs as expected based on the Pod spec", func() {
|
||||||
cpuCap, cpuAlloc, _ = getLocalNodeCPUDetails(f)
|
cpuCap, cpuAlloc, _ = getLocalNodeCPUDetails(f)
|
||||||
@ -281,104 +560,13 @@ func runCPUManagerTests(f *framework.Framework) {
|
|||||||
oldCfg = enableCPUManagerInKubelet(f, true)
|
oldCfg = enableCPUManagerInKubelet(f, true)
|
||||||
|
|
||||||
ginkgo.By("running a non-Gu pod")
|
ginkgo.By("running a non-Gu pod")
|
||||||
ctnAttrs = []ctnAttribute{
|
runNonGuPodTest(f, cpuCap)
|
||||||
{
|
|
||||||
ctnName: "non-gu-container",
|
|
||||||
cpuRequest: "100m",
|
|
||||||
cpuLimit: "200m",
|
|
||||||
},
|
|
||||||
}
|
|
||||||
pod = makeCPUManagerPod("non-gu-pod", ctnAttrs)
|
|
||||||
pod = f.PodClient().CreateSync(pod)
|
|
||||||
|
|
||||||
ginkgo.By("checking if the expected cpuset was assigned")
|
|
||||||
expAllowedCPUsListRegex = fmt.Sprintf("^0-%d\n$", cpuCap-1)
|
|
||||||
err = f.PodClient().MatchContainerOutput(pod.Name, pod.Spec.Containers[0].Name, expAllowedCPUsListRegex)
|
|
||||||
framework.ExpectNoError(err, "expected log not found in container [%s] of pod [%s]",
|
|
||||||
pod.Spec.Containers[0].Name, pod.Name)
|
|
||||||
|
|
||||||
ginkgo.By("by deleting the pods and waiting for container removal")
|
|
||||||
deletePods(f, []string{pod.Name})
|
|
||||||
waitForContainerRemoval(pod.Spec.Containers[0].Name, pod.Name, pod.Namespace)
|
|
||||||
|
|
||||||
ginkgo.By("running a Gu pod")
|
ginkgo.By("running a Gu pod")
|
||||||
ctnAttrs = []ctnAttribute{
|
runGuPodTest(f)
|
||||||
{
|
|
||||||
ctnName: "gu-container",
|
|
||||||
cpuRequest: "1000m",
|
|
||||||
cpuLimit: "1000m",
|
|
||||||
},
|
|
||||||
}
|
|
||||||
pod = makeCPUManagerPod("gu-pod", ctnAttrs)
|
|
||||||
pod = f.PodClient().CreateSync(pod)
|
|
||||||
|
|
||||||
ginkgo.By("checking if the expected cpuset was assigned")
|
|
||||||
cpu1 = 1
|
|
||||||
if isHTEnabled() {
|
|
||||||
cpuList = cpuset.MustParse(getCPUSiblingList(0)).ToSlice()
|
|
||||||
cpu1 = cpuList[1]
|
|
||||||
} else if isMultiNUMA() {
|
|
||||||
cpuList = cpuset.MustParse(getCoreSiblingList(0)).ToSlice()
|
|
||||||
cpu1 = cpuList[1]
|
|
||||||
}
|
|
||||||
expAllowedCPUsListRegex = fmt.Sprintf("^%d\n$", cpu1)
|
|
||||||
err = f.PodClient().MatchContainerOutput(pod.Name, pod.Spec.Containers[0].Name, expAllowedCPUsListRegex)
|
|
||||||
framework.ExpectNoError(err, "expected log not found in container [%s] of pod [%s]",
|
|
||||||
pod.Spec.Containers[0].Name, pod.Name)
|
|
||||||
|
|
||||||
ginkgo.By("by deleting the pods and waiting for container removal")
|
|
||||||
deletePods(f, []string{pod.Name})
|
|
||||||
waitForContainerRemoval(pod.Spec.Containers[0].Name, pod.Name, pod.Namespace)
|
|
||||||
|
|
||||||
ginkgo.By("running multiple Gu and non-Gu pods")
|
ginkgo.By("running multiple Gu and non-Gu pods")
|
||||||
ctnAttrs = []ctnAttribute{
|
runMultipleGuNonGuPods(f, cpuCap, cpuAlloc)
|
||||||
{
|
|
||||||
ctnName: "gu-container",
|
|
||||||
cpuRequest: "1000m",
|
|
||||||
cpuLimit: "1000m",
|
|
||||||
},
|
|
||||||
}
|
|
||||||
pod1 = makeCPUManagerPod("gu-pod", ctnAttrs)
|
|
||||||
pod1 = f.PodClient().CreateSync(pod1)
|
|
||||||
|
|
||||||
ctnAttrs = []ctnAttribute{
|
|
||||||
{
|
|
||||||
ctnName: "non-gu-container",
|
|
||||||
cpuRequest: "200m",
|
|
||||||
cpuLimit: "300m",
|
|
||||||
},
|
|
||||||
}
|
|
||||||
pod2 = makeCPUManagerPod("non-gu-pod", ctnAttrs)
|
|
||||||
pod2 = f.PodClient().CreateSync(pod2)
|
|
||||||
|
|
||||||
ginkgo.By("checking if the expected cpuset was assigned")
|
|
||||||
cpu1 = 1
|
|
||||||
if isHTEnabled() {
|
|
||||||
cpuList = cpuset.MustParse(getCPUSiblingList(0)).ToSlice()
|
|
||||||
cpu1 = cpuList[1]
|
|
||||||
} else if isMultiNUMA() {
|
|
||||||
cpuList = cpuset.MustParse(getCoreSiblingList(0)).ToSlice()
|
|
||||||
cpu1 = cpuList[1]
|
|
||||||
}
|
|
||||||
expAllowedCPUsListRegex = fmt.Sprintf("^%d\n$", cpu1)
|
|
||||||
err = f.PodClient().MatchContainerOutput(pod1.Name, pod1.Spec.Containers[0].Name, expAllowedCPUsListRegex)
|
|
||||||
framework.ExpectNoError(err, "expected log not found in container [%s] of pod [%s]",
|
|
||||||
pod1.Spec.Containers[0].Name, pod1.Name)
|
|
||||||
|
|
||||||
cpuListString = "0"
|
|
||||||
if cpuAlloc > 2 {
|
|
||||||
cset = cpuset.MustParse(fmt.Sprintf("0-%d", cpuCap-1))
|
|
||||||
cpuListString = fmt.Sprintf("%s", cset.Difference(cpuset.NewCPUSet(cpu1)))
|
|
||||||
}
|
|
||||||
expAllowedCPUsListRegex = fmt.Sprintf("^%s\n$", cpuListString)
|
|
||||||
err = f.PodClient().MatchContainerOutput(pod2.Name, pod2.Spec.Containers[0].Name, expAllowedCPUsListRegex)
|
|
||||||
framework.ExpectNoError(err, "expected log not found in container [%s] of pod [%s]",
|
|
||||||
pod2.Spec.Containers[0].Name, pod2.Name)
|
|
||||||
|
|
||||||
ginkgo.By("by deleting the pods and waiting for container removal")
|
|
||||||
deletePods(f, []string{pod1.Name, pod2.Name})
|
|
||||||
waitForContainerRemoval(pod1.Spec.Containers[0].Name, pod1.Name, pod1.Namespace)
|
|
||||||
waitForContainerRemoval(pod2.Spec.Containers[0].Name, pod2.Name, pod2.Namespace)
|
|
||||||
|
|
||||||
// Skip rest of the tests if CPU capacity < 3.
|
// Skip rest of the tests if CPU capacity < 3.
|
||||||
if cpuCap < 3 {
|
if cpuCap < 3 {
|
||||||
@ -386,138 +574,13 @@ func runCPUManagerTests(f *framework.Framework) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
ginkgo.By("running a Gu pod requesting multiple CPUs")
|
ginkgo.By("running a Gu pod requesting multiple CPUs")
|
||||||
ctnAttrs = []ctnAttribute{
|
runMultipleCPUGuPod(f)
|
||||||
{
|
|
||||||
ctnName: "gu-container",
|
|
||||||
cpuRequest: "2000m",
|
|
||||||
cpuLimit: "2000m",
|
|
||||||
},
|
|
||||||
}
|
|
||||||
pod = makeCPUManagerPod("gu-pod", ctnAttrs)
|
|
||||||
pod = f.PodClient().CreateSync(pod)
|
|
||||||
|
|
||||||
ginkgo.By("checking if the expected cpuset was assigned")
|
|
||||||
cpuListString = "1-2"
|
|
||||||
if isMultiNUMA() {
|
|
||||||
cpuList = cpuset.MustParse(getCoreSiblingList(0)).ToSlice()
|
|
||||||
if !isHTEnabled() {
|
|
||||||
cset = cpuset.MustParse(fmt.Sprintf("%d,%d", cpuList[1], cpuList[2]))
|
|
||||||
} else {
|
|
||||||
cset = cpuset.MustParse(getCPUSiblingList(int64(cpuList[1])))
|
|
||||||
}
|
|
||||||
cpuListString = fmt.Sprintf("%s", cset)
|
|
||||||
} else if isHTEnabled() {
|
|
||||||
cpuListString = "2-3"
|
|
||||||
cpuList = cpuset.MustParse(getCPUSiblingList(0)).ToSlice()
|
|
||||||
if cpuList[1] != 1 {
|
|
||||||
cset = cpuset.MustParse(getCPUSiblingList(1))
|
|
||||||
cpuListString = fmt.Sprintf("%s", cset)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
expAllowedCPUsListRegex = fmt.Sprintf("^%s\n$", cpuListString)
|
|
||||||
err = f.PodClient().MatchContainerOutput(pod.Name, pod.Spec.Containers[0].Name, expAllowedCPUsListRegex)
|
|
||||||
framework.ExpectNoError(err, "expected log not found in container [%s] of pod [%s]",
|
|
||||||
pod.Spec.Containers[0].Name, pod.Name)
|
|
||||||
|
|
||||||
ginkgo.By("by deleting the pods and waiting for container removal")
|
|
||||||
deletePods(f, []string{pod.Name})
|
|
||||||
waitForContainerRemoval(pod.Spec.Containers[0].Name, pod.Name, pod.Namespace)
|
|
||||||
|
|
||||||
ginkgo.By("running a Gu pod with multiple containers requesting integer CPUs")
|
ginkgo.By("running a Gu pod with multiple containers requesting integer CPUs")
|
||||||
ctnAttrs = []ctnAttribute{
|
runMultipleCPUContainersGuPod(f)
|
||||||
{
|
|
||||||
ctnName: "gu-container1",
|
|
||||||
cpuRequest: "1000m",
|
|
||||||
cpuLimit: "1000m",
|
|
||||||
},
|
|
||||||
{
|
|
||||||
ctnName: "gu-container2",
|
|
||||||
cpuRequest: "1000m",
|
|
||||||
cpuLimit: "1000m",
|
|
||||||
},
|
|
||||||
}
|
|
||||||
pod = makeCPUManagerPod("gu-pod", ctnAttrs)
|
|
||||||
pod = f.PodClient().CreateSync(pod)
|
|
||||||
|
|
||||||
ginkgo.By("checking if the expected cpuset was assigned")
|
|
||||||
cpu1, cpu2 = 1, 2
|
|
||||||
if isHTEnabled() {
|
|
||||||
cpuList = cpuset.MustParse(getCPUSiblingList(0)).ToSlice()
|
|
||||||
if cpuList[1] != 1 {
|
|
||||||
cpu1, cpu2 = cpuList[1], 1
|
|
||||||
}
|
|
||||||
if isMultiNUMA() {
|
|
||||||
cpuList = cpuset.MustParse(getCoreSiblingList(0)).ToSlice()
|
|
||||||
cpu2 = cpuList[1]
|
|
||||||
}
|
|
||||||
} else if isMultiNUMA() {
|
|
||||||
cpuList = cpuset.MustParse(getCoreSiblingList(0)).ToSlice()
|
|
||||||
cpu1, cpu2 = cpuList[1], cpuList[2]
|
|
||||||
}
|
|
||||||
expAllowedCPUsListRegex = fmt.Sprintf("^%d|%d\n$", cpu1, cpu2)
|
|
||||||
err = f.PodClient().MatchContainerOutput(pod.Name, pod.Spec.Containers[0].Name, expAllowedCPUsListRegex)
|
|
||||||
framework.ExpectNoError(err, "expected log not found in container [%s] of pod [%s]",
|
|
||||||
pod.Spec.Containers[0].Name, pod.Name)
|
|
||||||
|
|
||||||
err = f.PodClient().MatchContainerOutput(pod.Name, pod.Spec.Containers[1].Name, expAllowedCPUsListRegex)
|
|
||||||
framework.ExpectNoError(err, "expected log not found in container [%s] of pod [%s]",
|
|
||||||
pod.Spec.Containers[1].Name, pod.Name)
|
|
||||||
|
|
||||||
ginkgo.By("by deleting the pods and waiting for container removal")
|
|
||||||
deletePods(f, []string{pod.Name})
|
|
||||||
waitForContainerRemoval(pod.Spec.Containers[0].Name, pod.Name, pod.Namespace)
|
|
||||||
waitForContainerRemoval(pod.Spec.Containers[1].Name, pod.Name, pod.Namespace)
|
|
||||||
|
|
||||||
ginkgo.By("running multiple Gu pods")
|
ginkgo.By("running multiple Gu pods")
|
||||||
ctnAttrs = []ctnAttribute{
|
runMultipleGuPods(f)
|
||||||
{
|
|
||||||
ctnName: "gu-container1",
|
|
||||||
cpuRequest: "1000m",
|
|
||||||
cpuLimit: "1000m",
|
|
||||||
},
|
|
||||||
}
|
|
||||||
pod1 = makeCPUManagerPod("gu-pod1", ctnAttrs)
|
|
||||||
pod1 = f.PodClient().CreateSync(pod1)
|
|
||||||
|
|
||||||
ctnAttrs = []ctnAttribute{
|
|
||||||
{
|
|
||||||
ctnName: "gu-container2",
|
|
||||||
cpuRequest: "1000m",
|
|
||||||
cpuLimit: "1000m",
|
|
||||||
},
|
|
||||||
}
|
|
||||||
pod2 = makeCPUManagerPod("gu-pod2", ctnAttrs)
|
|
||||||
pod2 = f.PodClient().CreateSync(pod2)
|
|
||||||
|
|
||||||
ginkgo.By("checking if the expected cpuset was assigned")
|
|
||||||
cpu1, cpu2 = 1, 2
|
|
||||||
if isHTEnabled() {
|
|
||||||
cpuList = cpuset.MustParse(getCPUSiblingList(0)).ToSlice()
|
|
||||||
if cpuList[1] != 1 {
|
|
||||||
cpu1, cpu2 = cpuList[1], 1
|
|
||||||
}
|
|
||||||
if isMultiNUMA() {
|
|
||||||
cpuList = cpuset.MustParse(getCoreSiblingList(0)).ToSlice()
|
|
||||||
cpu2 = cpuList[1]
|
|
||||||
}
|
|
||||||
} else if isMultiNUMA() {
|
|
||||||
cpuList = cpuset.MustParse(getCoreSiblingList(0)).ToSlice()
|
|
||||||
cpu1, cpu2 = cpuList[1], cpuList[2]
|
|
||||||
}
|
|
||||||
expAllowedCPUsListRegex = fmt.Sprintf("^%d\n$", cpu1)
|
|
||||||
err = f.PodClient().MatchContainerOutput(pod1.Name, pod1.Spec.Containers[0].Name, expAllowedCPUsListRegex)
|
|
||||||
framework.ExpectNoError(err, "expected log not found in container [%s] of pod [%s]",
|
|
||||||
pod1.Spec.Containers[0].Name, pod1.Name)
|
|
||||||
|
|
||||||
expAllowedCPUsListRegex = fmt.Sprintf("^%d\n$", cpu2)
|
|
||||||
err = f.PodClient().MatchContainerOutput(pod2.Name, pod2.Spec.Containers[0].Name, expAllowedCPUsListRegex)
|
|
||||||
framework.ExpectNoError(err, "expected log not found in container [%s] of pod [%s]",
|
|
||||||
pod2.Spec.Containers[0].Name, pod2.Name)
|
|
||||||
|
|
||||||
ginkgo.By("by deleting the pods and waiting for container removal")
|
|
||||||
deletePods(f, []string{pod1.Name, pod2.Name})
|
|
||||||
waitForContainerRemoval(pod1.Spec.Containers[0].Name, pod1.Name, pod1.Namespace)
|
|
||||||
waitForContainerRemoval(pod2.Spec.Containers[0].Name, pod2.Name, pod2.Namespace)
|
|
||||||
|
|
||||||
ginkgo.By("test for automatically remove inactive pods from cpumanager state file.")
|
ginkgo.By("test for automatically remove inactive pods from cpumanager state file.")
|
||||||
// First running a Gu Pod,
|
// First running a Gu Pod,
|
||||||
|
Loading…
Reference in New Issue
Block a user