mirror of
https://github.com/k3s-io/kubernetes.git
synced 2025-07-25 20:53:33 +00:00
Merge pull request #115926 from ffromani/e2e-node-remove-kubevirt-device-plugin
e2e: node remove: kubevirt device plugin
This commit is contained in:
commit
3702411ef9
@ -31,7 +31,6 @@ import (
|
|||||||
"k8s.io/apimachinery/pkg/runtime"
|
"k8s.io/apimachinery/pkg/runtime"
|
||||||
"k8s.io/apimachinery/pkg/runtime/serializer"
|
"k8s.io/apimachinery/pkg/runtime/serializer"
|
||||||
kubeletdevicepluginv1beta1 "k8s.io/kubelet/pkg/apis/deviceplugin/v1beta1"
|
kubeletdevicepluginv1beta1 "k8s.io/kubelet/pkg/apis/deviceplugin/v1beta1"
|
||||||
e2etestfiles "k8s.io/kubernetes/test/e2e/framework/testfiles"
|
|
||||||
admissionapi "k8s.io/pod-security-admission/api"
|
admissionapi "k8s.io/pod-security-admission/api"
|
||||||
|
|
||||||
"k8s.io/apimachinery/pkg/api/resource"
|
"k8s.io/apimachinery/pkg/api/resource"
|
||||||
@ -44,17 +43,6 @@ import (
|
|||||||
e2epod "k8s.io/kubernetes/test/e2e/framework/pod"
|
e2epod "k8s.io/kubernetes/test/e2e/framework/pod"
|
||||||
)
|
)
|
||||||
|
|
||||||
const (
|
|
||||||
// sampleResourceName is the name of the example resource which is used in the e2e test
|
|
||||||
sampleResourceName = "example.com/resource"
|
|
||||||
// sampleDevicePluginName is the name of the device plugin pod
|
|
||||||
sampleDevicePluginName = "sample-device-plugin"
|
|
||||||
|
|
||||||
// fake resource name
|
|
||||||
resourceName = "example.com/resource"
|
|
||||||
envVarNamePluginSockDir = "PLUGIN_SOCK_DIR"
|
|
||||||
)
|
|
||||||
|
|
||||||
var (
|
var (
|
||||||
appsScheme = runtime.NewScheme()
|
appsScheme = runtime.NewScheme()
|
||||||
appsCodecs = serializer.NewCodecFactory(appsScheme)
|
appsCodecs = serializer.NewCodecFactory(appsScheme)
|
||||||
@ -67,17 +55,6 @@ var _ = SIGDescribe("Device Plugin [Feature:DevicePluginProbe][NodeFeature:Devic
|
|||||||
testDevicePlugin(f, kubeletdevicepluginv1beta1.DevicePluginPath)
|
testDevicePlugin(f, kubeletdevicepluginv1beta1.DevicePluginPath)
|
||||||
})
|
})
|
||||||
|
|
||||||
// numberOfSampleResources returns the number of resources advertised by a node.
|
|
||||||
func numberOfSampleResources(node *v1.Node) int64 {
|
|
||||||
val, ok := node.Status.Capacity[sampleResourceName]
|
|
||||||
|
|
||||||
if !ok {
|
|
||||||
return 0
|
|
||||||
}
|
|
||||||
|
|
||||||
return val.Value()
|
|
||||||
}
|
|
||||||
|
|
||||||
// readDaemonSetV1OrDie reads daemonset object from bytes. Panics on error.
|
// readDaemonSetV1OrDie reads daemonset object from bytes. Panics on error.
|
||||||
func readDaemonSetV1OrDie(objBytes []byte) *appsv1.DaemonSet {
|
func readDaemonSetV1OrDie(objBytes []byte) *appsv1.DaemonSet {
|
||||||
appsv1.AddToScheme(appsScheme)
|
appsv1.AddToScheme(appsScheme)
|
||||||
@ -133,31 +110,14 @@ func testDevicePlugin(f *framework.Framework, pluginSockDir string) {
|
|||||||
}, f.Timeouts.PodDelete, f.Timeouts.Poll).Should(gomega.Succeed())
|
}, f.Timeouts.PodDelete, f.Timeouts.Poll).Should(gomega.Succeed())
|
||||||
|
|
||||||
ginkgo.By("Scheduling a sample device plugin pod")
|
ginkgo.By("Scheduling a sample device plugin pod")
|
||||||
data, err := e2etestfiles.Read(SampleDevicePluginDSYAML)
|
dp := getSampleDevicePluginPod(pluginSockDir)
|
||||||
if err != nil {
|
|
||||||
framework.Fail(err.Error())
|
|
||||||
}
|
|
||||||
ds := readDaemonSetV1OrDie(data)
|
|
||||||
|
|
||||||
dp := &v1.Pod{
|
|
||||||
ObjectMeta: metav1.ObjectMeta{
|
|
||||||
Name: sampleDevicePluginName,
|
|
||||||
},
|
|
||||||
Spec: ds.Spec.Template.Spec,
|
|
||||||
}
|
|
||||||
|
|
||||||
for i := range dp.Spec.Containers[0].Env {
|
|
||||||
if dp.Spec.Containers[0].Env[i].Name == envVarNamePluginSockDir {
|
|
||||||
dp.Spec.Containers[0].Env[i].Value = pluginSockDir
|
|
||||||
}
|
|
||||||
}
|
|
||||||
dptemplate = dp.DeepCopy()
|
dptemplate = dp.DeepCopy()
|
||||||
devicePluginPod = e2epod.NewPodClient(f).CreateSync(ctx, dp)
|
devicePluginPod = e2epod.NewPodClient(f).CreateSync(ctx, dp)
|
||||||
|
|
||||||
ginkgo.By("Waiting for devices to become available on the local node")
|
ginkgo.By("Waiting for devices to become available on the local node")
|
||||||
gomega.Eventually(ctx, func(ctx context.Context) bool {
|
gomega.Eventually(ctx, func(ctx context.Context) bool {
|
||||||
node, ready := getLocalTestNode(ctx, f)
|
node, ready := getLocalTestNode(ctx, f)
|
||||||
return ready && numberOfSampleResources(node) > 0
|
return ready && CountSampleDeviceCapacity(node) > 0
|
||||||
}, 5*time.Minute, framework.Poll).Should(gomega.BeTrue())
|
}, 5*time.Minute, framework.Poll).Should(gomega.BeTrue())
|
||||||
framework.Logf("Successfully created device plugin pod")
|
framework.Logf("Successfully created device plugin pod")
|
||||||
|
|
||||||
@ -165,8 +125,8 @@ func testDevicePlugin(f *framework.Framework, pluginSockDir string) {
|
|||||||
gomega.Eventually(ctx, func(ctx context.Context) bool {
|
gomega.Eventually(ctx, func(ctx context.Context) bool {
|
||||||
node, ready := getLocalTestNode(ctx, f)
|
node, ready := getLocalTestNode(ctx, f)
|
||||||
return ready &&
|
return ready &&
|
||||||
numberOfDevicesCapacity(node, resourceName) == devsLen &&
|
CountSampleDeviceCapacity(node) == devsLen &&
|
||||||
numberOfDevicesAllocatable(node, resourceName) == devsLen
|
CountSampleDeviceAllocatable(node) == devsLen
|
||||||
}, 30*time.Second, framework.Poll).Should(gomega.BeTrue())
|
}, 30*time.Second, framework.Poll).Should(gomega.BeTrue())
|
||||||
})
|
})
|
||||||
|
|
||||||
@ -191,7 +151,7 @@ func testDevicePlugin(f *framework.Framework, pluginSockDir string) {
|
|||||||
ginkgo.By("Waiting for devices to become unavailable on the local node")
|
ginkgo.By("Waiting for devices to become unavailable on the local node")
|
||||||
gomega.Eventually(ctx, func(ctx context.Context) bool {
|
gomega.Eventually(ctx, func(ctx context.Context) bool {
|
||||||
node, ready := getLocalTestNode(ctx, f)
|
node, ready := getLocalTestNode(ctx, f)
|
||||||
return ready && numberOfSampleResources(node) <= 0
|
return ready && CountSampleDeviceCapacity(node) <= 0
|
||||||
}, 5*time.Minute, framework.Poll).Should(gomega.BeTrue())
|
}, 5*time.Minute, framework.Poll).Should(gomega.BeTrue())
|
||||||
|
|
||||||
ginkgo.By("devices now unavailable on the local node")
|
ginkgo.By("devices now unavailable on the local node")
|
||||||
@ -199,7 +159,7 @@ func testDevicePlugin(f *framework.Framework, pluginSockDir string) {
|
|||||||
|
|
||||||
ginkgo.It("Can schedule a pod that requires a device", func(ctx context.Context) {
|
ginkgo.It("Can schedule a pod that requires a device", func(ctx context.Context) {
|
||||||
podRECMD := "devs=$(ls /tmp/ | egrep '^Dev-[0-9]+$') && echo stub devices: $devs && sleep 60"
|
podRECMD := "devs=$(ls /tmp/ | egrep '^Dev-[0-9]+$') && echo stub devices: $devs && sleep 60"
|
||||||
pod1 := e2epod.NewPodClient(f).CreateSync(ctx, makeBusyboxPod(resourceName, podRECMD))
|
pod1 := e2epod.NewPodClient(f).CreateSync(ctx, makeBusyboxPod(SampleDeviceResourceName, podRECMD))
|
||||||
deviceIDRE := "stub devices: (Dev-[0-9]+)"
|
deviceIDRE := "stub devices: (Dev-[0-9]+)"
|
||||||
devID1 := parseLog(ctx, f, pod1.Name, pod1.Name, deviceIDRE)
|
devID1 := parseLog(ctx, f, pod1.Name, pod1.Name, deviceIDRE)
|
||||||
gomega.Expect(devID1).To(gomega.Not(gomega.Equal("")))
|
gomega.Expect(devID1).To(gomega.Not(gomega.Equal("")))
|
||||||
@ -250,8 +210,8 @@ func testDevicePlugin(f *framework.Framework, pluginSockDir string) {
|
|||||||
framework.ExpectEqual(len(v1alphaResourcesForOurPod.Containers[0].Devices), 1)
|
framework.ExpectEqual(len(v1alphaResourcesForOurPod.Containers[0].Devices), 1)
|
||||||
framework.ExpectEqual(len(v1ResourcesForOurPod.Containers[0].Devices), 1)
|
framework.ExpectEqual(len(v1ResourcesForOurPod.Containers[0].Devices), 1)
|
||||||
|
|
||||||
framework.ExpectEqual(v1alphaResourcesForOurPod.Containers[0].Devices[0].ResourceName, resourceName)
|
framework.ExpectEqual(v1alphaResourcesForOurPod.Containers[0].Devices[0].ResourceName, SampleDeviceResourceName)
|
||||||
framework.ExpectEqual(v1ResourcesForOurPod.Containers[0].Devices[0].ResourceName, resourceName)
|
framework.ExpectEqual(v1ResourcesForOurPod.Containers[0].Devices[0].ResourceName, SampleDeviceResourceName)
|
||||||
|
|
||||||
framework.ExpectEqual(len(v1alphaResourcesForOurPod.Containers[0].Devices[0].DeviceIds), 1)
|
framework.ExpectEqual(len(v1alphaResourcesForOurPod.Containers[0].Devices[0].DeviceIds), 1)
|
||||||
framework.ExpectEqual(len(v1ResourcesForOurPod.Containers[0].Devices[0].DeviceIds), 1)
|
framework.ExpectEqual(len(v1ResourcesForOurPod.Containers[0].Devices[0].DeviceIds), 1)
|
||||||
@ -259,7 +219,7 @@ func testDevicePlugin(f *framework.Framework, pluginSockDir string) {
|
|||||||
|
|
||||||
ginkgo.It("Keeps device plugin assignments across pod and kubelet restarts", func(ctx context.Context) {
|
ginkgo.It("Keeps device plugin assignments across pod and kubelet restarts", func(ctx context.Context) {
|
||||||
podRECMD := "devs=$(ls /tmp/ | egrep '^Dev-[0-9]+$') && echo stub devices: $devs && sleep 60"
|
podRECMD := "devs=$(ls /tmp/ | egrep '^Dev-[0-9]+$') && echo stub devices: $devs && sleep 60"
|
||||||
pod1 := e2epod.NewPodClient(f).CreateSync(ctx, makeBusyboxPod(resourceName, podRECMD))
|
pod1 := e2epod.NewPodClient(f).CreateSync(ctx, makeBusyboxPod(SampleDeviceResourceName, podRECMD))
|
||||||
deviceIDRE := "stub devices: (Dev-[0-9]+)"
|
deviceIDRE := "stub devices: (Dev-[0-9]+)"
|
||||||
devID1 := parseLog(ctx, f, pod1.Name, pod1.Name, deviceIDRE)
|
devID1 := parseLog(ctx, f, pod1.Name, pod1.Name, deviceIDRE)
|
||||||
gomega.Expect(devID1).To(gomega.Not(gomega.Equal("")))
|
gomega.Expect(devID1).To(gomega.Not(gomega.Equal("")))
|
||||||
@ -288,7 +248,7 @@ func testDevicePlugin(f *framework.Framework, pluginSockDir string) {
|
|||||||
|
|
||||||
ginkgo.It("Keeps device plugin assignments after the device plugin has been re-registered", func(ctx context.Context) {
|
ginkgo.It("Keeps device plugin assignments after the device plugin has been re-registered", func(ctx context.Context) {
|
||||||
podRECMD := "devs=$(ls /tmp/ | egrep '^Dev-[0-9]+$') && echo stub devices: $devs && sleep 60"
|
podRECMD := "devs=$(ls /tmp/ | egrep '^Dev-[0-9]+$') && echo stub devices: $devs && sleep 60"
|
||||||
pod1 := e2epod.NewPodClient(f).CreateSync(ctx, makeBusyboxPod(resourceName, podRECMD))
|
pod1 := e2epod.NewPodClient(f).CreateSync(ctx, makeBusyboxPod(SampleDeviceResourceName, podRECMD))
|
||||||
deviceIDRE := "stub devices: (Dev-[0-9]+)"
|
deviceIDRE := "stub devices: (Dev-[0-9]+)"
|
||||||
devID1 := parseLog(ctx, f, pod1.Name, pod1.Name, deviceIDRE)
|
devID1 := parseLog(ctx, f, pod1.Name, pod1.Name, deviceIDRE)
|
||||||
gomega.Expect(devID1).To(gomega.Not(gomega.Equal("")))
|
gomega.Expect(devID1).To(gomega.Not(gomega.Equal("")))
|
||||||
@ -322,12 +282,12 @@ func testDevicePlugin(f *framework.Framework, pluginSockDir string) {
|
|||||||
gomega.Eventually(ctx, func() bool {
|
gomega.Eventually(ctx, func() bool {
|
||||||
node, ready := getLocalTestNode(ctx, f)
|
node, ready := getLocalTestNode(ctx, f)
|
||||||
return ready &&
|
return ready &&
|
||||||
numberOfDevicesCapacity(node, resourceName) == devsLen &&
|
CountSampleDeviceCapacity(node) == devsLen &&
|
||||||
numberOfDevicesAllocatable(node, resourceName) == devsLen
|
CountSampleDeviceAllocatable(node) == devsLen
|
||||||
}, 30*time.Second, framework.Poll).Should(gomega.BeTrue())
|
}, 30*time.Second, framework.Poll).Should(gomega.BeTrue())
|
||||||
|
|
||||||
ginkgo.By("Creating another pod")
|
ginkgo.By("Creating another pod")
|
||||||
pod2 := e2epod.NewPodClient(f).CreateSync(ctx, makeBusyboxPod(resourceName, podRECMD))
|
pod2 := e2epod.NewPodClient(f).CreateSync(ctx, makeBusyboxPod(SampleDeviceResourceName, podRECMD))
|
||||||
|
|
||||||
ginkgo.By("Checking that pod got a different fake device")
|
ginkgo.By("Checking that pod got a different fake device")
|
||||||
devID2 := parseLog(ctx, f, pod2.Name, pod2.Name, deviceIDRE)
|
devID2 := parseLog(ctx, f, pod2.Name, pod2.Name, deviceIDRE)
|
||||||
@ -338,10 +298,10 @@ func testDevicePlugin(f *framework.Framework, pluginSockDir string) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// makeBusyboxPod returns a simple Pod spec with a busybox container
|
// makeBusyboxPod returns a simple Pod spec with a busybox container
|
||||||
// that requests resourceName and runs the specified command.
|
// that requests SampleDeviceResourceName and runs the specified command.
|
||||||
func makeBusyboxPod(resourceName, cmd string) *v1.Pod {
|
func makeBusyboxPod(SampleDeviceResourceName, cmd string) *v1.Pod {
|
||||||
podName := "device-plugin-test-" + string(uuid.NewUUID())
|
podName := "device-plugin-test-" + string(uuid.NewUUID())
|
||||||
rl := v1.ResourceList{v1.ResourceName(resourceName): *resource.NewQuantity(1, resource.DecimalSI)}
|
rl := v1.ResourceList{v1.ResourceName(SampleDeviceResourceName): *resource.NewQuantity(1, resource.DecimalSI)}
|
||||||
|
|
||||||
return &v1.Pod{
|
return &v1.Pod{
|
||||||
ObjectMeta: metav1.ObjectMeta{Name: podName},
|
ObjectMeta: metav1.ObjectMeta{Name: podName},
|
||||||
@ -397,23 +357,3 @@ func parseLog(ctx context.Context, f *framework.Framework, podName string, contN
|
|||||||
|
|
||||||
return matches[1]
|
return matches[1]
|
||||||
}
|
}
|
||||||
|
|
||||||
// numberOfDevicesCapacity returns the number of devices of resourceName advertised by a node capacity
|
|
||||||
func numberOfDevicesCapacity(node *v1.Node, resourceName string) int64 {
|
|
||||||
val, ok := node.Status.Capacity[v1.ResourceName(resourceName)]
|
|
||||||
if !ok {
|
|
||||||
return 0
|
|
||||||
}
|
|
||||||
|
|
||||||
return val.Value()
|
|
||||||
}
|
|
||||||
|
|
||||||
// numberOfDevicesAllocatable returns the number of devices of resourceName advertised by a node allocatable
|
|
||||||
func numberOfDevicesAllocatable(node *v1.Node, resourceName string) int64 {
|
|
||||||
val, ok := node.Status.Allocatable[v1.ResourceName(resourceName)]
|
|
||||||
if !ok {
|
|
||||||
return 0
|
|
||||||
}
|
|
||||||
|
|
||||||
return val.Value()
|
|
||||||
}
|
|
||||||
|
@ -24,13 +24,16 @@ import (
|
|||||||
"sync"
|
"sync"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
|
v1 "k8s.io/api/core/v1"
|
||||||
"k8s.io/klog/v2"
|
"k8s.io/klog/v2"
|
||||||
|
|
||||||
|
metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
|
||||||
utilerrors "k8s.io/apimachinery/pkg/util/errors"
|
utilerrors "k8s.io/apimachinery/pkg/util/errors"
|
||||||
"k8s.io/apimachinery/pkg/util/sets"
|
"k8s.io/apimachinery/pkg/util/sets"
|
||||||
internalapi "k8s.io/cri-api/pkg/apis"
|
internalapi "k8s.io/cri-api/pkg/apis"
|
||||||
runtimeapi "k8s.io/cri-api/pkg/apis/runtime/v1"
|
runtimeapi "k8s.io/cri-api/pkg/apis/runtime/v1"
|
||||||
commontest "k8s.io/kubernetes/test/e2e/common"
|
commontest "k8s.io/kubernetes/test/e2e/common"
|
||||||
|
"k8s.io/kubernetes/test/e2e/framework"
|
||||||
e2egpu "k8s.io/kubernetes/test/e2e/framework/gpu"
|
e2egpu "k8s.io/kubernetes/test/e2e/framework/gpu"
|
||||||
e2emanifest "k8s.io/kubernetes/test/e2e/framework/manifest"
|
e2emanifest "k8s.io/kubernetes/test/e2e/framework/manifest"
|
||||||
e2epod "k8s.io/kubernetes/test/e2e/framework/pod"
|
e2epod "k8s.io/kubernetes/test/e2e/framework/pod"
|
||||||
@ -45,9 +48,6 @@ const (
|
|||||||
imagePullRetryDelay = time.Second
|
imagePullRetryDelay = time.Second
|
||||||
// Number of parallel count to pull images.
|
// Number of parallel count to pull images.
|
||||||
maxParallelImagePullCount = 5
|
maxParallelImagePullCount = 5
|
||||||
|
|
||||||
// SampleDevicePluginDSYAML is the path of the daemonset template of the sample device plugin. // TODO: Parametrize it by making it a feature in TestFramework.
|
|
||||||
SampleDevicePluginDSYAML = "test/e2e/testing-manifests/sample-device-plugin.yaml"
|
|
||||||
)
|
)
|
||||||
|
|
||||||
// NodePrePullImageList is a list of images used in node e2e test. These images will be prepulled
|
// NodePrePullImageList is a list of images used in node e2e test. These images will be prepulled
|
||||||
@ -87,11 +87,6 @@ func updateImageAllowList(ctx context.Context) {
|
|||||||
} else {
|
} else {
|
||||||
e2epod.ImagePrePullList.Insert(gpuDevicePluginImage)
|
e2epod.ImagePrePullList.Insert(gpuDevicePluginImage)
|
||||||
}
|
}
|
||||||
if kubeVirtPluginImage, err := getKubeVirtDevicePluginImage(); err != nil {
|
|
||||||
klog.Errorln(err)
|
|
||||||
} else {
|
|
||||||
e2epod.ImagePrePullList.Insert(kubeVirtPluginImage)
|
|
||||||
}
|
|
||||||
if samplePluginImage, err := getSampleDevicePluginImage(); err != nil {
|
if samplePluginImage, err := getSampleDevicePluginImage(); err != nil {
|
||||||
klog.Errorln(err)
|
klog.Errorln(err)
|
||||||
} else {
|
} else {
|
||||||
@ -244,6 +239,29 @@ func getSampleDevicePluginImage() (string, error) {
|
|||||||
return ds.Spec.Template.Spec.Containers[0].Image, nil
|
return ds.Spec.Template.Spec.Containers[0].Image, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// getSampleDevicePluginPod returns the Sample Device Plugin pod to be used e2e tests.
|
||||||
|
func getSampleDevicePluginPod(pluginSockDir string) *v1.Pod {
|
||||||
|
data, err := e2etestfiles.Read(SampleDevicePluginDSYAML)
|
||||||
|
if err != nil {
|
||||||
|
framework.Fail(err.Error())
|
||||||
|
}
|
||||||
|
|
||||||
|
ds := readDaemonSetV1OrDie(data)
|
||||||
|
dp := &v1.Pod{
|
||||||
|
ObjectMeta: metav1.ObjectMeta{
|
||||||
|
Name: SampleDevicePluginName,
|
||||||
|
},
|
||||||
|
Spec: ds.Spec.Template.Spec,
|
||||||
|
}
|
||||||
|
for i := range dp.Spec.Containers[0].Env {
|
||||||
|
if dp.Spec.Containers[0].Env[i].Name == SampleDeviceEnvVarNamePluginSockDir {
|
||||||
|
dp.Spec.Containers[0].Env[i].Value = pluginSockDir
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return dp
|
||||||
|
}
|
||||||
|
|
||||||
// getSRIOVDevicePluginImage returns the image of SRIOV device plugin.
|
// getSRIOVDevicePluginImage returns the image of SRIOV device plugin.
|
||||||
func getSRIOVDevicePluginImage() (string, error) {
|
func getSRIOVDevicePluginImage() (string, error) {
|
||||||
data, err := e2etestfiles.Read(SRIOVDevicePluginDSYAML)
|
data, err := e2etestfiles.Read(SRIOVDevicePluginDSYAML)
|
||||||
@ -262,23 +280,3 @@ func getSRIOVDevicePluginImage() (string, error) {
|
|||||||
}
|
}
|
||||||
return ds.Spec.Template.Spec.Containers[0].Image, nil
|
return ds.Spec.Template.Spec.Containers[0].Image, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
// TODO generilize this function with above one
|
|
||||||
// getKubeVirtDevicePluginImage returns the image of SRIOV device plugin.
|
|
||||||
func getKubeVirtDevicePluginImage() (string, error) {
|
|
||||||
data, err := e2etestfiles.Read(KubeVirtDevicePluginDSYAML)
|
|
||||||
if err != nil {
|
|
||||||
return "", fmt.Errorf("failed to read the device plugin manifest: %w", err)
|
|
||||||
}
|
|
||||||
ds, err := e2emanifest.DaemonSetFromData(data)
|
|
||||||
if err != nil {
|
|
||||||
return "", fmt.Errorf("failed to parse the device plugin image: %w", err)
|
|
||||||
}
|
|
||||||
if ds == nil {
|
|
||||||
return "", fmt.Errorf("failed to parse the device plugin image: the extracted DaemonSet is nil")
|
|
||||||
}
|
|
||||||
if len(ds.Spec.Template.Spec.Containers) < 1 {
|
|
||||||
return "", fmt.Errorf("failed to parse the device plugin image: cannot extract the container from YAML")
|
|
||||||
}
|
|
||||||
return ds.Spec.Template.Spec.Containers[0].Image, nil
|
|
||||||
}
|
|
||||||
|
@ -18,7 +18,6 @@ package e2enode
|
|||||||
|
|
||||||
import (
|
import (
|
||||||
"context"
|
"context"
|
||||||
"errors"
|
|
||||||
"fmt"
|
"fmt"
|
||||||
"os"
|
"os"
|
||||||
"strings"
|
"strings"
|
||||||
@ -27,6 +26,7 @@ import (
|
|||||||
v1 "k8s.io/api/core/v1"
|
v1 "k8s.io/api/core/v1"
|
||||||
"k8s.io/apimachinery/pkg/api/resource"
|
"k8s.io/apimachinery/pkg/api/resource"
|
||||||
metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
|
metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
|
||||||
|
kubeletdevicepluginv1beta1 "k8s.io/kubelet/pkg/apis/deviceplugin/v1beta1"
|
||||||
kubeletpodresourcesv1 "k8s.io/kubelet/pkg/apis/podresources/v1"
|
kubeletpodresourcesv1 "k8s.io/kubelet/pkg/apis/podresources/v1"
|
||||||
kubefeatures "k8s.io/kubernetes/pkg/features"
|
kubefeatures "k8s.io/kubernetes/pkg/features"
|
||||||
kubeletconfig "k8s.io/kubernetes/pkg/kubelet/apis/config"
|
kubeletconfig "k8s.io/kubernetes/pkg/kubelet/apis/config"
|
||||||
@ -46,7 +46,10 @@ import (
|
|||||||
e2enode "k8s.io/kubernetes/test/e2e/framework/node"
|
e2enode "k8s.io/kubernetes/test/e2e/framework/node"
|
||||||
e2epod "k8s.io/kubernetes/test/e2e/framework/pod"
|
e2epod "k8s.io/kubernetes/test/e2e/framework/pod"
|
||||||
e2eskipper "k8s.io/kubernetes/test/e2e/framework/skipper"
|
e2eskipper "k8s.io/kubernetes/test/e2e/framework/skipper"
|
||||||
e2etestfiles "k8s.io/kubernetes/test/e2e/framework/testfiles"
|
)
|
||||||
|
|
||||||
|
const (
|
||||||
|
defaultTopologyUnawareResourceName = "example.com/resource"
|
||||||
)
|
)
|
||||||
|
|
||||||
type podDesc struct {
|
type podDesc struct {
|
||||||
@ -136,9 +139,11 @@ func logPodResources(podIdx int, pr *kubeletpodresourcesv1.PodResources) {
|
|||||||
|
|
||||||
type podResMap map[string]map[string]kubeletpodresourcesv1.ContainerResources
|
type podResMap map[string]map[string]kubeletpodresourcesv1.ContainerResources
|
||||||
|
|
||||||
func getPodResources(ctx context.Context, cli kubeletpodresourcesv1.PodResourcesListerClient) podResMap {
|
func getPodResourcesValues(ctx context.Context, cli kubeletpodresourcesv1.PodResourcesListerClient) (podResMap, error) {
|
||||||
resp, err := cli.List(ctx, &kubeletpodresourcesv1.ListPodResourcesRequest{})
|
resp, err := cli.List(ctx, &kubeletpodresourcesv1.ListPodResourcesRequest{})
|
||||||
framework.ExpectNoError(err)
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
|
||||||
res := make(map[string]map[string]kubeletpodresourcesv1.ContainerResources)
|
res := make(map[string]map[string]kubeletpodresourcesv1.ContainerResources)
|
||||||
for idx, podResource := range resp.GetPodResources() {
|
for idx, podResource := range resp.GetPodResources() {
|
||||||
@ -151,7 +156,7 @@ func getPodResources(ctx context.Context, cli kubeletpodresourcesv1.PodResources
|
|||||||
}
|
}
|
||||||
res[podResource.GetName()] = cnts
|
res[podResource.GetName()] = cnts
|
||||||
}
|
}
|
||||||
return res
|
return res, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
type testPodData struct {
|
type testPodData struct {
|
||||||
@ -232,10 +237,10 @@ func matchPodDescWithResources(expected []podDesc, found podResMap) error {
|
|||||||
return fmt.Errorf("pod %q container %q expected no resources, got %v", podReq.podName, podReq.cntName, devs)
|
return fmt.Errorf("pod %q container %q expected no resources, got %v", podReq.podName, podReq.cntName, devs)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
if cnts, ok := found[KubeVirtResourceName]; ok {
|
if cnts, ok := found[defaultTopologyUnawareResourceName]; ok {
|
||||||
for _, cnt := range cnts {
|
for _, cnt := range cnts {
|
||||||
for _, cd := range cnt.GetDevices() {
|
for _, cd := range cnt.GetDevices() {
|
||||||
if cd.ResourceName != KubeVirtResourceName {
|
if cd.ResourceName != defaultTopologyUnawareResourceName {
|
||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
if cd.Topology != nil {
|
if cd.Topology != nil {
|
||||||
@ -252,7 +257,10 @@ func matchPodDescWithResources(expected []podDesc, found podResMap) error {
|
|||||||
|
|
||||||
func expectPodResources(ctx context.Context, offset int, cli kubeletpodresourcesv1.PodResourcesListerClient, expected []podDesc) {
|
func expectPodResources(ctx context.Context, offset int, cli kubeletpodresourcesv1.PodResourcesListerClient, expected []podDesc) {
|
||||||
gomega.EventuallyWithOffset(1+offset, ctx, func(ctx context.Context) error {
|
gomega.EventuallyWithOffset(1+offset, ctx, func(ctx context.Context) error {
|
||||||
found := getPodResources(ctx, cli)
|
found, err := getPodResourcesValues(ctx, cli)
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
return matchPodDescWithResources(expected, found)
|
return matchPodDescWithResources(expected, found)
|
||||||
}, time.Minute, 10*time.Second).Should(gomega.Succeed())
|
}, time.Minute, 10*time.Second).Should(gomega.Succeed())
|
||||||
}
|
}
|
||||||
@ -280,8 +288,10 @@ func podresourcesListTests(ctx context.Context, f *framework.Framework, cli kube
|
|||||||
expectedBasePods = 1 // sriovdp
|
expectedBasePods = 1 // sriovdp
|
||||||
}
|
}
|
||||||
|
|
||||||
|
var err error
|
||||||
ginkgo.By("checking the output when no pods are present")
|
ginkgo.By("checking the output when no pods are present")
|
||||||
found = getPodResources(ctx, cli)
|
found, err = getPodResourcesValues(ctx, cli)
|
||||||
|
framework.ExpectNoError(err)
|
||||||
gomega.ExpectWithOffset(1, found).To(gomega.HaveLen(expectedBasePods), "base pod expectation mismatch")
|
gomega.ExpectWithOffset(1, found).To(gomega.HaveLen(expectedBasePods), "base pod expectation mismatch")
|
||||||
|
|
||||||
tpd = newTestPodData()
|
tpd = newTestPodData()
|
||||||
@ -732,14 +742,7 @@ var _ = SIGDescribe("POD Resources [Serial] [Feature:PodResources][NodeFeature:P
|
|||||||
})
|
})
|
||||||
})
|
})
|
||||||
|
|
||||||
ginkgo.Context("with KubeVirt device plugin, which reports resources w/o hardware topology", func() {
|
ginkgo.Context("with a topology-unaware device plugin, which reports resources w/o hardware topology", func() {
|
||||||
ginkgo.BeforeEach(func() {
|
|
||||||
_, err := os.Stat("/dev/kvm")
|
|
||||||
if errors.Is(err, os.ErrNotExist) {
|
|
||||||
e2eskipper.Skipf("KubeVirt device plugin could work only in kvm based environment")
|
|
||||||
}
|
|
||||||
})
|
|
||||||
|
|
||||||
ginkgo.Context("with CPU manager Static policy", func() {
|
ginkgo.Context("with CPU manager Static policy", func() {
|
||||||
ginkgo.BeforeEach(func(ctx context.Context) {
|
ginkgo.BeforeEach(func(ctx context.Context) {
|
||||||
// this is a very rough check. We just want to rule out system that does NOT have enough resources
|
// this is a very rough check. We just want to rule out system that does NOT have enough resources
|
||||||
@ -765,10 +768,10 @@ var _ = SIGDescribe("POD Resources [Serial] [Feature:PodResources][NodeFeature:P
|
|||||||
})
|
})
|
||||||
|
|
||||||
ginkgo.It("should return proper podresources the same as before the restart of kubelet", func(ctx context.Context) {
|
ginkgo.It("should return proper podresources the same as before the restart of kubelet", func(ctx context.Context) {
|
||||||
dpPod := setupKubeVirtDevicePluginOrFail(ctx, f)
|
dpPod := setupSampleDevicePluginOrFail(ctx, f)
|
||||||
ginkgo.DeferCleanup(teardownKubeVirtDevicePluginOrFail, f, dpPod)
|
ginkgo.DeferCleanup(teardownSampleDevicePluginOrFail, f, dpPod)
|
||||||
|
|
||||||
waitForKubeVirtResources(ctx, f, dpPod)
|
waitForTopologyUnawareResources(ctx, f)
|
||||||
|
|
||||||
endpoint, err := util.LocalEndpoint(defaultPodResourcesPath, podresources.Socket)
|
endpoint, err := util.LocalEndpoint(defaultPodResourcesPath, podresources.Socket)
|
||||||
framework.ExpectNoError(err)
|
framework.ExpectNoError(err)
|
||||||
@ -777,22 +780,21 @@ var _ = SIGDescribe("POD Resources [Serial] [Feature:PodResources][NodeFeature:P
|
|||||||
framework.ExpectNoError(err)
|
framework.ExpectNoError(err)
|
||||||
defer conn.Close()
|
defer conn.Close()
|
||||||
|
|
||||||
ginkgo.By("checking List and resources kubevirt resource should be without topology")
|
ginkgo.By("checking List and resources topology unaware resource should be without topology")
|
||||||
|
|
||||||
allocatableResponse, _ := cli.GetAllocatableResources(ctx, &kubeletpodresourcesv1.AllocatableResourcesRequest{})
|
allocatableResponse, _ := cli.GetAllocatableResources(ctx, &kubeletpodresourcesv1.AllocatableResourcesRequest{})
|
||||||
for _, dev := range allocatableResponse.GetDevices() {
|
for _, dev := range allocatableResponse.GetDevices() {
|
||||||
if dev.ResourceName != KubeVirtResourceName {
|
if dev.ResourceName != defaultTopologyUnawareResourceName {
|
||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
|
|
||||||
framework.ExpectEqual(dev.Topology == nil, true, "Topology is expected to be empty for kubevirt resources")
|
framework.ExpectEqual(dev.Topology == nil, true, "Topology is expected to be empty for topology unaware resources")
|
||||||
}
|
}
|
||||||
|
|
||||||
// Run pod which requires KubeVirtResourceName
|
|
||||||
desc := podDesc{
|
desc := podDesc{
|
||||||
podName: "pod-01",
|
podName: "pod-01",
|
||||||
cntName: "cnt-01",
|
cntName: "cnt-01",
|
||||||
resourceName: KubeVirtResourceName,
|
resourceName: defaultTopologyUnawareResourceName,
|
||||||
resourceAmount: 1,
|
resourceAmount: 1,
|
||||||
cpuRequest: 1000,
|
cpuRequest: 1000,
|
||||||
}
|
}
|
||||||
@ -804,23 +806,13 @@ var _ = SIGDescribe("POD Resources [Serial] [Feature:PodResources][NodeFeature:P
|
|||||||
|
|
||||||
expectPodResources(ctx, 1, cli, []podDesc{desc})
|
expectPodResources(ctx, 1, cli, []podDesc{desc})
|
||||||
|
|
||||||
restartTime := time.Now()
|
|
||||||
ginkgo.By("Restarting Kubelet")
|
ginkgo.By("Restarting Kubelet")
|
||||||
restartKubelet(true)
|
restartKubelet(true)
|
||||||
|
|
||||||
// we need to wait for the node to be reported ready before we can safely query
|
// we need to wait for the node to be reported ready before we can safely query
|
||||||
// the podresources endpoint again. Otherwise we will have false negatives.
|
// the podresources endpoint again. Otherwise we will have false negatives.
|
||||||
ginkgo.By("Wait for node to be ready")
|
ginkgo.By("Wait for node to be ready")
|
||||||
gomega.Eventually(ctx, func() bool {
|
waitForTopologyUnawareResources(ctx, f)
|
||||||
node, err := f.ClientSet.CoreV1().Nodes().Get(ctx, framework.TestContext.NodeName, metav1.GetOptions{})
|
|
||||||
framework.ExpectNoError(err)
|
|
||||||
for _, cond := range node.Status.Conditions {
|
|
||||||
if cond.Type == v1.NodeReady && cond.Status == v1.ConditionTrue && cond.LastHeartbeatTime.After(restartTime) {
|
|
||||||
return true
|
|
||||||
}
|
|
||||||
}
|
|
||||||
return false
|
|
||||||
}, 5*time.Minute, framework.Poll).Should(gomega.BeTrue())
|
|
||||||
|
|
||||||
expectPodResources(ctx, 1, cli, []podDesc{desc})
|
expectPodResources(ctx, 1, cli, []podDesc{desc})
|
||||||
tpd.deletePodsForTest(ctx, f)
|
tpd.deletePodsForTest(ctx, f)
|
||||||
@ -887,80 +879,47 @@ func getOnlineCPUs() (cpuset.CPUSet, error) {
|
|||||||
return cpuset.Parse(strings.TrimSpace(string(onlineCPUList)))
|
return cpuset.Parse(strings.TrimSpace(string(onlineCPUList)))
|
||||||
}
|
}
|
||||||
|
|
||||||
func setupKubeVirtDevicePluginOrFail(ctx context.Context, f *framework.Framework) *v1.Pod {
|
func setupSampleDevicePluginOrFail(ctx context.Context, f *framework.Framework) *v1.Pod {
|
||||||
e2enode.WaitForNodeToBeReady(ctx, f.ClientSet, framework.TestContext.NodeName, 5*time.Minute)
|
e2enode.WaitForNodeToBeReady(ctx, f.ClientSet, framework.TestContext.NodeName, 5*time.Minute)
|
||||||
|
|
||||||
dp := getKubeVirtDevicePluginPod()
|
dp := getSampleDevicePluginPod(kubeletdevicepluginv1beta1.DevicePluginPath)
|
||||||
dp.Spec.NodeName = framework.TestContext.NodeName
|
dp.Spec.NodeName = framework.TestContext.NodeName
|
||||||
|
|
||||||
ginkgo.By("Create KubeVirt device plugin pod")
|
ginkgo.By("Create the sample device plugin pod")
|
||||||
|
|
||||||
dpPod, err := f.ClientSet.CoreV1().Pods(metav1.NamespaceSystem).Create(ctx, dp, metav1.CreateOptions{})
|
dpPod := e2epod.NewPodClient(f).CreateSync(ctx, dp)
|
||||||
framework.ExpectNoError(err)
|
|
||||||
|
|
||||||
if err = e2epod.WaitForPodCondition(ctx, f.ClientSet, metav1.NamespaceSystem, dp.Name, "Ready", 120*time.Second, testutils.PodRunningReady); err != nil {
|
err := e2epod.WaitForPodCondition(ctx, f.ClientSet, dpPod.Namespace, dpPod.Name, "Ready", 120*time.Second, testutils.PodRunningReady)
|
||||||
framework.Logf("KubeVirt Pod %v took too long to enter running/ready: %v", dp.Name, err)
|
if err != nil {
|
||||||
|
framework.Logf("Sample Device Pod %v took too long to enter running/ready: %v", dp.Name, err)
|
||||||
}
|
}
|
||||||
framework.ExpectNoError(err)
|
framework.ExpectNoError(err)
|
||||||
|
|
||||||
return dpPod
|
return dpPod
|
||||||
}
|
}
|
||||||
|
|
||||||
func teardownKubeVirtDevicePluginOrFail(ctx context.Context, f *framework.Framework, pod *v1.Pod) {
|
func teardownSampleDevicePluginOrFail(ctx context.Context, f *framework.Framework, pod *v1.Pod) {
|
||||||
gp := int64(0)
|
gp := int64(0)
|
||||||
deleteOptions := metav1.DeleteOptions{
|
deleteOptions := metav1.DeleteOptions{
|
||||||
GracePeriodSeconds: &gp,
|
GracePeriodSeconds: &gp,
|
||||||
}
|
}
|
||||||
ginkgo.By(fmt.Sprintf("Delete KubeVirt device plugin pod %s/%s", pod.Namespace, pod.Name))
|
ginkgo.By(fmt.Sprintf("Delete sample device plugin pod %s/%s", pod.Namespace, pod.Name))
|
||||||
err := f.ClientSet.CoreV1().Pods(pod.Namespace).Delete(ctx, pod.Name, deleteOptions)
|
err := f.ClientSet.CoreV1().Pods(pod.Namespace).Delete(ctx, pod.Name, deleteOptions)
|
||||||
|
|
||||||
framework.ExpectNoError(err)
|
framework.ExpectNoError(err)
|
||||||
waitForAllContainerRemoval(ctx, pod.Name, pod.Namespace)
|
waitForAllContainerRemoval(ctx, pod.Name, pod.Namespace)
|
||||||
}
|
}
|
||||||
|
|
||||||
func findKubeVirtResource(node *v1.Node) int64 {
|
func waitForTopologyUnawareResources(ctx context.Context, f *framework.Framework) {
|
||||||
framework.Logf("Node status allocatable: %v", node.Status.Allocatable)
|
ginkgo.By(fmt.Sprintf("Waiting for %q resources to become available on the local node", defaultTopologyUnawareResourceName))
|
||||||
for key, val := range node.Status.Allocatable {
|
|
||||||
if string(key) == KubeVirtResourceName {
|
|
||||||
v := val.Value()
|
|
||||||
if v > 0 {
|
|
||||||
return v
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
return 0
|
|
||||||
}
|
|
||||||
|
|
||||||
func waitForKubeVirtResources(ctx context.Context, f *framework.Framework, pod *v1.Pod) {
|
|
||||||
ginkgo.By("Waiting for kubevirt resources to become available on the local node")
|
|
||||||
|
|
||||||
gomega.Eventually(ctx, func(ctx context.Context) bool {
|
gomega.Eventually(ctx, func(ctx context.Context) bool {
|
||||||
node := getLocalNode(ctx, f)
|
node := getLocalNode(ctx, f)
|
||||||
kubeVirtResourceAmount := findKubeVirtResource(node)
|
resourceAmount := CountSampleDeviceAllocatable(node)
|
||||||
return kubeVirtResourceAmount != 0
|
return resourceAmount > 0
|
||||||
}, 2*time.Minute, framework.Poll).Should(gomega.BeTrue())
|
}, 2*time.Minute, framework.Poll).Should(gomega.BeTrue())
|
||||||
}
|
}
|
||||||
|
|
||||||
// getKubeVirtDevicePluginPod returns the Device Plugin pod for kube resources in e2e tests.
|
|
||||||
func getKubeVirtDevicePluginPod() *v1.Pod {
|
|
||||||
data, err := e2etestfiles.Read(KubeVirtDevicePluginDSYAML)
|
|
||||||
if err != nil {
|
|
||||||
framework.Fail(err.Error())
|
|
||||||
}
|
|
||||||
|
|
||||||
ds := readDaemonSetV1OrDie(data)
|
|
||||||
p := &v1.Pod{
|
|
||||||
ObjectMeta: metav1.ObjectMeta{
|
|
||||||
Name: KubeVirtDevicePluginName,
|
|
||||||
Namespace: metav1.NamespaceSystem,
|
|
||||||
},
|
|
||||||
|
|
||||||
Spec: ds.Spec.Template.Spec,
|
|
||||||
}
|
|
||||||
|
|
||||||
return p
|
|
||||||
}
|
|
||||||
|
|
||||||
func getPodResourcesMetrics(ctx context.Context) (e2emetrics.KubeletMetrics, error) {
|
func getPodResourcesMetrics(ctx context.Context) (e2emetrics.KubeletMetrics, error) {
|
||||||
// we are running out of good names, so we need to be unnecessarily specific to avoid clashes
|
// we are running out of good names, so we need to be unnecessarily specific to avoid clashes
|
||||||
ginkgo.By("getting Pod Resources metrics from the metrics API")
|
ginkgo.By("getting Pod Resources metrics from the metrics API")
|
||||||
|
@ -1,28 +0,0 @@
|
|||||||
apiVersion: apps/v1
|
|
||||||
kind: DaemonSet
|
|
||||||
metadata:
|
|
||||||
labels:
|
|
||||||
name: kubevirt-kvm-device-plugin
|
|
||||||
name: kubevirt-kvm-device-plugin
|
|
||||||
spec:
|
|
||||||
selector:
|
|
||||||
matchLabels:
|
|
||||||
name: kubevirt-kvm-device-plugin
|
|
||||||
template:
|
|
||||||
metadata:
|
|
||||||
labels:
|
|
||||||
name: kubevirt-kvm-device-plugin
|
|
||||||
spec:
|
|
||||||
containers:
|
|
||||||
- name: kubevirt-kvm-device-plugin
|
|
||||||
image: quay.io/kubevirt/device-plugin-kvm
|
|
||||||
args: ["-v", "3", "-logtostderr"]
|
|
||||||
securityContext:
|
|
||||||
privileged: true
|
|
||||||
volumeMounts:
|
|
||||||
- name: device-plugin
|
|
||||||
mountPath: /var/lib/kubelet/device-plugins
|
|
||||||
volumes:
|
|
||||||
- name: device-plugin
|
|
||||||
hostPath:
|
|
||||||
path: /var/lib/kubelet/device-plugins
|
|
@ -1,27 +0,0 @@
|
|||||||
/*
|
|
||||||
Copyright 2021 The Kubernetes Authors.
|
|
||||||
|
|
||||||
Licensed under the Apache License, Version 2.0 (the "License");
|
|
||||||
you may not use this file except in compliance with the License.
|
|
||||||
You may obtain a copy of the License at
|
|
||||||
|
|
||||||
http://www.apache.org/licenses/LICENSE-2.0
|
|
||||||
|
|
||||||
Unless required by applicable law or agreed to in writing, software
|
|
||||||
distributed under the License is distributed on an "AS IS" BASIS,
|
|
||||||
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
||||||
See the License for the specific language governing permissions and
|
|
||||||
limitations under the License.
|
|
||||||
*/
|
|
||||||
|
|
||||||
package e2enode
|
|
||||||
|
|
||||||
const (
|
|
||||||
KubeVirtDevicePluginDSYAML = "test/e2e_node/testing-manifests/kubevirt-kvm-ds.yaml"
|
|
||||||
|
|
||||||
// KubeVirtDevicePluginName is the name of the device plugin pod
|
|
||||||
KubeVirtDevicePluginName = "kubevirt-device-plugin"
|
|
||||||
|
|
||||||
// KubeVirtResourceName is the name of the resource provided by kubevirt device plugin
|
|
||||||
KubeVirtResourceName = "devices.kubevirt.io/kvm"
|
|
||||||
)
|
|
52
test/e2e_node/util_sampledevice.go
Normal file
52
test/e2e_node/util_sampledevice.go
Normal file
@ -0,0 +1,52 @@
|
|||||||
|
/*
|
||||||
|
Copyright 2023 The Kubernetes Authors.
|
||||||
|
|
||||||
|
Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
you may not use this file except in compliance with the License.
|
||||||
|
You may obtain a copy of the License at
|
||||||
|
|
||||||
|
http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
|
||||||
|
Unless required by applicable law or agreed to in writing, software
|
||||||
|
distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
See the License for the specific language governing permissions and
|
||||||
|
limitations under the License.
|
||||||
|
*/
|
||||||
|
|
||||||
|
package e2enode
|
||||||
|
|
||||||
|
import (
|
||||||
|
v1 "k8s.io/api/core/v1"
|
||||||
|
)
|
||||||
|
|
||||||
|
const (
|
||||||
|
// SampleDevicePluginDSYAML is the path of the daemonset template of the sample device plugin. // TODO: Parametrize it by making it a feature in TestFramework.
|
||||||
|
SampleDevicePluginDSYAML = "test/e2e/testing-manifests/sample-device-plugin.yaml"
|
||||||
|
|
||||||
|
// SampleDevicePluginName is the name of the device plugin pod
|
||||||
|
SampleDevicePluginName = "sample-device-plugin"
|
||||||
|
|
||||||
|
// SampleDeviceResourceName is the name of the resource provided by the sample device plugin
|
||||||
|
SampleDeviceResourceName = "example.com/resource"
|
||||||
|
|
||||||
|
SampleDeviceEnvVarNamePluginSockDir = "PLUGIN_SOCK_DIR"
|
||||||
|
)
|
||||||
|
|
||||||
|
// CountSampleDeviceCapacity returns the number of devices of SampleDeviceResourceName advertised by a node capacity
|
||||||
|
func CountSampleDeviceCapacity(node *v1.Node) int64 {
|
||||||
|
val, ok := node.Status.Capacity[v1.ResourceName(SampleDeviceResourceName)]
|
||||||
|
if !ok {
|
||||||
|
return 0
|
||||||
|
}
|
||||||
|
return val.Value()
|
||||||
|
}
|
||||||
|
|
||||||
|
// CountSampleDeviceAllocatable returns the number of devices of SampleDeviceResourceName advertised by a node allocatable
|
||||||
|
func CountSampleDeviceAllocatable(node *v1.Node) int64 {
|
||||||
|
val, ok := node.Status.Allocatable[v1.ResourceName(SampleDeviceResourceName)]
|
||||||
|
if !ok {
|
||||||
|
return 0
|
||||||
|
}
|
||||||
|
return val.Value()
|
||||||
|
}
|
Loading…
Reference in New Issue
Block a user