mirror of
https://github.com/k3s-io/kubernetes.git
synced 2025-07-20 18:31:15 +00:00
Merge pull request #116189 from marosset/windows-hyperv-basic-e2e-test
Adding e2e test to verify hyperv container is running inside a VM on Windows
This commit is contained in:
commit
949bee0118
@ -673,21 +673,7 @@ var _ = SIGDescribe("[Feature:WindowsHostProcessContainers] [MinimumKubeletVersi
|
|||||||
|
|
||||||
ginkgo.It("should support querying api-server using in-cluster config", func(ctx context.Context) {
|
ginkgo.It("should support querying api-server using in-cluster config", func(ctx context.Context) {
|
||||||
// This functionality is only support on containerd v1.7+
|
// This functionality is only support on containerd v1.7+
|
||||||
ginkgo.By("Ensuring Windows nodes are running containerd v1.7+")
|
skipUnlessContainerdOneSevenOrGreater(ctx, f)
|
||||||
windowsNode, err := findWindowsNode(ctx, f)
|
|
||||||
framework.ExpectNoError(err, "error finding Windows node")
|
|
||||||
r, v, err := getNodeContainerRuntimeAndVersion(windowsNode)
|
|
||||||
framework.ExpectNoError(err, "error getting node container runtime and version")
|
|
||||||
framework.Logf("Got runtime: %s, version %v for node %s", r, v, windowsNode.Name)
|
|
||||||
|
|
||||||
if !strings.EqualFold(r, "containerd") {
|
|
||||||
e2eskipper.Skipf("container runtime is not containerd")
|
|
||||||
}
|
|
||||||
|
|
||||||
v1dot7 := semver.MustParse("1.7.0")
|
|
||||||
if v.LT(v1dot7) {
|
|
||||||
e2eskipper.Skipf("container runtime is < 1.7.0")
|
|
||||||
}
|
|
||||||
|
|
||||||
ginkgo.By("Scheduling a pod that runs agnhost inclusterclient")
|
ginkgo.By("Scheduling a pod that runs agnhost inclusterclient")
|
||||||
podName := "host-process-agnhost-icc"
|
podName := "host-process-agnhost-icc"
|
||||||
@ -752,21 +738,7 @@ var _ = SIGDescribe("[Feature:WindowsHostProcessContainers] [MinimumKubeletVersi
|
|||||||
|
|
||||||
ginkgo.It("should run as localgroup accounts", func(ctx context.Context) {
|
ginkgo.It("should run as localgroup accounts", func(ctx context.Context) {
|
||||||
// This functionality is only supported on containerd v1.7+
|
// This functionality is only supported on containerd v1.7+
|
||||||
ginkgo.By("Ensuring Windows nodes are running containerd v1.7+")
|
skipUnlessContainerdOneSevenOrGreater(ctx, f)
|
||||||
windowsNode, err := findWindowsNode(ctx, f)
|
|
||||||
framework.ExpectNoError(err, "error finding Windows node")
|
|
||||||
r, v, err := getNodeContainerRuntimeAndVersion(windowsNode)
|
|
||||||
framework.ExpectNoError(err, "error getting node container runtime and version")
|
|
||||||
framework.Logf("Got runtime: %s, version %v for node %s", r, v, windowsNode.Name)
|
|
||||||
|
|
||||||
if !strings.EqualFold(r, "containerd") {
|
|
||||||
e2eskipper.Skipf("container runtime is not containerd")
|
|
||||||
}
|
|
||||||
|
|
||||||
v1dot7 := semver.MustParse("1.7.0")
|
|
||||||
if v.LT(v1dot7) {
|
|
||||||
e2eskipper.Skipf("container runtime is < 1.7.0")
|
|
||||||
}
|
|
||||||
|
|
||||||
ginkgo.By("Scheduling a pod that creates a localgroup from an init container then starts a container using that group")
|
ginkgo.By("Scheduling a pod that creates a localgroup from an init container then starts a container using that group")
|
||||||
localGroupName := getRandomUserGrounName()
|
localGroupName := getRandomUserGrounName()
|
||||||
|
145
test/e2e/windows/hyperv.go
Normal file
145
test/e2e/windows/hyperv.go
Normal file
@ -0,0 +1,145 @@
|
|||||||
|
/*
|
||||||
|
Copyright 2023 The Kubernetes Authors.
|
||||||
|
|
||||||
|
Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
you may not use this file except in compliance with the License.
|
||||||
|
You may obtain a copy of the License at
|
||||||
|
|
||||||
|
http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
|
||||||
|
Unless required by applicable law or agreed to in writing, software
|
||||||
|
distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
See the License for the specific language governing permissions and
|
||||||
|
limitations under the License.
|
||||||
|
*/
|
||||||
|
|
||||||
|
package windows
|
||||||
|
|
||||||
|
import (
|
||||||
|
"context"
|
||||||
|
"time"
|
||||||
|
|
||||||
|
"github.com/onsi/ginkgo/v2"
|
||||||
|
v1 "k8s.io/api/core/v1"
|
||||||
|
metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
|
||||||
|
"k8s.io/kubernetes/test/e2e/framework"
|
||||||
|
e2epod "k8s.io/kubernetes/test/e2e/framework/pod"
|
||||||
|
e2eskipper "k8s.io/kubernetes/test/e2e/framework/skipper"
|
||||||
|
imageutils "k8s.io/kubernetes/test/utils/image"
|
||||||
|
admissionapi "k8s.io/pod-security-admission/api"
|
||||||
|
)
|
||||||
|
|
||||||
|
var (
|
||||||
|
WindowsHyperVContainerRuntimeClass = "runhcs-wcow-hypervisor"
|
||||||
|
)
|
||||||
|
|
||||||
|
var _ = SIGDescribe("[Feature:WindowsHyperVContainers] HyperV containers", func() {
|
||||||
|
ginkgo.BeforeEach(func() {
|
||||||
|
e2eskipper.SkipUnlessNodeOSDistroIs("windows")
|
||||||
|
})
|
||||||
|
|
||||||
|
f := framework.NewDefaultFramework("windows-hyperv-test")
|
||||||
|
f.NamespacePodSecurityEnforceLevel = admissionapi.LevelPrivileged
|
||||||
|
|
||||||
|
ginkgo.It("should start a hyperv isolated container", func(ctx context.Context) {
|
||||||
|
|
||||||
|
// HyperV isolated containers are only supported on containerd 1.7+
|
||||||
|
skipUnlessContainerdOneSevenOrGreater(ctx, f)
|
||||||
|
|
||||||
|
// check if hyperv runtime class is on node and skip otherwise
|
||||||
|
// Note: the runtime class is expected to be added to a cluster before running this test.
|
||||||
|
// see https://github.com/kubernetes-sigs/windows-testing/tree/master/helpers/hyper-v-mutating-webhook/hyperv-runtimeclass.yaml
|
||||||
|
// for an example.
|
||||||
|
_, err := f.ClientSet.NodeV1().RuntimeClasses().Get(ctx, WindowsHyperVContainerRuntimeClass, metav1.GetOptions{})
|
||||||
|
if err != nil {
|
||||||
|
framework.Logf("error getting runtime class: %v", err)
|
||||||
|
e2eskipper.Skipf("skipping test because runhcs-wcow-hypervisor runtime class is not present")
|
||||||
|
}
|
||||||
|
|
||||||
|
ginkgo.By("selecting a Windows node")
|
||||||
|
targetNode, err := findWindowsNode(ctx, f)
|
||||||
|
framework.ExpectNoError(err, "error finding Windows node")
|
||||||
|
framework.Logf("Using node: %v", targetNode.Name)
|
||||||
|
|
||||||
|
ginkgo.By("schedule a pod to that node")
|
||||||
|
image := imageutils.GetE2EImage(imageutils.BusyBox)
|
||||||
|
hypervPodName := "hyperv-test-pod"
|
||||||
|
hypervPod := &v1.Pod{
|
||||||
|
ObjectMeta: metav1.ObjectMeta{
|
||||||
|
Name: hypervPodName,
|
||||||
|
},
|
||||||
|
Spec: v1.PodSpec{
|
||||||
|
Containers: []v1.Container{
|
||||||
|
{
|
||||||
|
Image: image,
|
||||||
|
Name: "busybox-1",
|
||||||
|
Command: []string{"powershell.exe", "-Command", "Write-Host 'Hello'; sleep -Seconds 600"},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Image: image,
|
||||||
|
Name: "busybox-2",
|
||||||
|
Command: []string{"powershell.exe", "-Command", "Write-Host 'Hello'; sleep -Seconds 600"},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
RestartPolicy: v1.RestartPolicyNever,
|
||||||
|
RuntimeClassName: &WindowsHyperVContainerRuntimeClass,
|
||||||
|
NodeName: targetNode.Name,
|
||||||
|
},
|
||||||
|
}
|
||||||
|
|
||||||
|
pc := e2epod.NewPodClient(f)
|
||||||
|
|
||||||
|
pc.Create(ctx, hypervPod)
|
||||||
|
ginkgo.By("waiting for the pod to be running")
|
||||||
|
timeout := 3 * time.Minute
|
||||||
|
e2epod.WaitForPodsRunningReady(ctx, f.ClientSet, f.Namespace.Name, 1, 0, timeout)
|
||||||
|
|
||||||
|
ginkgo.By("creating a host process container in another pod to verify the pod is running hyperv isolated containers")
|
||||||
|
|
||||||
|
// Note: each pod runs in a separate UVM so even though we are scheduling 2 containers in the test pod
|
||||||
|
// we should only expect a single UVM to be running on the host.
|
||||||
|
podName := "validation-pod"
|
||||||
|
pod := &v1.Pod{
|
||||||
|
ObjectMeta: metav1.ObjectMeta{
|
||||||
|
Name: podName,
|
||||||
|
},
|
||||||
|
Spec: v1.PodSpec{
|
||||||
|
SecurityContext: &v1.PodSecurityContext{
|
||||||
|
WindowsOptions: &v1.WindowsSecurityContextOptions{
|
||||||
|
HostProcess: &trueVar,
|
||||||
|
RunAsUserName: &User_NTAuthoritySystem,
|
||||||
|
},
|
||||||
|
},
|
||||||
|
HostNetwork: true,
|
||||||
|
Containers: []v1.Container{
|
||||||
|
{
|
||||||
|
Image: image,
|
||||||
|
Name: "container",
|
||||||
|
Command: []string{"powershell.exe", "-Command", "$vms = Get-ComputeProcess | Where-Object { ($_.Type -EQ 'VirtualMachine') -and ($_.Owner -EQ 'containerd-shim-runhcs-v1.exe') } ; if ($vms.Length -le 0) { throw 'error' }"},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
RestartPolicy: v1.RestartPolicyNever,
|
||||||
|
NodeName: targetNode.Name,
|
||||||
|
},
|
||||||
|
}
|
||||||
|
|
||||||
|
pc.Create(ctx, pod)
|
||||||
|
ginkgo.By("waiting for the pod to be run")
|
||||||
|
pc.WaitForFinish(ctx, podName, timeout)
|
||||||
|
|
||||||
|
ginkgo.By("then ensuring pod finished running successfully")
|
||||||
|
p, err := f.ClientSet.CoreV1().Pods(f.Namespace.Name).Get(ctx, podName, metav1.GetOptions{})
|
||||||
|
framework.ExpectNoError(err, "error getting pod")
|
||||||
|
|
||||||
|
if p.Status.Phase != v1.PodSucceeded {
|
||||||
|
logs, err := e2epod.GetPodLogs(ctx, f.ClientSet, f.Namespace.Name, podName, "container")
|
||||||
|
if err != nil {
|
||||||
|
framework.Logf("Error pulling logs: %v", err)
|
||||||
|
}
|
||||||
|
framework.Logf("Pod phase: %v\nlogs:\n%s", p.Status.Phase, logs)
|
||||||
|
}
|
||||||
|
|
||||||
|
framework.ExpectEqual(p.Status.Phase, v1.PodSucceeded, "pod should have succeeded")
|
||||||
|
})
|
||||||
|
})
|
@ -17,17 +17,20 @@ limitations under the License.
|
|||||||
package windows
|
package windows
|
||||||
|
|
||||||
import (
|
import (
|
||||||
|
"context"
|
||||||
"fmt"
|
"fmt"
|
||||||
"math/rand"
|
"math/rand"
|
||||||
"strings"
|
"strings"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
|
"github.com/onsi/ginkgo/v2"
|
||||||
appsv1 "k8s.io/api/apps/v1"
|
appsv1 "k8s.io/api/apps/v1"
|
||||||
v1 "k8s.io/api/core/v1"
|
v1 "k8s.io/api/core/v1"
|
||||||
apierrors "k8s.io/apimachinery/pkg/api/errors"
|
apierrors "k8s.io/apimachinery/pkg/api/errors"
|
||||||
"k8s.io/apimachinery/pkg/util/wait"
|
"k8s.io/apimachinery/pkg/util/wait"
|
||||||
"k8s.io/kubernetes/pkg/controller/deployment/util"
|
"k8s.io/kubernetes/pkg/controller/deployment/util"
|
||||||
"k8s.io/kubernetes/test/e2e/framework"
|
"k8s.io/kubernetes/test/e2e/framework"
|
||||||
|
e2eskipper "k8s.io/kubernetes/test/e2e/framework/skipper"
|
||||||
|
|
||||||
semver "github.com/blang/semver/v4"
|
semver "github.com/blang/semver/v4"
|
||||||
)
|
)
|
||||||
@ -78,3 +81,21 @@ func getRandomUserGrounName() string {
|
|||||||
|
|
||||||
return "hpc-" + string(s)
|
return "hpc-" + string(s)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func skipUnlessContainerdOneSevenOrGreater(ctx context.Context, f *framework.Framework) {
|
||||||
|
ginkgo.By("Ensuring Windows nodes are running containerd v1.7+")
|
||||||
|
windowsNode, err := findWindowsNode(ctx, f)
|
||||||
|
framework.ExpectNoError(err, "error finding Windows node")
|
||||||
|
r, v, err := getNodeContainerRuntimeAndVersion(windowsNode)
|
||||||
|
framework.ExpectNoError(err, "error getting node container runtime and version")
|
||||||
|
framework.Logf("Got runtime: %s, version %v for node %s", r, v, windowsNode.Name)
|
||||||
|
|
||||||
|
if !strings.EqualFold(r, "containerd") {
|
||||||
|
e2eskipper.Skipf("container runtime is not containerd")
|
||||||
|
}
|
||||||
|
|
||||||
|
v1dot7 := semver.MustParse("1.7.0-alpha.1")
|
||||||
|
if v.LT(v1dot7) {
|
||||||
|
e2eskipper.Skipf("container runtime is < 1.7.0")
|
||||||
|
}
|
||||||
|
}
|
||||||
|
Loading…
Reference in New Issue
Block a user