mirror of
				https://github.com/k3s-io/kubernetes.git
				synced 2025-10-30 21:30:16 +00:00 
			
		
		
		
	Use the same pattern everywhere in the e2e test harness, use busybox (from dockerhub) instead of using the one from k8s.gcr.io registry. Change-Id: I57c3b867408c1f9478a8909c26744ea0368ff003
		
			
				
	
	
		
			586 lines
		
	
	
		
			19 KiB
		
	
	
	
		
			Go
		
	
	
	
	
	
			
		
		
	
	
			586 lines
		
	
	
		
			19 KiB
		
	
	
	
		
			Go
		
	
	
	
	
	
| /*
 | |
| Copyright 2017 The Kubernetes Authors.
 | |
| 
 | |
| Licensed under the Apache License, Version 2.0 (the "License");
 | |
| you may not use this file except in compliance with the License.
 | |
| You may obtain a copy of the License at
 | |
| 
 | |
|     http://www.apache.org/licenses/LICENSE-2.0
 | |
| 
 | |
| Unless required by applicable law or agreed to in writing, software
 | |
| distributed under the License is distributed on an "AS IS" BASIS,
 | |
| WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 | |
| See the License for the specific language governing permissions and
 | |
| limitations under the License.
 | |
| */
 | |
| 
 | |
| package e2e_node
 | |
| 
 | |
| import (
 | |
| 	"fmt"
 | |
| 	"net"
 | |
| 	"os/exec"
 | |
| 	"strings"
 | |
| 
 | |
| 	"k8s.io/api/core/v1"
 | |
| 	metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
 | |
| 	"k8s.io/apimachinery/pkg/util/sets"
 | |
| 	"k8s.io/apimachinery/pkg/util/uuid"
 | |
| 	utilfeature "k8s.io/apiserver/pkg/util/feature"
 | |
| 	"k8s.io/kubernetes/pkg/features"
 | |
| 	"k8s.io/kubernetes/test/e2e/framework"
 | |
| 
 | |
| 	. "github.com/onsi/ginkgo"
 | |
| 	imageutils "k8s.io/kubernetes/test/utils/image"
 | |
| )
 | |
| 
 | |
| var _ = framework.KubeDescribe("Security Context", func() {
 | |
| 	f := framework.NewDefaultFramework("security-context-test")
 | |
| 	var podClient *framework.PodClient
 | |
| 	BeforeEach(func() {
 | |
| 		podClient = f.PodClient()
 | |
| 	})
 | |
| 
 | |
| 	Context("when pod PID namespace is configurable [Feature:ShareProcessNamespace][NodeAlphaFeature:ShareProcessNamespace]", func() {
 | |
| 		It("containers in pods using isolated PID namespaces should all receive PID 1", func() {
 | |
| 			By("Create a pod with isolated PID namespaces.")
 | |
| 			f.PodClient().CreateSync(&v1.Pod{
 | |
| 				ObjectMeta: metav1.ObjectMeta{Name: "isolated-pid-ns-test-pod"},
 | |
| 				Spec: v1.PodSpec{
 | |
| 					Containers: []v1.Container{
 | |
| 						{
 | |
| 							Name:    "test-container-1",
 | |
| 							Image:   imageutils.GetE2EImage(imageutils.BusyBox),
 | |
| 							Command: []string{"/bin/top"},
 | |
| 						},
 | |
| 						{
 | |
| 							Name:    "test-container-2",
 | |
| 							Image:   imageutils.GetE2EImage(imageutils.BusyBox),
 | |
| 							Command: []string{"/bin/sleep"},
 | |
| 							Args:    []string{"10000"},
 | |
| 						},
 | |
| 					},
 | |
| 				},
 | |
| 			})
 | |
| 
 | |
| 			By("Check if both containers receive PID 1.")
 | |
| 			pid1 := f.ExecCommandInContainer("isolated-pid-ns-test-pod", "test-container-1", "/bin/pidof", "top")
 | |
| 			pid2 := f.ExecCommandInContainer("isolated-pid-ns-test-pod", "test-container-2", "/bin/pidof", "sleep")
 | |
| 			if pid1 != "1" || pid2 != "1" {
 | |
| 				framework.Failf("PIDs of different containers are not all 1: test-container-1=%v, test-container-2=%v", pid1, pid2)
 | |
| 			}
 | |
| 		})
 | |
| 
 | |
| 		It("processes in containers sharing a pod namespace should be able to see each other [Alpha]", func() {
 | |
| 			By("Check whether shared PID namespace is supported.")
 | |
| 			isEnabled, err := isSharedPIDNamespaceSupported()
 | |
| 			framework.ExpectNoError(err)
 | |
| 			if !isEnabled {
 | |
| 				framework.Skipf("Skipped because shared PID namespace is not supported by this docker version.")
 | |
| 			}
 | |
| 			// It's not enough to set this flag in the kubelet because the apiserver needs it too
 | |
| 			if !utilfeature.DefaultFeatureGate.Enabled(features.PodShareProcessNamespace) {
 | |
| 				framework.Skipf("run test with --feature-gates=PodShareProcessNamespace=true to test PID namespace sharing")
 | |
| 			}
 | |
| 
 | |
| 			By("Create a pod with shared PID namespace.")
 | |
| 			f.PodClient().CreateSync(&v1.Pod{
 | |
| 				ObjectMeta: metav1.ObjectMeta{Name: "shared-pid-ns-test-pod"},
 | |
| 				Spec: v1.PodSpec{
 | |
| 					ShareProcessNamespace: &[]bool{true}[0],
 | |
| 					Containers: []v1.Container{
 | |
| 						{
 | |
| 							Name:    "test-container-1",
 | |
| 							Image:   imageutils.GetE2EImage(imageutils.BusyBox),
 | |
| 							Command: []string{"/bin/top"},
 | |
| 						},
 | |
| 						{
 | |
| 							Name:    "test-container-2",
 | |
| 							Image:   imageutils.GetE2EImage(imageutils.BusyBox),
 | |
| 							Command: []string{"/bin/sleep"},
 | |
| 							Args:    []string{"10000"},
 | |
| 						},
 | |
| 					},
 | |
| 				},
 | |
| 			})
 | |
| 
 | |
| 			By("Check if the process in one container is visible to the process in the other.")
 | |
| 			pid1 := f.ExecCommandInContainer("shared-pid-ns-test-pod", "test-container-1", "/bin/pidof", "top")
 | |
| 			pid2 := f.ExecCommandInContainer("shared-pid-ns-test-pod", "test-container-2", "/bin/pidof", "top")
 | |
| 			if pid1 != pid2 {
 | |
| 				framework.Failf("PIDs are not the same in different containers: test-container-1=%v, test-container-2=%v", pid1, pid2)
 | |
| 			}
 | |
| 		})
 | |
| 	})
 | |
| 
 | |
| 	Context("when creating a pod in the host PID namespace", func() {
 | |
| 		makeHostPidPod := func(podName, image string, command []string, hostPID bool) *v1.Pod {
 | |
| 			return &v1.Pod{
 | |
| 				ObjectMeta: metav1.ObjectMeta{
 | |
| 					Name: podName,
 | |
| 				},
 | |
| 				Spec: v1.PodSpec{
 | |
| 					RestartPolicy: v1.RestartPolicyNever,
 | |
| 					HostPID:       hostPID,
 | |
| 					Containers: []v1.Container{
 | |
| 						{
 | |
| 							Image:   image,
 | |
| 							Name:    podName,
 | |
| 							Command: command,
 | |
| 						},
 | |
| 					},
 | |
| 				},
 | |
| 			}
 | |
| 		}
 | |
| 		createAndWaitHostPidPod := func(podName string, hostPID bool) {
 | |
| 			podClient.Create(makeHostPidPod(podName,
 | |
| 				busyboxImage,
 | |
| 				[]string{"sh", "-c", "pidof nginx || true"},
 | |
| 				hostPID,
 | |
| 			))
 | |
| 
 | |
| 			podClient.WaitForSuccess(podName, framework.PodStartTimeout)
 | |
| 		}
 | |
| 
 | |
| 		nginxPid := ""
 | |
| 		BeforeEach(func() {
 | |
| 			nginxPodName := "nginx-hostpid-" + string(uuid.NewUUID())
 | |
| 			podClient.CreateSync(makeHostPidPod(nginxPodName,
 | |
| 				imageutils.GetE2EImage(imageutils.Nginx),
 | |
| 				nil,
 | |
| 				true,
 | |
| 			))
 | |
| 
 | |
| 			output := f.ExecShellInContainer(nginxPodName, nginxPodName,
 | |
| 				"cat /var/run/nginx.pid")
 | |
| 			nginxPid = strings.TrimSpace(output)
 | |
| 		})
 | |
| 
 | |
| 		It("should show its pid in the host PID namespace [NodeFeature:HostAccess]", func() {
 | |
| 			busyboxPodName := "busybox-hostpid-" + string(uuid.NewUUID())
 | |
| 			createAndWaitHostPidPod(busyboxPodName, true)
 | |
| 			logs, err := framework.GetPodLogs(f.ClientSet, f.Namespace.Name, busyboxPodName, busyboxPodName)
 | |
| 			if err != nil {
 | |
| 				framework.Failf("GetPodLogs for pod %q failed: %v", busyboxPodName, err)
 | |
| 			}
 | |
| 
 | |
| 			pids := strings.TrimSpace(logs)
 | |
| 			framework.Logf("Got nginx's pid %q from pod %q", pids, busyboxPodName)
 | |
| 			if pids == "" {
 | |
| 				framework.Failf("nginx's pid should be seen by hostpid containers")
 | |
| 			}
 | |
| 
 | |
| 			pidSets := sets.NewString(strings.Split(pids, " ")...)
 | |
| 			if !pidSets.Has(nginxPid) {
 | |
| 				framework.Failf("nginx's pid should be seen by hostpid containers")
 | |
| 			}
 | |
| 		})
 | |
| 
 | |
| 		It("should not show its pid in the non-hostpid containers [NodeFeature:HostAccess]", func() {
 | |
| 			busyboxPodName := "busybox-non-hostpid-" + string(uuid.NewUUID())
 | |
| 			createAndWaitHostPidPod(busyboxPodName, false)
 | |
| 			logs, err := framework.GetPodLogs(f.ClientSet, f.Namespace.Name, busyboxPodName, busyboxPodName)
 | |
| 			if err != nil {
 | |
| 				framework.Failf("GetPodLogs for pod %q failed: %v", busyboxPodName, err)
 | |
| 			}
 | |
| 
 | |
| 			pids := strings.TrimSpace(logs)
 | |
| 			framework.Logf("Got nginx's pid %q from pod %q", pids, busyboxPodName)
 | |
| 			pidSets := sets.NewString(strings.Split(pids, " ")...)
 | |
| 			if pidSets.Has(nginxPid) {
 | |
| 				framework.Failf("nginx's pid should not be seen by non-hostpid containers")
 | |
| 			}
 | |
| 		})
 | |
| 	})
 | |
| 
 | |
| 	Context("when creating a pod in the host IPC namespace", func() {
 | |
| 		makeHostIPCPod := func(podName, image string, command []string, hostIPC bool) *v1.Pod {
 | |
| 			return &v1.Pod{
 | |
| 				ObjectMeta: metav1.ObjectMeta{
 | |
| 					Name: podName,
 | |
| 				},
 | |
| 				Spec: v1.PodSpec{
 | |
| 					RestartPolicy: v1.RestartPolicyNever,
 | |
| 					HostIPC:       hostIPC,
 | |
| 					Containers: []v1.Container{
 | |
| 						{
 | |
| 							Image:   image,
 | |
| 							Name:    podName,
 | |
| 							Command: command,
 | |
| 						},
 | |
| 					},
 | |
| 				},
 | |
| 			}
 | |
| 		}
 | |
| 		createAndWaitHostIPCPod := func(podName string, hostNetwork bool) {
 | |
| 			podClient.Create(makeHostIPCPod(podName,
 | |
| 				imageutils.GetE2EImage(imageutils.IpcUtils),
 | |
| 				[]string{"sh", "-c", "ipcs -m | awk '{print $2}'"},
 | |
| 				hostNetwork,
 | |
| 			))
 | |
| 
 | |
| 			podClient.WaitForSuccess(podName, framework.PodStartTimeout)
 | |
| 		}
 | |
| 
 | |
| 		hostSharedMemoryID := ""
 | |
| 		BeforeEach(func() {
 | |
| 			output, err := exec.Command("sh", "-c", "ipcmk -M 1048576 | awk '{print $NF}'").Output()
 | |
| 			if err != nil {
 | |
| 				framework.Failf("Failed to create the shared memory on the host: %v", err)
 | |
| 			}
 | |
| 			hostSharedMemoryID = strings.TrimSpace(string(output))
 | |
| 			framework.Logf("Got host shared memory ID %q", hostSharedMemoryID)
 | |
| 		})
 | |
| 
 | |
| 		It("should show the shared memory ID in the host IPC containers [NodeFeature:HostAccess]", func() {
 | |
| 			ipcutilsPodName := "ipcutils-hostipc-" + string(uuid.NewUUID())
 | |
| 			createAndWaitHostIPCPod(ipcutilsPodName, true)
 | |
| 			logs, err := framework.GetPodLogs(f.ClientSet, f.Namespace.Name, ipcutilsPodName, ipcutilsPodName)
 | |
| 			if err != nil {
 | |
| 				framework.Failf("GetPodLogs for pod %q failed: %v", ipcutilsPodName, err)
 | |
| 			}
 | |
| 
 | |
| 			podSharedMemoryIDs := strings.TrimSpace(logs)
 | |
| 			framework.Logf("Got shared memory IDs %q from pod %q", podSharedMemoryIDs, ipcutilsPodName)
 | |
| 			if !strings.Contains(podSharedMemoryIDs, hostSharedMemoryID) {
 | |
| 				framework.Failf("hostIPC container should show shared memory IDs on host")
 | |
| 			}
 | |
| 		})
 | |
| 
 | |
| 		It("should not show the shared memory ID in the non-hostIPC containers [NodeFeature:HostAccess]", func() {
 | |
| 			ipcutilsPodName := "ipcutils-non-hostipc-" + string(uuid.NewUUID())
 | |
| 			createAndWaitHostIPCPod(ipcutilsPodName, false)
 | |
| 			logs, err := framework.GetPodLogs(f.ClientSet, f.Namespace.Name, ipcutilsPodName, ipcutilsPodName)
 | |
| 			if err != nil {
 | |
| 				framework.Failf("GetPodLogs for pod %q failed: %v", ipcutilsPodName, err)
 | |
| 			}
 | |
| 
 | |
| 			podSharedMemoryIDs := strings.TrimSpace(logs)
 | |
| 			framework.Logf("Got shared memory IDs %q from pod %q", podSharedMemoryIDs, ipcutilsPodName)
 | |
| 			if strings.Contains(podSharedMemoryIDs, hostSharedMemoryID) {
 | |
| 				framework.Failf("non-hostIPC container should not show shared memory IDs on host")
 | |
| 			}
 | |
| 		})
 | |
| 
 | |
| 		AfterEach(func() {
 | |
| 			if hostSharedMemoryID != "" {
 | |
| 				_, err := exec.Command("sh", "-c", fmt.Sprintf("ipcrm -m %q", hostSharedMemoryID)).Output()
 | |
| 				if err != nil {
 | |
| 					framework.Failf("Failed to remove shared memory %q on the host: %v", hostSharedMemoryID, err)
 | |
| 				}
 | |
| 			}
 | |
| 		})
 | |
| 	})
 | |
| 
 | |
| 	Context("when creating a pod in the host network namespace", func() {
 | |
| 		makeHostNetworkPod := func(podName, image string, command []string, hostNetwork bool) *v1.Pod {
 | |
| 			return &v1.Pod{
 | |
| 				ObjectMeta: metav1.ObjectMeta{
 | |
| 					Name: podName,
 | |
| 				},
 | |
| 				Spec: v1.PodSpec{
 | |
| 					RestartPolicy: v1.RestartPolicyNever,
 | |
| 					HostNetwork:   hostNetwork,
 | |
| 					Containers: []v1.Container{
 | |
| 						{
 | |
| 							Image:   image,
 | |
| 							Name:    podName,
 | |
| 							Command: command,
 | |
| 						},
 | |
| 					},
 | |
| 				},
 | |
| 			}
 | |
| 		}
 | |
| 		listListeningPortsCommand := []string{"sh", "-c", "netstat -ln"}
 | |
| 		createAndWaitHostNetworkPod := func(podName string, hostNetwork bool) {
 | |
| 			podClient.Create(makeHostNetworkPod(podName,
 | |
| 				busyboxImage,
 | |
| 				listListeningPortsCommand,
 | |
| 				hostNetwork,
 | |
| 			))
 | |
| 
 | |
| 			podClient.WaitForSuccess(podName, framework.PodStartTimeout)
 | |
| 		}
 | |
| 
 | |
| 		listeningPort := ""
 | |
| 		var l net.Listener
 | |
| 		var err error
 | |
| 		BeforeEach(func() {
 | |
| 			l, err = net.Listen("tcp", ":0")
 | |
| 			if err != nil {
 | |
| 				framework.Failf("Failed to open a new tcp port: %v", err)
 | |
| 			}
 | |
| 			addr := strings.Split(l.Addr().String(), ":")
 | |
| 			listeningPort = addr[len(addr)-1]
 | |
| 			framework.Logf("Opened a new tcp port %q", listeningPort)
 | |
| 		})
 | |
| 
 | |
| 		It("should listen on same port in the host network containers [NodeFeature:HostAccess]", func() {
 | |
| 			busyboxPodName := "busybox-hostnetwork-" + string(uuid.NewUUID())
 | |
| 			createAndWaitHostNetworkPod(busyboxPodName, true)
 | |
| 			logs, err := framework.GetPodLogs(f.ClientSet, f.Namespace.Name, busyboxPodName, busyboxPodName)
 | |
| 			if err != nil {
 | |
| 				framework.Failf("GetPodLogs for pod %q failed: %v", busyboxPodName, err)
 | |
| 			}
 | |
| 
 | |
| 			framework.Logf("Got logs for pod %q: %q", busyboxPodName, logs)
 | |
| 			if !strings.Contains(logs, listeningPort) {
 | |
| 				framework.Failf("host-networked container should listening on same port as host")
 | |
| 			}
 | |
| 		})
 | |
| 
 | |
| 		It("shouldn't show the same port in the non-hostnetwork containers [NodeFeature:HostAccess]", func() {
 | |
| 			busyboxPodName := "busybox-non-hostnetwork-" + string(uuid.NewUUID())
 | |
| 			createAndWaitHostNetworkPod(busyboxPodName, false)
 | |
| 			logs, err := framework.GetPodLogs(f.ClientSet, f.Namespace.Name, busyboxPodName, busyboxPodName)
 | |
| 			if err != nil {
 | |
| 				framework.Failf("GetPodLogs for pod %q failed: %v", busyboxPodName, err)
 | |
| 			}
 | |
| 
 | |
| 			framework.Logf("Got logs for pod %q: %q", busyboxPodName, logs)
 | |
| 			if strings.Contains(logs, listeningPort) {
 | |
| 				framework.Failf("non-hostnetworked container shouldn't show the same port as host")
 | |
| 			}
 | |
| 		})
 | |
| 
 | |
| 		AfterEach(func() {
 | |
| 			if l != nil {
 | |
| 				l.Close()
 | |
| 			}
 | |
| 		})
 | |
| 	})
 | |
| 
 | |
| 	Context("When creating a container with runAsUser", func() {
 | |
| 		makeUserPod := func(podName, image string, command []string, userid int64) *v1.Pod {
 | |
| 			return &v1.Pod{
 | |
| 				ObjectMeta: metav1.ObjectMeta{
 | |
| 					Name: podName,
 | |
| 				},
 | |
| 				Spec: v1.PodSpec{
 | |
| 					RestartPolicy: v1.RestartPolicyNever,
 | |
| 					Containers: []v1.Container{
 | |
| 						{
 | |
| 							Image:   image,
 | |
| 							Name:    podName,
 | |
| 							Command: command,
 | |
| 							SecurityContext: &v1.SecurityContext{
 | |
| 								RunAsUser: &userid,
 | |
| 							},
 | |
| 						},
 | |
| 					},
 | |
| 				},
 | |
| 			}
 | |
| 		}
 | |
| 		createAndWaitUserPod := func(userid int64) {
 | |
| 			podName := fmt.Sprintf("busybox-user-%d-%s", userid, uuid.NewUUID())
 | |
| 			podClient.Create(makeUserPod(podName,
 | |
| 				busyboxImage,
 | |
| 				[]string{"sh", "-c", fmt.Sprintf("test $(id -u) -eq %d", userid)},
 | |
| 				userid,
 | |
| 			))
 | |
| 
 | |
| 			podClient.WaitForSuccess(podName, framework.PodStartTimeout)
 | |
| 		}
 | |
| 
 | |
| 		It("should run the container with uid 65534 [NodeConformance]", func() {
 | |
| 			createAndWaitUserPod(65534)
 | |
| 		})
 | |
| 
 | |
| 		It("should run the container with uid 0 [NodeConformance]", func() {
 | |
| 			createAndWaitUserPod(0)
 | |
| 		})
 | |
| 	})
 | |
| 
 | |
| 	Context("When creating a pod with readOnlyRootFilesystem", func() {
 | |
| 		makeUserPod := func(podName, image string, command []string, readOnlyRootFilesystem bool) *v1.Pod {
 | |
| 			return &v1.Pod{
 | |
| 				ObjectMeta: metav1.ObjectMeta{
 | |
| 					Name: podName,
 | |
| 				},
 | |
| 				Spec: v1.PodSpec{
 | |
| 					RestartPolicy: v1.RestartPolicyNever,
 | |
| 					Containers: []v1.Container{
 | |
| 						{
 | |
| 							Image:   image,
 | |
| 							Name:    podName,
 | |
| 							Command: command,
 | |
| 							SecurityContext: &v1.SecurityContext{
 | |
| 								ReadOnlyRootFilesystem: &readOnlyRootFilesystem,
 | |
| 							},
 | |
| 						},
 | |
| 					},
 | |
| 				},
 | |
| 			}
 | |
| 		}
 | |
| 		createAndWaitUserPod := func(readOnlyRootFilesystem bool) string {
 | |
| 			podName := fmt.Sprintf("busybox-readonly-%v-%s", readOnlyRootFilesystem, uuid.NewUUID())
 | |
| 			podClient.Create(makeUserPod(podName,
 | |
| 				"busybox",
 | |
| 				[]string{"sh", "-c", "touch checkfile"},
 | |
| 				readOnlyRootFilesystem,
 | |
| 			))
 | |
| 
 | |
| 			if readOnlyRootFilesystem {
 | |
| 				podClient.WaitForFailure(podName, framework.PodStartTimeout)
 | |
| 			} else {
 | |
| 				podClient.WaitForSuccess(podName, framework.PodStartTimeout)
 | |
| 			}
 | |
| 
 | |
| 			return podName
 | |
| 		}
 | |
| 
 | |
| 		It("should run the container with readonly rootfs when readOnlyRootFilesystem=true [NodeConformance]", func() {
 | |
| 			createAndWaitUserPod(true)
 | |
| 		})
 | |
| 
 | |
| 		It("should run the container with writable rootfs when readOnlyRootFilesystem=false [NodeConformance]", func() {
 | |
| 			createAndWaitUserPod(false)
 | |
| 		})
 | |
| 	})
 | |
| 
 | |
| 	Context("when creating containers with AllowPrivilegeEscalation", func() {
 | |
| 
 | |
| 		BeforeEach(func() {
 | |
| 			if framework.TestContext.ContainerRuntime == "docker" {
 | |
| 				isSupported, err := isDockerNoNewPrivilegesSupported()
 | |
| 				framework.ExpectNoError(err)
 | |
| 				if !isSupported {
 | |
| 					framework.Skipf("Skipping because no_new_privs is not supported in this docker")
 | |
| 				}
 | |
| 				// It turns out SELinux policy in RHEL 7 does not play well with
 | |
| 				// the "NoNewPrivileges" flag. So let's skip this test when running
 | |
| 				// with SELinux support enabled.
 | |
| 				//
 | |
| 				// TODO(filbranden): Remove this after the fix for
 | |
| 				// https://github.com/projectatomic/container-selinux/issues/45
 | |
| 				// has been backported to RHEL 7 (expected on RHEL 7.5)
 | |
| 				selinuxEnabled, err := isDockerSELinuxSupportEnabled()
 | |
| 				framework.ExpectNoError(err)
 | |
| 				if selinuxEnabled {
 | |
| 					framework.Skipf("Skipping because Docker daemon is running with SELinux support enabled")
 | |
| 				}
 | |
| 			}
 | |
| 		})
 | |
| 
 | |
| 		makeAllowPrivilegeEscalationPod := func(podName string, allowPrivilegeEscalation *bool, uid int64) *v1.Pod {
 | |
| 			return &v1.Pod{
 | |
| 				ObjectMeta: metav1.ObjectMeta{
 | |
| 					Name: podName,
 | |
| 				},
 | |
| 				Spec: v1.PodSpec{
 | |
| 					RestartPolicy: v1.RestartPolicyNever,
 | |
| 					Containers: []v1.Container{
 | |
| 						{
 | |
| 							Image: imageutils.GetE2EImage(imageutils.Nonewprivs),
 | |
| 							Name:  podName,
 | |
| 							SecurityContext: &v1.SecurityContext{
 | |
| 								AllowPrivilegeEscalation: allowPrivilegeEscalation,
 | |
| 								RunAsUser:                &uid,
 | |
| 							},
 | |
| 						},
 | |
| 					},
 | |
| 				},
 | |
| 			}
 | |
| 		}
 | |
| 		createAndMatchOutput := func(podName, output string, allowPrivilegeEscalation *bool, uid int64) error {
 | |
| 			podClient.Create(makeAllowPrivilegeEscalationPod(podName,
 | |
| 				allowPrivilegeEscalation,
 | |
| 				uid,
 | |
| 			))
 | |
| 
 | |
| 			podClient.WaitForSuccess(podName, framework.PodStartTimeout)
 | |
| 
 | |
| 			if err := podClient.MatchContainerOutput(podName, podName, output); err != nil {
 | |
| 				return err
 | |
| 			}
 | |
| 
 | |
| 			return nil
 | |
| 		}
 | |
| 
 | |
| 		It("should allow privilege escalation when not explicitly set and uid != 0 [NodeConformance]", func() {
 | |
| 			podName := "alpine-nnp-nil-" + string(uuid.NewUUID())
 | |
| 			if err := createAndMatchOutput(podName, "Effective uid: 0", nil, 1000); err != nil {
 | |
| 				framework.Failf("Match output for pod %q failed: %v", podName, err)
 | |
| 			}
 | |
| 		})
 | |
| 
 | |
| 		It("should not allow privilege escalation when false [NodeConformance]", func() {
 | |
| 			podName := "alpine-nnp-false-" + string(uuid.NewUUID())
 | |
| 			apeFalse := false
 | |
| 			if err := createAndMatchOutput(podName, "Effective uid: 1000", &apeFalse, 1000); err != nil {
 | |
| 				framework.Failf("Match output for pod %q failed: %v", podName, err)
 | |
| 			}
 | |
| 		})
 | |
| 
 | |
| 		It("should allow privilege escalation when true [NodeConformance]", func() {
 | |
| 			podName := "alpine-nnp-true-" + string(uuid.NewUUID())
 | |
| 			apeTrue := true
 | |
| 			if err := createAndMatchOutput(podName, "Effective uid: 0", &apeTrue, 1000); err != nil {
 | |
| 				framework.Failf("Match output for pod %q failed: %v", podName, err)
 | |
| 			}
 | |
| 		})
 | |
| 	})
 | |
| 
 | |
| 	Context("When creating a pod with privileged", func() {
 | |
| 		makeUserPod := func(podName, image string, command []string, privileged bool) *v1.Pod {
 | |
| 			return &v1.Pod{
 | |
| 				ObjectMeta: metav1.ObjectMeta{
 | |
| 					Name: podName,
 | |
| 				},
 | |
| 				Spec: v1.PodSpec{
 | |
| 					RestartPolicy: v1.RestartPolicyNever,
 | |
| 					Containers: []v1.Container{
 | |
| 						{
 | |
| 							Image:   image,
 | |
| 							Name:    podName,
 | |
| 							Command: command,
 | |
| 							SecurityContext: &v1.SecurityContext{
 | |
| 								Privileged: &privileged,
 | |
| 							},
 | |
| 						},
 | |
| 					},
 | |
| 				},
 | |
| 			}
 | |
| 		}
 | |
| 		createAndWaitUserPod := func(privileged bool) string {
 | |
| 			podName := fmt.Sprintf("busybox-privileged-%v-%s", privileged, uuid.NewUUID())
 | |
| 			podClient.Create(makeUserPod(podName,
 | |
| 				busyboxImage,
 | |
| 				[]string{"sh", "-c", "ip link add dummy0 type dummy || true"},
 | |
| 				privileged,
 | |
| 			))
 | |
| 
 | |
| 			podClient.WaitForSuccess(podName, framework.PodStartTimeout)
 | |
| 
 | |
| 			return podName
 | |
| 		}
 | |
| 
 | |
| 		It("should run the container as privileged when true [NodeFeature:HostAccess]", func() {
 | |
| 			podName := createAndWaitUserPod(true)
 | |
| 			logs, err := framework.GetPodLogs(f.ClientSet, f.Namespace.Name, podName, podName)
 | |
| 			if err != nil {
 | |
| 				framework.Failf("GetPodLogs for pod %q failed: %v", podName, err)
 | |
| 			}
 | |
| 
 | |
| 			framework.Logf("Got logs for pod %q: %q", podName, logs)
 | |
| 			if strings.Contains(logs, "Operation not permitted") {
 | |
| 				framework.Failf("privileged container should be able to create dummy device")
 | |
| 			}
 | |
| 		})
 | |
| 
 | |
| 		It("should run the container as unprivileged when false [NodeConformance]", func() {
 | |
| 			podName := createAndWaitUserPod(false)
 | |
| 			logs, err := framework.GetPodLogs(f.ClientSet, f.Namespace.Name, podName, podName)
 | |
| 			if err != nil {
 | |
| 				framework.Failf("GetPodLogs for pod %q failed: %v", podName, err)
 | |
| 			}
 | |
| 
 | |
| 			framework.Logf("Got logs for pod %q: %q", podName, logs)
 | |
| 			if !strings.Contains(logs, "Operation not permitted") {
 | |
| 				framework.Failf("unprivileged container shouldn't be able to create dummy device")
 | |
| 			}
 | |
| 		})
 | |
| 	})
 | |
| 
 | |
| })
 |