mirror of
				https://github.com/k3s-io/kubernetes.git
				synced 2025-10-30 21:30:16 +00:00 
			
		
		
		
	
		
			
				
	
	
		
			720 lines
		
	
	
		
			29 KiB
		
	
	
	
		
			Go
		
	
	
	
	
	
			
		
		
	
	
			720 lines
		
	
	
		
			29 KiB
		
	
	
	
		
			Go
		
	
	
	
	
	
| /*
 | |
| Copyright 2015 The Kubernetes Authors.
 | |
| 
 | |
| Licensed under the Apache License, Version 2.0 (the "License");
 | |
| you may not use this file except in compliance with the License.
 | |
| You may obtain a copy of the License at
 | |
| 
 | |
|     http://www.apache.org/licenses/LICENSE-2.0
 | |
| 
 | |
| Unless required by applicable law or agreed to in writing, software
 | |
| distributed under the License is distributed on an "AS IS" BASIS,
 | |
| WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 | |
| See the License for the specific language governing permissions and
 | |
| limitations under the License.
 | |
| */
 | |
| 
 | |
| package storage
 | |
| 
 | |
| import (
 | |
| 	"context"
 | |
| 	"fmt"
 | |
| 	"math/rand"
 | |
| 	"strings"
 | |
| 	"time"
 | |
| 
 | |
| 	"google.golang.org/api/googleapi"
 | |
| 
 | |
| 	"github.com/aws/aws-sdk-go/aws"
 | |
| 	"github.com/aws/aws-sdk-go/aws/session"
 | |
| 	"github.com/aws/aws-sdk-go/service/ec2"
 | |
| 	"github.com/onsi/ginkgo"
 | |
| 	v1 "k8s.io/api/core/v1"
 | |
| 	policyv1 "k8s.io/api/policy/v1"
 | |
| 	"k8s.io/apimachinery/pkg/api/resource"
 | |
| 	metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
 | |
| 	"k8s.io/apimachinery/pkg/labels"
 | |
| 	"k8s.io/apimachinery/pkg/types"
 | |
| 	"k8s.io/apimachinery/pkg/util/uuid"
 | |
| 	"k8s.io/apimachinery/pkg/util/wait"
 | |
| 	clientset "k8s.io/client-go/kubernetes"
 | |
| 	v1core "k8s.io/client-go/kubernetes/typed/core/v1"
 | |
| 	"k8s.io/kubernetes/test/e2e/framework"
 | |
| 	e2ekubectl "k8s.io/kubernetes/test/e2e/framework/kubectl"
 | |
| 	e2enode "k8s.io/kubernetes/test/e2e/framework/node"
 | |
| 	e2epod "k8s.io/kubernetes/test/e2e/framework/pod"
 | |
| 	"k8s.io/kubernetes/test/e2e/framework/providers/gce"
 | |
| 	e2epv "k8s.io/kubernetes/test/e2e/framework/pv"
 | |
| 	e2eskipper "k8s.io/kubernetes/test/e2e/framework/skipper"
 | |
| 	"k8s.io/kubernetes/test/e2e/storage/utils"
 | |
| 	imageutils "k8s.io/kubernetes/test/utils/image"
 | |
| )
 | |
| 
 | |
| const (
 | |
| 	gcePDDetachTimeout  = 10 * time.Minute
 | |
| 	gcePDDetachPollTime = 10 * time.Second
 | |
| 	nodeStatusTimeout   = 10 * time.Minute
 | |
| 	nodeStatusPollTime  = 1 * time.Second
 | |
| 	podEvictTimeout     = 2 * time.Minute
 | |
| 	minNodes            = 2
 | |
| )
 | |
| 
 | |
| var _ = utils.SIGDescribe("Pod Disks", func() {
 | |
| 	var (
 | |
| 		ns         string
 | |
| 		cs         clientset.Interface
 | |
| 		podClient  v1core.PodInterface
 | |
| 		nodeClient v1core.NodeInterface
 | |
| 		host0Name  types.NodeName
 | |
| 		host1Name  types.NodeName
 | |
| 		nodes      *v1.NodeList
 | |
| 	)
 | |
| 	f := framework.NewDefaultFramework("pod-disks")
 | |
| 
 | |
| 	ginkgo.BeforeEach(func() {
 | |
| 		e2eskipper.SkipUnlessNodeCountIsAtLeast(minNodes)
 | |
| 		cs = f.ClientSet
 | |
| 		ns = f.Namespace.Name
 | |
| 
 | |
| 		e2eskipper.SkipIfMultizone(cs)
 | |
| 
 | |
| 		podClient = cs.CoreV1().Pods(ns)
 | |
| 		nodeClient = cs.CoreV1().Nodes()
 | |
| 		var err error
 | |
| 		nodes, err = e2enode.GetReadySchedulableNodes(cs)
 | |
| 		framework.ExpectNoError(err)
 | |
| 		if len(nodes.Items) < minNodes {
 | |
| 			e2eskipper.Skipf("The test requires %d schedulable nodes, got only %d", minNodes, len(nodes.Items))
 | |
| 		}
 | |
| 		host0Name = types.NodeName(nodes.Items[0].ObjectMeta.Name)
 | |
| 		host1Name = types.NodeName(nodes.Items[1].ObjectMeta.Name)
 | |
| 	})
 | |
| 
 | |
| 	ginkgo.Context("schedule pods each with a PD, delete pod and verify detach [Slow]", func() {
 | |
| 		const (
 | |
| 			podDefaultGrace   = "default (30s)"
 | |
| 			podImmediateGrace = "immediate (0s)"
 | |
| 		)
 | |
| 		var readOnlyMap = map[bool]string{
 | |
| 			true:  "read-only",
 | |
| 			false: "RW",
 | |
| 		}
 | |
| 		type testT struct {
 | |
| 			descr     string               // It description
 | |
| 			readOnly  bool                 // true means pd is read-only
 | |
| 			deleteOpt metav1.DeleteOptions // pod delete option
 | |
| 		}
 | |
| 		tests := []testT{
 | |
| 			{
 | |
| 				descr:     podImmediateGrace,
 | |
| 				readOnly:  false,
 | |
| 				deleteOpt: *metav1.NewDeleteOptions(0),
 | |
| 			},
 | |
| 			{
 | |
| 				descr:     podDefaultGrace,
 | |
| 				readOnly:  false,
 | |
| 				deleteOpt: metav1.DeleteOptions{},
 | |
| 			},
 | |
| 			{
 | |
| 				descr:     podImmediateGrace,
 | |
| 				readOnly:  true,
 | |
| 				deleteOpt: *metav1.NewDeleteOptions(0),
 | |
| 			},
 | |
| 			{
 | |
| 				descr:     podDefaultGrace,
 | |
| 				readOnly:  true,
 | |
| 				deleteOpt: metav1.DeleteOptions{},
 | |
| 			},
 | |
| 		}
 | |
| 
 | |
| 		for _, t := range tests {
 | |
| 			podDelOpt := t.deleteOpt
 | |
| 			readOnly := t.readOnly
 | |
| 			readOnlyTxt := readOnlyMap[readOnly]
 | |
| 
 | |
| 			ginkgo.It(fmt.Sprintf("for %s PD with pod delete grace period of %q", readOnlyTxt, t.descr), func() {
 | |
| 				e2eskipper.SkipUnlessProviderIs("gce", "gke", "aws")
 | |
| 				if readOnly {
 | |
| 					e2eskipper.SkipIfProviderIs("aws")
 | |
| 				}
 | |
| 
 | |
| 				ginkgo.By("creating PD")
 | |
| 				diskName, err := e2epv.CreatePDWithRetry()
 | |
| 				framework.ExpectNoError(err, "Error creating PD")
 | |
| 
 | |
| 				var fmtPod *v1.Pod
 | |
| 				if readOnly {
 | |
| 					// if all test pods are RO then need a RW pod to format pd
 | |
| 					ginkgo.By("creating RW fmt Pod to ensure PD is formatted")
 | |
| 					fmtPod = testPDPod([]string{diskName}, host0Name, false, 1)
 | |
| 					_, err = podClient.Create(context.TODO(), fmtPod, metav1.CreateOptions{})
 | |
| 					framework.ExpectNoError(err, "Failed to create fmtPod")
 | |
| 					framework.ExpectNoError(e2epod.WaitTimeoutForPodRunningInNamespace(f.ClientSet, fmtPod.Name, f.Namespace.Name, f.Timeouts.PodStartSlow))
 | |
| 
 | |
| 					ginkgo.By("deleting the fmtPod")
 | |
| 					framework.ExpectNoError(podClient.Delete(context.TODO(), fmtPod.Name, *metav1.NewDeleteOptions(0)), "Failed to delete fmtPod")
 | |
| 					framework.Logf("deleted fmtPod %q", fmtPod.Name)
 | |
| 					ginkgo.By("waiting for PD to detach")
 | |
| 					framework.ExpectNoError(waitForPDDetach(diskName, host0Name))
 | |
| 				}
 | |
| 
 | |
| 				// prepare to create two test pods on separate nodes
 | |
| 				host0Pod := testPDPod([]string{diskName}, host0Name, readOnly, 1)
 | |
| 				host1Pod := testPDPod([]string{diskName}, host1Name, readOnly, 1)
 | |
| 
 | |
| 				defer func() {
 | |
| 					// Teardown should do nothing unless test failed
 | |
| 					ginkgo.By("defer: cleaning up PD-RW test environment")
 | |
| 					framework.Logf("defer cleanup errors can usually be ignored")
 | |
| 					if fmtPod != nil {
 | |
| 						podClient.Delete(context.TODO(), fmtPod.Name, podDelOpt)
 | |
| 					}
 | |
| 					podClient.Delete(context.TODO(), host0Pod.Name, podDelOpt)
 | |
| 					podClient.Delete(context.TODO(), host1Pod.Name, podDelOpt)
 | |
| 					detachAndDeletePDs(diskName, []types.NodeName{host0Name, host1Name})
 | |
| 				}()
 | |
| 
 | |
| 				ginkgo.By("creating host0Pod on node0")
 | |
| 				_, err = podClient.Create(context.TODO(), host0Pod, metav1.CreateOptions{})
 | |
| 				framework.ExpectNoError(err, fmt.Sprintf("Failed to create host0Pod: %v", err))
 | |
| 				framework.ExpectNoError(e2epod.WaitTimeoutForPodRunningInNamespace(f.ClientSet, host0Pod.Name, f.Namespace.Name, f.Timeouts.PodStartSlow))
 | |
| 				framework.Logf("host0Pod: %q, node0: %q", host0Pod.Name, host0Name)
 | |
| 
 | |
| 				var containerName, testFile, testFileContents string
 | |
| 				if !readOnly {
 | |
| 					ginkgo.By("writing content to host0Pod on node0")
 | |
| 					containerName = "mycontainer"
 | |
| 					testFile = "/testpd1/tracker"
 | |
| 					testFileContents = fmt.Sprintf("%v", rand.Int())
 | |
| 					tk := e2ekubectl.NewTestKubeconfig(framework.TestContext.CertDir, framework.TestContext.Host, framework.TestContext.KubeConfig, framework.TestContext.KubeContext, framework.TestContext.KubectlPath, ns)
 | |
| 					framework.ExpectNoError(tk.WriteFileViaContainer(host0Pod.Name, containerName, testFile, testFileContents))
 | |
| 					framework.Logf("wrote %q to file %q in pod %q on node %q", testFileContents, testFile, host0Pod.Name, host0Name)
 | |
| 					ginkgo.By("verifying PD is present in node0's VolumeInUse list")
 | |
| 					framework.ExpectNoError(waitForPDInVolumesInUse(nodeClient, diskName, host0Name, nodeStatusTimeout, true /* shouldExist */))
 | |
| 					ginkgo.By("deleting host0Pod") // delete this pod before creating next pod
 | |
| 					framework.ExpectNoError(podClient.Delete(context.TODO(), host0Pod.Name, podDelOpt), "Failed to delete host0Pod")
 | |
| 					framework.Logf("deleted host0Pod %q", host0Pod.Name)
 | |
| 					e2epod.WaitForPodToDisappear(cs, host0Pod.Namespace, host0Pod.Name, labels.Everything(), framework.Poll, framework.PodDeleteTimeout)
 | |
| 					framework.Logf("deleted host0Pod %q disappeared", host0Pod.Name)
 | |
| 				}
 | |
| 
 | |
| 				ginkgo.By("creating host1Pod on node1")
 | |
| 				_, err = podClient.Create(context.TODO(), host1Pod, metav1.CreateOptions{})
 | |
| 				framework.ExpectNoError(err, "Failed to create host1Pod")
 | |
| 				framework.ExpectNoError(e2epod.WaitTimeoutForPodRunningInNamespace(f.ClientSet, host1Pod.Name, f.Namespace.Name, f.Timeouts.PodStartSlow))
 | |
| 				framework.Logf("host1Pod: %q, node1: %q", host1Pod.Name, host1Name)
 | |
| 
 | |
| 				if readOnly {
 | |
| 					ginkgo.By("deleting host0Pod")
 | |
| 					framework.ExpectNoError(podClient.Delete(context.TODO(), host0Pod.Name, podDelOpt), "Failed to delete host0Pod")
 | |
| 					framework.Logf("deleted host0Pod %q", host0Pod.Name)
 | |
| 				} else {
 | |
| 					ginkgo.By("verifying PD contents in host1Pod")
 | |
| 					verifyPDContentsViaContainer(ns, f, host1Pod.Name, containerName, map[string]string{testFile: testFileContents})
 | |
| 					framework.Logf("verified PD contents in pod %q", host1Pod.Name)
 | |
| 					ginkgo.By("verifying PD is removed from node0")
 | |
| 					framework.ExpectNoError(waitForPDInVolumesInUse(nodeClient, diskName, host0Name, nodeStatusTimeout, false /* shouldExist */))
 | |
| 					framework.Logf("PD %q removed from node %q's VolumeInUse list", diskName, host1Pod.Name)
 | |
| 				}
 | |
| 
 | |
| 				ginkgo.By("deleting host1Pod")
 | |
| 				framework.ExpectNoError(podClient.Delete(context.TODO(), host1Pod.Name, podDelOpt), "Failed to delete host1Pod")
 | |
| 				framework.Logf("deleted host1Pod %q", host1Pod.Name)
 | |
| 
 | |
| 				ginkgo.By("Test completed successfully, waiting for PD to detach from both nodes")
 | |
| 				waitForPDDetach(diskName, host0Name)
 | |
| 				waitForPDDetach(diskName, host1Name)
 | |
| 			})
 | |
| 		}
 | |
| 	})
 | |
| 
 | |
| 	ginkgo.Context("schedule a pod w/ RW PD(s) mounted to 1 or more containers, write to PD, verify content, delete pod, and repeat in rapid succession [Slow]", func() {
 | |
| 		type testT struct {
 | |
| 			numContainers int
 | |
| 			numPDs        int
 | |
| 			repeatCnt     int
 | |
| 		}
 | |
| 		tests := []testT{
 | |
| 			{
 | |
| 				numContainers: 4,
 | |
| 				numPDs:        1,
 | |
| 				repeatCnt:     3,
 | |
| 			},
 | |
| 			{
 | |
| 				numContainers: 1,
 | |
| 				numPDs:        2,
 | |
| 				repeatCnt:     3,
 | |
| 			},
 | |
| 		}
 | |
| 
 | |
| 		for _, t := range tests {
 | |
| 			numPDs := t.numPDs
 | |
| 			numContainers := t.numContainers
 | |
| 
 | |
| 			ginkgo.It(fmt.Sprintf("using %d containers and %d PDs", numContainers, numPDs), func() {
 | |
| 				e2eskipper.SkipUnlessProviderIs("gce", "gke", "aws")
 | |
| 				var host0Pod *v1.Pod
 | |
| 				var err error
 | |
| 				fileAndContentToVerify := make(map[string]string)
 | |
| 				diskNames := make([]string, 0, numPDs)
 | |
| 
 | |
| 				ginkgo.By(fmt.Sprintf("creating %d PD(s)", numPDs))
 | |
| 				for i := 0; i < numPDs; i++ {
 | |
| 					name, err := e2epv.CreatePDWithRetry()
 | |
| 					framework.ExpectNoError(err, fmt.Sprintf("Error creating PD %d", i))
 | |
| 					diskNames = append(diskNames, name)
 | |
| 				}
 | |
| 
 | |
| 				defer func() {
 | |
| 					// Teardown should do nothing unless test failed.
 | |
| 					ginkgo.By("defer: cleaning up PD-RW test environment")
 | |
| 					framework.Logf("defer cleanup errors can usually be ignored")
 | |
| 					if host0Pod != nil {
 | |
| 						podClient.Delete(context.TODO(), host0Pod.Name, *metav1.NewDeleteOptions(0))
 | |
| 					}
 | |
| 					for _, diskName := range diskNames {
 | |
| 						detachAndDeletePDs(diskName, []types.NodeName{host0Name})
 | |
| 					}
 | |
| 				}()
 | |
| 
 | |
| 				for i := 0; i < t.repeatCnt; i++ { // "rapid" repeat loop
 | |
| 					framework.Logf("PD Read/Writer Iteration #%v", i)
 | |
| 					ginkgo.By(fmt.Sprintf("creating host0Pod with %d containers on node0", numContainers))
 | |
| 					host0Pod = testPDPod(diskNames, host0Name, false /* readOnly */, numContainers)
 | |
| 					_, err = podClient.Create(context.TODO(), host0Pod, metav1.CreateOptions{})
 | |
| 					framework.ExpectNoError(err, fmt.Sprintf("Failed to create host0Pod: %v", err))
 | |
| 					framework.ExpectNoError(e2epod.WaitTimeoutForPodRunningInNamespace(f.ClientSet, host0Pod.Name, f.Namespace.Name, f.Timeouts.PodStartSlow))
 | |
| 
 | |
| 					ginkgo.By(fmt.Sprintf("writing %d file(s) via a container", numPDs))
 | |
| 					containerName := "mycontainer"
 | |
| 					if numContainers > 1 {
 | |
| 						containerName = fmt.Sprintf("mycontainer%v", rand.Intn(numContainers)+1)
 | |
| 					}
 | |
| 					for x := 1; x <= numPDs; x++ {
 | |
| 						testFile := fmt.Sprintf("/testpd%d/tracker%d", x, i)
 | |
| 						testFileContents := fmt.Sprintf("%v", rand.Int())
 | |
| 						fileAndContentToVerify[testFile] = testFileContents
 | |
| 						tk := e2ekubectl.NewTestKubeconfig(framework.TestContext.CertDir, framework.TestContext.Host, framework.TestContext.KubeConfig, framework.TestContext.KubeContext, framework.TestContext.KubectlPath, ns)
 | |
| 						framework.ExpectNoError(tk.WriteFileViaContainer(host0Pod.Name, containerName, testFile, testFileContents))
 | |
| 						framework.Logf("wrote %q to file %q in pod %q (container %q) on node %q", testFileContents, testFile, host0Pod.Name, containerName, host0Name)
 | |
| 					}
 | |
| 
 | |
| 					ginkgo.By("verifying PD contents via a container")
 | |
| 					if numContainers > 1 {
 | |
| 						containerName = fmt.Sprintf("mycontainer%v", rand.Intn(numContainers)+1)
 | |
| 					}
 | |
| 					verifyPDContentsViaContainer(ns, f, host0Pod.Name, containerName, fileAndContentToVerify)
 | |
| 
 | |
| 					ginkgo.By("deleting host0Pod")
 | |
| 					framework.ExpectNoError(podClient.Delete(context.TODO(), host0Pod.Name, *metav1.NewDeleteOptions(0)), "Failed to delete host0Pod")
 | |
| 				}
 | |
| 				ginkgo.By(fmt.Sprintf("Test completed successfully, waiting for %d PD(s) to detach from node0", numPDs))
 | |
| 				for _, diskName := range diskNames {
 | |
| 					waitForPDDetach(diskName, host0Name)
 | |
| 				}
 | |
| 			})
 | |
| 		}
 | |
| 	})
 | |
| 
 | |
| 	ginkgo.Context("detach in a disrupted environment [Slow] [Disruptive]", func() {
 | |
| 		const (
 | |
| 			deleteNode    = 1 // delete physical node
 | |
| 			deleteNodeObj = 2 // delete node's api object only
 | |
| 			evictPod      = 3 // evict host0Pod on node0
 | |
| 		)
 | |
| 		type testT struct {
 | |
| 			descr     string // It description
 | |
| 			disruptOp int    // disruptive operation performed on target node
 | |
| 		}
 | |
| 		tests := []testT{
 | |
| 			// https://github.com/kubernetes/kubernetes/issues/85972
 | |
| 			// This test case is flawed. Disabling for now.
 | |
| 			// {
 | |
| 			//		descr:     "node is deleted",
 | |
| 			//		disruptOp: deleteNode,
 | |
| 			// },
 | |
| 			{
 | |
| 				descr:     "node's API object is deleted",
 | |
| 				disruptOp: deleteNodeObj,
 | |
| 			},
 | |
| 			{
 | |
| 				descr:     "pod is evicted",
 | |
| 				disruptOp: evictPod,
 | |
| 			},
 | |
| 		}
 | |
| 
 | |
| 		for _, t := range tests {
 | |
| 			disruptOp := t.disruptOp
 | |
| 			ginkgo.It(fmt.Sprintf("when %s", t.descr), func() {
 | |
| 				e2eskipper.SkipUnlessProviderIs("gce")
 | |
| 				origNodeCnt := len(nodes.Items) // healhy nodes running kubelet
 | |
| 
 | |
| 				ginkgo.By("creating a pd")
 | |
| 				diskName, err := e2epv.CreatePDWithRetry()
 | |
| 				framework.ExpectNoError(err, "Error creating a pd")
 | |
| 
 | |
| 				targetNode := &nodes.Items[0] // for node delete ops
 | |
| 				host0Pod := testPDPod([]string{diskName}, host0Name, false, 1)
 | |
| 				containerName := "mycontainer"
 | |
| 
 | |
| 				defer func() {
 | |
| 					ginkgo.By("defer: cleaning up PD-RW test env")
 | |
| 					framework.Logf("defer cleanup errors can usually be ignored")
 | |
| 					ginkgo.By("defer: delete host0Pod")
 | |
| 					podClient.Delete(context.TODO(), host0Pod.Name, *metav1.NewDeleteOptions(0))
 | |
| 					ginkgo.By("defer: detach and delete PDs")
 | |
| 					detachAndDeletePDs(diskName, []types.NodeName{host0Name})
 | |
| 					if disruptOp == deleteNode || disruptOp == deleteNodeObj {
 | |
| 						if disruptOp == deleteNodeObj {
 | |
| 							targetNode.ObjectMeta.SetResourceVersion("0")
 | |
| 							// need to set the resource version or else the Create() fails
 | |
| 							ginkgo.By("defer: re-create host0 node object")
 | |
| 							_, err := nodeClient.Create(context.TODO(), targetNode, metav1.CreateOptions{})
 | |
| 							framework.ExpectNoError(err, fmt.Sprintf("defer: Unable to re-create the deleted node object %q", targetNode.Name))
 | |
| 						}
 | |
| 						ginkgo.By("defer: verify the number of ready nodes")
 | |
| 						numNodes := countReadyNodes(cs, host0Name)
 | |
| 						// if this defer is reached due to an Expect then nested
 | |
| 						// Expects are lost, so use Failf here
 | |
| 						if numNodes != origNodeCnt {
 | |
| 							framework.Failf("defer: Requires current node count (%d) to return to original node count (%d)", numNodes, origNodeCnt)
 | |
| 						}
 | |
| 					}
 | |
| 				}()
 | |
| 
 | |
| 				ginkgo.By("creating host0Pod on node0")
 | |
| 				_, err = podClient.Create(context.TODO(), host0Pod, metav1.CreateOptions{})
 | |
| 				framework.ExpectNoError(err, fmt.Sprintf("Failed to create host0Pod: %v", err))
 | |
| 				ginkgo.By("waiting for host0Pod to be running")
 | |
| 				framework.ExpectNoError(e2epod.WaitTimeoutForPodRunningInNamespace(f.ClientSet, host0Pod.Name, f.Namespace.Name, f.Timeouts.PodStartSlow))
 | |
| 
 | |
| 				ginkgo.By("writing content to host0Pod")
 | |
| 				testFile := "/testpd1/tracker"
 | |
| 				testFileContents := fmt.Sprintf("%v", rand.Int())
 | |
| 				tk := e2ekubectl.NewTestKubeconfig(framework.TestContext.CertDir, framework.TestContext.Host, framework.TestContext.KubeConfig, framework.TestContext.KubeContext, framework.TestContext.KubectlPath, ns)
 | |
| 				framework.ExpectNoError(tk.WriteFileViaContainer(host0Pod.Name, containerName, testFile, testFileContents))
 | |
| 				framework.Logf("wrote %q to file %q in pod %q on node %q", testFileContents, testFile, host0Pod.Name, host0Name)
 | |
| 
 | |
| 				ginkgo.By("verifying PD is present in node0's VolumeInUse list")
 | |
| 				framework.ExpectNoError(waitForPDInVolumesInUse(nodeClient, diskName, host0Name, nodeStatusTimeout, true /* should exist*/))
 | |
| 
 | |
| 				if disruptOp == deleteNode {
 | |
| 					ginkgo.By("getting gce instances")
 | |
| 					gceCloud, err := gce.GetGCECloud()
 | |
| 					framework.ExpectNoError(err, fmt.Sprintf("Unable to create gcloud client err=%v", err))
 | |
| 					output, err := gceCloud.ListInstanceNames(framework.TestContext.CloudConfig.ProjectID, framework.TestContext.CloudConfig.Zone)
 | |
| 					framework.ExpectNoError(err, fmt.Sprintf("Unable to get list of node instances err=%v output=%s", err, output))
 | |
| 					framework.ExpectEqual(true, strings.Contains(string(output), string(host0Name)))
 | |
| 
 | |
| 					ginkgo.By("deleting host0")
 | |
| 					err = gceCloud.DeleteInstance(framework.TestContext.CloudConfig.ProjectID, framework.TestContext.CloudConfig.Zone, string(host0Name))
 | |
| 					framework.ExpectNoError(err, fmt.Sprintf("Failed to delete host0Pod: err=%v", err))
 | |
| 					ginkgo.By("expecting host0 node to be re-created")
 | |
| 					numNodes := countReadyNodes(cs, host0Name)
 | |
| 					framework.ExpectEqual(numNodes, origNodeCnt, fmt.Sprintf("Requires current node count (%d) to return to original node count (%d)", numNodes, origNodeCnt))
 | |
| 					output, err = gceCloud.ListInstanceNames(framework.TestContext.CloudConfig.ProjectID, framework.TestContext.CloudConfig.Zone)
 | |
| 					framework.ExpectNoError(err, fmt.Sprintf("Unable to get list of node instances err=%v output=%s", err, output))
 | |
| 					framework.ExpectEqual(true, strings.Contains(string(output), string(host0Name)))
 | |
| 
 | |
| 				} else if disruptOp == deleteNodeObj {
 | |
| 					ginkgo.By("deleting host0's node api object")
 | |
| 					framework.ExpectNoError(nodeClient.Delete(context.TODO(), string(host0Name), *metav1.NewDeleteOptions(0)), "Unable to delete host0's node object")
 | |
| 					ginkgo.By("deleting host0Pod")
 | |
| 					framework.ExpectNoError(podClient.Delete(context.TODO(), host0Pod.Name, *metav1.NewDeleteOptions(0)), "Unable to delete host0Pod")
 | |
| 
 | |
| 				} else if disruptOp == evictPod {
 | |
| 					evictTarget := &policyv1.Eviction{
 | |
| 						ObjectMeta: metav1.ObjectMeta{
 | |
| 							Name:      host0Pod.Name,
 | |
| 							Namespace: ns,
 | |
| 						},
 | |
| 					}
 | |
| 					ginkgo.By("evicting host0Pod")
 | |
| 					err = wait.PollImmediate(framework.Poll, podEvictTimeout, func() (bool, error) {
 | |
| 						if err := cs.CoreV1().Pods(ns).EvictV1(context.TODO(), evictTarget); err != nil {
 | |
| 							framework.Logf("Failed to evict host0Pod, ignoring error: %v", err)
 | |
| 							return false, nil
 | |
| 						}
 | |
| 						return true, nil
 | |
| 					})
 | |
| 					framework.ExpectNoError(err, "failed to evict host0Pod after %v", podEvictTimeout)
 | |
| 				}
 | |
| 
 | |
| 				ginkgo.By("waiting for pd to detach from host0")
 | |
| 				waitForPDDetach(diskName, host0Name)
 | |
| 			})
 | |
| 		}
 | |
| 	})
 | |
| 
 | |
| 	ginkgo.It("should be able to delete a non-existent PD without error", func() {
 | |
| 		e2eskipper.SkipUnlessProviderIs("gce")
 | |
| 
 | |
| 		ginkgo.By("delete a PD")
 | |
| 		framework.ExpectNoError(e2epv.DeletePDWithRetry("non-exist"))
 | |
| 	})
 | |
| 
 | |
| 	// This test is marked to run as serial so as device selection on AWS does not
 | |
| 	// conflict with other concurrent attach operations.
 | |
| 	ginkgo.It("[Serial] attach on previously attached volumes should work", func() {
 | |
| 		e2eskipper.SkipUnlessProviderIs("gce", "gke", "aws")
 | |
| 		ginkgo.By("creating PD")
 | |
| 		diskName, err := e2epv.CreatePDWithRetry()
 | |
| 		framework.ExpectNoError(err, "Error creating PD")
 | |
| 
 | |
| 		// this should be safe to do because if attach fails then detach will be considered
 | |
| 		// successful and we will delete the volume.
 | |
| 		defer func() {
 | |
| 			detachAndDeletePDs(diskName, []types.NodeName{host0Name})
 | |
| 		}()
 | |
| 
 | |
| 		ginkgo.By("Attaching volume to a node")
 | |
| 		err = attachPD(host0Name, diskName)
 | |
| 		framework.ExpectNoError(err, "Error attaching PD")
 | |
| 
 | |
| 		pod := testPDPod([]string{diskName}, host0Name /*readOnly*/, false, 1)
 | |
| 		ginkgo.By("Creating test pod with same volume")
 | |
| 		_, err = podClient.Create(context.TODO(), pod, metav1.CreateOptions{})
 | |
| 		framework.ExpectNoError(err, "Failed to create pod")
 | |
| 		framework.ExpectNoError(e2epod.WaitTimeoutForPodRunningInNamespace(f.ClientSet, pod.Name, f.Namespace.Name, f.Timeouts.PodStartSlow))
 | |
| 
 | |
| 		ginkgo.By("deleting the pod")
 | |
| 		framework.ExpectNoError(podClient.Delete(context.TODO(), pod.Name, *metav1.NewDeleteOptions(0)), "Failed to delete pod")
 | |
| 		framework.Logf("deleted pod %q", pod.Name)
 | |
| 		ginkgo.By("waiting for PD to detach")
 | |
| 		framework.ExpectNoError(waitForPDDetach(diskName, host0Name))
 | |
| 	})
 | |
| })
 | |
| 
 | |
| func countReadyNodes(c clientset.Interface, hostName types.NodeName) int {
 | |
| 	e2enode.WaitForNodeToBeReady(c, string(hostName), nodeStatusTimeout)
 | |
| 	framework.WaitForAllNodesSchedulable(c, nodeStatusTimeout)
 | |
| 	nodes, err := e2enode.GetReadySchedulableNodes(c)
 | |
| 	framework.ExpectNoError(err)
 | |
| 	return len(nodes.Items)
 | |
| }
 | |
| 
 | |
| func verifyPDContentsViaContainer(namespace string, f *framework.Framework, podName, containerName string, fileAndContentToVerify map[string]string) {
 | |
| 	for filePath, expectedContents := range fileAndContentToVerify {
 | |
| 		// No retry loop as there should not be temporal based failures
 | |
| 		tk := e2ekubectl.NewTestKubeconfig(framework.TestContext.CertDir, framework.TestContext.Host, framework.TestContext.KubeConfig, framework.TestContext.KubeContext, framework.TestContext.KubectlPath, namespace)
 | |
| 		v, err := tk.ReadFileViaContainer(podName, containerName, filePath)
 | |
| 		framework.ExpectNoError(err, "Error reading file %s via container %s", filePath, containerName)
 | |
| 		framework.Logf("Read file %q with content: %v", filePath, v)
 | |
| 		if strings.TrimSpace(v) != strings.TrimSpace(expectedContents) {
 | |
| 			framework.Failf("Read content <%q> does not match execpted content <%q>.", v, expectedContents)
 | |
| 		}
 | |
| 	}
 | |
| }
 | |
| 
 | |
| // TODO: move detachPD to standard cloudprovider functions so as these tests can run on other cloudproviders too
 | |
| func detachPD(nodeName types.NodeName, pdName string) error {
 | |
| 	if framework.TestContext.Provider == "gce" || framework.TestContext.Provider == "gke" {
 | |
| 		gceCloud, err := gce.GetGCECloud()
 | |
| 		if err != nil {
 | |
| 			return err
 | |
| 		}
 | |
| 		err = gceCloud.DetachDisk(pdName, nodeName)
 | |
| 		if err != nil {
 | |
| 			if gerr, ok := err.(*googleapi.Error); ok && strings.Contains(gerr.Message, "Invalid value for field 'disk'") {
 | |
| 				// PD already detached, ignore error.
 | |
| 				return nil
 | |
| 			}
 | |
| 			framework.Logf("Error detaching PD %q: %v", pdName, err)
 | |
| 		}
 | |
| 		return err
 | |
| 
 | |
| 	} else if framework.TestContext.Provider == "aws" {
 | |
| 		awsSession, err := session.NewSession()
 | |
| 		if err != nil {
 | |
| 			return fmt.Errorf("error creating session: %v", err)
 | |
| 		}
 | |
| 		client := ec2.New(awsSession)
 | |
| 		tokens := strings.Split(pdName, "/")
 | |
| 		awsVolumeID := tokens[len(tokens)-1]
 | |
| 		request := ec2.DetachVolumeInput{
 | |
| 			VolumeId: aws.String(awsVolumeID),
 | |
| 		}
 | |
| 		_, err = client.DetachVolume(&request)
 | |
| 		if err != nil {
 | |
| 			return fmt.Errorf("error detaching EBS volume: %v", err)
 | |
| 		}
 | |
| 		return nil
 | |
| 
 | |
| 	} else {
 | |
| 		return fmt.Errorf("Provider does not support volume detaching")
 | |
| 	}
 | |
| }
 | |
| 
 | |
| // TODO: move attachPD to standard cloudprovider functions so as these tests can run on other cloudproviders too
 | |
| func attachPD(nodeName types.NodeName, pdName string) error {
 | |
| 	if framework.TestContext.Provider == "gce" || framework.TestContext.Provider == "gke" {
 | |
| 		gceCloud, err := gce.GetGCECloud()
 | |
| 		if err != nil {
 | |
| 			return err
 | |
| 		}
 | |
| 		err = gceCloud.AttachDisk(pdName, nodeName, false /*readOnly*/, false /*regional*/)
 | |
| 		if err != nil {
 | |
| 			framework.Logf("Error attaching PD %q: %v", pdName, err)
 | |
| 		}
 | |
| 		return err
 | |
| 
 | |
| 	} else if framework.TestContext.Provider == "aws" {
 | |
| 		awsSession, err := session.NewSession()
 | |
| 		if err != nil {
 | |
| 			return fmt.Errorf("error creating session: %v", err)
 | |
| 		}
 | |
| 		client := ec2.New(awsSession)
 | |
| 		tokens := strings.Split(pdName, "/")
 | |
| 		awsVolumeID := tokens[len(tokens)-1]
 | |
| 		ebsUtil := utils.NewEBSUtil(client)
 | |
| 		err = ebsUtil.AttachDisk(awsVolumeID, string(nodeName))
 | |
| 		if err != nil {
 | |
| 			return fmt.Errorf("error attaching volume %s to node %s: %v", awsVolumeID, nodeName, err)
 | |
| 		}
 | |
| 		return nil
 | |
| 	} else {
 | |
| 		return fmt.Errorf("Provider does not support volume attaching")
 | |
| 	}
 | |
| }
 | |
| 
 | |
| // Returns pod spec suitable for api Create call. Handles gce, gke and aws providers only and
 | |
| // escapes if a different provider is supplied.
 | |
| // The first container name is hard-coded to "mycontainer". Subsequent containers are named:
 | |
| // "mycontainer<number> where <number> is 1..numContainers. Note if there is only one container it's
 | |
| // name has no number.
 | |
| // Container's volumeMounts are hard-coded to "/testpd<number>" where <number> is 1..len(diskNames).
 | |
| func testPDPod(diskNames []string, targetNode types.NodeName, readOnly bool, numContainers int) *v1.Pod {
 | |
| 	// escape if not a supported provider
 | |
| 	if !(framework.TestContext.Provider == "gce" || framework.TestContext.Provider == "gke" ||
 | |
| 		framework.TestContext.Provider == "aws") {
 | |
| 		framework.Failf(fmt.Sprintf("func `testPDPod` only supports gce, gke, and aws providers, not %v", framework.TestContext.Provider))
 | |
| 	}
 | |
| 
 | |
| 	containers := make([]v1.Container, numContainers)
 | |
| 	for i := range containers {
 | |
| 		containers[i].Name = "mycontainer"
 | |
| 		if numContainers > 1 {
 | |
| 			containers[i].Name = fmt.Sprintf("mycontainer%v", i+1)
 | |
| 		}
 | |
| 		containers[i].Image = e2epod.GetTestImage(imageutils.BusyBox)
 | |
| 		containers[i].Command = []string{"sleep", "6000"}
 | |
| 		containers[i].VolumeMounts = make([]v1.VolumeMount, len(diskNames))
 | |
| 		for k := range diskNames {
 | |
| 			containers[i].VolumeMounts[k].Name = fmt.Sprintf("testpd%v", k+1)
 | |
| 			containers[i].VolumeMounts[k].MountPath = fmt.Sprintf("/testpd%v", k+1)
 | |
| 		}
 | |
| 		containers[i].Resources.Limits = v1.ResourceList{}
 | |
| 		containers[i].Resources.Limits[v1.ResourceCPU] = *resource.NewQuantity(int64(0), resource.DecimalSI)
 | |
| 	}
 | |
| 
 | |
| 	pod := &v1.Pod{
 | |
| 		TypeMeta: metav1.TypeMeta{
 | |
| 			Kind:       "Pod",
 | |
| 			APIVersion: "v1",
 | |
| 		},
 | |
| 		ObjectMeta: metav1.ObjectMeta{
 | |
| 			Name: "pd-test-" + string(uuid.NewUUID()),
 | |
| 		},
 | |
| 		Spec: v1.PodSpec{
 | |
| 			Containers: containers,
 | |
| 			NodeName:   string(targetNode),
 | |
| 		},
 | |
| 	}
 | |
| 
 | |
| 	pod.Spec.Volumes = make([]v1.Volume, len(diskNames))
 | |
| 	for k, diskName := range diskNames {
 | |
| 		pod.Spec.Volumes[k].Name = fmt.Sprintf("testpd%v", k+1)
 | |
| 		if framework.TestContext.Provider == "aws" {
 | |
| 			pod.Spec.Volumes[k].VolumeSource = v1.VolumeSource{
 | |
| 				AWSElasticBlockStore: &v1.AWSElasticBlockStoreVolumeSource{
 | |
| 					VolumeID: diskName,
 | |
| 					FSType:   "ext4",
 | |
| 					ReadOnly: readOnly,
 | |
| 				},
 | |
| 			}
 | |
| 		} else { // "gce" or "gke"
 | |
| 			pod.Spec.Volumes[k].VolumeSource = v1.VolumeSource{
 | |
| 				GCEPersistentDisk: &v1.GCEPersistentDiskVolumeSource{
 | |
| 					PDName:   diskName,
 | |
| 					FSType:   e2epv.GetDefaultFSType(),
 | |
| 					ReadOnly: readOnly,
 | |
| 				},
 | |
| 			}
 | |
| 		}
 | |
| 	}
 | |
| 	return pod
 | |
| }
 | |
| 
 | |
| // Waits for specified PD to detach from specified hostName
 | |
| func waitForPDDetach(diskName string, nodeName types.NodeName) error {
 | |
| 	if framework.TestContext.Provider == "gce" || framework.TestContext.Provider == "gke" {
 | |
| 		framework.Logf("Waiting for GCE PD %q to detach from node %q.", diskName, nodeName)
 | |
| 		gceCloud, err := gce.GetGCECloud()
 | |
| 		if err != nil {
 | |
| 			return err
 | |
| 		}
 | |
| 		for start := time.Now(); time.Since(start) < gcePDDetachTimeout; time.Sleep(gcePDDetachPollTime) {
 | |
| 			diskAttached, err := gceCloud.DiskIsAttached(diskName, nodeName)
 | |
| 			if err != nil {
 | |
| 				framework.Logf("Error waiting for PD %q to detach from node %q. 'DiskIsAttached(...)' failed with %v", diskName, nodeName, err)
 | |
| 				return err
 | |
| 			}
 | |
| 			if !diskAttached {
 | |
| 				// Specified disk does not appear to be attached to specified node
 | |
| 				framework.Logf("GCE PD %q appears to have successfully detached from %q.", diskName, nodeName)
 | |
| 				return nil
 | |
| 			}
 | |
| 			framework.Logf("Waiting for GCE PD %q to detach from %q.", diskName, nodeName)
 | |
| 		}
 | |
| 		return fmt.Errorf("Gave up waiting for GCE PD %q to detach from %q after %v", diskName, nodeName, gcePDDetachTimeout)
 | |
| 	}
 | |
| 	return nil
 | |
| }
 | |
| 
 | |
| func detachAndDeletePDs(diskName string, hosts []types.NodeName) {
 | |
| 	for _, host := range hosts {
 | |
| 		framework.Logf("Detaching GCE PD %q from node %q.", diskName, host)
 | |
| 		detachPD(host, diskName)
 | |
| 		ginkgo.By(fmt.Sprintf("Waiting for PD %q to detach from %q", diskName, host))
 | |
| 		waitForPDDetach(diskName, host)
 | |
| 	}
 | |
| 	ginkgo.By(fmt.Sprintf("Deleting PD %q", diskName))
 | |
| 	framework.ExpectNoError(e2epv.DeletePDWithRetry(diskName))
 | |
| }
 | |
| 
 | |
| func waitForPDInVolumesInUse(
 | |
| 	nodeClient v1core.NodeInterface,
 | |
| 	diskName string,
 | |
| 	nodeName types.NodeName,
 | |
| 	timeout time.Duration,
 | |
| 	shouldExist bool) error {
 | |
| 	logStr := "to contain"
 | |
| 	if !shouldExist {
 | |
| 		logStr = "to NOT contain"
 | |
| 	}
 | |
| 	framework.Logf("Waiting for node %s's VolumesInUse Status %s PD %q", nodeName, logStr, diskName)
 | |
| 	for start := time.Now(); time.Since(start) < timeout; time.Sleep(nodeStatusPollTime) {
 | |
| 		nodeObj, err := nodeClient.Get(context.TODO(), string(nodeName), metav1.GetOptions{})
 | |
| 		if err != nil || nodeObj == nil {
 | |
| 			framework.Logf("Failed to fetch node object %q from API server. err=%v", nodeName, err)
 | |
| 			continue
 | |
| 		}
 | |
| 		exists := false
 | |
| 		for _, volumeInUse := range nodeObj.Status.VolumesInUse {
 | |
| 			volumeInUseStr := string(volumeInUse)
 | |
| 			if strings.Contains(volumeInUseStr, diskName) {
 | |
| 				if shouldExist {
 | |
| 					framework.Logf("Found PD %q in node %q's VolumesInUse Status: %q", diskName, nodeName, volumeInUseStr)
 | |
| 					return nil
 | |
| 				}
 | |
| 				exists = true
 | |
| 			}
 | |
| 		}
 | |
| 		if !shouldExist && !exists {
 | |
| 			framework.Logf("Verified PD %q does not exist in node %q's VolumesInUse Status.", diskName, nodeName)
 | |
| 			return nil
 | |
| 		}
 | |
| 	}
 | |
| 	return fmt.Errorf("Timed out waiting for node %s VolumesInUse Status %s diskName %q", nodeName, logStr, diskName)
 | |
| }
 |