mirror of
https://github.com/k3s-io/kubernetes.git
synced 2025-08-04 18:00:08 +00:00
Add PV E2E: disruptive environment tests
Tweaked 'delete pod during stopped kubelet' test to check for mount Refactor start, stop, restart funcs into 1 func
This commit is contained in:
parent
1bfd4f0a71
commit
7d5305718d
263
test/e2e/persistent_volumes-disruptive.go
Normal file
263
test/e2e/persistent_volumes-disruptive.go
Normal file
@ -0,0 +1,263 @@
|
|||||||
|
/*
|
||||||
|
Copyright 2015 The Kubernetes Authors.
|
||||||
|
|
||||||
|
Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
you may not use this file except in compliance with the License.
|
||||||
|
You may obtain a copy of the License at
|
||||||
|
|
||||||
|
http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
|
||||||
|
Unless required by applicable law or agreed to in writing, software
|
||||||
|
distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
See the License for the specific language governing permissions and
|
||||||
|
limitations under the License.
|
||||||
|
*/
|
||||||
|
|
||||||
|
// This test references
|
||||||
|
// persistent_volumes.go
|
||||||
|
// volumes.go
|
||||||
|
|
||||||
|
package e2e
|
||||||
|
|
||||||
|
import (
|
||||||
|
"time"
|
||||||
|
|
||||||
|
. "github.com/onsi/ginkgo"
|
||||||
|
. "github.com/onsi/gomega"
|
||||||
|
"k8s.io/kubernetes/pkg/api/v1"
|
||||||
|
metav1 "k8s.io/kubernetes/pkg/apis/meta/v1"
|
||||||
|
"k8s.io/kubernetes/pkg/client/clientset_generated/clientset"
|
||||||
|
"k8s.io/kubernetes/test/e2e/framework"
|
||||||
|
|
||||||
|
"strings"
|
||||||
|
)
|
||||||
|
|
||||||
|
type testBody func(c clientset.Interface, f *framework.Framework, clientPod *v1.Pod, pvc *v1.PersistentVolumeClaim, pv *v1.PersistentVolume)
|
||||||
|
type disruptiveTest struct {
|
||||||
|
testItStmt string
|
||||||
|
runTest testBody
|
||||||
|
}
|
||||||
|
type kubeletOpt string
|
||||||
|
|
||||||
|
const (
|
||||||
|
MIN_NODES = 2
|
||||||
|
NODE_STATE_WAIT = 2 * time.Minute
|
||||||
|
kStart kubeletOpt = "start"
|
||||||
|
kStop kubeletOpt = "stop"
|
||||||
|
kRestart kubeletOpt = "restart"
|
||||||
|
)
|
||||||
|
|
||||||
|
var _ = framework.KubeDescribe("PersistentVolumes:Disruptive", func() {
|
||||||
|
|
||||||
|
f := framework.NewDefaultFramework("disruptive-pv")
|
||||||
|
var (
|
||||||
|
c clientset.Interface
|
||||||
|
ns string
|
||||||
|
nfsServerPod *v1.Pod
|
||||||
|
nfsPVconfig persistentVolumeConfig
|
||||||
|
nfsServerIP, clientNodeIP string
|
||||||
|
clientNode *v1.Node
|
||||||
|
)
|
||||||
|
|
||||||
|
nfsServerConfig := VolumeTestConfig{
|
||||||
|
namespace: v1.NamespaceDefault,
|
||||||
|
prefix: "nfs",
|
||||||
|
serverImage: "gcr.io/google_containers/volume-nfs:0.7",
|
||||||
|
serverPorts: []int{2049},
|
||||||
|
serverArgs: []string{"-G", "777", "/exports"},
|
||||||
|
}
|
||||||
|
|
||||||
|
BeforeEach(func() {
|
||||||
|
// To protect the NFS volume pod from the kubelet restart, we isolate it on its own node.
|
||||||
|
framework.SkipUnlessNodeCountIsAtLeast(MIN_NODES)
|
||||||
|
c = f.ClientSet
|
||||||
|
ns = f.Namespace.Name
|
||||||
|
|
||||||
|
// Start the NFS server pod.
|
||||||
|
if nfsServerPod == nil {
|
||||||
|
framework.Logf("[BeforeEach] Initializing NFS Server Pod")
|
||||||
|
nfsServerPod = startVolumeServer(c, nfsServerConfig)
|
||||||
|
|
||||||
|
framework.Logf("[BeforeEach] Configuring PersistentVolume")
|
||||||
|
nfsServerIP = nfsServerPod.Status.PodIP
|
||||||
|
nfsPVconfig = persistentVolumeConfig{
|
||||||
|
namePrefix: "nfs-",
|
||||||
|
pvSource: v1.PersistentVolumeSource{
|
||||||
|
NFS: &v1.NFSVolumeSource{
|
||||||
|
Server: nfsServerIP,
|
||||||
|
Path: "/exports",
|
||||||
|
ReadOnly: false,
|
||||||
|
},
|
||||||
|
},
|
||||||
|
}
|
||||||
|
}
|
||||||
|
// Get the first ready node IP that is not hosting the NFS pod.
|
||||||
|
if clientNodeIP == "" {
|
||||||
|
framework.Logf("Designating test node")
|
||||||
|
nodes := framework.GetReadySchedulableNodesOrDie(c)
|
||||||
|
for _, node := range nodes.Items {
|
||||||
|
if node.Name != nfsServerPod.Spec.NodeName {
|
||||||
|
clientNode = &node
|
||||||
|
break
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
})
|
||||||
|
|
||||||
|
AddCleanupAction(func() {
|
||||||
|
if nfsServerPod != nil && c != nil {
|
||||||
|
By("Deleting NFS server pod")
|
||||||
|
nfsServerPodCleanup(c, nfsServerConfig)
|
||||||
|
nfsServerPod = nil
|
||||||
|
}
|
||||||
|
})
|
||||||
|
|
||||||
|
Context("when kubelet restarts", func() {
|
||||||
|
|
||||||
|
var (
|
||||||
|
clientPod *v1.Pod
|
||||||
|
pv *v1.PersistentVolume
|
||||||
|
pvc *v1.PersistentVolumeClaim
|
||||||
|
)
|
||||||
|
|
||||||
|
BeforeEach(func() {
|
||||||
|
framework.Logf("Initializing test spec")
|
||||||
|
clientPod, pv, pvc = initTestCase(f, c, nfsPVconfig, ns, clientNode.Name)
|
||||||
|
})
|
||||||
|
|
||||||
|
AfterEach(func() {
|
||||||
|
framework.Logf("Tearing down test spec")
|
||||||
|
tearDownTestCase(c, f, ns, clientPod, pvc, pv)
|
||||||
|
})
|
||||||
|
|
||||||
|
// Test table housing the It() title string and test spec. runTest is type testBody, defined at
|
||||||
|
// the start of this file. To add tests, define a function mirroring the testBody signature and assign
|
||||||
|
// to runTest.
|
||||||
|
disruptiveTestTable := []disruptiveTest{
|
||||||
|
{
|
||||||
|
testItStmt: "Should test that a file written to the mount before kubelet restart is stat-able after restart.",
|
||||||
|
runTest: testKubeletRestartsAndRestoresMount,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
testItStmt: "Should test that a volume mount to a pod that is deleted while the kubelet is down unmounts when the kubelet returns.",
|
||||||
|
runTest: testVolumeUnmountsFromDeletedPod,
|
||||||
|
},
|
||||||
|
}
|
||||||
|
|
||||||
|
// Test loop executes each disruptiveTest iteratively.
|
||||||
|
for _, test := range disruptiveTestTable {
|
||||||
|
func(t disruptiveTest) {
|
||||||
|
It(t.testItStmt+" [Disruptive]", func() {
|
||||||
|
By("Executing Spec")
|
||||||
|
t.runTest(c, f, clientPod, pvc, pv)
|
||||||
|
})
|
||||||
|
}(test)
|
||||||
|
}
|
||||||
|
})
|
||||||
|
})
|
||||||
|
|
||||||
|
// SPEC: Test that a volume mounted to a pod remains mounted after a kubelet restarts
|
||||||
|
func testKubeletRestartsAndRestoresMount(c clientset.Interface, f *framework.Framework, clientPod *v1.Pod, pvc *v1.PersistentVolumeClaim, pv *v1.PersistentVolume) {
|
||||||
|
file := "/mnt/_SUCCESS"
|
||||||
|
_, err := podExec(clientPod, "touch "+file)
|
||||||
|
Expect(err).NotTo(HaveOccurred())
|
||||||
|
|
||||||
|
kubeletCommand(kRestart, c, clientPod)
|
||||||
|
|
||||||
|
_, err = podExec(clientPod, "cat "+file)
|
||||||
|
Expect(err).NotTo(HaveOccurred())
|
||||||
|
framework.Logf("Pod %s detected %s after kubelet restart", clientPod.Name, file)
|
||||||
|
}
|
||||||
|
|
||||||
|
// SPEC: Test that a volume unmounts if the client pod was deleted while the kubelet was down.
|
||||||
|
func testVolumeUnmountsFromDeletedPod(c clientset.Interface, f *framework.Framework, clientPod *v1.Pod, pvc *v1.PersistentVolumeClaim, pv *v1.PersistentVolume) {
|
||||||
|
|
||||||
|
nodeIP, err := framework.GetHostExternalAddress(c, clientPod)
|
||||||
|
nodeIP = nodeIP + ":22"
|
||||||
|
Expect(err).NotTo(HaveOccurred())
|
||||||
|
|
||||||
|
result, err := framework.SSH("mount", nodeIP, framework.TestContext.Provider)
|
||||||
|
Expect(err).NotTo(HaveOccurred())
|
||||||
|
if strings.Contains(result.Stdout, string(clientPod.UID)) {
|
||||||
|
framework.Logf("Sanity Check: Client UID %s found in `mount` output", clientPod.UID)
|
||||||
|
} else {
|
||||||
|
framework.Failf("Sanity Check: Client UID %s NOT found in `mount` output prior to pod deletion. Something has gone wrong.", clientPod.UID)
|
||||||
|
}
|
||||||
|
|
||||||
|
file := "/mnt/_SUCCESS"
|
||||||
|
_, err = podExec(clientPod, "touch "+file)
|
||||||
|
Expect(err).NotTo(HaveOccurred())
|
||||||
|
|
||||||
|
kubeletCommand(kStop, c, clientPod)
|
||||||
|
deletePod(f, c, clientPod.Namespace, clientPod)
|
||||||
|
kubeletCommand(kStart, c, clientPod)
|
||||||
|
|
||||||
|
result, err = framework.SSH("mount", nodeIP, framework.TestContext.Provider)
|
||||||
|
Expect(err).NotTo(HaveOccurred())
|
||||||
|
if strings.Contains(result.Stdout, string(clientPod.UID)) {
|
||||||
|
framework.Failf("Client UID %s found in `mount` output. Volume failed to unmount.", clientPod.UID)
|
||||||
|
} else {
|
||||||
|
framework.Logf("Client UID %s not found in `mount` output. Volume has unmounted.", clientPod.UID)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Initialize a test spec with a pv, pvc, and nfs client pod
|
||||||
|
func initTestCase(f *framework.Framework, c clientset.Interface, pvConfig persistentVolumeConfig, ns, nodeName string) (pod *v1.Pod, pv *v1.PersistentVolume, pvc *v1.PersistentVolumeClaim) {
|
||||||
|
pv, pvc = createPVPVC(c, pvConfig, ns, false)
|
||||||
|
pod = makePod(ns, pvc.Name)
|
||||||
|
pod.Spec.NodeName = nodeName
|
||||||
|
framework.Logf("Creating nfs client Pod %s on node %s", pod.Name, nodeName)
|
||||||
|
pod, err := c.Core().Pods(ns).Create(pod)
|
||||||
|
Expect(err).NotTo(HaveOccurred())
|
||||||
|
err = framework.WaitForPodRunningInNamespace(c, pod)
|
||||||
|
Expect(err).NotTo(HaveOccurred())
|
||||||
|
|
||||||
|
pod, err = c.Core().Pods(ns).Get(pod.Name, metav1.GetOptions{})
|
||||||
|
Expect(err).NotTo(HaveOccurred())
|
||||||
|
pvc, err = c.Core().PersistentVolumeClaims(ns).Get(pvc.Name, metav1.GetOptions{})
|
||||||
|
Expect(err).NotTo(HaveOccurred())
|
||||||
|
pv, err = c.Core().PersistentVolumes().Get(pv.Name, metav1.GetOptions{})
|
||||||
|
Expect(err).NotTo(HaveOccurred())
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
// Post-spec clean up
|
||||||
|
func tearDownTestCase(c clientset.Interface, f *framework.Framework, ns string, pod *v1.Pod, pvc *v1.PersistentVolumeClaim, pv *v1.PersistentVolume) {
|
||||||
|
deletePod(f, c, ns, pod)
|
||||||
|
deletePersistentVolumeClaim(c, pvc.Name, ns)
|
||||||
|
deletePersistentVolume(c, pv.Name)
|
||||||
|
}
|
||||||
|
|
||||||
|
// Start, Restart, or Stop the kubelet running on the node of the target pod.
|
||||||
|
// Allowed kubeltOps are `kStart`, `kStop`, and `kRestart`
|
||||||
|
func kubeletCommand(kOp kubeletOpt, c clientset.Interface, pod *v1.Pod) {
|
||||||
|
nodeIP, err := framework.GetHostExternalAddress(c, pod)
|
||||||
|
Expect(err).NotTo(HaveOccurred())
|
||||||
|
nodeIP = nodeIP + ":22"
|
||||||
|
sshResult, err := framework.SSH("/etc/init.d/kubelet "+string(kOp), nodeIP, framework.TestContext.Provider)
|
||||||
|
Expect(err).NotTo(HaveOccurred())
|
||||||
|
framework.LogSSHResult(sshResult)
|
||||||
|
|
||||||
|
// On restart, waiting for node NotReady prevents a race condition where the node takes a few moments to leave the
|
||||||
|
// Ready state which in turn short circuits WaitForNodeToBeReady()
|
||||||
|
if kOp == kStop || kOp == kRestart {
|
||||||
|
if ok := framework.WaitForNodeToBeNotReady(c, pod.Spec.NodeName, NODE_STATE_WAIT); !ok {
|
||||||
|
framework.Failf("Node %s failed to enter NotReady state", pod.Spec.NodeName)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if kOp == kStart || kOp == kRestart {
|
||||||
|
if ok := framework.WaitForNodeToBeReady(c, pod.Spec.NodeName, NODE_STATE_WAIT); !ok {
|
||||||
|
framework.Failf("Node %s failed to enter NotReady state", pod.Spec.NodeName)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
}
|
||||||
|
|
||||||
|
// Wraps RunKubectl to execute a bash cmd in target pod
|
||||||
|
func podExec(pod *v1.Pod, bashExec string) (string, error) {
|
||||||
|
args := strings.Split(bashExec, " ")
|
||||||
|
cmd := []string{"exec", "--namespace=" + pod.Namespace, pod.Name}
|
||||||
|
cmd = append(cmd, args...)
|
||||||
|
return framework.RunKubectl(cmd...)
|
||||||
|
}
|
Loading…
Reference in New Issue
Block a user