mirror of
				https://github.com/k3s-io/kubernetes.git
				synced 2025-11-03 23:40:03 +00:00 
			
		
		
		
	
		
			
				
	
	
		
			480 lines
		
	
	
		
			20 KiB
		
	
	
	
		
			Go
		
	
	
	
	
	
			
		
		
	
	
			480 lines
		
	
	
		
			20 KiB
		
	
	
	
		
			Go
		
	
	
	
	
	
/*
 | 
						|
Copyright 2016 The Kubernetes Authors.
 | 
						|
 | 
						|
Licensed under the Apache License, Version 2.0 (the "License");
 | 
						|
you may not use this file except in compliance with the License.
 | 
						|
You may obtain a copy of the License at
 | 
						|
 | 
						|
    http://www.apache.org/licenses/LICENSE-2.0
 | 
						|
 | 
						|
Unless required by applicable law or agreed to in writing, software
 | 
						|
distributed under the License is distributed on an "AS IS" BASIS,
 | 
						|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 | 
						|
See the License for the specific language governing permissions and
 | 
						|
limitations under the License.
 | 
						|
*/
 | 
						|
 | 
						|
package e2e
 | 
						|
 | 
						|
import (
 | 
						|
	"fmt"
 | 
						|
	"strings"
 | 
						|
	"time"
 | 
						|
 | 
						|
	metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
 | 
						|
	"k8s.io/apimachinery/pkg/labels"
 | 
						|
	"k8s.io/apimachinery/pkg/runtime"
 | 
						|
	"k8s.io/apimachinery/pkg/util/wait"
 | 
						|
	"k8s.io/apimachinery/pkg/watch"
 | 
						|
	"k8s.io/kubernetes/pkg/api"
 | 
						|
	"k8s.io/kubernetes/pkg/api/v1"
 | 
						|
	"k8s.io/kubernetes/pkg/client/cache"
 | 
						|
	"k8s.io/kubernetes/pkg/client/clientset_generated/clientset"
 | 
						|
	"k8s.io/kubernetes/pkg/fields"
 | 
						|
	"k8s.io/kubernetes/test/e2e/framework"
 | 
						|
	testutils "k8s.io/kubernetes/test/utils"
 | 
						|
 | 
						|
	. "github.com/onsi/ginkgo"
 | 
						|
	. "github.com/onsi/gomega"
 | 
						|
)
 | 
						|
 | 
						|
// Blocks outgoing network traffic on 'node'. Then runs testFunc and returns its status.
 | 
						|
// At the end (even in case of errors), the network traffic is brought back to normal.
 | 
						|
// This function executes commands on a node so it will work only for some
 | 
						|
// environments.
 | 
						|
func testUnderTemporaryNetworkFailure(c clientset.Interface, ns string, node *v1.Node, testFunc func()) {
 | 
						|
	host := framework.GetNodeExternalIP(node)
 | 
						|
	master := framework.GetMasterAddress(c)
 | 
						|
	By(fmt.Sprintf("block network traffic from node %s to the master", node.Name))
 | 
						|
	defer func() {
 | 
						|
		// This code will execute even if setting the iptables rule failed.
 | 
						|
		// It is on purpose because we may have an error even if the new rule
 | 
						|
		// had been inserted. (yes, we could look at the error code and ssh error
 | 
						|
		// separately, but I prefer to stay on the safe side).
 | 
						|
		By(fmt.Sprintf("Unblock network traffic from node %s to the master", node.Name))
 | 
						|
		framework.UnblockNetwork(host, master)
 | 
						|
	}()
 | 
						|
 | 
						|
	framework.Logf("Waiting %v to ensure node %s is ready before beginning test...", resizeNodeReadyTimeout, node.Name)
 | 
						|
	if !framework.WaitForNodeToBe(c, node.Name, v1.NodeReady, true, resizeNodeReadyTimeout) {
 | 
						|
		framework.Failf("Node %s did not become ready within %v", node.Name, resizeNodeReadyTimeout)
 | 
						|
	}
 | 
						|
	framework.BlockNetwork(host, master)
 | 
						|
 | 
						|
	framework.Logf("Waiting %v for node %s to be not ready after simulated network failure", resizeNodeNotReadyTimeout, node.Name)
 | 
						|
	if !framework.WaitForNodeToBe(c, node.Name, v1.NodeReady, false, resizeNodeNotReadyTimeout) {
 | 
						|
		framework.Failf("Node %s did not become not-ready within %v", node.Name, resizeNodeNotReadyTimeout)
 | 
						|
	}
 | 
						|
 | 
						|
	testFunc()
 | 
						|
	// network traffic is unblocked in a deferred function
 | 
						|
}
 | 
						|
 | 
						|
func expectNodeReadiness(isReady bool, newNode chan *v1.Node) {
 | 
						|
	timeout := false
 | 
						|
	expected := false
 | 
						|
	timer := time.After(nodeReadinessTimeout)
 | 
						|
	for !expected && !timeout {
 | 
						|
		select {
 | 
						|
		case n := <-newNode:
 | 
						|
			if framework.IsNodeConditionSetAsExpected(n, v1.NodeReady, isReady) {
 | 
						|
				expected = true
 | 
						|
			} else {
 | 
						|
				framework.Logf("Observed node ready status is NOT %v as expected", isReady)
 | 
						|
			}
 | 
						|
		case <-timer:
 | 
						|
			timeout = true
 | 
						|
		}
 | 
						|
	}
 | 
						|
	if !expected {
 | 
						|
		framework.Failf("Failed to observe node ready status change to %v", isReady)
 | 
						|
	}
 | 
						|
}
 | 
						|
 | 
						|
func podOnNode(podName, nodeName string, image string) *v1.Pod {
 | 
						|
	return &v1.Pod{
 | 
						|
		ObjectMeta: v1.ObjectMeta{
 | 
						|
			Name: podName,
 | 
						|
			Labels: map[string]string{
 | 
						|
				"name": podName,
 | 
						|
			},
 | 
						|
		},
 | 
						|
		Spec: v1.PodSpec{
 | 
						|
			Containers: []v1.Container{
 | 
						|
				{
 | 
						|
					Name:  podName,
 | 
						|
					Image: image,
 | 
						|
					Ports: []v1.ContainerPort{{ContainerPort: 9376}},
 | 
						|
				},
 | 
						|
			},
 | 
						|
			NodeName:      nodeName,
 | 
						|
			RestartPolicy: v1.RestartPolicyNever,
 | 
						|
		},
 | 
						|
	}
 | 
						|
}
 | 
						|
 | 
						|
func newPodOnNode(c clientset.Interface, namespace, podName, nodeName string) error {
 | 
						|
	pod, err := c.Core().Pods(namespace).Create(podOnNode(podName, nodeName, serveHostnameImage))
 | 
						|
	if err == nil {
 | 
						|
		framework.Logf("Created pod %s on node %s", pod.ObjectMeta.Name, nodeName)
 | 
						|
	} else {
 | 
						|
		framework.Logf("Failed to create pod %s on node %s: %v", podName, nodeName, err)
 | 
						|
	}
 | 
						|
	return err
 | 
						|
}
 | 
						|
 | 
						|
var _ = framework.KubeDescribe("Network Partition [Disruptive] [Slow]", func() {
 | 
						|
	f := framework.NewDefaultFramework("network-partition")
 | 
						|
	var systemPodsNo int32
 | 
						|
	var c clientset.Interface
 | 
						|
	var ns string
 | 
						|
	ignoreLabels := framework.ImagePullerLabels
 | 
						|
	var group string
 | 
						|
 | 
						|
	BeforeEach(func() {
 | 
						|
		c = f.ClientSet
 | 
						|
		ns = f.Namespace.Name
 | 
						|
		systemPods, err := framework.GetPodsInNamespace(c, ns, ignoreLabels)
 | 
						|
		Expect(err).NotTo(HaveOccurred())
 | 
						|
		systemPodsNo = int32(len(systemPods))
 | 
						|
		if strings.Index(framework.TestContext.CloudConfig.NodeInstanceGroup, ",") >= 0 {
 | 
						|
			framework.Failf("Test dose not support cluster setup with more than one MIG: %s", framework.TestContext.CloudConfig.NodeInstanceGroup)
 | 
						|
		} else {
 | 
						|
			group = framework.TestContext.CloudConfig.NodeInstanceGroup
 | 
						|
		}
 | 
						|
	})
 | 
						|
 | 
						|
	framework.KubeDescribe("Pods", func() {
 | 
						|
		Context("should return to running and ready state after network partition is healed", func() {
 | 
						|
			BeforeEach(func() {
 | 
						|
				framework.SkipUnlessProviderIs("gce", "gke", "aws")
 | 
						|
				framework.SkipUnlessNodeCountIsAtLeast(2)
 | 
						|
			})
 | 
						|
 | 
						|
			// What happens in this test:
 | 
						|
			//	Network traffic from a node to master is cut off to simulate network partition
 | 
						|
			// Expect to observe:
 | 
						|
			// 1. Node is marked NotReady after timeout by nodecontroller (40seconds)
 | 
						|
			// 2. All pods on node are marked NotReady shortly after #1
 | 
						|
			// 3. Node and pods return to Ready after connectivivty recovers
 | 
						|
			It("All pods on the unreachable node should be marked as NotReady upon the node turn NotReady "+
 | 
						|
				"AND all pods should be mark back to Ready when the node get back to Ready before pod eviction timeout", func() {
 | 
						|
				By("choose a node - we will block all network traffic on this node")
 | 
						|
				var podOpts v1.ListOptions
 | 
						|
				nodeOpts := v1.ListOptions{}
 | 
						|
				nodes, err := c.Core().Nodes().List(nodeOpts)
 | 
						|
				Expect(err).NotTo(HaveOccurred())
 | 
						|
				framework.FilterNodes(nodes, func(node v1.Node) bool {
 | 
						|
					if !framework.IsNodeConditionSetAsExpected(&node, v1.NodeReady, true) {
 | 
						|
						return false
 | 
						|
					}
 | 
						|
					podOpts = v1.ListOptions{FieldSelector: fields.OneTermEqualSelector(api.PodHostField, node.Name).String()}
 | 
						|
					pods, err := c.Core().Pods(v1.NamespaceAll).List(podOpts)
 | 
						|
					if err != nil || len(pods.Items) <= 0 {
 | 
						|
						return false
 | 
						|
					}
 | 
						|
					return true
 | 
						|
				})
 | 
						|
				if len(nodes.Items) <= 0 {
 | 
						|
					framework.Failf("No eligible node were found: %d", len(nodes.Items))
 | 
						|
				}
 | 
						|
				node := nodes.Items[0]
 | 
						|
				podOpts = v1.ListOptions{FieldSelector: fields.OneTermEqualSelector(api.PodHostField, node.Name).String()}
 | 
						|
				if err = framework.WaitForMatchPodsCondition(c, podOpts, "Running and Ready", podReadyTimeout, testutils.PodRunningReady); err != nil {
 | 
						|
					framework.Failf("Pods on node %s are not ready and running within %v: %v", node.Name, podReadyTimeout, err)
 | 
						|
				}
 | 
						|
 | 
						|
				By("Set up watch on node status")
 | 
						|
				nodeSelector := fields.OneTermEqualSelector("metadata.name", node.Name)
 | 
						|
				stopCh := make(chan struct{})
 | 
						|
				newNode := make(chan *v1.Node)
 | 
						|
				var controller cache.Controller
 | 
						|
				_, controller = cache.NewInformer(
 | 
						|
					&cache.ListWatch{
 | 
						|
						ListFunc: func(options v1.ListOptions) (runtime.Object, error) {
 | 
						|
							options.FieldSelector = nodeSelector.String()
 | 
						|
							obj, err := f.ClientSet.Core().Nodes().List(options)
 | 
						|
							return runtime.Object(obj), err
 | 
						|
						},
 | 
						|
						WatchFunc: func(options v1.ListOptions) (watch.Interface, error) {
 | 
						|
							options.FieldSelector = nodeSelector.String()
 | 
						|
							return f.ClientSet.Core().Nodes().Watch(options)
 | 
						|
						},
 | 
						|
					},
 | 
						|
					&v1.Node{},
 | 
						|
					0,
 | 
						|
					cache.ResourceEventHandlerFuncs{
 | 
						|
						UpdateFunc: func(oldObj, newObj interface{}) {
 | 
						|
							n, ok := newObj.(*v1.Node)
 | 
						|
							Expect(ok).To(Equal(true))
 | 
						|
							newNode <- n
 | 
						|
 | 
						|
						},
 | 
						|
					},
 | 
						|
				)
 | 
						|
 | 
						|
				defer func() {
 | 
						|
					// Will not explicitly close newNode channel here due to
 | 
						|
					// race condition where stopCh and newNode are closed but informer onUpdate still executes.
 | 
						|
					close(stopCh)
 | 
						|
				}()
 | 
						|
				go controller.Run(stopCh)
 | 
						|
 | 
						|
				By(fmt.Sprintf("Block traffic from node %s to the master", node.Name))
 | 
						|
				host := framework.GetNodeExternalIP(&node)
 | 
						|
				master := framework.GetMasterAddress(c)
 | 
						|
				defer func() {
 | 
						|
					By(fmt.Sprintf("Unblock traffic from node %s to the master", node.Name))
 | 
						|
					framework.UnblockNetwork(host, master)
 | 
						|
 | 
						|
					if CurrentGinkgoTestDescription().Failed {
 | 
						|
						return
 | 
						|
					}
 | 
						|
 | 
						|
					By("Expect to observe node and pod status change from NotReady to Ready after network connectivity recovers")
 | 
						|
					expectNodeReadiness(true, newNode)
 | 
						|
					if err = framework.WaitForMatchPodsCondition(c, podOpts, "Running and Ready", podReadyTimeout, testutils.PodRunningReady); err != nil {
 | 
						|
						framework.Failf("Pods on node %s did not become ready and running within %v: %v", node.Name, podReadyTimeout, err)
 | 
						|
					}
 | 
						|
				}()
 | 
						|
 | 
						|
				framework.BlockNetwork(host, master)
 | 
						|
 | 
						|
				By("Expect to observe node and pod status change from Ready to NotReady after network partition")
 | 
						|
				expectNodeReadiness(false, newNode)
 | 
						|
				if err = framework.WaitForMatchPodsCondition(c, podOpts, "NotReady", podNotReadyTimeout, testutils.PodNotReady); err != nil {
 | 
						|
					framework.Failf("Pods on node %s did not become NotReady within %v: %v", node.Name, podNotReadyTimeout, err)
 | 
						|
				}
 | 
						|
			})
 | 
						|
		})
 | 
						|
	})
 | 
						|
 | 
						|
	framework.KubeDescribe("[ReplicationController]", func() {
 | 
						|
		It("should recreate pods scheduled on the unreachable node "+
 | 
						|
			"AND allow scheduling of pods on a node after it rejoins the cluster", func() {
 | 
						|
 | 
						|
			// Create a replication controller for a service that serves its hostname.
 | 
						|
			// The source for the Docker container kubernetes/serve_hostname is in contrib/for-demos/serve_hostname
 | 
						|
			name := "my-hostname-net"
 | 
						|
			newSVCByName(c, ns, name)
 | 
						|
			replicas := int32(framework.TestContext.CloudConfig.NumNodes)
 | 
						|
			newRCByName(c, ns, name, replicas, nil)
 | 
						|
			err := framework.VerifyPods(c, ns, name, true, replicas)
 | 
						|
			Expect(err).NotTo(HaveOccurred(), "Each pod should start running and responding")
 | 
						|
 | 
						|
			By("choose a node with at least one pod - we will block some network traffic on this node")
 | 
						|
			label := labels.SelectorFromSet(labels.Set(map[string]string{"name": name}))
 | 
						|
			options := v1.ListOptions{LabelSelector: label.String()}
 | 
						|
			pods, err := c.Core().Pods(ns).List(options) // list pods after all have been scheduled
 | 
						|
			Expect(err).NotTo(HaveOccurred())
 | 
						|
			nodeName := pods.Items[0].Spec.NodeName
 | 
						|
 | 
						|
			node, err := c.Core().Nodes().Get(nodeName, metav1.GetOptions{})
 | 
						|
			Expect(err).NotTo(HaveOccurred())
 | 
						|
 | 
						|
			// This creates a temporary network partition, verifies that 'podNameToDisappear',
 | 
						|
			// that belongs to replication controller 'rcName', really disappeared (because its
 | 
						|
			// grace period is set to 0).
 | 
						|
			// Finally, it checks that the replication controller recreates the
 | 
						|
			// pods on another node and that now the number of replicas is equal 'replicas'.
 | 
						|
			By(fmt.Sprintf("blocking network traffic from node %s", node.Name))
 | 
						|
			testUnderTemporaryNetworkFailure(c, ns, node, func() {
 | 
						|
				framework.Logf("Waiting for pod %s to be removed", pods.Items[0].Name)
 | 
						|
				err := framework.WaitForRCPodToDisappear(c, ns, name, pods.Items[0].Name)
 | 
						|
				Expect(err).NotTo(HaveOccurred())
 | 
						|
 | 
						|
				By("verifying whether the pod from the unreachable node is recreated")
 | 
						|
				err = framework.VerifyPods(c, ns, name, true, replicas)
 | 
						|
				Expect(err).NotTo(HaveOccurred())
 | 
						|
			})
 | 
						|
 | 
						|
			framework.Logf("Waiting %v for node %s to be ready once temporary network failure ends", resizeNodeReadyTimeout, node.Name)
 | 
						|
			if !framework.WaitForNodeToBeReady(c, node.Name, resizeNodeReadyTimeout) {
 | 
						|
				framework.Failf("Node %s did not become ready within %v", node.Name, resizeNodeReadyTimeout)
 | 
						|
			}
 | 
						|
 | 
						|
			// sleep a bit, to allow Watch in NodeController to catch up.
 | 
						|
			time.Sleep(5 * time.Second)
 | 
						|
 | 
						|
			By("verify whether new pods can be created on the re-attached node")
 | 
						|
			// increasing the RC size is not a valid way to test this
 | 
						|
			// since we have no guarantees the pod will be scheduled on our node.
 | 
						|
			additionalPod := "additionalpod"
 | 
						|
			err = newPodOnNode(c, ns, additionalPod, node.Name)
 | 
						|
			Expect(err).NotTo(HaveOccurred())
 | 
						|
			err = framework.VerifyPods(c, ns, additionalPod, true, 1)
 | 
						|
			Expect(err).NotTo(HaveOccurred())
 | 
						|
 | 
						|
			// verify that it is really on the requested node
 | 
						|
			{
 | 
						|
				pod, err := c.Core().Pods(ns).Get(additionalPod, metav1.GetOptions{})
 | 
						|
				Expect(err).NotTo(HaveOccurred())
 | 
						|
				if pod.Spec.NodeName != node.Name {
 | 
						|
					framework.Logf("Pod %s found on invalid node: %s instead of %s", pod.Name, pod.Spec.NodeName, node.Name)
 | 
						|
				}
 | 
						|
			}
 | 
						|
		})
 | 
						|
 | 
						|
		It("should eagerly create replacement pod during network partition when termination grace is non-zero", func() {
 | 
						|
			// Create a replication controller for a service that serves its hostname.
 | 
						|
			// The source for the Docker container kubernetes/serve_hostname is in contrib/for-demos/serve_hostname
 | 
						|
			name := "my-hostname-net"
 | 
						|
			gracePeriod := int64(30)
 | 
						|
 | 
						|
			newSVCByName(c, ns, name)
 | 
						|
			replicas := int32(framework.TestContext.CloudConfig.NumNodes)
 | 
						|
			newRCByName(c, ns, name, replicas, &gracePeriod)
 | 
						|
			err := framework.VerifyPods(c, ns, name, true, replicas)
 | 
						|
			Expect(err).NotTo(HaveOccurred(), "Each pod should start running and responding")
 | 
						|
 | 
						|
			By("choose a node with at least one pod - we will block some network traffic on this node")
 | 
						|
			label := labels.SelectorFromSet(labels.Set(map[string]string{"name": name}))
 | 
						|
			options := v1.ListOptions{LabelSelector: label.String()}
 | 
						|
			pods, err := c.Core().Pods(ns).List(options) // list pods after all have been scheduled
 | 
						|
			Expect(err).NotTo(HaveOccurred())
 | 
						|
			nodeName := pods.Items[0].Spec.NodeName
 | 
						|
 | 
						|
			node, err := c.Core().Nodes().Get(nodeName, metav1.GetOptions{})
 | 
						|
			Expect(err).NotTo(HaveOccurred())
 | 
						|
 | 
						|
			// This creates a temporary network partition, verifies that 'podNameToDisappear',
 | 
						|
			// that belongs to replication controller 'rcName', did not disappear (because its
 | 
						|
			// grace period is set to 30).
 | 
						|
			// Finally, it checks that the replication controller recreates the
 | 
						|
			// pods on another node and that now the number of replicas is equal 'replicas + 1'.
 | 
						|
			By(fmt.Sprintf("blocking network traffic from node %s", node.Name))
 | 
						|
			testUnderTemporaryNetworkFailure(c, ns, node, func() {
 | 
						|
				framework.Logf("Waiting for pod %s to be removed", pods.Items[0].Name)
 | 
						|
				err := framework.WaitForRCPodToDisappear(c, ns, name, pods.Items[0].Name)
 | 
						|
				Expect(err).To(Equal(wait.ErrWaitTimeout), "Pod was not deleted during network partition.")
 | 
						|
 | 
						|
				By(fmt.Sprintf("verifying that there are %v running pods during partition", replicas))
 | 
						|
				_, err = framework.PodsCreated(c, ns, name, replicas)
 | 
						|
				Expect(err).NotTo(HaveOccurred())
 | 
						|
			})
 | 
						|
 | 
						|
			framework.Logf("Waiting %v for node %s to be ready once temporary network failure ends", resizeNodeReadyTimeout, node.Name)
 | 
						|
			if !framework.WaitForNodeToBeReady(c, node.Name, resizeNodeReadyTimeout) {
 | 
						|
				framework.Failf("Node %s did not become ready within %v", node.Name, resizeNodeReadyTimeout)
 | 
						|
			}
 | 
						|
		})
 | 
						|
	})
 | 
						|
 | 
						|
	framework.KubeDescribe("[StatefulSet]", func() {
 | 
						|
		psName := "ss"
 | 
						|
		labels := map[string]string{
 | 
						|
			"foo": "bar",
 | 
						|
		}
 | 
						|
		headlessSvcName := "test"
 | 
						|
 | 
						|
		BeforeEach(func() {
 | 
						|
			framework.SkipUnlessProviderIs("gce", "gke")
 | 
						|
			By("creating service " + headlessSvcName + " in namespace " + f.Namespace.Name)
 | 
						|
			headlessService := createServiceSpec(headlessSvcName, "", true, labels)
 | 
						|
			_, err := f.ClientSet.Core().Services(f.Namespace.Name).Create(headlessService)
 | 
						|
			framework.ExpectNoError(err)
 | 
						|
			c = f.ClientSet
 | 
						|
			ns = f.Namespace.Name
 | 
						|
		})
 | 
						|
 | 
						|
		AfterEach(func() {
 | 
						|
			if CurrentGinkgoTestDescription().Failed {
 | 
						|
				dumpDebugInfo(c, ns)
 | 
						|
			}
 | 
						|
			framework.Logf("Deleting all stateful set in ns %v", ns)
 | 
						|
			deleteAllStatefulSets(c, ns)
 | 
						|
		})
 | 
						|
 | 
						|
		It("should come back up if node goes down [Slow] [Disruptive]", func() {
 | 
						|
			petMounts := []v1.VolumeMount{{Name: "datadir", MountPath: "/data/"}}
 | 
						|
			podMounts := []v1.VolumeMount{{Name: "home", MountPath: "/home"}}
 | 
						|
			ps := newStatefulSet(psName, ns, headlessSvcName, 3, petMounts, podMounts, labels)
 | 
						|
			_, err := c.Apps().StatefulSets(ns).Create(ps)
 | 
						|
			Expect(err).NotTo(HaveOccurred())
 | 
						|
 | 
						|
			pst := statefulSetTester{c: c}
 | 
						|
 | 
						|
			nn := framework.TestContext.CloudConfig.NumNodes
 | 
						|
			nodeNames, err := framework.CheckNodesReady(f.ClientSet, framework.NodeReadyInitialTimeout, nn)
 | 
						|
			framework.ExpectNoError(err)
 | 
						|
			restartNodes(f, nodeNames)
 | 
						|
 | 
						|
			By("waiting for pods to be running again")
 | 
						|
			pst.waitForRunningAndReady(*ps.Spec.Replicas, ps)
 | 
						|
		})
 | 
						|
 | 
						|
		It("should not reschedule stateful pods if there is a network partition [Slow] [Disruptive]", func() {
 | 
						|
			ps := newStatefulSet(psName, ns, headlessSvcName, 3, []v1.VolumeMount{}, []v1.VolumeMount{}, labels)
 | 
						|
			_, err := c.Apps().StatefulSets(ns).Create(ps)
 | 
						|
			Expect(err).NotTo(HaveOccurred())
 | 
						|
 | 
						|
			pst := statefulSetTester{c: c}
 | 
						|
			pst.waitForRunningAndReady(*ps.Spec.Replicas, ps)
 | 
						|
 | 
						|
			pod := pst.getPodList(ps).Items[0]
 | 
						|
			node, err := c.Core().Nodes().Get(pod.Spec.NodeName, metav1.GetOptions{})
 | 
						|
			framework.ExpectNoError(err)
 | 
						|
 | 
						|
			// Blocks outgoing network traffic on 'node'. Then verifies that 'podNameToDisappear',
 | 
						|
			// that belongs to StatefulSet 'statefulSetName', **does not** disappear due to forced deletion from the apiserver.
 | 
						|
			// The grace period on the stateful pods is set to a value > 0.
 | 
						|
			testUnderTemporaryNetworkFailure(c, ns, node, func() {
 | 
						|
				framework.Logf("Checking that the NodeController does not force delete stateful pods %v", pod.Name)
 | 
						|
				err := framework.WaitTimeoutForPodNoLongerRunningInNamespace(c, pod.Name, ns, pod.ResourceVersion, 10*time.Minute)
 | 
						|
				Expect(err).To(Equal(wait.ErrWaitTimeout), "Pod was not deleted during network partition.")
 | 
						|
			})
 | 
						|
 | 
						|
			framework.Logf("Waiting %v for node %s to be ready once temporary network failure ends", resizeNodeReadyTimeout, node.Name)
 | 
						|
			if !framework.WaitForNodeToBeReady(c, node.Name, resizeNodeReadyTimeout) {
 | 
						|
				framework.Failf("Node %s did not become ready within %v", node.Name, resizeNodeReadyTimeout)
 | 
						|
			}
 | 
						|
 | 
						|
			By("waiting for pods to be running again")
 | 
						|
			pst.waitForRunningAndReady(*ps.Spec.Replicas, ps)
 | 
						|
		})
 | 
						|
	})
 | 
						|
 | 
						|
	framework.KubeDescribe("[Job]", func() {
 | 
						|
		It("should create new pods when node is partitioned", func() {
 | 
						|
			parallelism := int32(2)
 | 
						|
			completions := int32(4)
 | 
						|
 | 
						|
			job := newTestJob("notTerminate", "network-partition", v1.RestartPolicyNever, parallelism, completions)
 | 
						|
			job, err := createJob(f.ClientSet, f.Namespace.Name, job)
 | 
						|
			Expect(err).NotTo(HaveOccurred())
 | 
						|
			label := labels.SelectorFromSet(labels.Set(map[string]string{jobSelectorKey: job.Name}))
 | 
						|
 | 
						|
			By(fmt.Sprintf("verifying that there are now %v running pods", parallelism))
 | 
						|
			_, err = framework.PodsCreatedByLabel(c, ns, job.Name, parallelism, label)
 | 
						|
			Expect(err).NotTo(HaveOccurred())
 | 
						|
 | 
						|
			By("choose a node with at least one pod - we will block some network traffic on this node")
 | 
						|
			options := v1.ListOptions{LabelSelector: label.String()}
 | 
						|
			pods, err := c.Core().Pods(ns).List(options) // list pods after all have been scheduled
 | 
						|
			Expect(err).NotTo(HaveOccurred())
 | 
						|
			nodeName := pods.Items[0].Spec.NodeName
 | 
						|
 | 
						|
			node, err := c.Core().Nodes().Get(nodeName, metav1.GetOptions{})
 | 
						|
			Expect(err).NotTo(HaveOccurred())
 | 
						|
 | 
						|
			// This creates a temporary network partition, verifies that the job has 'parallelism' number of
 | 
						|
			// running pods after the node-controller detects node unreachable.
 | 
						|
			By(fmt.Sprintf("blocking network traffic from node %s", node.Name))
 | 
						|
			testUnderTemporaryNetworkFailure(c, ns, node, func() {
 | 
						|
				framework.Logf("Waiting for pod %s to be removed", pods.Items[0].Name)
 | 
						|
				err := framework.WaitForPodToDisappear(c, ns, pods.Items[0].Name, label, 20*time.Second, 10*time.Minute)
 | 
						|
				Expect(err).To(Equal(wait.ErrWaitTimeout), "Pod was not deleted during network partition.")
 | 
						|
 | 
						|
				By(fmt.Sprintf("verifying that there are now %v running pods", parallelism))
 | 
						|
				_, err = framework.PodsCreatedByLabel(c, ns, job.Name, parallelism, label)
 | 
						|
				Expect(err).NotTo(HaveOccurred())
 | 
						|
			})
 | 
						|
 | 
						|
			framework.Logf("Waiting %v for node %s to be ready once temporary network failure ends", resizeNodeReadyTimeout, node.Name)
 | 
						|
			if !framework.WaitForNodeToBeReady(c, node.Name, resizeNodeReadyTimeout) {
 | 
						|
				framework.Failf("Node %s did not become ready within %v", node.Name, resizeNodeReadyTimeout)
 | 
						|
			}
 | 
						|
		})
 | 
						|
	})
 | 
						|
})
 |