mirror of
https://github.com/k3s-io/kubernetes.git
synced 2025-07-25 12:43:23 +00:00
Merge pull request #77890 from oomichi/use-ExpectNoError-scalability
Use ExpectNoError() for scalability and scheduling
This commit is contained in:
commit
c2633d8969
@ -347,7 +347,7 @@ func createClients(numberOfClients int) ([]clientset.Interface, []scaleclient.Sc
|
|||||||
|
|
||||||
for i := 0; i < numberOfClients; i++ {
|
for i := 0; i < numberOfClients; i++ {
|
||||||
config, err := framework.LoadConfig()
|
config, err := framework.LoadConfig()
|
||||||
Expect(err).NotTo(HaveOccurred())
|
framework.ExpectNoError(err)
|
||||||
config.QPS = 100
|
config.QPS = 100
|
||||||
config.Burst = 200
|
config.Burst = 200
|
||||||
if framework.TestContext.KubeAPIContentType != "" {
|
if framework.TestContext.KubeAPIContentType != "" {
|
||||||
|
@ -62,7 +62,7 @@ var _ = framework.KubeDescribe("EquivalenceCache [Serial]", func() {
|
|||||||
// cannot be run in parallel with any other test that touches Nodes or Pods.
|
// cannot be run in parallel with any other test that touches Nodes or Pods.
|
||||||
// It is so because we need to have precise control on what's running in the cluster.
|
// It is so because we need to have precise control on what's running in the cluster.
|
||||||
systemPods, err := framework.GetPodsInNamespace(cs, ns, map[string]string{})
|
systemPods, err := framework.GetPodsInNamespace(cs, ns, map[string]string{})
|
||||||
gomega.Expect(err).NotTo(gomega.HaveOccurred())
|
framework.ExpectNoError(err)
|
||||||
systemPodsNo = 0
|
systemPodsNo = 0
|
||||||
for _, pod := range systemPods {
|
for _, pod := range systemPods {
|
||||||
if !masterNodes.Has(pod.Spec.NodeName) && pod.DeletionTimestamp == nil {
|
if !masterNodes.Has(pod.Spec.NodeName) && pod.DeletionTimestamp == nil {
|
||||||
@ -71,7 +71,7 @@ var _ = framework.KubeDescribe("EquivalenceCache [Serial]", func() {
|
|||||||
}
|
}
|
||||||
|
|
||||||
err = framework.WaitForPodsRunningReady(cs, api.NamespaceSystem, int32(systemPodsNo), int32(systemPodsNo), framework.PodReadyBeforeTimeout, map[string]string{})
|
err = framework.WaitForPodsRunningReady(cs, api.NamespaceSystem, int32(systemPodsNo), int32(systemPodsNo), framework.PodReadyBeforeTimeout, map[string]string{})
|
||||||
gomega.Expect(err).NotTo(gomega.HaveOccurred())
|
framework.ExpectNoError(err)
|
||||||
|
|
||||||
for _, node := range nodeList.Items {
|
for _, node := range nodeList.Items {
|
||||||
e2elog.Logf("\nLogging pods the kubelet thinks is on node %v before test", node.Name)
|
e2elog.Logf("\nLogging pods the kubelet thinks is on node %v before test", node.Name)
|
||||||
|
@ -58,18 +58,18 @@ var _ = SIGDescribe("LimitRange", func() {
|
|||||||
selector := labels.SelectorFromSet(labels.Set(map[string]string{"name": limitRange.Name}))
|
selector := labels.SelectorFromSet(labels.Set(map[string]string{"name": limitRange.Name}))
|
||||||
options := metav1.ListOptions{LabelSelector: selector.String()}
|
options := metav1.ListOptions{LabelSelector: selector.String()}
|
||||||
limitRanges, err := f.ClientSet.CoreV1().LimitRanges(f.Namespace.Name).List(options)
|
limitRanges, err := f.ClientSet.CoreV1().LimitRanges(f.Namespace.Name).List(options)
|
||||||
gomega.Expect(err).NotTo(gomega.HaveOccurred(), "failed to query for limitRanges")
|
framework.ExpectNoError(err, "failed to query for limitRanges")
|
||||||
gomega.Expect(len(limitRanges.Items)).To(gomega.Equal(0))
|
gomega.Expect(len(limitRanges.Items)).To(gomega.Equal(0))
|
||||||
options = metav1.ListOptions{
|
options = metav1.ListOptions{
|
||||||
LabelSelector: selector.String(),
|
LabelSelector: selector.String(),
|
||||||
ResourceVersion: limitRanges.ListMeta.ResourceVersion,
|
ResourceVersion: limitRanges.ListMeta.ResourceVersion,
|
||||||
}
|
}
|
||||||
w, err := f.ClientSet.CoreV1().LimitRanges(f.Namespace.Name).Watch(metav1.ListOptions{})
|
w, err := f.ClientSet.CoreV1().LimitRanges(f.Namespace.Name).Watch(metav1.ListOptions{})
|
||||||
gomega.Expect(err).NotTo(gomega.HaveOccurred(), "failed to set up watch")
|
framework.ExpectNoError(err, "failed to set up watch")
|
||||||
|
|
||||||
ginkgo.By("Submitting a LimitRange")
|
ginkgo.By("Submitting a LimitRange")
|
||||||
limitRange, err = f.ClientSet.CoreV1().LimitRanges(f.Namespace.Name).Create(limitRange)
|
limitRange, err = f.ClientSet.CoreV1().LimitRanges(f.Namespace.Name).Create(limitRange)
|
||||||
gomega.Expect(err).NotTo(gomega.HaveOccurred())
|
framework.ExpectNoError(err)
|
||||||
|
|
||||||
ginkgo.By("Verifying LimitRange creation was observed")
|
ginkgo.By("Verifying LimitRange creation was observed")
|
||||||
select {
|
select {
|
||||||
@ -83,37 +83,37 @@ var _ = SIGDescribe("LimitRange", func() {
|
|||||||
|
|
||||||
ginkgo.By("Fetching the LimitRange to ensure it has proper values")
|
ginkgo.By("Fetching the LimitRange to ensure it has proper values")
|
||||||
limitRange, err = f.ClientSet.CoreV1().LimitRanges(f.Namespace.Name).Get(limitRange.Name, metav1.GetOptions{})
|
limitRange, err = f.ClientSet.CoreV1().LimitRanges(f.Namespace.Name).Get(limitRange.Name, metav1.GetOptions{})
|
||||||
gomega.Expect(err).NotTo(gomega.HaveOccurred())
|
framework.ExpectNoError(err)
|
||||||
expected := v1.ResourceRequirements{Requests: defaultRequest, Limits: defaultLimit}
|
expected := v1.ResourceRequirements{Requests: defaultRequest, Limits: defaultLimit}
|
||||||
actual := v1.ResourceRequirements{Requests: limitRange.Spec.Limits[0].DefaultRequest, Limits: limitRange.Spec.Limits[0].Default}
|
actual := v1.ResourceRequirements{Requests: limitRange.Spec.Limits[0].DefaultRequest, Limits: limitRange.Spec.Limits[0].Default}
|
||||||
err = equalResourceRequirement(expected, actual)
|
err = equalResourceRequirement(expected, actual)
|
||||||
gomega.Expect(err).NotTo(gomega.HaveOccurred())
|
framework.ExpectNoError(err)
|
||||||
|
|
||||||
ginkgo.By("Creating a Pod with no resource requirements")
|
ginkgo.By("Creating a Pod with no resource requirements")
|
||||||
pod := f.NewTestPod("pod-no-resources", v1.ResourceList{}, v1.ResourceList{})
|
pod := f.NewTestPod("pod-no-resources", v1.ResourceList{}, v1.ResourceList{})
|
||||||
pod, err = f.ClientSet.CoreV1().Pods(f.Namespace.Name).Create(pod)
|
pod, err = f.ClientSet.CoreV1().Pods(f.Namespace.Name).Create(pod)
|
||||||
gomega.Expect(err).NotTo(gomega.HaveOccurred())
|
framework.ExpectNoError(err)
|
||||||
|
|
||||||
ginkgo.By("Ensuring Pod has resource requirements applied from LimitRange")
|
ginkgo.By("Ensuring Pod has resource requirements applied from LimitRange")
|
||||||
pod, err = f.ClientSet.CoreV1().Pods(f.Namespace.Name).Get(pod.Name, metav1.GetOptions{})
|
pod, err = f.ClientSet.CoreV1().Pods(f.Namespace.Name).Get(pod.Name, metav1.GetOptions{})
|
||||||
gomega.Expect(err).NotTo(gomega.HaveOccurred())
|
framework.ExpectNoError(err)
|
||||||
for i := range pod.Spec.Containers {
|
for i := range pod.Spec.Containers {
|
||||||
err = equalResourceRequirement(expected, pod.Spec.Containers[i].Resources)
|
err = equalResourceRequirement(expected, pod.Spec.Containers[i].Resources)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
// Print the pod to help in debugging.
|
// Print the pod to help in debugging.
|
||||||
e2elog.Logf("Pod %+v does not have the expected requirements", pod)
|
e2elog.Logf("Pod %+v does not have the expected requirements", pod)
|
||||||
gomega.Expect(err).NotTo(gomega.HaveOccurred())
|
framework.ExpectNoError(err)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
ginkgo.By("Creating a Pod with partial resource requirements")
|
ginkgo.By("Creating a Pod with partial resource requirements")
|
||||||
pod = f.NewTestPod("pod-partial-resources", getResourceList("", "150Mi", "150Gi"), getResourceList("300m", "", ""))
|
pod = f.NewTestPod("pod-partial-resources", getResourceList("", "150Mi", "150Gi"), getResourceList("300m", "", ""))
|
||||||
pod, err = f.ClientSet.CoreV1().Pods(f.Namespace.Name).Create(pod)
|
pod, err = f.ClientSet.CoreV1().Pods(f.Namespace.Name).Create(pod)
|
||||||
gomega.Expect(err).NotTo(gomega.HaveOccurred())
|
framework.ExpectNoError(err)
|
||||||
|
|
||||||
ginkgo.By("Ensuring Pod has merged resource requirements applied from LimitRange")
|
ginkgo.By("Ensuring Pod has merged resource requirements applied from LimitRange")
|
||||||
pod, err = f.ClientSet.CoreV1().Pods(f.Namespace.Name).Get(pod.Name, metav1.GetOptions{})
|
pod, err = f.ClientSet.CoreV1().Pods(f.Namespace.Name).Get(pod.Name, metav1.GetOptions{})
|
||||||
gomega.Expect(err).NotTo(gomega.HaveOccurred())
|
framework.ExpectNoError(err)
|
||||||
// This is an interesting case, so it's worth a comment
|
// This is an interesting case, so it's worth a comment
|
||||||
// If you specify a Limit, and no Request, the Limit will default to the Request
|
// If you specify a Limit, and no Request, the Limit will default to the Request
|
||||||
// This means that the LimitRange.DefaultRequest will ONLY take affect if a container.resources.limit is not supplied
|
// This means that the LimitRange.DefaultRequest will ONLY take affect if a container.resources.limit is not supplied
|
||||||
@ -123,7 +123,7 @@ var _ = SIGDescribe("LimitRange", func() {
|
|||||||
if err != nil {
|
if err != nil {
|
||||||
// Print the pod to help in debugging.
|
// Print the pod to help in debugging.
|
||||||
e2elog.Logf("Pod %+v does not have the expected requirements", pod)
|
e2elog.Logf("Pod %+v does not have the expected requirements", pod)
|
||||||
gomega.Expect(err).NotTo(gomega.HaveOccurred())
|
framework.ExpectNoError(err)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -141,19 +141,20 @@ var _ = SIGDescribe("LimitRange", func() {
|
|||||||
newMin := getResourceList("9m", "49Mi", "49Gi")
|
newMin := getResourceList("9m", "49Mi", "49Gi")
|
||||||
limitRange.Spec.Limits[0].Min = newMin
|
limitRange.Spec.Limits[0].Min = newMin
|
||||||
limitRange, err = f.ClientSet.CoreV1().LimitRanges(f.Namespace.Name).Update(limitRange)
|
limitRange, err = f.ClientSet.CoreV1().LimitRanges(f.Namespace.Name).Update(limitRange)
|
||||||
gomega.Expect(err).NotTo(gomega.HaveOccurred())
|
framework.ExpectNoError(err)
|
||||||
|
|
||||||
ginkgo.By("Verifying LimitRange updating is effective")
|
ginkgo.By("Verifying LimitRange updating is effective")
|
||||||
gomega.Expect(wait.Poll(time.Second*2, time.Second*20, func() (bool, error) {
|
err = wait.Poll(time.Second*2, time.Second*20, func() (bool, error) {
|
||||||
limitRange, err = f.ClientSet.CoreV1().LimitRanges(f.Namespace.Name).Get(limitRange.Name, metav1.GetOptions{})
|
limitRange, err = f.ClientSet.CoreV1().LimitRanges(f.Namespace.Name).Get(limitRange.Name, metav1.GetOptions{})
|
||||||
gomega.Expect(err).NotTo(gomega.HaveOccurred())
|
gomega.Expect(err).NotTo(gomega.HaveOccurred())
|
||||||
return reflect.DeepEqual(limitRange.Spec.Limits[0].Min, newMin), nil
|
return reflect.DeepEqual(limitRange.Spec.Limits[0].Min, newMin), nil
|
||||||
})).NotTo(gomega.HaveOccurred())
|
})
|
||||||
|
framework.ExpectNoError(err)
|
||||||
|
|
||||||
ginkgo.By("Creating a Pod with less than former min resources")
|
ginkgo.By("Creating a Pod with less than former min resources")
|
||||||
pod = f.NewTestPod(podName, getResourceList("10m", "50Mi", "50Gi"), v1.ResourceList{})
|
pod = f.NewTestPod(podName, getResourceList("10m", "50Mi", "50Gi"), v1.ResourceList{})
|
||||||
pod, err = f.ClientSet.CoreV1().Pods(f.Namespace.Name).Create(pod)
|
pod, err = f.ClientSet.CoreV1().Pods(f.Namespace.Name).Create(pod)
|
||||||
gomega.Expect(err).NotTo(gomega.HaveOccurred())
|
framework.ExpectNoError(err)
|
||||||
|
|
||||||
ginkgo.By("Failing to create a Pod with more than max resources")
|
ginkgo.By("Failing to create a Pod with more than max resources")
|
||||||
pod = f.NewTestPod(podName, getResourceList("600m", "600Mi", "600Gi"), v1.ResourceList{})
|
pod = f.NewTestPod(podName, getResourceList("600m", "600Mi", "600Gi"), v1.ResourceList{})
|
||||||
@ -162,10 +163,10 @@ var _ = SIGDescribe("LimitRange", func() {
|
|||||||
|
|
||||||
ginkgo.By("Deleting a LimitRange")
|
ginkgo.By("Deleting a LimitRange")
|
||||||
err = f.ClientSet.CoreV1().LimitRanges(f.Namespace.Name).Delete(limitRange.Name, metav1.NewDeleteOptions(30))
|
err = f.ClientSet.CoreV1().LimitRanges(f.Namespace.Name).Delete(limitRange.Name, metav1.NewDeleteOptions(30))
|
||||||
gomega.Expect(err).NotTo(gomega.HaveOccurred())
|
framework.ExpectNoError(err)
|
||||||
|
|
||||||
ginkgo.By("Verifying the LimitRange was deleted")
|
ginkgo.By("Verifying the LimitRange was deleted")
|
||||||
gomega.Expect(wait.Poll(time.Second*5, time.Second*30, func() (bool, error) {
|
err = wait.Poll(time.Second*5, time.Second*30, func() (bool, error) {
|
||||||
selector := labels.SelectorFromSet(labels.Set(map[string]string{"name": limitRange.Name}))
|
selector := labels.SelectorFromSet(labels.Set(map[string]string{"name": limitRange.Name}))
|
||||||
options := metav1.ListOptions{LabelSelector: selector.String()}
|
options := metav1.ListOptions{LabelSelector: selector.String()}
|
||||||
limitRanges, err := f.ClientSet.CoreV1().LimitRanges(f.Namespace.Name).List(options)
|
limitRanges, err := f.ClientSet.CoreV1().LimitRanges(f.Namespace.Name).List(options)
|
||||||
@ -190,12 +191,13 @@ var _ = SIGDescribe("LimitRange", func() {
|
|||||||
|
|
||||||
return false, nil
|
return false, nil
|
||||||
|
|
||||||
})).NotTo(gomega.HaveOccurred(), "kubelet never observed the termination notice")
|
})
|
||||||
|
framework.ExpectNoError(err, "kubelet never observed the termination notice")
|
||||||
|
|
||||||
ginkgo.By("Creating a Pod with more than former max resources")
|
ginkgo.By("Creating a Pod with more than former max resources")
|
||||||
pod = f.NewTestPod(podName+"2", getResourceList("600m", "600Mi", "600Gi"), v1.ResourceList{})
|
pod = f.NewTestPod(podName+"2", getResourceList("600m", "600Mi", "600Gi"), v1.ResourceList{})
|
||||||
pod, err = f.ClientSet.CoreV1().Pods(f.Namespace.Name).Create(pod)
|
pod, err = f.ClientSet.CoreV1().Pods(f.Namespace.Name).Create(pod)
|
||||||
gomega.Expect(err).NotTo(gomega.HaveOccurred())
|
framework.ExpectNoError(err)
|
||||||
})
|
})
|
||||||
|
|
||||||
})
|
})
|
||||||
|
@ -713,7 +713,7 @@ func WaitForSchedulerAfterAction(f *framework.Framework, action common.Action, n
|
|||||||
predicate = scheduleSuccessEvent(ns, podName, "" /* any node */)
|
predicate = scheduleSuccessEvent(ns, podName, "" /* any node */)
|
||||||
}
|
}
|
||||||
success, err := common.ObserveEventAfterAction(f, predicate, action)
|
success, err := common.ObserveEventAfterAction(f, predicate, action)
|
||||||
gomega.Expect(err).NotTo(gomega.HaveOccurred())
|
framework.ExpectNoError(err)
|
||||||
gomega.Expect(success).To(gomega.Equal(true))
|
gomega.Expect(success).To(gomega.Equal(true))
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -82,7 +82,7 @@ var _ = SIGDescribe("SchedulerPriorities [Serial]", func() {
|
|||||||
err := framework.CheckTestingNSDeletedExcept(cs, ns)
|
err := framework.CheckTestingNSDeletedExcept(cs, ns)
|
||||||
framework.ExpectNoError(err)
|
framework.ExpectNoError(err)
|
||||||
err = framework.WaitForPodsRunningReady(cs, metav1.NamespaceSystem, int32(systemPodsNo), 0, framework.PodReadyBeforeTimeout, map[string]string{})
|
err = framework.WaitForPodsRunningReady(cs, metav1.NamespaceSystem, int32(systemPodsNo), 0, framework.PodReadyBeforeTimeout, map[string]string{})
|
||||||
gomega.Expect(err).NotTo(gomega.HaveOccurred())
|
framework.ExpectNoError(err)
|
||||||
})
|
})
|
||||||
|
|
||||||
ginkgo.It("Pod should be scheduled to node that don't match the PodAntiAffinity terms", func() {
|
ginkgo.It("Pod should be scheduled to node that don't match the PodAntiAffinity terms", func() {
|
||||||
@ -191,7 +191,7 @@ var _ = SIGDescribe("SchedulerPriorities [Serial]", func() {
|
|||||||
return node.Annotations[v1.PreferAvoidPodsAnnotationKey] == string(val)
|
return node.Annotations[v1.PreferAvoidPodsAnnotationKey] == string(val)
|
||||||
}
|
}
|
||||||
success, err := common.ObserveNodeUpdateAfterAction(f, nodeName, predicate, action)
|
success, err := common.ObserveNodeUpdateAfterAction(f, nodeName, predicate, action)
|
||||||
gomega.Expect(err).NotTo(gomega.HaveOccurred())
|
framework.ExpectNoError(err)
|
||||||
gomega.Expect(success).To(gomega.Equal(true))
|
gomega.Expect(success).To(gomega.Equal(true))
|
||||||
|
|
||||||
defer framework.RemoveAvoidPodsOffNode(cs, nodeName)
|
defer framework.RemoveAvoidPodsOffNode(cs, nodeName)
|
||||||
@ -202,7 +202,7 @@ var _ = SIGDescribe("SchedulerPriorities [Serial]", func() {
|
|||||||
testPods, err := cs.CoreV1().Pods(ns).List(metav1.ListOptions{
|
testPods, err := cs.CoreV1().Pods(ns).List(metav1.ListOptions{
|
||||||
LabelSelector: "name=scheduler-priority-avoid-pod",
|
LabelSelector: "name=scheduler-priority-avoid-pod",
|
||||||
})
|
})
|
||||||
gomega.Expect(err).NotTo(gomega.HaveOccurred())
|
framework.ExpectNoError(err)
|
||||||
ginkgo.By(fmt.Sprintf("Verify the pods should not scheduled to the node: %s", nodeName))
|
ginkgo.By(fmt.Sprintf("Verify the pods should not scheduled to the node: %s", nodeName))
|
||||||
for _, pod := range testPods.Items {
|
for _, pod := range testPods.Items {
|
||||||
gomega.Expect(pod.Spec.NodeName).NotTo(gomega.Equal(nodeName))
|
gomega.Expect(pod.Spec.NodeName).NotTo(gomega.Equal(nodeName))
|
||||||
@ -235,7 +235,7 @@ var _ = SIGDescribe("SchedulerPriorities [Serial]", func() {
|
|||||||
|
|
||||||
ginkgo.By("Pod should prefer scheduled to the node don't have the taint.")
|
ginkgo.By("Pod should prefer scheduled to the node don't have the taint.")
|
||||||
tolePod, err := cs.CoreV1().Pods(ns).Get(tolerationPodName, metav1.GetOptions{})
|
tolePod, err := cs.CoreV1().Pods(ns).Get(tolerationPodName, metav1.GetOptions{})
|
||||||
gomega.Expect(err).NotTo(gomega.HaveOccurred())
|
framework.ExpectNoError(err)
|
||||||
gomega.Expect(tolePod.Spec.NodeName).To(gomega.Equal(nodeName))
|
gomega.Expect(tolePod.Spec.NodeName).To(gomega.Equal(nodeName))
|
||||||
|
|
||||||
ginkgo.By("Trying to apply 10 taint on the first node.")
|
ginkgo.By("Trying to apply 10 taint on the first node.")
|
||||||
@ -255,7 +255,7 @@ var _ = SIGDescribe("SchedulerPriorities [Serial]", func() {
|
|||||||
|
|
||||||
ginkgo.By("Pod should prefer scheduled to the node that pod can tolerate.")
|
ginkgo.By("Pod should prefer scheduled to the node that pod can tolerate.")
|
||||||
tolePod, err = cs.CoreV1().Pods(ns).Get(tolerationPodName, metav1.GetOptions{})
|
tolePod, err = cs.CoreV1().Pods(ns).Get(tolerationPodName, metav1.GetOptions{})
|
||||||
gomega.Expect(err).NotTo(gomega.HaveOccurred())
|
framework.ExpectNoError(err)
|
||||||
gomega.Expect(tolePod.Spec.NodeName).To(gomega.Equal(nodeName))
|
gomega.Expect(tolePod.Spec.NodeName).To(gomega.Equal(nodeName))
|
||||||
})
|
})
|
||||||
})
|
})
|
||||||
@ -400,7 +400,7 @@ func createRC(ns, rsName string, replicas int32, rcPodLabels map[string]string,
|
|||||||
},
|
},
|
||||||
}
|
}
|
||||||
rc, err := f.ClientSet.CoreV1().ReplicationControllers(ns).Create(rc)
|
rc, err := f.ClientSet.CoreV1().ReplicationControllers(ns).Create(rc)
|
||||||
gomega.Expect(err).NotTo(gomega.HaveOccurred())
|
framework.ExpectNoError(err)
|
||||||
return rc
|
return rc
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -43,7 +43,7 @@ var _ = SIGDescribe("Multi-AZ Clusters", func() {
|
|||||||
framework.SkipUnlessProviderIs("gce", "gke", "aws")
|
framework.SkipUnlessProviderIs("gce", "gke", "aws")
|
||||||
if zoneCount <= 0 {
|
if zoneCount <= 0 {
|
||||||
zoneCount, err = getZoneCount(f.ClientSet)
|
zoneCount, err = getZoneCount(f.ClientSet)
|
||||||
gomega.Expect(err).NotTo(gomega.HaveOccurred())
|
framework.ExpectNoError(err)
|
||||||
}
|
}
|
||||||
ginkgo.By(fmt.Sprintf("Checking for multi-zone cluster. Zone count = %d", zoneCount))
|
ginkgo.By(fmt.Sprintf("Checking for multi-zone cluster. Zone count = %d", zoneCount))
|
||||||
msg := fmt.Sprintf("Zone count is %d, only run for multi-zone clusters, skipping test", zoneCount)
|
msg := fmt.Sprintf("Zone count is %d, only run for multi-zone clusters, skipping test", zoneCount)
|
||||||
@ -80,7 +80,7 @@ func SpreadServiceOrFail(f *framework.Framework, replicaCount int, image string)
|
|||||||
},
|
},
|
||||||
}
|
}
|
||||||
_, err := f.ClientSet.CoreV1().Services(f.Namespace.Name).Create(serviceSpec)
|
_, err := f.ClientSet.CoreV1().Services(f.Namespace.Name).Create(serviceSpec)
|
||||||
gomega.Expect(err).NotTo(gomega.HaveOccurred())
|
framework.ExpectNoError(err)
|
||||||
|
|
||||||
// Now create some pods behind the service
|
// Now create some pods behind the service
|
||||||
podSpec := &v1.Pod{
|
podSpec := &v1.Pod{
|
||||||
@ -107,11 +107,11 @@ func SpreadServiceOrFail(f *framework.Framework, replicaCount int, image string)
|
|||||||
// Wait for all of them to be scheduled
|
// Wait for all of them to be scheduled
|
||||||
selector := labels.SelectorFromSet(labels.Set(map[string]string{"service": serviceName}))
|
selector := labels.SelectorFromSet(labels.Set(map[string]string{"service": serviceName}))
|
||||||
pods, err := framework.WaitForPodsWithLabelScheduled(f.ClientSet, f.Namespace.Name, selector)
|
pods, err := framework.WaitForPodsWithLabelScheduled(f.ClientSet, f.Namespace.Name, selector)
|
||||||
gomega.Expect(err).NotTo(gomega.HaveOccurred())
|
framework.ExpectNoError(err)
|
||||||
|
|
||||||
// Now make sure they're spread across zones
|
// Now make sure they're spread across zones
|
||||||
zoneNames, err := framework.GetClusterZones(f.ClientSet)
|
zoneNames, err := framework.GetClusterZones(f.ClientSet)
|
||||||
gomega.Expect(err).NotTo(gomega.HaveOccurred())
|
framework.ExpectNoError(err)
|
||||||
gomega.Expect(checkZoneSpreading(f.ClientSet, pods, zoneNames.List())).To(gomega.Equal(true))
|
gomega.Expect(checkZoneSpreading(f.ClientSet, pods, zoneNames.List())).To(gomega.Equal(true))
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -139,7 +139,7 @@ func getZoneCount(c clientset.Interface) (int, error) {
|
|||||||
func getZoneNameForPod(c clientset.Interface, pod v1.Pod) (string, error) {
|
func getZoneNameForPod(c clientset.Interface, pod v1.Pod) (string, error) {
|
||||||
ginkgo.By(fmt.Sprintf("Getting zone name for pod %s, on node %s", pod.Name, pod.Spec.NodeName))
|
ginkgo.By(fmt.Sprintf("Getting zone name for pod %s, on node %s", pod.Name, pod.Spec.NodeName))
|
||||||
node, err := c.CoreV1().Nodes().Get(pod.Spec.NodeName, metav1.GetOptions{})
|
node, err := c.CoreV1().Nodes().Get(pod.Spec.NodeName, metav1.GetOptions{})
|
||||||
gomega.Expect(err).NotTo(gomega.HaveOccurred())
|
framework.ExpectNoError(err)
|
||||||
return getZoneNameForNode(*node)
|
return getZoneNameForNode(*node)
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -155,7 +155,7 @@ func checkZoneSpreading(c clientset.Interface, pods *v1.PodList, zoneNames []str
|
|||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
zoneName, err := getZoneNameForPod(c, pod)
|
zoneName, err := getZoneNameForPod(c, pod)
|
||||||
gomega.Expect(err).NotTo(gomega.HaveOccurred())
|
framework.ExpectNoError(err)
|
||||||
podsPerZone[zoneName] = podsPerZone[zoneName] + 1
|
podsPerZone[zoneName] = podsPerZone[zoneName] + 1
|
||||||
}
|
}
|
||||||
minPodsPerZone := math.MaxInt32
|
minPodsPerZone := math.MaxInt32
|
||||||
@ -205,7 +205,7 @@ func SpreadRCOrFail(f *framework.Framework, replicaCount int32, image string) {
|
|||||||
},
|
},
|
||||||
},
|
},
|
||||||
})
|
})
|
||||||
gomega.Expect(err).NotTo(gomega.HaveOccurred())
|
framework.ExpectNoError(err)
|
||||||
// Cleanup the replication controller when we are done.
|
// Cleanup the replication controller when we are done.
|
||||||
defer func() {
|
defer func() {
|
||||||
// Resize the replication controller to zero to get rid of pods.
|
// Resize the replication controller to zero to get rid of pods.
|
||||||
@ -216,15 +216,15 @@ func SpreadRCOrFail(f *framework.Framework, replicaCount int32, image string) {
|
|||||||
// List the pods, making sure we observe all the replicas.
|
// List the pods, making sure we observe all the replicas.
|
||||||
selector := labels.SelectorFromSet(labels.Set(map[string]string{"name": name}))
|
selector := labels.SelectorFromSet(labels.Set(map[string]string{"name": name}))
|
||||||
pods, err := framework.PodsCreated(f.ClientSet, f.Namespace.Name, name, replicaCount)
|
pods, err := framework.PodsCreated(f.ClientSet, f.Namespace.Name, name, replicaCount)
|
||||||
gomega.Expect(err).NotTo(gomega.HaveOccurred())
|
framework.ExpectNoError(err)
|
||||||
|
|
||||||
// Wait for all of them to be scheduled
|
// Wait for all of them to be scheduled
|
||||||
ginkgo.By(fmt.Sprintf("Waiting for %d replicas of %s to be scheduled. Selector: %v", replicaCount, name, selector))
|
ginkgo.By(fmt.Sprintf("Waiting for %d replicas of %s to be scheduled. Selector: %v", replicaCount, name, selector))
|
||||||
pods, err = framework.WaitForPodsWithLabelScheduled(f.ClientSet, f.Namespace.Name, selector)
|
pods, err = framework.WaitForPodsWithLabelScheduled(f.ClientSet, f.Namespace.Name, selector)
|
||||||
gomega.Expect(err).NotTo(gomega.HaveOccurred())
|
framework.ExpectNoError(err)
|
||||||
|
|
||||||
// Now make sure they're spread across zones
|
// Now make sure they're spread across zones
|
||||||
zoneNames, err := framework.GetClusterZones(f.ClientSet)
|
zoneNames, err := framework.GetClusterZones(f.ClientSet)
|
||||||
gomega.Expect(err).NotTo(gomega.HaveOccurred())
|
framework.ExpectNoError(err)
|
||||||
gomega.Expect(checkZoneSpreading(f.ClientSet, pods, zoneNames.List())).To(gomega.Equal(true))
|
gomega.Expect(checkZoneSpreading(f.ClientSet, pods, zoneNames.List())).To(gomega.Equal(true))
|
||||||
}
|
}
|
||||||
|
@ -42,7 +42,7 @@ var _ = SIGDescribe("Multi-AZ Cluster Volumes [sig-storage]", func() {
|
|||||||
framework.SkipUnlessProviderIs("gce", "gke")
|
framework.SkipUnlessProviderIs("gce", "gke")
|
||||||
if zoneCount <= 0 {
|
if zoneCount <= 0 {
|
||||||
zoneCount, err = getZoneCount(f.ClientSet)
|
zoneCount, err = getZoneCount(f.ClientSet)
|
||||||
gomega.Expect(err).NotTo(gomega.HaveOccurred())
|
framework.ExpectNoError(err)
|
||||||
}
|
}
|
||||||
ginkgo.By(fmt.Sprintf("Checking for multi-zone cluster. Zone count = %d", zoneCount))
|
ginkgo.By(fmt.Sprintf("Checking for multi-zone cluster. Zone count = %d", zoneCount))
|
||||||
msg := fmt.Sprintf("Zone count is %d, only run for multi-zone clusters, skipping test", zoneCount)
|
msg := fmt.Sprintf("Zone count is %d, only run for multi-zone clusters, skipping test", zoneCount)
|
||||||
@ -61,17 +61,17 @@ var _ = SIGDescribe("Multi-AZ Cluster Volumes [sig-storage]", func() {
|
|||||||
// OnlyAllowNodeZones tests that GetAllCurrentZones returns only zones with Nodes
|
// OnlyAllowNodeZones tests that GetAllCurrentZones returns only zones with Nodes
|
||||||
func OnlyAllowNodeZones(f *framework.Framework, zoneCount int, image string) {
|
func OnlyAllowNodeZones(f *framework.Framework, zoneCount int, image string) {
|
||||||
gceCloud, err := gce.GetGCECloud()
|
gceCloud, err := gce.GetGCECloud()
|
||||||
gomega.Expect(err).NotTo(gomega.HaveOccurred())
|
framework.ExpectNoError(err)
|
||||||
|
|
||||||
// Get all the zones that the nodes are in
|
// Get all the zones that the nodes are in
|
||||||
expectedZones, err := gceCloud.GetAllZonesFromCloudProvider()
|
expectedZones, err := gceCloud.GetAllZonesFromCloudProvider()
|
||||||
gomega.Expect(err).NotTo(gomega.HaveOccurred())
|
framework.ExpectNoError(err)
|
||||||
e2elog.Logf("Expected zones: %v", expectedZones)
|
e2elog.Logf("Expected zones: %v", expectedZones)
|
||||||
|
|
||||||
// Get all the zones in this current region
|
// Get all the zones in this current region
|
||||||
region := gceCloud.Region()
|
region := gceCloud.Region()
|
||||||
allZonesInRegion, err := gceCloud.ListZonesInRegion(region)
|
allZonesInRegion, err := gceCloud.ListZonesInRegion(region)
|
||||||
gomega.Expect(err).NotTo(gomega.HaveOccurred())
|
framework.ExpectNoError(err)
|
||||||
|
|
||||||
var extraZone string
|
var extraZone string
|
||||||
for _, zone := range allZonesInRegion {
|
for _, zone := range allZonesInRegion {
|
||||||
@ -117,13 +117,13 @@ func OnlyAllowNodeZones(f *framework.Framework, zoneCount int, image string) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
err = gceCloud.InsertInstance(project, zone, rb)
|
err = gceCloud.InsertInstance(project, zone, rb)
|
||||||
gomega.Expect(err).NotTo(gomega.HaveOccurred())
|
framework.ExpectNoError(err)
|
||||||
|
|
||||||
defer func() {
|
defer func() {
|
||||||
// Teardown of the compute instance
|
// Teardown of the compute instance
|
||||||
e2elog.Logf("Deleting compute resource: %v", name)
|
e2elog.Logf("Deleting compute resource: %v", name)
|
||||||
err := gceCloud.DeleteInstance(project, zone, name)
|
err := gceCloud.DeleteInstance(project, zone, name)
|
||||||
gomega.Expect(err).NotTo(gomega.HaveOccurred())
|
framework.ExpectNoError(err)
|
||||||
}()
|
}()
|
||||||
|
|
||||||
ginkgo.By("Creating zoneCount+1 PVCs and making sure PDs are only provisioned in zones with nodes")
|
ginkgo.By("Creating zoneCount+1 PVCs and making sure PDs are only provisioned in zones with nodes")
|
||||||
@ -136,7 +136,7 @@ func OnlyAllowNodeZones(f *framework.Framework, zoneCount int, image string) {
|
|||||||
for index := 1; index <= zoneCount+1; index++ {
|
for index := 1; index <= zoneCount+1; index++ {
|
||||||
pvc := newNamedDefaultClaim(ns, index)
|
pvc := newNamedDefaultClaim(ns, index)
|
||||||
pvc, err = framework.CreatePVC(c, ns, pvc)
|
pvc, err = framework.CreatePVC(c, ns, pvc)
|
||||||
gomega.Expect(err).NotTo(gomega.HaveOccurred())
|
framework.ExpectNoError(err)
|
||||||
pvcList = append(pvcList, pvc)
|
pvcList = append(pvcList, pvc)
|
||||||
|
|
||||||
// Defer the cleanup
|
// Defer the cleanup
|
||||||
@ -152,7 +152,7 @@ func OnlyAllowNodeZones(f *framework.Framework, zoneCount int, image string) {
|
|||||||
// Wait for all claims bound
|
// Wait for all claims bound
|
||||||
for _, claim := range pvcList {
|
for _, claim := range pvcList {
|
||||||
err = framework.WaitForPersistentVolumeClaimPhase(v1.ClaimBound, c, claim.Namespace, claim.Name, framework.Poll, framework.ClaimProvisionTimeout)
|
err = framework.WaitForPersistentVolumeClaimPhase(v1.ClaimBound, c, claim.Namespace, claim.Name, framework.Poll, framework.ClaimProvisionTimeout)
|
||||||
gomega.Expect(err).NotTo(gomega.HaveOccurred())
|
framework.ExpectNoError(err)
|
||||||
}
|
}
|
||||||
|
|
||||||
pvZones := sets.NewString()
|
pvZones := sets.NewString()
|
||||||
@ -160,11 +160,12 @@ func OnlyAllowNodeZones(f *framework.Framework, zoneCount int, image string) {
|
|||||||
for _, claim := range pvcList {
|
for _, claim := range pvcList {
|
||||||
// Get a new copy of the claim to have all fields populated
|
// Get a new copy of the claim to have all fields populated
|
||||||
claim, err = c.CoreV1().PersistentVolumeClaims(claim.Namespace).Get(claim.Name, metav1.GetOptions{})
|
claim, err = c.CoreV1().PersistentVolumeClaims(claim.Namespace).Get(claim.Name, metav1.GetOptions{})
|
||||||
gomega.Expect(err).NotTo(gomega.HaveOccurred())
|
framework.ExpectNoError(err)
|
||||||
|
|
||||||
// Get the related PV
|
// Get the related PV
|
||||||
pv, err := c.CoreV1().PersistentVolumes().Get(claim.Spec.VolumeName, metav1.GetOptions{})
|
pv, err := c.CoreV1().PersistentVolumes().Get(claim.Spec.VolumeName, metav1.GetOptions{})
|
||||||
gomega.Expect(err).NotTo(gomega.HaveOccurred())
|
gomega.Expect(err).NotTo(gomega.HaveOccurred())
|
||||||
|
framework.ExpectNoError(err)
|
||||||
|
|
||||||
pvZone, ok := pv.ObjectMeta.Labels[v1.LabelZoneFailureDomain]
|
pvZone, ok := pv.ObjectMeta.Labels[v1.LabelZoneFailureDomain]
|
||||||
gomega.Expect(ok).To(gomega.BeTrue(), "PV has no LabelZone to be found")
|
gomega.Expect(ok).To(gomega.BeTrue(), "PV has no LabelZone to be found")
|
||||||
@ -188,7 +189,7 @@ func PodsUseStaticPVsOrFail(f *framework.Framework, podCount int, image string)
|
|||||||
ns := f.Namespace.Name
|
ns := f.Namespace.Name
|
||||||
|
|
||||||
zones, err := framework.GetClusterZones(c)
|
zones, err := framework.GetClusterZones(c)
|
||||||
gomega.Expect(err).NotTo(gomega.HaveOccurred())
|
framework.ExpectNoError(err)
|
||||||
zonelist := zones.List()
|
zonelist := zones.List()
|
||||||
ginkgo.By("Creating static PVs across zones")
|
ginkgo.By("Creating static PVs across zones")
|
||||||
configs := make([]*staticPVTestConfig, podCount)
|
configs := make([]*staticPVTestConfig, podCount)
|
||||||
@ -205,14 +206,14 @@ func PodsUseStaticPVsOrFail(f *framework.Framework, podCount int, image string)
|
|||||||
framework.WaitForPodNoLongerRunningInNamespace(c, config.pod.Name, ns)
|
framework.WaitForPodNoLongerRunningInNamespace(c, config.pod.Name, ns)
|
||||||
framework.PVPVCCleanup(c, ns, config.pv, config.pvc)
|
framework.PVPVCCleanup(c, ns, config.pv, config.pvc)
|
||||||
err = framework.DeletePVSource(config.pvSource)
|
err = framework.DeletePVSource(config.pvSource)
|
||||||
gomega.Expect(err).NotTo(gomega.HaveOccurred())
|
framework.ExpectNoError(err)
|
||||||
}
|
}
|
||||||
}()
|
}()
|
||||||
|
|
||||||
for i, config := range configs {
|
for i, config := range configs {
|
||||||
zone := zonelist[i%len(zones)]
|
zone := zonelist[i%len(zones)]
|
||||||
config.pvSource, err = framework.CreatePVSource(zone)
|
config.pvSource, err = framework.CreatePVSource(zone)
|
||||||
gomega.Expect(err).NotTo(gomega.HaveOccurred())
|
framework.ExpectNoError(err)
|
||||||
|
|
||||||
pvConfig := framework.PersistentVolumeConfig{
|
pvConfig := framework.PersistentVolumeConfig{
|
||||||
NamePrefix: "multizone-pv",
|
NamePrefix: "multizone-pv",
|
||||||
@ -223,7 +224,7 @@ func PodsUseStaticPVsOrFail(f *framework.Framework, podCount int, image string)
|
|||||||
pvcConfig := framework.PersistentVolumeClaimConfig{StorageClassName: &className}
|
pvcConfig := framework.PersistentVolumeClaimConfig{StorageClassName: &className}
|
||||||
|
|
||||||
config.pv, config.pvc, err = framework.CreatePVPVC(c, pvConfig, pvcConfig, ns, true)
|
config.pv, config.pvc, err = framework.CreatePVPVC(c, pvConfig, pvcConfig, ns, true)
|
||||||
gomega.Expect(err).NotTo(gomega.HaveOccurred())
|
framework.ExpectNoError(err)
|
||||||
}
|
}
|
||||||
|
|
||||||
ginkgo.By("Waiting for all PVCs to be bound")
|
ginkgo.By("Waiting for all PVCs to be bound")
|
||||||
@ -235,13 +236,13 @@ func PodsUseStaticPVsOrFail(f *framework.Framework, podCount int, image string)
|
|||||||
for _, config := range configs {
|
for _, config := range configs {
|
||||||
podConfig := framework.MakePod(ns, nil, []*v1.PersistentVolumeClaim{config.pvc}, false, "")
|
podConfig := framework.MakePod(ns, nil, []*v1.PersistentVolumeClaim{config.pvc}, false, "")
|
||||||
config.pod, err = c.CoreV1().Pods(ns).Create(podConfig)
|
config.pod, err = c.CoreV1().Pods(ns).Create(podConfig)
|
||||||
gomega.Expect(err).NotTo(gomega.HaveOccurred())
|
framework.ExpectNoError(err)
|
||||||
}
|
}
|
||||||
|
|
||||||
ginkgo.By("Waiting for all pods to be running")
|
ginkgo.By("Waiting for all pods to be running")
|
||||||
for _, config := range configs {
|
for _, config := range configs {
|
||||||
err = framework.WaitForPodRunningInNamespace(c, config.pod)
|
err = framework.WaitForPodRunningInNamespace(c, config.pod)
|
||||||
gomega.Expect(err).NotTo(gomega.HaveOccurred())
|
framework.ExpectNoError(err)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
Loading…
Reference in New Issue
Block a user