mirror of
https://github.com/k3s-io/kubernetes.git
synced 2025-08-11 21:12:07 +00:00
Merge pull request #48006 from bsalamat/e2e_test_cleanup
Automatic merge from submit-queue (batch tested with PRs 47650, 47936, 47939, 47986, 48006) Remove e2e test that checked scheduler priority function for spreading of ReplicationController pods **What this PR does / why we need it**: Remove e2e test that checked scheduler priority function for spreading of ReplicationController pods. The test is flaky in our test clusters, but passes in our local clusters. It looks like our e2e test clusters are not deterministic enough for checking all of the scheduler priority functions. Besides, this functionality is checked extensively by our unit tests. **Which issue this PR fixes** *(optional, in `fixes #<issue number>(, fixes #<issue_number>, ...)` format, will close that issue when PR gets merged)*: fixes #47769 **Special notes for your reviewer**: **Release note**: ```release-note NONE ```
This commit is contained in:
commit
d39115a938
@ -84,52 +84,6 @@ var _ = framework.KubeDescribe("SchedulerPriorities [Serial]", func() {
|
||||
Expect(err).NotTo(HaveOccurred())
|
||||
})
|
||||
|
||||
It("Pods created by ReplicationController should spread to different node", func() {
|
||||
By("Create a pod for each node to make the nodes have almost same cpu/mem usage ratio")
|
||||
|
||||
createBalancedPodForNodes(f, cs, ns, nodeList.Items, podRequestedResource, 0.6)
|
||||
|
||||
By("Create an RC with the same number of replicas as the schedualble nodes. One pod should be scheduled on each node.")
|
||||
config := testutils.RCConfig{
|
||||
Client: f.ClientSet,
|
||||
InternalClient: f.InternalClientset,
|
||||
Name: "scheduler-priority-selector-spreading",
|
||||
Namespace: ns,
|
||||
Image: framework.GetPauseImageName(f.ClientSet),
|
||||
Replicas: len(nodeList.Items),
|
||||
CreatedPods: &[]*v1.Pod{},
|
||||
Labels: map[string]string{"name": "scheduler-priority-selector-spreading"},
|
||||
CpuRequest: podRequestedResource.Requests.Cpu().MilliValue(),
|
||||
MemRequest: podRequestedResource.Requests.Memory().Value(),
|
||||
}
|
||||
Expect(framework.RunRC(config)).NotTo(HaveOccurred())
|
||||
// Cleanup the replication controller when we are done.
|
||||
defer func() {
|
||||
// Resize the replication controller to zero to get rid of pods.
|
||||
if err := framework.DeleteRCAndPods(f.ClientSet, f.InternalClientset, f.Namespace.Name, "scheduler-priority-selector-spreading"); err != nil {
|
||||
framework.Logf("Failed to cleanup replication controller %v: %v.", "scheduler-priority-selector-spreading", err)
|
||||
}
|
||||
}()
|
||||
pods, err := framework.PodsCreated(f.ClientSet, f.Namespace.Name, "scheduler-priority-selector-spreading", int32(len(nodeList.Items)))
|
||||
Expect(err).NotTo(HaveOccurred())
|
||||
By("Ensuring each pod is running")
|
||||
|
||||
result := map[string]bool{}
|
||||
for _, pod := range pods.Items {
|
||||
if pod.DeletionTimestamp != nil {
|
||||
continue
|
||||
}
|
||||
err = f.WaitForPodRunning(pod.Name)
|
||||
Expect(err).NotTo(HaveOccurred())
|
||||
result[pod.Spec.NodeName] = true
|
||||
framework.PrintAllKubeletPods(cs, pod.Spec.NodeName)
|
||||
}
|
||||
By("Verify the pods were scheduled to each node")
|
||||
if len(nodeList.Items) != len(result) {
|
||||
framework.Failf("Pods are not spread to each node.")
|
||||
}
|
||||
})
|
||||
|
||||
It("Pod should be prefer scheduled to node that satisify the NodeAffinity", func() {
|
||||
nodeName := GetNodeThatCanRunPod(f)
|
||||
By("Trying to apply a label on the found node.")
|
||||
|
Loading…
Reference in New Issue
Block a user