mirror of
https://github.com/k3s-io/kubernetes.git
synced 2025-07-26 21:17:23 +00:00
Merge pull request #46665 from yujuhong/fail-clearly
Automatic merge from submit-queue (batch tested with PRs 47113, 46665, 47189) Improve the e2e node restart test This commit includes the following two changes: * Move pre-test checks (pods/nodes ready) to BeforeEach() so that it's clear whether the test has run or not. * Dumping logs for unready pods.
This commit is contained in:
commit
6b7e7f7797
@ -58,13 +58,37 @@ func filterIrrelevantPods(pods []*v1.Pod) []*v1.Pod {
|
|||||||
var _ = framework.KubeDescribe("Restart [Disruptive]", func() {
|
var _ = framework.KubeDescribe("Restart [Disruptive]", func() {
|
||||||
f := framework.NewDefaultFramework("restart")
|
f := framework.NewDefaultFramework("restart")
|
||||||
var ps *testutils.PodStore
|
var ps *testutils.PodStore
|
||||||
|
var originalNodeNames []string
|
||||||
|
var originalPodNames []string
|
||||||
|
var numNodes int
|
||||||
|
var systemNamespace string
|
||||||
|
|
||||||
BeforeEach(func() {
|
BeforeEach(func() {
|
||||||
// This test requires the ability to restart all nodes, so the provider
|
// This test requires the ability to restart all nodes, so the provider
|
||||||
// check must be identical to that call.
|
// check must be identical to that call.
|
||||||
framework.SkipUnlessProviderIs("gce", "gke")
|
framework.SkipUnlessProviderIs("gce", "gke")
|
||||||
|
|
||||||
ps = testutils.NewPodStore(f.ClientSet, metav1.NamespaceSystem, labels.Everything(), fields.Everything())
|
ps = testutils.NewPodStore(f.ClientSet, metav1.NamespaceSystem, labels.Everything(), fields.Everything())
|
||||||
|
numNodes = framework.TestContext.CloudConfig.NumNodes
|
||||||
|
systemNamespace = metav1.NamespaceSystem
|
||||||
|
|
||||||
|
By("ensuring all nodes are ready")
|
||||||
|
var err error
|
||||||
|
originalNodeNames, err = framework.CheckNodesReady(f.ClientSet, framework.NodeReadyInitialTimeout, numNodes)
|
||||||
|
Expect(err).NotTo(HaveOccurred())
|
||||||
|
framework.Logf("Got the following nodes before restart: %v", originalNodeNames)
|
||||||
|
|
||||||
|
By("ensuring all pods are running and ready")
|
||||||
|
allPods := ps.List()
|
||||||
|
pods := filterIrrelevantPods(allPods)
|
||||||
|
|
||||||
|
originalPodNames = make([]string, len(pods))
|
||||||
|
for i, p := range pods {
|
||||||
|
originalPodNames[i] = p.ObjectMeta.Name
|
||||||
|
}
|
||||||
|
if !framework.CheckPodsRunningReadyOrSucceeded(f.ClientSet, systemNamespace, originalPodNames, framework.PodReadyBeforeTimeout) {
|
||||||
|
printStatusAndLogsForNotReadyPods(f.ClientSet, systemNamespace, originalPodNames, pods)
|
||||||
|
framework.Failf("At least one pod wasn't running and ready or succeeded at test start.")
|
||||||
|
}
|
||||||
})
|
})
|
||||||
|
|
||||||
AfterEach(func() {
|
AfterEach(func() {
|
||||||
@ -74,41 +98,21 @@ var _ = framework.KubeDescribe("Restart [Disruptive]", func() {
|
|||||||
})
|
})
|
||||||
|
|
||||||
It("should restart all nodes and ensure all nodes and pods recover", func() {
|
It("should restart all nodes and ensure all nodes and pods recover", func() {
|
||||||
nn := framework.TestContext.CloudConfig.NumNodes
|
|
||||||
|
|
||||||
By("ensuring all nodes are ready")
|
|
||||||
nodeNamesBefore, err := framework.CheckNodesReady(f.ClientSet, framework.NodeReadyInitialTimeout, nn)
|
|
||||||
Expect(err).NotTo(HaveOccurred())
|
|
||||||
framework.Logf("Got the following nodes before restart: %v", nodeNamesBefore)
|
|
||||||
|
|
||||||
By("ensuring all pods are running and ready")
|
|
||||||
allPods := ps.List()
|
|
||||||
pods := filterIrrelevantPods(allPods)
|
|
||||||
|
|
||||||
podNamesBefore := make([]string, len(pods))
|
|
||||||
for i, p := range pods {
|
|
||||||
podNamesBefore[i] = p.ObjectMeta.Name
|
|
||||||
}
|
|
||||||
ns := metav1.NamespaceSystem
|
|
||||||
if !framework.CheckPodsRunningReadyOrSucceeded(f.ClientSet, ns, podNamesBefore, framework.PodReadyBeforeTimeout) {
|
|
||||||
framework.Failf("At least one pod wasn't running and ready or succeeded at test start.")
|
|
||||||
}
|
|
||||||
|
|
||||||
By("restarting all of the nodes")
|
By("restarting all of the nodes")
|
||||||
err = restartNodes(f, nodeNamesBefore)
|
err := restartNodes(f, originalNodeNames)
|
||||||
Expect(err).NotTo(HaveOccurred())
|
Expect(err).NotTo(HaveOccurred())
|
||||||
|
|
||||||
By("ensuring all nodes are ready after the restart")
|
By("ensuring all nodes are ready after the restart")
|
||||||
nodeNamesAfter, err := framework.CheckNodesReady(f.ClientSet, framework.RestartNodeReadyAgainTimeout, nn)
|
nodeNamesAfter, err := framework.CheckNodesReady(f.ClientSet, framework.RestartNodeReadyAgainTimeout, numNodes)
|
||||||
Expect(err).NotTo(HaveOccurred())
|
Expect(err).NotTo(HaveOccurred())
|
||||||
framework.Logf("Got the following nodes after restart: %v", nodeNamesAfter)
|
framework.Logf("Got the following nodes after restart: %v", nodeNamesAfter)
|
||||||
|
|
||||||
// Make sure that we have the same number of nodes. We're not checking
|
// Make sure that we have the same number of nodes. We're not checking
|
||||||
// that the names match because that's implementation specific.
|
// that the names match because that's implementation specific.
|
||||||
By("ensuring the same number of nodes exist after the restart")
|
By("ensuring the same number of nodes exist after the restart")
|
||||||
if len(nodeNamesBefore) != len(nodeNamesAfter) {
|
if len(originalNodeNames) != len(nodeNamesAfter) {
|
||||||
framework.Failf("Had %d nodes before nodes were restarted, but now only have %d",
|
framework.Failf("Had %d nodes before nodes were restarted, but now only have %d",
|
||||||
len(nodeNamesBefore), len(nodeNamesAfter))
|
len(originalNodeNames), len(nodeNamesAfter))
|
||||||
}
|
}
|
||||||
|
|
||||||
// Make sure that we have the same number of pods. We're not checking
|
// Make sure that we have the same number of pods. We're not checking
|
||||||
@ -116,10 +120,12 @@ var _ = framework.KubeDescribe("Restart [Disruptive]", func() {
|
|||||||
// across node restarts.
|
// across node restarts.
|
||||||
By("ensuring the same number of pods are running and ready after restart")
|
By("ensuring the same number of pods are running and ready after restart")
|
||||||
podCheckStart := time.Now()
|
podCheckStart := time.Now()
|
||||||
podNamesAfter, err := waitForNPods(ps, len(podNamesBefore), framework.RestartPodReadyAgainTimeout)
|
podNamesAfter, err := waitForNPods(ps, len(originalPodNames), framework.RestartPodReadyAgainTimeout)
|
||||||
Expect(err).NotTo(HaveOccurred())
|
Expect(err).NotTo(HaveOccurred())
|
||||||
remaining := framework.RestartPodReadyAgainTimeout - time.Since(podCheckStart)
|
remaining := framework.RestartPodReadyAgainTimeout - time.Since(podCheckStart)
|
||||||
if !framework.CheckPodsRunningReadyOrSucceeded(f.ClientSet, ns, podNamesAfter, remaining) {
|
if !framework.CheckPodsRunningReadyOrSucceeded(f.ClientSet, systemNamespace, podNamesAfter, remaining) {
|
||||||
|
pods := ps.List()
|
||||||
|
printStatusAndLogsForNotReadyPods(f.ClientSet, systemNamespace, podNamesAfter, pods)
|
||||||
framework.Failf("At least one pod wasn't running and ready after the restart.")
|
framework.Failf("At least one pod wasn't running and ready after the restart.")
|
||||||
}
|
}
|
||||||
})
|
})
|
||||||
|
Loading…
Reference in New Issue
Block a user