mirror of
https://github.com/k3s-io/kubernetes.git
synced 2025-07-23 03:41:45 +00:00
fix typo in e2e test
This commit is contained in:
parent
5935fcd704
commit
c4d0191466
@ -51,7 +51,7 @@ import (
|
|||||||
const (
|
const (
|
||||||
postTestConditionMonitoringPeriod = 1 * time.Minute
|
postTestConditionMonitoringPeriod = 1 * time.Minute
|
||||||
evictionPollInterval = 2 * time.Second
|
evictionPollInterval = 2 * time.Second
|
||||||
pressureDissapearTimeout = 1 * time.Minute
|
pressureDisappearTimeout = 1 * time.Minute
|
||||||
// pressure conditions often surface after evictions because the kubelet only updates
|
// pressure conditions often surface after evictions because the kubelet only updates
|
||||||
// node conditions periodically.
|
// node conditions periodically.
|
||||||
// we wait this period after evictions to make sure that we wait out this delay
|
// we wait this period after evictions to make sure that we wait out this delay
|
||||||
@ -464,7 +464,7 @@ func runEvictionTest(f *framework.Framework, pressureTimeout time.Duration, expe
|
|||||||
// Nodes do not immediately report local storage capacity
|
// Nodes do not immediately report local storage capacity
|
||||||
// Sleep so that pods requesting local storage do not fail to schedule
|
// Sleep so that pods requesting local storage do not fail to schedule
|
||||||
time.Sleep(30 * time.Second)
|
time.Sleep(30 * time.Second)
|
||||||
ginkgo.By("seting up pods to be used by tests")
|
ginkgo.By("setting up pods to be used by tests")
|
||||||
pods := []*v1.Pod{}
|
pods := []*v1.Pod{}
|
||||||
for _, spec := range testSpecs {
|
for _, spec := range testSpecs {
|
||||||
pods = append(pods, spec.pod)
|
pods = append(pods, spec.pod)
|
||||||
@ -507,15 +507,15 @@ func runEvictionTest(f *framework.Framework, pressureTimeout time.Duration, expe
|
|||||||
logFunc()
|
logFunc()
|
||||||
logKubeletLatencyMetrics(kubeletmetrics.EvictionStatsAgeKey)
|
logKubeletLatencyMetrics(kubeletmetrics.EvictionStatsAgeKey)
|
||||||
if expectedNodeCondition != noPressure && hasNodeCondition(f, expectedNodeCondition) {
|
if expectedNodeCondition != noPressure && hasNodeCondition(f, expectedNodeCondition) {
|
||||||
return fmt.Errorf("Conditions havent returned to normal, node still has %s", expectedNodeCondition)
|
return fmt.Errorf("Conditions haven't returned to normal, node still has %s", expectedNodeCondition)
|
||||||
}
|
}
|
||||||
return nil
|
return nil
|
||||||
}, pressureDissapearTimeout, evictionPollInterval).Should(gomega.BeNil())
|
}, pressureDisappearTimeout, evictionPollInterval).Should(gomega.BeNil())
|
||||||
|
|
||||||
ginkgo.By("checking for stable, pressure-free condition without unexpected pod failures")
|
ginkgo.By("checking for stable, pressure-free condition without unexpected pod failures")
|
||||||
gomega.Consistently(func() error {
|
gomega.Consistently(func() error {
|
||||||
if expectedNodeCondition != noPressure && hasNodeCondition(f, expectedNodeCondition) {
|
if expectedNodeCondition != noPressure && hasNodeCondition(f, expectedNodeCondition) {
|
||||||
return fmt.Errorf("%s dissappeared and then reappeared", expectedNodeCondition)
|
return fmt.Errorf("%s disappeared and then reappeared", expectedNodeCondition)
|
||||||
}
|
}
|
||||||
logFunc()
|
logFunc()
|
||||||
logKubeletLatencyMetrics(kubeletmetrics.EvictionStatsAgeKey)
|
logKubeletLatencyMetrics(kubeletmetrics.EvictionStatsAgeKey)
|
||||||
@ -548,10 +548,10 @@ func runEvictionTest(f *framework.Framework, pressureTimeout time.Duration, expe
|
|||||||
ginkgo.By(fmt.Sprintf("making sure NodeCondition %s no longer exist on the node", expectedNodeCondition))
|
ginkgo.By(fmt.Sprintf("making sure NodeCondition %s no longer exist on the node", expectedNodeCondition))
|
||||||
gomega.Eventually(func() error {
|
gomega.Eventually(func() error {
|
||||||
if expectedNodeCondition != noPressure && hasNodeCondition(f, expectedNodeCondition) {
|
if expectedNodeCondition != noPressure && hasNodeCondition(f, expectedNodeCondition) {
|
||||||
return fmt.Errorf("Conditions havent returned to normal, node still has %s", expectedNodeCondition)
|
return fmt.Errorf("Conditions haven't returned to normal, node still has %s", expectedNodeCondition)
|
||||||
}
|
}
|
||||||
return nil
|
return nil
|
||||||
}, pressureDissapearTimeout, evictionPollInterval).Should(gomega.BeNil())
|
}, pressureDisappearTimeout, evictionPollInterval).Should(gomega.BeNil())
|
||||||
|
|
||||||
reduceAllocatableMemoryUsage()
|
reduceAllocatableMemoryUsage()
|
||||||
ginkgo.By("making sure we have all the required images for testing")
|
ginkgo.By("making sure we have all the required images for testing")
|
||||||
|
@ -113,10 +113,10 @@ var _ = framework.KubeDescribe("SystemNodeCriticalPod [Slow] [Serial] [Disruptiv
|
|||||||
ginkgo.By("making sure that node no longer has DiskPressure")
|
ginkgo.By("making sure that node no longer has DiskPressure")
|
||||||
gomega.Eventually(func() error {
|
gomega.Eventually(func() error {
|
||||||
if hasNodeCondition(f, v1.NodeDiskPressure) {
|
if hasNodeCondition(f, v1.NodeDiskPressure) {
|
||||||
return fmt.Errorf("Conditions havent returned to normal, node still has DiskPressure")
|
return fmt.Errorf("Conditions haven't returned to normal, node still has DiskPressure")
|
||||||
}
|
}
|
||||||
return nil
|
return nil
|
||||||
}, pressureDissapearTimeout, evictionPollInterval).Should(gomega.BeNil())
|
}, pressureDisappearTimeout, evictionPollInterval).Should(gomega.BeNil())
|
||||||
})
|
})
|
||||||
})
|
})
|
||||||
})
|
})
|
||||||
|
Loading…
Reference in New Issue
Block a user