mirror of
https://github.com/k3s-io/kubernetes.git
synced 2025-07-19 09:52:49 +00:00
test/e2e/framework:remove unused code and move const
This commit is contained in:
parent
3ed0f1bec1
commit
119f4df9c5
@ -60,6 +60,14 @@ import (
|
|||||||
_ "k8s.io/kubernetes/test/e2e/framework/providers/vsphere"
|
_ "k8s.io/kubernetes/test/e2e/framework/providers/vsphere"
|
||||||
)
|
)
|
||||||
|
|
||||||
|
const (
|
||||||
|
// namespaceCleanupTimeout is how long to wait for the namespace to be deleted.
|
||||||
|
// If there are any orphaned namespaces to clean up, this test is running
|
||||||
|
// on a long lived cluster. A long wait here is preferably to spurious test
|
||||||
|
// failures caused by leaked resources from a previous test run.
|
||||||
|
namespaceCleanupTimeout = 15 * time.Minute
|
||||||
|
)
|
||||||
|
|
||||||
var _ = ginkgo.SynchronizedBeforeSuite(func() []byte {
|
var _ = ginkgo.SynchronizedBeforeSuite(func() []byte {
|
||||||
// Reference common test to make the import valid.
|
// Reference common test to make the import valid.
|
||||||
commontest.CurrentSuite = commontest.E2E
|
commontest.CurrentSuite = commontest.E2E
|
||||||
@ -232,7 +240,7 @@ func setupSuite() {
|
|||||||
framework.Failf("Error deleting orphaned namespaces: %v", err)
|
framework.Failf("Error deleting orphaned namespaces: %v", err)
|
||||||
}
|
}
|
||||||
klog.Infof("Waiting for deletion of the following namespaces: %v", deleted)
|
klog.Infof("Waiting for deletion of the following namespaces: %v", deleted)
|
||||||
if err := framework.WaitForNamespacesDeleted(c, deleted, framework.NamespaceCleanupTimeout); err != nil {
|
if err := framework.WaitForNamespacesDeleted(c, deleted, namespaceCleanupTimeout); err != nil {
|
||||||
framework.Failf("Failed to delete orphaned namespaces %v: %v", deleted, err)
|
framework.Failf("Failed to delete orphaned namespaces %v: %v", deleted, err)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -45,7 +45,6 @@ import (
|
|||||||
|
|
||||||
const (
|
const (
|
||||||
dynamicConsumptionTimeInSeconds = 30
|
dynamicConsumptionTimeInSeconds = 30
|
||||||
staticConsumptionTimeInSeconds = 3600
|
|
||||||
dynamicRequestSizeInMillicores = 100
|
dynamicRequestSizeInMillicores = 100
|
||||||
dynamicRequestSizeInMegabytes = 100
|
dynamicRequestSizeInMegabytes = 100
|
||||||
dynamicRequestSizeCustomMetric = 10
|
dynamicRequestSizeCustomMetric = 10
|
||||||
|
@ -34,6 +34,12 @@ import (
|
|||||||
testutils "k8s.io/kubernetes/test/utils"
|
testutils "k8s.io/kubernetes/test/utils"
|
||||||
)
|
)
|
||||||
|
|
||||||
|
const (
|
||||||
|
// recreateNodeReadyAgainTimeout is how long a node is allowed to become "Ready" after it is recreated before
|
||||||
|
// the test is considered failed.
|
||||||
|
recreateNodeReadyAgainTimeout = 10 * time.Minute
|
||||||
|
)
|
||||||
|
|
||||||
func nodeNames(nodes []v1.Node) []string {
|
func nodeNames(nodes []v1.Node) []string {
|
||||||
result := make([]string, 0, len(nodes))
|
result := make([]string, 0, len(nodes))
|
||||||
for i := range nodes {
|
for i := range nodes {
|
||||||
@ -102,9 +108,9 @@ func testRecreate(c clientset.Interface, ps *testutils.PodStore, systemNamespace
|
|||||||
framework.Failf("Test failed; failed to start the restart instance group command.")
|
framework.Failf("Test failed; failed to start the restart instance group command.")
|
||||||
}
|
}
|
||||||
|
|
||||||
err = WaitForNodeBootIdsToChange(c, nodes, framework.RecreateNodeReadyAgainTimeout)
|
err = WaitForNodeBootIdsToChange(c, nodes, recreateNodeReadyAgainTimeout)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
framework.Failf("Test failed; failed to recreate at least one node in %v.", framework.RecreateNodeReadyAgainTimeout)
|
framework.Failf("Test failed; failed to recreate at least one node in %v.", recreateNodeReadyAgainTimeout)
|
||||||
}
|
}
|
||||||
|
|
||||||
nodesAfter, err := e2enode.CheckReady(c, len(nodes), framework.RestartNodeReadyAgainTimeout)
|
nodesAfter, err := e2enode.CheckReady(c, len(nodes), framework.RestartNodeReadyAgainTimeout)
|
||||||
|
@ -103,12 +103,6 @@ const (
|
|||||||
// PodEventTimeout is how much we wait for a pod event to occur.
|
// PodEventTimeout is how much we wait for a pod event to occur.
|
||||||
PodEventTimeout = 2 * time.Minute
|
PodEventTimeout = 2 * time.Minute
|
||||||
|
|
||||||
// NamespaceCleanupTimeout is how long to wait for the namespace to be deleted.
|
|
||||||
// If there are any orphaned namespaces to clean up, this test is running
|
|
||||||
// on a long lived cluster. A long wait here is preferably to spurious test
|
|
||||||
// failures caused by leaked resources from a previous test run.
|
|
||||||
NamespaceCleanupTimeout = 15 * time.Minute
|
|
||||||
|
|
||||||
// ServiceStartTimeout is how long to wait for a service endpoint to be resolvable.
|
// ServiceStartTimeout is how long to wait for a service endpoint to be resolvable.
|
||||||
ServiceStartTimeout = 3 * time.Minute
|
ServiceStartTimeout = 3 * time.Minute
|
||||||
|
|
||||||
@ -138,12 +132,8 @@ const (
|
|||||||
// Use it case by case when we are sure this timeout is enough.
|
// Use it case by case when we are sure this timeout is enough.
|
||||||
ClaimProvisionShortTimeout = 1 * time.Minute
|
ClaimProvisionShortTimeout = 1 * time.Minute
|
||||||
|
|
||||||
// ClaimDeletingTimeout is How long claims have to become deleted.
|
// ClaimProvisionTimeout is how long claims have to become dynamically provisioned.
|
||||||
ClaimDeletingTimeout = 3 * time.Minute
|
ClaimProvisionTimeout = 5 * time.Minute
|
||||||
|
|
||||||
// RecreateNodeReadyAgainTimeout is how long a node is allowed to become "Ready" after it is recreated before
|
|
||||||
// the test is considered failed.
|
|
||||||
RecreateNodeReadyAgainTimeout = 10 * time.Minute
|
|
||||||
|
|
||||||
// RestartNodeReadyAgainTimeout is how long a node is allowed to become "Ready" after it is restarted before
|
// RestartNodeReadyAgainTimeout is how long a node is allowed to become "Ready" after it is restarted before
|
||||||
// the test is considered failed.
|
// the test is considered failed.
|
||||||
@ -171,9 +161,6 @@ const (
|
|||||||
)
|
)
|
||||||
|
|
||||||
var (
|
var (
|
||||||
// ClaimProvisionTimeout is how long claims have to become dynamically provisioned.
|
|
||||||
ClaimProvisionTimeout = 5 * time.Minute
|
|
||||||
|
|
||||||
// BusyBoxImage is the image URI of BusyBox.
|
// BusyBoxImage is the image URI of BusyBox.
|
||||||
BusyBoxImage = imageutils.GetE2EImage(imageutils.BusyBox)
|
BusyBoxImage = imageutils.GetE2EImage(imageutils.BusyBox)
|
||||||
|
|
||||||
|
@ -36,6 +36,11 @@ import (
|
|||||||
"k8s.io/kubernetes/test/e2e/storage/utils"
|
"k8s.io/kubernetes/test/e2e/storage/utils"
|
||||||
)
|
)
|
||||||
|
|
||||||
|
const (
|
||||||
|
// claimDeletingTimeout is How long claims have to become deleted.
|
||||||
|
claimDeletingTimeout = 3 * time.Minute
|
||||||
|
)
|
||||||
|
|
||||||
// waitForPersistentVolumeClaimDeleted waits for a PersistentVolumeClaim to be removed from the system until timeout occurs, whichever comes first.
|
// waitForPersistentVolumeClaimDeleted waits for a PersistentVolumeClaim to be removed from the system until timeout occurs, whichever comes first.
|
||||||
func waitForPersistentVolumeClaimDeleted(c clientset.Interface, ns string, pvcName string, Poll, timeout time.Duration) error {
|
func waitForPersistentVolumeClaimDeleted(c clientset.Interface, ns string, pvcName string, Poll, timeout time.Duration) error {
|
||||||
framework.Logf("Waiting up to %v for PersistentVolumeClaim %s to be removed", timeout, pvcName)
|
framework.Logf("Waiting up to %v for PersistentVolumeClaim %s to be removed", timeout, pvcName)
|
||||||
@ -112,7 +117,7 @@ var _ = utils.SIGDescribe("PVC Protection", func() {
|
|||||||
ginkgo.By("Deleting the PVC")
|
ginkgo.By("Deleting the PVC")
|
||||||
err = client.CoreV1().PersistentVolumeClaims(pvc.Namespace).Delete(context.TODO(), pvc.Name, metav1.NewDeleteOptions(0))
|
err = client.CoreV1().PersistentVolumeClaims(pvc.Namespace).Delete(context.TODO(), pvc.Name, metav1.NewDeleteOptions(0))
|
||||||
framework.ExpectNoError(err, "Error deleting PVC")
|
framework.ExpectNoError(err, "Error deleting PVC")
|
||||||
waitForPersistentVolumeClaimDeleted(client, pvc.Namespace, pvc.Name, framework.Poll, framework.ClaimDeletingTimeout)
|
waitForPersistentVolumeClaimDeleted(client, pvc.Namespace, pvc.Name, framework.Poll, claimDeletingTimeout)
|
||||||
pvcCreatedAndNotDeleted = false
|
pvcCreatedAndNotDeleted = false
|
||||||
})
|
})
|
||||||
|
|
||||||
@ -131,7 +136,7 @@ var _ = utils.SIGDescribe("PVC Protection", func() {
|
|||||||
framework.ExpectNoError(err, "Error terminating and deleting pod")
|
framework.ExpectNoError(err, "Error terminating and deleting pod")
|
||||||
|
|
||||||
ginkgo.By("Checking that the PVC is automatically removed from the system because it's no longer in active use by a pod")
|
ginkgo.By("Checking that the PVC is automatically removed from the system because it's no longer in active use by a pod")
|
||||||
waitForPersistentVolumeClaimDeleted(client, pvc.Namespace, pvc.Name, framework.Poll, framework.ClaimDeletingTimeout)
|
waitForPersistentVolumeClaimDeleted(client, pvc.Namespace, pvc.Name, framework.Poll, claimDeletingTimeout)
|
||||||
pvcCreatedAndNotDeleted = false
|
pvcCreatedAndNotDeleted = false
|
||||||
})
|
})
|
||||||
|
|
||||||
@ -163,7 +168,7 @@ var _ = utils.SIGDescribe("PVC Protection", func() {
|
|||||||
framework.ExpectNoError(err, "Error terminating and deleting pod")
|
framework.ExpectNoError(err, "Error terminating and deleting pod")
|
||||||
|
|
||||||
ginkgo.By("Checking that the PVC is automatically removed from the system because it's no longer in active use by a pod")
|
ginkgo.By("Checking that the PVC is automatically removed from the system because it's no longer in active use by a pod")
|
||||||
waitForPersistentVolumeClaimDeleted(client, pvc.Namespace, pvc.Name, framework.Poll, framework.ClaimDeletingTimeout)
|
waitForPersistentVolumeClaimDeleted(client, pvc.Namespace, pvc.Name, framework.Poll, claimDeletingTimeout)
|
||||||
pvcCreatedAndNotDeleted = false
|
pvcCreatedAndNotDeleted = false
|
||||||
})
|
})
|
||||||
})
|
})
|
||||||
|
Loading…
Reference in New Issue
Block a user