mirror of
https://github.com/k3s-io/kubernetes.git
synced 2025-08-05 18:24:07 +00:00
Merge pull request #126048 from jsafrane/remove-broken-tests
Remove static PV topology test
This commit is contained in:
commit
a39f42582f
@ -1,150 +0,0 @@
|
|||||||
/*
|
|
||||||
Copyright 2017 The Kubernetes Authors.
|
|
||||||
|
|
||||||
Licensed under the Apache License, Version 2.0 (the "License");
|
|
||||||
you may not use this file except in compliance with the License.
|
|
||||||
You may obtain a copy of the License at
|
|
||||||
|
|
||||||
http://www.apache.org/licenses/LICENSE-2.0
|
|
||||||
|
|
||||||
Unless required by applicable law or agreed to in writing, software
|
|
||||||
distributed under the License is distributed on an "AS IS" BASIS,
|
|
||||||
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
||||||
See the License for the specific language governing permissions and
|
|
||||||
limitations under the License.
|
|
||||||
*/
|
|
||||||
|
|
||||||
package storage
|
|
||||||
|
|
||||||
import (
|
|
||||||
"context"
|
|
||||||
"fmt"
|
|
||||||
"sync"
|
|
||||||
|
|
||||||
"github.com/onsi/ginkgo/v2"
|
|
||||||
v1 "k8s.io/api/core/v1"
|
|
||||||
metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
|
|
||||||
utilerrors "k8s.io/apimachinery/pkg/util/errors"
|
|
||||||
"k8s.io/apimachinery/pkg/util/sets"
|
|
||||||
clientset "k8s.io/client-go/kubernetes"
|
|
||||||
"k8s.io/kubernetes/test/e2e/framework"
|
|
||||||
e2enode "k8s.io/kubernetes/test/e2e/framework/node"
|
|
||||||
e2epod "k8s.io/kubernetes/test/e2e/framework/pod"
|
|
||||||
e2epv "k8s.io/kubernetes/test/e2e/framework/pv"
|
|
||||||
e2eskipper "k8s.io/kubernetes/test/e2e/framework/skipper"
|
|
||||||
"k8s.io/kubernetes/test/e2e/storage/utils"
|
|
||||||
imageutils "k8s.io/kubernetes/test/utils/image"
|
|
||||||
admissionapi "k8s.io/pod-security-admission/api"
|
|
||||||
)
|
|
||||||
|
|
||||||
var _ = utils.SIGDescribe("Multi-AZ Cluster Volumes", func() {
|
|
||||||
f := framework.NewDefaultFramework("multi-az")
|
|
||||||
f.NamespacePodSecurityLevel = admissionapi.LevelPrivileged
|
|
||||||
var zoneCount int
|
|
||||||
var err error
|
|
||||||
image := imageutils.GetE2EImage(imageutils.Agnhost)
|
|
||||||
ginkgo.BeforeEach(func(ctx context.Context) {
|
|
||||||
e2eskipper.SkipUnlessProviderIs("gce", "gke")
|
|
||||||
if zoneCount <= 0 {
|
|
||||||
zoneCount, err = getZoneCount(ctx, f.ClientSet)
|
|
||||||
framework.ExpectNoError(err)
|
|
||||||
}
|
|
||||||
ginkgo.By(fmt.Sprintf("Checking for multi-zone cluster. Zone count = %d", zoneCount))
|
|
||||||
msg := fmt.Sprintf("Zone count is %d, only run for multi-zone clusters, skipping test", zoneCount)
|
|
||||||
e2eskipper.SkipUnlessAtLeast(zoneCount, 2, msg)
|
|
||||||
// TODO: SkipUnlessDefaultScheduler() // Non-default schedulers might not spread
|
|
||||||
})
|
|
||||||
ginkgo.It("should schedule pods in the same zones as statically provisioned PVs", func(ctx context.Context) {
|
|
||||||
PodsUseStaticPVsOrFail(ctx, f, (2*zoneCount)+1, image)
|
|
||||||
})
|
|
||||||
})
|
|
||||||
|
|
||||||
// Return the number of zones in which we have nodes in this cluster.
|
|
||||||
func getZoneCount(ctx context.Context, c clientset.Interface) (int, error) {
|
|
||||||
zoneNames, err := e2enode.GetSchedulableClusterZones(ctx, c)
|
|
||||||
if err != nil {
|
|
||||||
return -1, err
|
|
||||||
}
|
|
||||||
return len(zoneNames), nil
|
|
||||||
}
|
|
||||||
|
|
||||||
type staticPVTestConfig struct {
|
|
||||||
pvSource *v1.PersistentVolumeSource
|
|
||||||
pv *v1.PersistentVolume
|
|
||||||
pvc *v1.PersistentVolumeClaim
|
|
||||||
pod *v1.Pod
|
|
||||||
}
|
|
||||||
|
|
||||||
// PodsUseStaticPVsOrFail Check that the pods using statically
|
|
||||||
// created PVs get scheduled to the same zone that the PV is in.
|
|
||||||
func PodsUseStaticPVsOrFail(ctx context.Context, f *framework.Framework, podCount int, image string) {
|
|
||||||
var err error
|
|
||||||
c := f.ClientSet
|
|
||||||
ns := f.Namespace.Name
|
|
||||||
|
|
||||||
zones, err := e2enode.GetSchedulableClusterZones(ctx, c)
|
|
||||||
framework.ExpectNoError(err)
|
|
||||||
zonelist := sets.List(zones)
|
|
||||||
ginkgo.By("Creating static PVs across zones")
|
|
||||||
configs := make([]*staticPVTestConfig, podCount)
|
|
||||||
for i := range configs {
|
|
||||||
configs[i] = &staticPVTestConfig{}
|
|
||||||
}
|
|
||||||
|
|
||||||
ginkgo.DeferCleanup(func(ctx context.Context) {
|
|
||||||
ginkgo.By("Cleaning up pods and PVs")
|
|
||||||
for _, config := range configs {
|
|
||||||
e2epod.DeletePodOrFail(ctx, c, ns, config.pod.Name)
|
|
||||||
}
|
|
||||||
var wg sync.WaitGroup
|
|
||||||
wg.Add(len(configs))
|
|
||||||
for i := range configs {
|
|
||||||
go func(config *staticPVTestConfig) {
|
|
||||||
defer ginkgo.GinkgoRecover()
|
|
||||||
defer wg.Done()
|
|
||||||
err := e2epod.WaitForPodNotFoundInNamespace(ctx, c, config.pod.Name, ns, f.Timeouts.PodDelete)
|
|
||||||
framework.ExpectNoError(err, "while waiting for pod to disappear")
|
|
||||||
errs := e2epv.PVPVCCleanup(ctx, c, ns, config.pv, config.pvc)
|
|
||||||
framework.ExpectNoError(utilerrors.NewAggregate(errs), "while cleaning up PVs and PVCs")
|
|
||||||
err = e2epv.DeletePVSource(ctx, config.pvSource)
|
|
||||||
framework.ExpectNoError(err, "while deleting PVSource")
|
|
||||||
}(configs[i])
|
|
||||||
}
|
|
||||||
wg.Wait()
|
|
||||||
})
|
|
||||||
|
|
||||||
for i, config := range configs {
|
|
||||||
zone := zonelist[i%len(zones)]
|
|
||||||
config.pvSource, err = e2epv.CreatePVSource(ctx, zone)
|
|
||||||
framework.ExpectNoError(err)
|
|
||||||
|
|
||||||
pvConfig := e2epv.PersistentVolumeConfig{
|
|
||||||
NamePrefix: "multizone-pv",
|
|
||||||
PVSource: *config.pvSource,
|
|
||||||
Prebind: nil,
|
|
||||||
}
|
|
||||||
className := ""
|
|
||||||
pvcConfig := e2epv.PersistentVolumeClaimConfig{StorageClassName: &className}
|
|
||||||
|
|
||||||
config.pv, config.pvc, err = e2epv.CreatePVPVC(ctx, c, f.Timeouts, pvConfig, pvcConfig, ns, true)
|
|
||||||
framework.ExpectNoError(err)
|
|
||||||
}
|
|
||||||
|
|
||||||
ginkgo.By("Waiting for all PVCs to be bound")
|
|
||||||
for _, config := range configs {
|
|
||||||
e2epv.WaitOnPVandPVC(ctx, c, f.Timeouts, ns, config.pv, config.pvc)
|
|
||||||
}
|
|
||||||
|
|
||||||
ginkgo.By("Creating pods for each static PV")
|
|
||||||
for _, config := range configs {
|
|
||||||
podConfig := e2epod.MakePod(ns, nil, []*v1.PersistentVolumeClaim{config.pvc}, f.NamespacePodSecurityLevel, "")
|
|
||||||
config.pod, err = c.CoreV1().Pods(ns).Create(ctx, podConfig, metav1.CreateOptions{})
|
|
||||||
framework.ExpectNoError(err)
|
|
||||||
}
|
|
||||||
|
|
||||||
ginkgo.By("Waiting for all pods to be running")
|
|
||||||
for _, config := range configs {
|
|
||||||
err = e2epod.WaitForPodRunningInNamespace(ctx, c, config.pod)
|
|
||||||
framework.ExpectNoError(err)
|
|
||||||
}
|
|
||||||
}
|
|
Loading…
Reference in New Issue
Block a user