mirror of
https://github.com/k3s-io/kubernetes.git
synced 2025-07-30 23:15:14 +00:00
Merge pull request #126048 from jsafrane/remove-broken-tests
Remove static PV topology test
This commit is contained in:
commit
a39f42582f
@ -1,150 +0,0 @@
|
||||
/*
|
||||
Copyright 2017 The Kubernetes Authors.
|
||||
|
||||
Licensed under the Apache License, Version 2.0 (the "License");
|
||||
you may not use this file except in compliance with the License.
|
||||
You may obtain a copy of the License at
|
||||
|
||||
http://www.apache.org/licenses/LICENSE-2.0
|
||||
|
||||
Unless required by applicable law or agreed to in writing, software
|
||||
distributed under the License is distributed on an "AS IS" BASIS,
|
||||
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
See the License for the specific language governing permissions and
|
||||
limitations under the License.
|
||||
*/
|
||||
|
||||
package storage
|
||||
|
||||
import (
|
||||
"context"
|
||||
"fmt"
|
||||
"sync"
|
||||
|
||||
"github.com/onsi/ginkgo/v2"
|
||||
v1 "k8s.io/api/core/v1"
|
||||
metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
|
||||
utilerrors "k8s.io/apimachinery/pkg/util/errors"
|
||||
"k8s.io/apimachinery/pkg/util/sets"
|
||||
clientset "k8s.io/client-go/kubernetes"
|
||||
"k8s.io/kubernetes/test/e2e/framework"
|
||||
e2enode "k8s.io/kubernetes/test/e2e/framework/node"
|
||||
e2epod "k8s.io/kubernetes/test/e2e/framework/pod"
|
||||
e2epv "k8s.io/kubernetes/test/e2e/framework/pv"
|
||||
e2eskipper "k8s.io/kubernetes/test/e2e/framework/skipper"
|
||||
"k8s.io/kubernetes/test/e2e/storage/utils"
|
||||
imageutils "k8s.io/kubernetes/test/utils/image"
|
||||
admissionapi "k8s.io/pod-security-admission/api"
|
||||
)
|
||||
|
||||
var _ = utils.SIGDescribe("Multi-AZ Cluster Volumes", func() {
|
||||
f := framework.NewDefaultFramework("multi-az")
|
||||
f.NamespacePodSecurityLevel = admissionapi.LevelPrivileged
|
||||
var zoneCount int
|
||||
var err error
|
||||
image := imageutils.GetE2EImage(imageutils.Agnhost)
|
||||
ginkgo.BeforeEach(func(ctx context.Context) {
|
||||
e2eskipper.SkipUnlessProviderIs("gce", "gke")
|
||||
if zoneCount <= 0 {
|
||||
zoneCount, err = getZoneCount(ctx, f.ClientSet)
|
||||
framework.ExpectNoError(err)
|
||||
}
|
||||
ginkgo.By(fmt.Sprintf("Checking for multi-zone cluster. Zone count = %d", zoneCount))
|
||||
msg := fmt.Sprintf("Zone count is %d, only run for multi-zone clusters, skipping test", zoneCount)
|
||||
e2eskipper.SkipUnlessAtLeast(zoneCount, 2, msg)
|
||||
// TODO: SkipUnlessDefaultScheduler() // Non-default schedulers might not spread
|
||||
})
|
||||
ginkgo.It("should schedule pods in the same zones as statically provisioned PVs", func(ctx context.Context) {
|
||||
PodsUseStaticPVsOrFail(ctx, f, (2*zoneCount)+1, image)
|
||||
})
|
||||
})
|
||||
|
||||
// Return the number of zones in which we have nodes in this cluster.
|
||||
func getZoneCount(ctx context.Context, c clientset.Interface) (int, error) {
|
||||
zoneNames, err := e2enode.GetSchedulableClusterZones(ctx, c)
|
||||
if err != nil {
|
||||
return -1, err
|
||||
}
|
||||
return len(zoneNames), nil
|
||||
}
|
||||
|
||||
type staticPVTestConfig struct {
|
||||
pvSource *v1.PersistentVolumeSource
|
||||
pv *v1.PersistentVolume
|
||||
pvc *v1.PersistentVolumeClaim
|
||||
pod *v1.Pod
|
||||
}
|
||||
|
||||
// PodsUseStaticPVsOrFail Check that the pods using statically
|
||||
// created PVs get scheduled to the same zone that the PV is in.
|
||||
func PodsUseStaticPVsOrFail(ctx context.Context, f *framework.Framework, podCount int, image string) {
|
||||
var err error
|
||||
c := f.ClientSet
|
||||
ns := f.Namespace.Name
|
||||
|
||||
zones, err := e2enode.GetSchedulableClusterZones(ctx, c)
|
||||
framework.ExpectNoError(err)
|
||||
zonelist := sets.List(zones)
|
||||
ginkgo.By("Creating static PVs across zones")
|
||||
configs := make([]*staticPVTestConfig, podCount)
|
||||
for i := range configs {
|
||||
configs[i] = &staticPVTestConfig{}
|
||||
}
|
||||
|
||||
ginkgo.DeferCleanup(func(ctx context.Context) {
|
||||
ginkgo.By("Cleaning up pods and PVs")
|
||||
for _, config := range configs {
|
||||
e2epod.DeletePodOrFail(ctx, c, ns, config.pod.Name)
|
||||
}
|
||||
var wg sync.WaitGroup
|
||||
wg.Add(len(configs))
|
||||
for i := range configs {
|
||||
go func(config *staticPVTestConfig) {
|
||||
defer ginkgo.GinkgoRecover()
|
||||
defer wg.Done()
|
||||
err := e2epod.WaitForPodNotFoundInNamespace(ctx, c, config.pod.Name, ns, f.Timeouts.PodDelete)
|
||||
framework.ExpectNoError(err, "while waiting for pod to disappear")
|
||||
errs := e2epv.PVPVCCleanup(ctx, c, ns, config.pv, config.pvc)
|
||||
framework.ExpectNoError(utilerrors.NewAggregate(errs), "while cleaning up PVs and PVCs")
|
||||
err = e2epv.DeletePVSource(ctx, config.pvSource)
|
||||
framework.ExpectNoError(err, "while deleting PVSource")
|
||||
}(configs[i])
|
||||
}
|
||||
wg.Wait()
|
||||
})
|
||||
|
||||
for i, config := range configs {
|
||||
zone := zonelist[i%len(zones)]
|
||||
config.pvSource, err = e2epv.CreatePVSource(ctx, zone)
|
||||
framework.ExpectNoError(err)
|
||||
|
||||
pvConfig := e2epv.PersistentVolumeConfig{
|
||||
NamePrefix: "multizone-pv",
|
||||
PVSource: *config.pvSource,
|
||||
Prebind: nil,
|
||||
}
|
||||
className := ""
|
||||
pvcConfig := e2epv.PersistentVolumeClaimConfig{StorageClassName: &className}
|
||||
|
||||
config.pv, config.pvc, err = e2epv.CreatePVPVC(ctx, c, f.Timeouts, pvConfig, pvcConfig, ns, true)
|
||||
framework.ExpectNoError(err)
|
||||
}
|
||||
|
||||
ginkgo.By("Waiting for all PVCs to be bound")
|
||||
for _, config := range configs {
|
||||
e2epv.WaitOnPVandPVC(ctx, c, f.Timeouts, ns, config.pv, config.pvc)
|
||||
}
|
||||
|
||||
ginkgo.By("Creating pods for each static PV")
|
||||
for _, config := range configs {
|
||||
podConfig := e2epod.MakePod(ns, nil, []*v1.PersistentVolumeClaim{config.pvc}, f.NamespacePodSecurityLevel, "")
|
||||
config.pod, err = c.CoreV1().Pods(ns).Create(ctx, podConfig, metav1.CreateOptions{})
|
||||
framework.ExpectNoError(err)
|
||||
}
|
||||
|
||||
ginkgo.By("Waiting for all pods to be running")
|
||||
for _, config := range configs {
|
||||
err = e2epod.WaitForPodRunningInNamespace(ctx, c, config.pod)
|
||||
framework.ExpectNoError(err)
|
||||
}
|
||||
}
|
Loading…
Reference in New Issue
Block a user