mirror of
https://github.com/k3s-io/kubernetes.git
synced 2025-07-19 01:40:13 +00:00
380 lines
12 KiB
Go
380 lines
12 KiB
Go
/*
|
|
Copyright 2018 The Kubernetes Authors.
|
|
|
|
Licensed under the Apache License, Version 2.0 (the "License");
|
|
you may not use this file except in compliance with the License.
|
|
You may obtain a copy of the License at
|
|
|
|
http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
Unless required by applicable law or agreed to in writing, software
|
|
distributed under the License is distributed on an "AS IS" BASIS,
|
|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
See the License for the specific language governing permissions and
|
|
limitations under the License.
|
|
*/
|
|
|
|
package statefulset
|
|
|
|
import (
|
|
"context"
|
|
"fmt"
|
|
"sync"
|
|
"testing"
|
|
"time"
|
|
|
|
appsv1 "k8s.io/api/apps/v1"
|
|
v1 "k8s.io/api/core/v1"
|
|
"k8s.io/apimachinery/pkg/api/resource"
|
|
metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
|
|
"k8s.io/apimachinery/pkg/labels"
|
|
"k8s.io/apimachinery/pkg/util/wait"
|
|
"k8s.io/apiserver/pkg/admission"
|
|
"k8s.io/client-go/informers"
|
|
clientset "k8s.io/client-go/kubernetes"
|
|
typedappsv1 "k8s.io/client-go/kubernetes/typed/apps/v1"
|
|
typedv1 "k8s.io/client-go/kubernetes/typed/core/v1"
|
|
restclient "k8s.io/client-go/rest"
|
|
"k8s.io/client-go/util/retry"
|
|
kubeapiservertesting "k8s.io/kubernetes/cmd/kube-apiserver/app/testing"
|
|
api "k8s.io/kubernetes/pkg/apis/core"
|
|
|
|
//svc "k8s.io/kubernetes/pkg/api/v1/service"
|
|
"k8s.io/kubernetes/pkg/controller/statefulset"
|
|
"k8s.io/kubernetes/test/integration/framework"
|
|
)
|
|
|
|
const (
|
|
pollInterval = 100 * time.Millisecond
|
|
pollTimeout = 60 * time.Second
|
|
)
|
|
|
|
func labelMap() map[string]string {
|
|
return map[string]string{"foo": "bar"}
|
|
}
|
|
|
|
// newService returns a service with a fake name for StatefulSet to be created soon
|
|
func newHeadlessService(namespace string) *v1.Service {
|
|
return &v1.Service{
|
|
TypeMeta: metav1.TypeMeta{
|
|
Kind: "Service",
|
|
APIVersion: "v1",
|
|
},
|
|
ObjectMeta: metav1.ObjectMeta{
|
|
Namespace: namespace,
|
|
Name: "fake-service-name",
|
|
},
|
|
Spec: v1.ServiceSpec{
|
|
ClusterIP: "None",
|
|
Ports: []v1.ServicePort{
|
|
{Port: 80, Name: "http", Protocol: "TCP"},
|
|
},
|
|
Selector: labelMap(),
|
|
},
|
|
}
|
|
}
|
|
|
|
// newSTS returns a StatefulSet with a fake container image
|
|
func newSTS(name, namespace string, replicas int) *appsv1.StatefulSet {
|
|
replicasCopy := int32(replicas)
|
|
return &appsv1.StatefulSet{
|
|
TypeMeta: metav1.TypeMeta{
|
|
Kind: "StatefulSet",
|
|
APIVersion: "apps/v1",
|
|
},
|
|
ObjectMeta: metav1.ObjectMeta{
|
|
Namespace: namespace,
|
|
Name: name,
|
|
},
|
|
Spec: appsv1.StatefulSetSpec{
|
|
PodManagementPolicy: appsv1.ParallelPodManagement,
|
|
Replicas: &replicasCopy,
|
|
Selector: &metav1.LabelSelector{
|
|
MatchLabels: labelMap(),
|
|
},
|
|
Template: v1.PodTemplateSpec{
|
|
ObjectMeta: metav1.ObjectMeta{
|
|
Labels: labelMap(),
|
|
},
|
|
Spec: v1.PodSpec{
|
|
Containers: []v1.Container{
|
|
{
|
|
Name: "fake-name",
|
|
Image: "fakeimage",
|
|
VolumeMounts: []v1.VolumeMount{
|
|
{Name: "datadir", MountPath: "/data/"},
|
|
{Name: "home", MountPath: "/home"},
|
|
},
|
|
},
|
|
},
|
|
Volumes: []v1.Volume{
|
|
{
|
|
Name: "datadir",
|
|
VolumeSource: v1.VolumeSource{
|
|
PersistentVolumeClaim: &v1.PersistentVolumeClaimVolumeSource{
|
|
ClaimName: "fake-pvc-name",
|
|
},
|
|
},
|
|
},
|
|
{
|
|
Name: "home",
|
|
VolumeSource: v1.VolumeSource{
|
|
HostPath: &v1.HostPathVolumeSource{
|
|
Path: fmt.Sprintf("/tmp/%v", "home"),
|
|
},
|
|
},
|
|
},
|
|
},
|
|
},
|
|
},
|
|
ServiceName: "fake-service-name",
|
|
UpdateStrategy: appsv1.StatefulSetUpdateStrategy{
|
|
Type: appsv1.RollingUpdateStatefulSetStrategyType,
|
|
},
|
|
VolumeClaimTemplates: []v1.PersistentVolumeClaim{
|
|
// for volume mount "datadir"
|
|
newStatefulSetPVC("fake-pvc-name"),
|
|
},
|
|
},
|
|
}
|
|
}
|
|
|
|
func newStatefulSetPVC(name string) v1.PersistentVolumeClaim {
|
|
return v1.PersistentVolumeClaim{
|
|
ObjectMeta: metav1.ObjectMeta{
|
|
Name: name,
|
|
Annotations: map[string]string{
|
|
"volume.alpha.kubernetes.io/storage-class": "anything",
|
|
},
|
|
},
|
|
Spec: v1.PersistentVolumeClaimSpec{
|
|
AccessModes: []v1.PersistentVolumeAccessMode{
|
|
v1.ReadWriteOnce,
|
|
},
|
|
Resources: v1.ResourceRequirements{
|
|
Requests: v1.ResourceList{
|
|
v1.ResourceStorage: *resource.NewQuantity(1, resource.BinarySI),
|
|
},
|
|
},
|
|
},
|
|
}
|
|
}
|
|
|
|
// scSetup sets up necessities for Statefulset integration test, including control plane, apiserver, informers, and clientset
|
|
func scSetup(ctx context.Context, t *testing.T) (kubeapiservertesting.TearDownFunc, *statefulset.StatefulSetController, informers.SharedInformerFactory, clientset.Interface) {
|
|
// Disable ServiceAccount admission plugin as we don't have serviceaccount controller running.
|
|
server := kubeapiservertesting.StartTestServerOrDie(t, nil, []string{"--disable-admission-plugins=ServiceAccount"}, framework.SharedEtcd())
|
|
|
|
config := restclient.CopyConfig(server.ClientConfig)
|
|
clientSet, err := clientset.NewForConfig(config)
|
|
if err != nil {
|
|
t.Fatalf("error in create clientset: %v", err)
|
|
}
|
|
resyncPeriod := 12 * time.Hour
|
|
informers := informers.NewSharedInformerFactory(clientset.NewForConfigOrDie(restclient.AddUserAgent(config, "statefulset-informers")), resyncPeriod)
|
|
|
|
sc := statefulset.NewStatefulSetController(
|
|
ctx,
|
|
informers.Core().V1().Pods(),
|
|
informers.Apps().V1().StatefulSets(),
|
|
informers.Core().V1().PersistentVolumeClaims(),
|
|
informers.Apps().V1().ControllerRevisions(),
|
|
clientset.NewForConfigOrDie(restclient.AddUserAgent(config, "statefulset-controller")),
|
|
)
|
|
|
|
return server.TearDownFn, sc, informers, clientSet
|
|
}
|
|
|
|
// Run STS controller and informers
|
|
func runControllerAndInformers(sc *statefulset.StatefulSetController, informers informers.SharedInformerFactory) context.CancelFunc {
|
|
ctx, cancel := context.WithCancel(context.Background())
|
|
informers.Start(ctx.Done())
|
|
go sc.Run(ctx, 5)
|
|
return cancel
|
|
}
|
|
|
|
func createHeadlessService(t *testing.T, clientSet clientset.Interface, headlessService *v1.Service) {
|
|
_, err := clientSet.CoreV1().Services(headlessService.Namespace).Create(context.TODO(), headlessService, metav1.CreateOptions{})
|
|
if err != nil {
|
|
t.Fatalf("failed creating headless service: %v", err)
|
|
}
|
|
}
|
|
|
|
func createSTSs(t *testing.T, clientSet clientset.Interface, stss []*appsv1.StatefulSet) []*appsv1.StatefulSet {
|
|
var createdSTSs []*appsv1.StatefulSet
|
|
for _, sts := range stss {
|
|
createdSTS, err := clientSet.AppsV1().StatefulSets(sts.Namespace).Create(context.TODO(), sts, metav1.CreateOptions{})
|
|
if err != nil {
|
|
t.Fatalf("failed to create sts %s: %v", sts.Name, err)
|
|
}
|
|
createdSTSs = append(createdSTSs, createdSTS)
|
|
}
|
|
return createdSTSs
|
|
}
|
|
|
|
func createPods(t *testing.T, clientSet clientset.Interface, pods []*v1.Pod) []*v1.Pod {
|
|
var createdPods []*v1.Pod
|
|
for _, pod := range pods {
|
|
createdPod, err := clientSet.CoreV1().Pods(pod.Namespace).Create(context.TODO(), pod, metav1.CreateOptions{})
|
|
if err != nil {
|
|
t.Fatalf("failed to create pod %s: %v", pod.Name, err)
|
|
}
|
|
createdPods = append(createdPods, createdPod)
|
|
}
|
|
|
|
return createdPods
|
|
}
|
|
|
|
func createSTSsPods(t *testing.T, clientSet clientset.Interface, stss []*appsv1.StatefulSet, pods []*v1.Pod) ([]*appsv1.StatefulSet, []*v1.Pod) {
|
|
return createSTSs(t, clientSet, stss), createPods(t, clientSet, pods)
|
|
}
|
|
|
|
// Verify .Status.Replicas is equal to .Spec.Replicas
|
|
func waitSTSStable(t *testing.T, clientSet clientset.Interface, sts *appsv1.StatefulSet) {
|
|
stsClient := clientSet.AppsV1().StatefulSets(sts.Namespace)
|
|
desiredGeneration := sts.Generation
|
|
if err := wait.PollImmediate(pollInterval, pollTimeout, func() (bool, error) {
|
|
newSTS, err := stsClient.Get(context.TODO(), sts.Name, metav1.GetOptions{})
|
|
if err != nil {
|
|
return false, err
|
|
}
|
|
return newSTS.Status.Replicas == *newSTS.Spec.Replicas && newSTS.Status.ObservedGeneration >= desiredGeneration, nil
|
|
}); err != nil {
|
|
t.Fatalf("failed to verify .Status.Replicas is equal to .Spec.Replicas for sts %s: %v", sts.Name, err)
|
|
}
|
|
}
|
|
|
|
func updatePod(t *testing.T, podClient typedv1.PodInterface, podName string, updateFunc func(*v1.Pod)) *v1.Pod {
|
|
var pod *v1.Pod
|
|
if err := retry.RetryOnConflict(retry.DefaultBackoff, func() error {
|
|
newPod, err := podClient.Get(context.TODO(), podName, metav1.GetOptions{})
|
|
if err != nil {
|
|
return err
|
|
}
|
|
updateFunc(newPod)
|
|
pod, err = podClient.Update(context.TODO(), newPod, metav1.UpdateOptions{})
|
|
return err
|
|
}); err != nil {
|
|
t.Fatalf("failed to update pod %s: %v", podName, err)
|
|
}
|
|
return pod
|
|
}
|
|
|
|
func updatePodStatus(t *testing.T, podClient typedv1.PodInterface, podName string, updateStatusFunc func(*v1.Pod)) *v1.Pod {
|
|
var pod *v1.Pod
|
|
if err := retry.RetryOnConflict(retry.DefaultBackoff, func() error {
|
|
newPod, err := podClient.Get(context.TODO(), podName, metav1.GetOptions{})
|
|
if err != nil {
|
|
return err
|
|
}
|
|
updateStatusFunc(newPod)
|
|
pod, err = podClient.UpdateStatus(context.TODO(), newPod, metav1.UpdateOptions{})
|
|
return err
|
|
}); err != nil {
|
|
t.Fatalf("failed to update status of pod %s: %v", podName, err)
|
|
}
|
|
return pod
|
|
}
|
|
|
|
func getPods(t *testing.T, podClient typedv1.PodInterface, labelMap map[string]string) *v1.PodList {
|
|
podSelector := labels.Set(labelMap).AsSelector()
|
|
options := metav1.ListOptions{LabelSelector: podSelector.String()}
|
|
pods, err := podClient.List(context.TODO(), options)
|
|
if err != nil {
|
|
t.Fatalf("failed obtaining a list of pods that match the pod labels %v: %v", labelMap, err)
|
|
}
|
|
if pods == nil {
|
|
t.Fatalf("obtained a nil list of pods")
|
|
}
|
|
return pods
|
|
}
|
|
|
|
func getStatefulSetPVCs(t *testing.T, pvcClient typedv1.PersistentVolumeClaimInterface, sts *appsv1.StatefulSet) []*v1.PersistentVolumeClaim {
|
|
pvcs := []*v1.PersistentVolumeClaim{}
|
|
for i := int32(0); i < *sts.Spec.Replicas; i++ {
|
|
pvcName := fmt.Sprintf("%s-%s-%d", sts.Spec.VolumeClaimTemplates[0].Name, sts.Name, i)
|
|
pvc, err := pvcClient.Get(context.TODO(), pvcName, metav1.GetOptions{})
|
|
if err != nil {
|
|
t.Fatalf("failed to get PVC %s: %v", pvcName, err)
|
|
}
|
|
pvcs = append(pvcs, pvc)
|
|
}
|
|
return pvcs
|
|
}
|
|
|
|
func verifyOwnerRef(t *testing.T, pvc *v1.PersistentVolumeClaim, kind string, expected bool) {
|
|
found := false
|
|
for _, ref := range pvc.GetOwnerReferences() {
|
|
if ref.Kind == kind {
|
|
if expected {
|
|
found = true
|
|
} else {
|
|
t.Fatalf("Found %s ref but expected none for PVC %s", kind, pvc.Name)
|
|
}
|
|
}
|
|
}
|
|
if expected && !found {
|
|
t.Fatalf("Expected %s ref but found none for PVC %s", kind, pvc.Name)
|
|
}
|
|
}
|
|
|
|
func updateSTS(t *testing.T, stsClient typedappsv1.StatefulSetInterface, stsName string, updateFunc func(*appsv1.StatefulSet)) *appsv1.StatefulSet {
|
|
var sts *appsv1.StatefulSet
|
|
if err := retry.RetryOnConflict(retry.DefaultBackoff, func() error {
|
|
newSTS, err := stsClient.Get(context.TODO(), stsName, metav1.GetOptions{})
|
|
if err != nil {
|
|
return err
|
|
}
|
|
updateFunc(newSTS)
|
|
sts, err = stsClient.Update(context.TODO(), newSTS, metav1.UpdateOptions{})
|
|
return err
|
|
}); err != nil {
|
|
t.Fatalf("failed to update sts %s: %v", stsName, err)
|
|
}
|
|
return sts
|
|
}
|
|
|
|
// Update .Spec.Replicas to replicas and verify .Status.Replicas is changed accordingly
|
|
func scaleSTS(t *testing.T, c clientset.Interface, sts *appsv1.StatefulSet, replicas int32) {
|
|
stsClient := c.AppsV1().StatefulSets(sts.Namespace)
|
|
if err := retry.RetryOnConflict(retry.DefaultBackoff, func() error {
|
|
newSTS, err := stsClient.Get(context.TODO(), sts.Name, metav1.GetOptions{})
|
|
if err != nil {
|
|
return err
|
|
}
|
|
*newSTS.Spec.Replicas = replicas
|
|
sts, err = stsClient.Update(context.TODO(), newSTS, metav1.UpdateOptions{})
|
|
return err
|
|
}); err != nil {
|
|
t.Fatalf("failed to update .Spec.Replicas to %d for sts %s: %v", replicas, sts.Name, err)
|
|
}
|
|
waitSTSStable(t, c, sts)
|
|
}
|
|
|
|
var _ admission.ValidationInterface = &fakePodFailAdmission{}
|
|
|
|
type fakePodFailAdmission struct {
|
|
lock sync.Mutex
|
|
limitedPodNumber int
|
|
succeedPodsCount int
|
|
}
|
|
|
|
func (f *fakePodFailAdmission) Handles(operation admission.Operation) bool {
|
|
return operation == admission.Create
|
|
}
|
|
|
|
func (f *fakePodFailAdmission) Validate(ctx context.Context, attr admission.Attributes, o admission.ObjectInterfaces) (err error) {
|
|
if attr.GetKind().GroupKind() != api.Kind("Pod") {
|
|
return nil
|
|
}
|
|
|
|
f.lock.Lock()
|
|
defer f.lock.Unlock()
|
|
|
|
if f.succeedPodsCount >= f.limitedPodNumber {
|
|
return fmt.Errorf("fakePodFailAdmission error")
|
|
}
|
|
f.succeedPodsCount++
|
|
return nil
|
|
}
|