mirror of
https://github.com/k3s-io/kubernetes.git
synced 2025-07-27 05:27:21 +00:00
Merge pull request #76328 from pontiyaraja/limitRangeCNF
removed flaky watch code and added NewIndexerInformerWatcher for LimitRange test case
This commit is contained in:
commit
d0dc9c84c0
@ -42,6 +42,7 @@ go_library(
|
|||||||
"//staging/src/k8s.io/apimachinery/pkg/watch:go_default_library",
|
"//staging/src/k8s.io/apimachinery/pkg/watch:go_default_library",
|
||||||
"//staging/src/k8s.io/client-go/kubernetes:go_default_library",
|
"//staging/src/k8s.io/client-go/kubernetes:go_default_library",
|
||||||
"//staging/src/k8s.io/client-go/tools/cache:go_default_library",
|
"//staging/src/k8s.io/client-go/tools/cache:go_default_library",
|
||||||
|
"//staging/src/k8s.io/client-go/tools/watch:go_default_library",
|
||||||
"//test/e2e/common:go_default_library",
|
"//test/e2e/common:go_default_library",
|
||||||
"//test/e2e/framework:go_default_library",
|
"//test/e2e/framework:go_default_library",
|
||||||
"//test/e2e/framework/gpu:go_default_library",
|
"//test/e2e/framework/gpu:go_default_library",
|
||||||
|
@ -19,18 +19,24 @@ package scheduling
|
|||||||
import (
|
import (
|
||||||
"fmt"
|
"fmt"
|
||||||
"reflect"
|
"reflect"
|
||||||
|
"strconv"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
"k8s.io/api/core/v1"
|
"k8s.io/api/core/v1"
|
||||||
"k8s.io/apimachinery/pkg/api/resource"
|
"k8s.io/apimachinery/pkg/api/resource"
|
||||||
metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
|
metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
|
||||||
"k8s.io/apimachinery/pkg/labels"
|
"k8s.io/apimachinery/pkg/labels"
|
||||||
|
"k8s.io/apimachinery/pkg/runtime"
|
||||||
|
"k8s.io/apimachinery/pkg/util/uuid"
|
||||||
"k8s.io/apimachinery/pkg/util/wait"
|
"k8s.io/apimachinery/pkg/util/wait"
|
||||||
"k8s.io/apimachinery/pkg/watch"
|
"k8s.io/apimachinery/pkg/watch"
|
||||||
|
"k8s.io/client-go/tools/cache"
|
||||||
|
watchtools "k8s.io/client-go/tools/watch"
|
||||||
"k8s.io/kubernetes/test/e2e/framework"
|
"k8s.io/kubernetes/test/e2e/framework"
|
||||||
e2elog "k8s.io/kubernetes/test/e2e/framework/log"
|
e2elog "k8s.io/kubernetes/test/e2e/framework/log"
|
||||||
|
|
||||||
"github.com/onsi/ginkgo"
|
"github.com/onsi/ginkgo"
|
||||||
|
"github.com/onsi/gomega"
|
||||||
)
|
)
|
||||||
|
|
||||||
const (
|
const (
|
||||||
@ -40,21 +46,28 @@ const (
|
|||||||
var _ = SIGDescribe("LimitRange", func() {
|
var _ = SIGDescribe("LimitRange", func() {
|
||||||
f := framework.NewDefaultFramework("limitrange")
|
f := framework.NewDefaultFramework("limitrange")
|
||||||
|
|
||||||
|
/*
|
||||||
|
Release : v1.15
|
||||||
|
Testname: LimitRange, resources
|
||||||
|
Description: Creating a Limitrange and verifying the creation of Limitrange, updating the Limitrange and validating the Limitrange. Creating Pods with resources and validate the pod resources are applied to the Limitrange
|
||||||
|
*/
|
||||||
|
//note: this test case can be promoted to conformance after verified the stability of the test case
|
||||||
ginkgo.It("should create a LimitRange with defaults and ensure pod has those defaults applied.", func() {
|
ginkgo.It("should create a LimitRange with defaults and ensure pod has those defaults applied.", func() {
|
||||||
ginkgo.By("Creating a LimitRange")
|
ginkgo.By("Creating a LimitRange")
|
||||||
|
|
||||||
min := getResourceList("50m", "100Mi", "100Gi")
|
min := getResourceList("50m", "100Mi", "100Gi")
|
||||||
max := getResourceList("500m", "500Mi", "500Gi")
|
max := getResourceList("500m", "500Mi", "500Gi")
|
||||||
defaultLimit := getResourceList("500m", "500Mi", "500Gi")
|
defaultLimit := getResourceList("500m", "500Mi", "500Gi")
|
||||||
defaultRequest := getResourceList("100m", "200Mi", "200Gi")
|
defaultRequest := getResourceList("100m", "200Mi", "200Gi")
|
||||||
maxLimitRequestRatio := v1.ResourceList{}
|
maxLimitRequestRatio := v1.ResourceList{}
|
||||||
limitRange := newLimitRange("limit-range", v1.LimitTypeContainer,
|
value := strconv.Itoa(time.Now().Nanosecond()) + string(uuid.NewUUID())
|
||||||
|
limitRange := newLimitRange("limit-range", value, v1.LimitTypeContainer,
|
||||||
min, max,
|
min, max,
|
||||||
defaultLimit, defaultRequest,
|
defaultLimit, defaultRequest,
|
||||||
maxLimitRequestRatio)
|
maxLimitRequestRatio)
|
||||||
|
|
||||||
ginkgo.By("Setting up watch")
|
ginkgo.By("Setting up watch")
|
||||||
selector := labels.SelectorFromSet(labels.Set(map[string]string{"name": limitRange.Name}))
|
selector := labels.SelectorFromSet(labels.Set(map[string]string{"time": value}))
|
||||||
|
|
||||||
options := metav1.ListOptions{LabelSelector: selector.String()}
|
options := metav1.ListOptions{LabelSelector: selector.String()}
|
||||||
limitRanges, err := f.ClientSet.CoreV1().LimitRanges(f.Namespace.Name).List(options)
|
limitRanges, err := f.ClientSet.CoreV1().LimitRanges(f.Namespace.Name).List(options)
|
||||||
framework.ExpectNoError(err, "failed to query for limitRanges")
|
framework.ExpectNoError(err, "failed to query for limitRanges")
|
||||||
@ -63,22 +76,49 @@ var _ = SIGDescribe("LimitRange", func() {
|
|||||||
LabelSelector: selector.String(),
|
LabelSelector: selector.String(),
|
||||||
ResourceVersion: limitRanges.ListMeta.ResourceVersion,
|
ResourceVersion: limitRanges.ListMeta.ResourceVersion,
|
||||||
}
|
}
|
||||||
w, err := f.ClientSet.CoreV1().LimitRanges(f.Namespace.Name).Watch(metav1.ListOptions{})
|
|
||||||
framework.ExpectNoError(err, "failed to set up watch")
|
listCompleted := make(chan bool, 1)
|
||||||
|
lw := &cache.ListWatch{
|
||||||
|
ListFunc: func(options metav1.ListOptions) (runtime.Object, error) {
|
||||||
|
options.LabelSelector = selector.String()
|
||||||
|
limitRanges, err := f.ClientSet.CoreV1().LimitRanges(f.Namespace.Name).List(options)
|
||||||
|
if err == nil {
|
||||||
|
select {
|
||||||
|
case listCompleted <- true:
|
||||||
|
e2elog.Logf("observed the limitRanges list")
|
||||||
|
return limitRanges, err
|
||||||
|
default:
|
||||||
|
e2elog.Logf("channel blocked")
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return limitRanges, err
|
||||||
|
},
|
||||||
|
WatchFunc: func(options metav1.ListOptions) (watch.Interface, error) {
|
||||||
|
options.LabelSelector = selector.String()
|
||||||
|
return f.ClientSet.CoreV1().LimitRanges(f.Namespace.Name).Watch(options)
|
||||||
|
},
|
||||||
|
}
|
||||||
|
_, _, w, _ := watchtools.NewIndexerInformerWatcher(lw, &v1.LimitRange{})
|
||||||
|
defer w.Stop()
|
||||||
|
|
||||||
ginkgo.By("Submitting a LimitRange")
|
ginkgo.By("Submitting a LimitRange")
|
||||||
limitRange, err = f.ClientSet.CoreV1().LimitRanges(f.Namespace.Name).Create(limitRange)
|
limitRange, err = f.ClientSet.CoreV1().LimitRanges(f.Namespace.Name).Create(limitRange)
|
||||||
framework.ExpectNoError(err)
|
framework.ExpectNoError(err)
|
||||||
|
|
||||||
ginkgo.By("Verifying LimitRange creation was observed")
|
ginkgo.By("Verifying LimitRange creation was observed")
|
||||||
|
select {
|
||||||
|
case <-listCompleted:
|
||||||
select {
|
select {
|
||||||
case event, _ := <-w.ResultChan():
|
case event, _ := <-w.ResultChan():
|
||||||
if event.Type != watch.Added {
|
if event.Type != watch.Added {
|
||||||
e2elog.Failf("Failed to observe pod creation: %v", event)
|
e2elog.Failf("Failed to observe limitRange creation : %v", event)
|
||||||
}
|
}
|
||||||
case <-time.After(framework.ServiceRespondingTimeout):
|
case <-time.After(framework.ServiceRespondingTimeout):
|
||||||
e2elog.Failf("Timeout while waiting for LimitRange creation")
|
e2elog.Failf("Timeout while waiting for LimitRange creation")
|
||||||
}
|
}
|
||||||
|
case <-time.After(framework.ServiceRespondingTimeout):
|
||||||
|
e2elog.Failf("Timeout while waiting for LimitRange list complete")
|
||||||
|
}
|
||||||
|
|
||||||
ginkgo.By("Fetching the LimitRange to ensure it has proper values")
|
ginkgo.By("Fetching the LimitRange to ensure it has proper values")
|
||||||
limitRange, err = f.ClientSet.CoreV1().LimitRanges(f.Namespace.Name).Get(limitRange.Name, metav1.GetOptions{})
|
limitRange, err = f.ClientSet.CoreV1().LimitRanges(f.Namespace.Name).Get(limitRange.Name, metav1.GetOptions{})
|
||||||
@ -165,7 +205,7 @@ var _ = SIGDescribe("LimitRange", func() {
|
|||||||
framework.ExpectNoError(err)
|
framework.ExpectNoError(err)
|
||||||
|
|
||||||
ginkgo.By("Verifying the LimitRange was deleted")
|
ginkgo.By("Verifying the LimitRange was deleted")
|
||||||
err = wait.Poll(time.Second*5, time.Second*30, func() (bool, error) {
|
gomega.Expect(wait.Poll(time.Second*5, framework.ServiceRespondingTimeout, func() (bool, error) {
|
||||||
selector := labels.SelectorFromSet(labels.Set(map[string]string{"name": limitRange.Name}))
|
selector := labels.SelectorFromSet(labels.Set(map[string]string{"name": limitRange.Name}))
|
||||||
options := metav1.ListOptions{LabelSelector: selector.String()}
|
options := metav1.ListOptions{LabelSelector: selector.String()}
|
||||||
limitRanges, err := f.ClientSet.CoreV1().LimitRanges(f.Namespace.Name).List(options)
|
limitRanges, err := f.ClientSet.CoreV1().LimitRanges(f.Namespace.Name).List(options)
|
||||||
@ -190,7 +230,8 @@ var _ = SIGDescribe("LimitRange", func() {
|
|||||||
|
|
||||||
return false, nil
|
return false, nil
|
||||||
|
|
||||||
})
|
}))
|
||||||
|
|
||||||
framework.ExpectNoError(err, "kubelet never observed the termination notice")
|
framework.ExpectNoError(err, "kubelet never observed the termination notice")
|
||||||
|
|
||||||
ginkgo.By("Creating a Pod with more than former max resources")
|
ginkgo.By("Creating a Pod with more than former max resources")
|
||||||
@ -241,13 +282,16 @@ func getResourceList(cpu, memory string, ephemeralStorage string) v1.ResourceLis
|
|||||||
}
|
}
|
||||||
|
|
||||||
// newLimitRange returns a limit range with specified data
|
// newLimitRange returns a limit range with specified data
|
||||||
func newLimitRange(name string, limitType v1.LimitType,
|
func newLimitRange(name, value string, limitType v1.LimitType,
|
||||||
min, max,
|
min, max,
|
||||||
defaultLimit, defaultRequest,
|
defaultLimit, defaultRequest,
|
||||||
maxLimitRequestRatio v1.ResourceList) *v1.LimitRange {
|
maxLimitRequestRatio v1.ResourceList) *v1.LimitRange {
|
||||||
return &v1.LimitRange{
|
return &v1.LimitRange{
|
||||||
ObjectMeta: metav1.ObjectMeta{
|
ObjectMeta: metav1.ObjectMeta{
|
||||||
Name: name,
|
Name: name,
|
||||||
|
Labels: map[string]string{
|
||||||
|
"time": value,
|
||||||
|
},
|
||||||
},
|
},
|
||||||
Spec: v1.LimitRangeSpec{
|
Spec: v1.LimitRangeSpec{
|
||||||
Limits: []v1.LimitRangeItem{
|
Limits: []v1.LimitRangeItem{
|
||||||
|
Loading…
Reference in New Issue
Block a user