mirror of
https://github.com/k3s-io/kubernetes.git
synced 2026-01-04 23:17:50 +00:00
Merge pull request #119503 from wojtek-t/pagination_ga
Graduate APIListChunking to GA
This commit is contained in:
@@ -19,7 +19,6 @@ package apimachinery
|
||||
import (
|
||||
"context"
|
||||
"fmt"
|
||||
"math/rand"
|
||||
"reflect"
|
||||
"time"
|
||||
|
||||
@@ -30,18 +29,12 @@ import (
|
||||
apierrors "k8s.io/apimachinery/pkg/api/errors"
|
||||
metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
|
||||
"k8s.io/apimachinery/pkg/util/wait"
|
||||
"k8s.io/apiserver/pkg/features"
|
||||
"k8s.io/apiserver/pkg/storage/storagebackend"
|
||||
utilfeature "k8s.io/apiserver/pkg/util/feature"
|
||||
"k8s.io/client-go/util/workqueue"
|
||||
"k8s.io/kubernetes/test/e2e/framework"
|
||||
admissionapi "k8s.io/pod-security-admission/api"
|
||||
)
|
||||
|
||||
func shouldCheckRemainingItem() bool {
|
||||
return utilfeature.DefaultFeatureGate.Enabled(features.RemainingItemCount)
|
||||
}
|
||||
|
||||
const numberOfTotalResources = 400
|
||||
|
||||
var _ = SIGDescribe("Servers with support for API chunking", func() {
|
||||
@@ -76,7 +69,7 @@ var _ = SIGDescribe("Servers with support for API chunking", func() {
|
||||
})
|
||||
})
|
||||
|
||||
ginkgo.It("should return chunks of results for list calls", func(ctx context.Context) {
|
||||
framework.ConformanceIt("should return chunks of results for list calls", func(ctx context.Context) {
|
||||
ns := f.Namespace.Name
|
||||
c := f.ClientSet
|
||||
client := c.CoreV1().PodTemplates(ns)
|
||||
@@ -86,7 +79,9 @@ var _ = SIGDescribe("Servers with support for API chunking", func() {
|
||||
found := 0
|
||||
var lastRV string
|
||||
for {
|
||||
opts.Limit = int64(rand.Int31n(numberOfTotalResources/10) + 1)
|
||||
// With numberOfTotalResources=400, we want to ensure that both
|
||||
// number of items per page and number of pages are non-trivial.
|
||||
opts.Limit = 17
|
||||
list, err := client.List(ctx, opts)
|
||||
framework.ExpectNoError(err, "failed to list pod templates in namespace: %s, given limit: %d", ns, opts.Limit)
|
||||
framework.Logf("Retrieved %d/%d results with rv %s and continue %s", len(list.Items), opts.Limit, list.ResourceVersion, list.Continue)
|
||||
@@ -96,13 +91,11 @@ var _ = SIGDescribe("Servers with support for API chunking", func() {
|
||||
lastRV = list.ResourceVersion
|
||||
}
|
||||
framework.ExpectEqual(list.ResourceVersion, lastRV)
|
||||
if shouldCheckRemainingItem() {
|
||||
if list.GetContinue() == "" {
|
||||
gomega.Expect(list.GetRemainingItemCount()).To(gomega.BeNil())
|
||||
} else {
|
||||
gomega.Expect(list.GetRemainingItemCount()).ToNot(gomega.BeNil())
|
||||
gomega.Expect(int(*list.GetRemainingItemCount()) + len(list.Items) + found).To(gomega.BeNumerically("==", numberOfTotalResources))
|
||||
}
|
||||
if list.GetContinue() == "" {
|
||||
gomega.Expect(list.GetRemainingItemCount()).To(gomega.BeNil())
|
||||
} else {
|
||||
gomega.Expect(list.GetRemainingItemCount()).ToNot(gomega.BeNil())
|
||||
gomega.Expect(int(*list.GetRemainingItemCount()) + len(list.Items) + found).To(gomega.BeNumerically("==", numberOfTotalResources))
|
||||
}
|
||||
for _, item := range list.Items {
|
||||
framework.ExpectEqual(item.Name, fmt.Sprintf("template-%04d", found))
|
||||
@@ -123,7 +116,7 @@ var _ = SIGDescribe("Servers with support for API chunking", func() {
|
||||
gomega.Expect(list.Items).To(gomega.HaveLen(numberOfTotalResources))
|
||||
})
|
||||
|
||||
ginkgo.It("should support continue listing from the last key if the original version has been compacted away, though the list is inconsistent [Slow]", func(ctx context.Context) {
|
||||
framework.ConformanceIt("should support continue listing from the last key if the original version has been compacted away, though the list is inconsistent [Slow]", func(ctx context.Context) {
|
||||
ns := f.Namespace.Name
|
||||
c := f.ClientSet
|
||||
client := c.CoreV1().PodTemplates(ns)
|
||||
@@ -136,13 +129,11 @@ var _ = SIGDescribe("Servers with support for API chunking", func() {
|
||||
framework.ExpectNoError(err, "failed to list pod templates in namespace: %s, given limit: %d", ns, opts.Limit)
|
||||
firstToken := list.Continue
|
||||
firstRV := list.ResourceVersion
|
||||
if shouldCheckRemainingItem() {
|
||||
if list.GetContinue() == "" {
|
||||
gomega.Expect(list.GetRemainingItemCount()).To(gomega.BeNil())
|
||||
} else {
|
||||
gomega.Expect(list.GetRemainingItemCount()).ToNot(gomega.BeNil())
|
||||
gomega.Expect(int(*list.GetRemainingItemCount()) + len(list.Items)).To(gomega.BeNumerically("==", numberOfTotalResources))
|
||||
}
|
||||
if list.GetContinue() == "" {
|
||||
gomega.Expect(list.GetRemainingItemCount()).To(gomega.BeNil())
|
||||
} else {
|
||||
gomega.Expect(list.GetRemainingItemCount()).ToNot(gomega.BeNil())
|
||||
gomega.Expect(int(*list.GetRemainingItemCount()) + len(list.Items)).To(gomega.BeNumerically("==", numberOfTotalResources))
|
||||
}
|
||||
framework.Logf("Retrieved %d/%d results with rv %s and continue %s", len(list.Items), opts.Limit, list.ResourceVersion, firstToken)
|
||||
|
||||
@@ -179,13 +170,11 @@ var _ = SIGDescribe("Servers with support for API chunking", func() {
|
||||
gomega.Expect(len(list.Items)).To(gomega.BeNumerically("==", opts.Limit))
|
||||
found := int(oneTenth)
|
||||
|
||||
if shouldCheckRemainingItem() {
|
||||
if list.GetContinue() == "" {
|
||||
gomega.Expect(list.GetRemainingItemCount()).To(gomega.BeNil())
|
||||
} else {
|
||||
gomega.Expect(list.GetRemainingItemCount()).ToNot(gomega.BeNil())
|
||||
gomega.Expect(int(*list.GetRemainingItemCount()) + len(list.Items) + found).To(gomega.BeNumerically("==", numberOfTotalResources))
|
||||
}
|
||||
if list.GetContinue() == "" {
|
||||
gomega.Expect(list.GetRemainingItemCount()).To(gomega.BeNil())
|
||||
} else {
|
||||
gomega.Expect(list.GetRemainingItemCount()).ToNot(gomega.BeNil())
|
||||
gomega.Expect(int(*list.GetRemainingItemCount()) + len(list.Items) + found).To(gomega.BeNumerically("==", numberOfTotalResources))
|
||||
}
|
||||
for _, item := range list.Items {
|
||||
framework.ExpectEqual(item.Name, fmt.Sprintf("template-%04d", found))
|
||||
@@ -198,13 +187,11 @@ var _ = SIGDescribe("Servers with support for API chunking", func() {
|
||||
for {
|
||||
list, err := client.List(ctx, opts)
|
||||
framework.ExpectNoError(err, "failed to list pod templates in namespace: %s, given limit: %d", ns, opts.Limit)
|
||||
if shouldCheckRemainingItem() {
|
||||
if list.GetContinue() == "" {
|
||||
gomega.Expect(list.GetRemainingItemCount()).To(gomega.BeNil())
|
||||
} else {
|
||||
gomega.Expect(list.GetRemainingItemCount()).ToNot(gomega.BeNil())
|
||||
gomega.Expect(int(*list.GetRemainingItemCount()) + len(list.Items) + found).To(gomega.BeNumerically("==", numberOfTotalResources))
|
||||
}
|
||||
if list.GetContinue() == "" {
|
||||
gomega.Expect(list.GetRemainingItemCount()).To(gomega.BeNil())
|
||||
} else {
|
||||
gomega.Expect(list.GetRemainingItemCount()).ToNot(gomega.BeNil())
|
||||
gomega.Expect(int(*list.GetRemainingItemCount()) + len(list.Items) + found).To(gomega.BeNumerically("==", numberOfTotalResources))
|
||||
}
|
||||
framework.Logf("Retrieved %d/%d results with rv %s and continue %s", len(list.Items), opts.Limit, list.ResourceVersion, list.Continue)
|
||||
gomega.Expect(len(list.Items)).To(gomega.BeNumerically("<=", opts.Limit))
|
||||
|
||||
@@ -53,16 +53,13 @@ import (
|
||||
"k8s.io/apimachinery/pkg/util/wait"
|
||||
"k8s.io/apimachinery/pkg/watch"
|
||||
"k8s.io/apiserver/pkg/endpoints/handlers"
|
||||
"k8s.io/apiserver/pkg/features"
|
||||
"k8s.io/apiserver/pkg/storage/storagebackend"
|
||||
utilfeature "k8s.io/apiserver/pkg/util/feature"
|
||||
"k8s.io/client-go/dynamic"
|
||||
clientset "k8s.io/client-go/kubernetes"
|
||||
appsv1 "k8s.io/client-go/kubernetes/typed/apps/v1"
|
||||
"k8s.io/client-go/metadata"
|
||||
restclient "k8s.io/client-go/rest"
|
||||
"k8s.io/client-go/tools/pager"
|
||||
featuregatetesting "k8s.io/component-base/featuregate/testing"
|
||||
"k8s.io/klog/v2"
|
||||
"k8s.io/kubernetes/cmd/kube-apiserver/app/options"
|
||||
kubeapiservertesting "k8s.io/kubernetes/cmd/kube-apiserver/app/testing"
|
||||
@@ -381,8 +378,6 @@ func TestListOptions(t *testing.T) {
|
||||
ctx, cancel := context.WithCancel(ctx)
|
||||
defer cancel()
|
||||
|
||||
defer featuregatetesting.SetFeatureGateDuringTest(t, utilfeature.DefaultFeatureGate, features.APIListChunking, true)()
|
||||
|
||||
var storageTransport *storagebackend.TransportConfig
|
||||
clientSet, _, tearDownFn := framework.StartTestServer(ctx, t, framework.TestServerSetup{
|
||||
ModifyServerRunOptions: func(opts *options.ServerRunOptions) {
|
||||
@@ -576,9 +571,8 @@ func testListOptionsCase(t *testing.T, rsClient appsv1.ReplicaSetInterface, watc
|
||||
|
||||
// Cacher.GetList defines this for logic to decide if the watch cache is skipped. We need to know it to know if
|
||||
// the limit is respected when testing here.
|
||||
pagingEnabled := utilfeature.DefaultFeatureGate.Enabled(features.APIListChunking)
|
||||
hasContinuation := pagingEnabled && len(opts.Continue) > 0
|
||||
hasLimit := pagingEnabled && opts.Limit > 0 && opts.ResourceVersion != "0"
|
||||
hasContinuation := len(opts.Continue) > 0
|
||||
hasLimit := opts.Limit > 0 && opts.ResourceVersion != "0"
|
||||
skipWatchCache := opts.ResourceVersion == "" || hasContinuation || hasLimit || isExact
|
||||
usingWatchCache := watchCacheEnabled && !skipWatchCache
|
||||
|
||||
@@ -627,8 +621,6 @@ func TestListResourceVersion0(t *testing.T) {
|
||||
ctx, cancel := context.WithCancel(ctx)
|
||||
defer cancel()
|
||||
|
||||
defer featuregatetesting.SetFeatureGateDuringTest(t, utilfeature.DefaultFeatureGate, features.APIListChunking, true)()
|
||||
|
||||
clientSet, _, tearDownFn := framework.StartTestServer(ctx, t, framework.TestServerSetup{
|
||||
ModifyServerRunOptions: func(opts *options.ServerRunOptions) {
|
||||
opts.Etcd.EnableWatchCache = tc.watchCacheEnabled
|
||||
@@ -685,7 +677,6 @@ func TestListResourceVersion0(t *testing.T) {
|
||||
}
|
||||
|
||||
func TestAPIListChunking(t *testing.T) {
|
||||
defer featuregatetesting.SetFeatureGateDuringTest(t, utilfeature.DefaultFeatureGate, features.APIListChunking, true)()
|
||||
ctx, clientSet, _, tearDownFn := setup(t)
|
||||
defer tearDownFn()
|
||||
|
||||
@@ -753,7 +744,6 @@ func TestAPIListChunking(t *testing.T) {
|
||||
}
|
||||
|
||||
func TestAPIListChunkingWithLabelSelector(t *testing.T) {
|
||||
defer featuregatetesting.SetFeatureGateDuringTest(t, utilfeature.DefaultFeatureGate, features.APIListChunking, true)()
|
||||
ctx, clientSet, _, tearDownFn := setup(t)
|
||||
defer tearDownFn()
|
||||
|
||||
|
||||
Reference in New Issue
Block a user