mirror of
https://github.com/k3s-io/kubernetes.git
synced 2025-07-29 06:27:05 +00:00
Change the kubelet --max-pods limit to 110 for test clusters
This allows resource usage monitoring test to launch 100 test pods per node, in addition to the add-on pods. Also reduce the test time length since the results over the shorter period are representative enough.
This commit is contained in:
parent
33eda2ffb5
commit
c1a0502487
@ -71,7 +71,7 @@ ENABLE_CLUSTER_MONITORING="${KUBE_ENABLE_CLUSTER_MONITORING:-influxdb}"
|
|||||||
TEST_CLUSTER_LOG_LEVEL="${TEST_CLUSTER_LOG_LEVEL:---v=4}"
|
TEST_CLUSTER_LOG_LEVEL="${TEST_CLUSTER_LOG_LEVEL:---v=4}"
|
||||||
TEST_CLUSTER_RESYNC_PERIOD="${TEST_CLUSTER_RESYNC_PERIOD:---min-resync-period=3m}"
|
TEST_CLUSTER_RESYNC_PERIOD="${TEST_CLUSTER_RESYNC_PERIOD:---min-resync-period=3m}"
|
||||||
|
|
||||||
KUBELET_TEST_ARGS="--max-pods=100 $TEST_CLUSTER_LOG_LEVEL"
|
KUBELET_TEST_ARGS="--max-pods=110 $TEST_CLUSTER_LOG_LEVEL"
|
||||||
APISERVER_TEST_ARGS="--runtime-config=extensions/v1beta1 ${TEST_CLUSTER_LOG_LEVEL}"
|
APISERVER_TEST_ARGS="--runtime-config=extensions/v1beta1 ${TEST_CLUSTER_LOG_LEVEL}"
|
||||||
CONTROLLER_MANAGER_TEST_ARGS="${TEST_CLUSTER_LOG_LEVEL} ${TEST_CLUSTER_RESYNC_PERIOD}"
|
CONTROLLER_MANAGER_TEST_ARGS="${TEST_CLUSTER_LOG_LEVEL} ${TEST_CLUSTER_RESYNC_PERIOD}"
|
||||||
SCHEDULER_TEST_ARGS="${TEST_CLUSTER_LOG_LEVEL}"
|
SCHEDULER_TEST_ARGS="${TEST_CLUSTER_LOG_LEVEL}"
|
||||||
|
@ -275,7 +275,7 @@ GKE_REQUIRED_SKIP_TESTS=(
|
|||||||
"Nodes"
|
"Nodes"
|
||||||
"Etcd\sFailure"
|
"Etcd\sFailure"
|
||||||
"MasterCerts"
|
"MasterCerts"
|
||||||
"experimental\sresource\susage\stracking" # Expect --max-pods=100
|
"experimental\sresource\susage\stracking" # Expect --max-pods=110
|
||||||
"ServiceLoadBalancer" # issue: #16602
|
"ServiceLoadBalancer" # issue: #16602
|
||||||
"Shell"
|
"Shell"
|
||||||
# Alpha features, remove from skip when these move to beta
|
# Alpha features, remove from skip when these move to beta
|
||||||
@ -298,7 +298,7 @@ GKE_DEFAULT_SKIP_TESTS=(
|
|||||||
|
|
||||||
# Tests which cannot be run on AWS.
|
# Tests which cannot be run on AWS.
|
||||||
AWS_REQUIRED_SKIP_TESTS=(
|
AWS_REQUIRED_SKIP_TESTS=(
|
||||||
"experimental\sresource\susage\stracking" # Expect --max-pods=100
|
"experimental\sresource\susage\stracking" # Expect --max-pods=110
|
||||||
"GCE\sL7\sLoadBalancer\sController" # GCE L7 loadbalancing
|
"GCE\sL7\sLoadBalancer\sController" # GCE L7 loadbalancing
|
||||||
)
|
)
|
||||||
|
|
||||||
|
@ -33,7 +33,7 @@ const (
|
|||||||
// Interval to poll /stats/container on a node
|
// Interval to poll /stats/container on a node
|
||||||
containerStatsPollingPeriod = 10 * time.Second
|
containerStatsPollingPeriod = 10 * time.Second
|
||||||
// The monitoring time for one test.
|
// The monitoring time for one test.
|
||||||
monitoringTime = 30 * time.Minute
|
monitoringTime = 20 * time.Minute
|
||||||
// The periodic reporting period.
|
// The periodic reporting period.
|
||||||
reportingPeriod = 5 * time.Minute
|
reportingPeriod = 5 * time.Minute
|
||||||
)
|
)
|
||||||
@ -118,7 +118,7 @@ var _ = Describe("Kubelet", func() {
|
|||||||
})
|
})
|
||||||
|
|
||||||
Describe("regular resource usage tracking", func() {
|
Describe("regular resource usage tracking", func() {
|
||||||
density := []int{0, 35}
|
density := []int{0, 40}
|
||||||
for i := range density {
|
for i := range density {
|
||||||
podsPerNode := density[i]
|
podsPerNode := density[i]
|
||||||
name := fmt.Sprintf(
|
name := fmt.Sprintf(
|
||||||
@ -129,7 +129,7 @@ var _ = Describe("Kubelet", func() {
|
|||||||
}
|
}
|
||||||
})
|
})
|
||||||
Describe("experimental resource usage tracking", func() {
|
Describe("experimental resource usage tracking", func() {
|
||||||
density := []int{50}
|
density := []int{100}
|
||||||
for i := range density {
|
for i := range density {
|
||||||
podsPerNode := density[i]
|
podsPerNode := density[i]
|
||||||
name := fmt.Sprintf(
|
name := fmt.Sprintf(
|
||||||
|
Loading…
Reference in New Issue
Block a user