mirror of
https://github.com/k3s-io/kubernetes.git
synced 2025-07-25 20:53:33 +00:00
Merge pull request #88459 from joakimr-axis/joakimr-axis_config-test.sh
Fix shellcheck warnings/errors in cluster/gce/config-test.sh
This commit is contained in:
commit
aed3f05495
@ -21,10 +21,9 @@ source "${KUBE_ROOT}/cluster/gce/config-common.sh"
|
|||||||
|
|
||||||
# Specifying KUBE_GCE_API_ENDPOINT will override the default GCE Compute API endpoint (https://www.googleapis.com/compute/v1/).
|
# Specifying KUBE_GCE_API_ENDPOINT will override the default GCE Compute API endpoint (https://www.googleapis.com/compute/v1/).
|
||||||
# This endpoint has to be pointing to v1 api. For example, https://www.googleapis.com/compute/staging_v1/
|
# This endpoint has to be pointing to v1 api. For example, https://www.googleapis.com/compute/staging_v1/
|
||||||
GCE_API_ENDPOINT=${KUBE_GCE_API_ENDPOINT:-}
|
export GCE_API_ENDPOINT=${KUBE_GCE_API_ENDPOINT:-}
|
||||||
GCLOUD=gcloud
|
|
||||||
ZONE=${KUBE_GCE_ZONE:-us-central1-b}
|
ZONE=${KUBE_GCE_ZONE:-us-central1-b}
|
||||||
REGION=${ZONE%-*}
|
export REGION=${ZONE%-*}
|
||||||
RELEASE_REGION_FALLBACK=${RELEASE_REGION_FALLBACK:-false}
|
RELEASE_REGION_FALLBACK=${RELEASE_REGION_FALLBACK:-false}
|
||||||
REGIONAL_KUBE_ADDONS=${REGIONAL_KUBE_ADDONS:-true}
|
REGIONAL_KUBE_ADDONS=${REGIONAL_KUBE_ADDONS:-true}
|
||||||
NODE_SIZE=${NODE_SIZE:-n1-standard-2}
|
NODE_SIZE=${NODE_SIZE:-n1-standard-2}
|
||||||
@ -32,27 +31,27 @@ NUM_NODES=${NUM_NODES:-3}
|
|||||||
NUM_WINDOWS_NODES=${NUM_WINDOWS_NODES:-0}
|
NUM_WINDOWS_NODES=${NUM_WINDOWS_NODES:-0}
|
||||||
MASTER_SIZE=${MASTER_SIZE:-n1-standard-$(get-master-size)}
|
MASTER_SIZE=${MASTER_SIZE:-n1-standard-$(get-master-size)}
|
||||||
MASTER_MIN_CPU_ARCHITECTURE=${MASTER_MIN_CPU_ARCHITECTURE:-} # To allow choosing better architectures.
|
MASTER_MIN_CPU_ARCHITECTURE=${MASTER_MIN_CPU_ARCHITECTURE:-} # To allow choosing better architectures.
|
||||||
MASTER_DISK_TYPE=pd-ssd
|
export MASTER_DISK_TYPE=pd-ssd
|
||||||
MASTER_DISK_SIZE=${MASTER_DISK_SIZE:-$(get-master-disk-size)}
|
MASTER_DISK_SIZE=${MASTER_DISK_SIZE:-$(get-master-disk-size)}
|
||||||
MASTER_ROOT_DISK_SIZE=${MASTER_ROOT_DISK_SIZE:-$(get-master-root-disk-size)}
|
MASTER_ROOT_DISK_SIZE=${MASTER_ROOT_DISK_SIZE:-$(get-master-root-disk-size)}
|
||||||
NODE_DISK_TYPE=${NODE_DISK_TYPE:-pd-standard}
|
NODE_DISK_TYPE=${NODE_DISK_TYPE:-pd-standard}
|
||||||
NODE_DISK_SIZE=${NODE_DISK_SIZE:-100GB}
|
NODE_DISK_SIZE=${NODE_DISK_SIZE:-100GB}
|
||||||
NODE_LOCAL_SSDS=${NODE_LOCAL_SSDS:-0}
|
NODE_LOCAL_SSDS=${NODE_LOCAL_SSDS:-0}
|
||||||
NODE_LABELS="${KUBE_NODE_LABELS:-}"
|
NODE_LABELS=${KUBE_NODE_LABELS:-}
|
||||||
WINDOWS_NODE_LABELS="${WINDOWS_NODE_LABELS:-}"
|
WINDOWS_NODE_LABELS=${WINDOWS_NODE_LABELS:-}
|
||||||
|
|
||||||
# KUBE_CREATE_NODES can be used to avoid creating nodes, while master will be sized for NUM_NODES nodes.
|
# KUBE_CREATE_NODES can be used to avoid creating nodes, while master will be sized for NUM_NODES nodes.
|
||||||
# Firewalls and node templates are still created.
|
# Firewalls and node templates are still created.
|
||||||
KUBE_CREATE_NODES="${KUBE_CREATE_NODES:-true}"
|
KUBE_CREATE_NODES=${KUBE_CREATE_NODES:-true}
|
||||||
|
|
||||||
# An extension to local SSDs allowing users to specify block/fs and SCSI/NVMe devices
|
# An extension to local SSDs allowing users to specify block/fs and SCSI/NVMe devices
|
||||||
# Format of this variable will be "#,scsi/nvme,block/fs" you can specify multiple
|
# Format of this variable will be "#,scsi/nvme,block/fs" you can specify multiple
|
||||||
# configurations by separating them by a semi-colon ex. "2,scsi,fs;1,nvme,block"
|
# configurations by separating them by a semi-colon ex. "2,scsi,fs;1,nvme,block"
|
||||||
# is a request for 2 SCSI formatted and mounted SSDs and 1 NVMe block device SSD.
|
# is a request for 2 SCSI formatted and mounted SSDs and 1 NVMe block device SSD.
|
||||||
NODE_LOCAL_SSDS_EXT=${NODE_LOCAL_SSDS_EXT:-}
|
NODE_LOCAL_SSDS_EXT=${NODE_LOCAL_SSDS_EXT:-}
|
||||||
NODE_ACCELERATORS=${NODE_ACCELERATORS:-""}
|
NODE_ACCELERATORS=${NODE_ACCELERATORS:-''}
|
||||||
REGISTER_MASTER_KUBELET=${REGISTER_MASTER:-true}
|
export REGISTER_MASTER_KUBELET=${REGISTER_MASTER:-true}
|
||||||
KUBE_APISERVER_REQUEST_TIMEOUT=300
|
export KUBE_APISERVER_REQUEST_TIMEOUT=300
|
||||||
# Increase initial delay for the apiserver liveness probe, to avoid prematurely tearing it down
|
# Increase initial delay for the apiserver liveness probe, to avoid prematurely tearing it down
|
||||||
KUBE_APISERVER_LIVENESS_PROBE_INITIAL_DELAY_SEC=${KUBE_APISERVER_LIVENESS_PROBE_INITIAL_DELAY_SEC:-45}
|
KUBE_APISERVER_LIVENESS_PROBE_INITIAL_DELAY_SEC=${KUBE_APISERVER_LIVENESS_PROBE_INITIAL_DELAY_SEC:-45}
|
||||||
# Also increase the initial delay for etcd just to be safe
|
# Also increase the initial delay for etcd just to be safe
|
||||||
@ -68,22 +67,22 @@ MASTER_OS_DISTRIBUTION=${KUBE_MASTER_OS_DISTRIBUTION:-${KUBE_OS_DISTRIBUTION:-gc
|
|||||||
NODE_OS_DISTRIBUTION=${KUBE_NODE_OS_DISTRIBUTION:-${KUBE_OS_DISTRIBUTION:-gci}}
|
NODE_OS_DISTRIBUTION=${KUBE_NODE_OS_DISTRIBUTION:-${KUBE_OS_DISTRIBUTION:-gci}}
|
||||||
WINDOWS_NODE_OS_DISTRIBUTION=${WINDOWS_NODE_OS_DISTRIBUTION:-win2019}
|
WINDOWS_NODE_OS_DISTRIBUTION=${WINDOWS_NODE_OS_DISTRIBUTION:-win2019}
|
||||||
|
|
||||||
if [[ "${MASTER_OS_DISTRIBUTION}" == "cos" ]]; then
|
if [[ "${MASTER_OS_DISTRIBUTION}" = 'cos' ]]; then
|
||||||
MASTER_OS_DISTRIBUTION="gci"
|
MASTER_OS_DISTRIBUTION='gci'
|
||||||
fi
|
fi
|
||||||
|
|
||||||
if [[ "${NODE_OS_DISTRIBUTION}" == "cos" ]]; then
|
if [[ "${NODE_OS_DISTRIBUTION}" = 'cos' ]]; then
|
||||||
NODE_OS_DISTRIBUTION="gci"
|
NODE_OS_DISTRIBUTION='gci'
|
||||||
fi
|
fi
|
||||||
|
|
||||||
# GPUs supported in GCE do not have compatible drivers in Debian 7.
|
# GPUs supported in GCE do not have compatible drivers in Debian 7.
|
||||||
if [[ "${NODE_OS_DISTRIBUTION}" == "debian" ]]; then
|
if [[ "${NODE_OS_DISTRIBUTION}" = 'debian' ]]; then
|
||||||
NODE_ACCELERATORS=""
|
NODE_ACCELERATORS=''
|
||||||
fi
|
fi
|
||||||
|
|
||||||
# To avoid failing large tests due to some flakes in starting nodes, allow
|
# To avoid failing large tests due to some flakes in starting nodes, allow
|
||||||
# for a small percentage of nodes to not start during cluster startup.
|
# for a small percentage of nodes to not start during cluster startup.
|
||||||
ALLOWED_NOTREADY_NODES="${ALLOWED_NOTREADY_NODES:-$(($(get-num-nodes) / 100))}"
|
ALLOWED_NOTREADY_NODES=${ALLOWED_NOTREADY_NODES:-$(($(get-num-nodes) / 100))}
|
||||||
|
|
||||||
# By default a cluster will be started with the master and nodes
|
# By default a cluster will be started with the master and nodes
|
||||||
# on Container-optimized OS (cos, previously known as gci). If
|
# on Container-optimized OS (cos, previously known as gci). If
|
||||||
@ -91,136 +90,139 @@ ALLOWED_NOTREADY_NODES="${ALLOWED_NOTREADY_NODES:-$(($(get-num-nodes) / 100))}"
|
|||||||
# Also please update corresponding image for node e2e at:
|
# Also please update corresponding image for node e2e at:
|
||||||
# https://github.com/kubernetes/kubernetes/blob/master/test/e2e_node/jenkins/image-config.yaml
|
# https://github.com/kubernetes/kubernetes/blob/master/test/e2e_node/jenkins/image-config.yaml
|
||||||
GCI_VERSION=${KUBE_GCI_VERSION:-cos-81-12871-59-0}
|
GCI_VERSION=${KUBE_GCI_VERSION:-cos-81-12871-59-0}
|
||||||
MASTER_IMAGE=${KUBE_GCE_MASTER_IMAGE:-}
|
export MASTER_IMAGE=${KUBE_GCE_MASTER_IMAGE:-}
|
||||||
MASTER_IMAGE_PROJECT=${KUBE_GCE_MASTER_PROJECT:-cos-cloud}
|
export MASTER_IMAGE_PROJECT=${KUBE_GCE_MASTER_PROJECT:-cos-cloud}
|
||||||
NODE_IMAGE=${KUBE_GCE_NODE_IMAGE:-${GCI_VERSION}}
|
export NODE_IMAGE=${KUBE_GCE_NODE_IMAGE:-${GCI_VERSION}}
|
||||||
NODE_IMAGE_PROJECT=${KUBE_GCE_NODE_PROJECT:-cos-cloud}
|
export NODE_IMAGE_PROJECT=${KUBE_GCE_NODE_PROJECT:-cos-cloud}
|
||||||
NODE_SERVICE_ACCOUNT=${KUBE_GCE_NODE_SERVICE_ACCOUNT:-default}
|
export NODE_SERVICE_ACCOUNT=${KUBE_GCE_NODE_SERVICE_ACCOUNT:-default}
|
||||||
|
|
||||||
CONTAINER_RUNTIME=${KUBE_CONTAINER_RUNTIME:-docker}
|
CONTAINER_RUNTIME=${KUBE_CONTAINER_RUNTIME:-docker}
|
||||||
CONTAINER_RUNTIME_ENDPOINT=${KUBE_CONTAINER_RUNTIME_ENDPOINT:-}
|
export CONTAINER_RUNTIME_ENDPOINT=${KUBE_CONTAINER_RUNTIME_ENDPOINT:-}
|
||||||
CONTAINER_RUNTIME_NAME=${KUBE_CONTAINER_RUNTIME_NAME:-}
|
CONTAINER_RUNTIME_NAME=${KUBE_CONTAINER_RUNTIME_NAME:-}
|
||||||
LOAD_IMAGE_COMMAND=${KUBE_LOAD_IMAGE_COMMAND:-}
|
LOAD_IMAGE_COMMAND=${KUBE_LOAD_IMAGE_COMMAND:-}
|
||||||
GCI_DOCKER_VERSION=${KUBE_GCI_DOCKER_VERSION:-}
|
export GCI_DOCKER_VERSION=${KUBE_GCI_DOCKER_VERSION:-}
|
||||||
if [[ "${CONTAINER_RUNTIME}" == "containerd" ]]; then
|
if [[ "${CONTAINER_RUNTIME}" = 'containerd' ]]; then
|
||||||
CONTAINER_RUNTIME_NAME=${KUBE_CONTAINER_RUNTIME_NAME:-containerd}
|
CONTAINER_RUNTIME_NAME=${KUBE_CONTAINER_RUNTIME_NAME:-containerd}
|
||||||
LOAD_IMAGE_COMMAND=${KUBE_LOAD_IMAGE_COMMAND:-ctr -n=k8s.io images import}
|
LOAD_IMAGE_COMMAND=${KUBE_LOAD_IMAGE_COMMAND:-ctr -n=k8s.io images import}
|
||||||
fi
|
fi
|
||||||
|
export CONTAINER_RUNTIME_NAME
|
||||||
|
export LOAD_IMAGE_COMMAND
|
||||||
|
|
||||||
# Ability to inject custom versions (Ubuntu OS images ONLY)
|
# Ability to inject custom versions (Ubuntu OS images ONLY)
|
||||||
# if KUBE_UBUNTU_INSTALL_CONTAINERD_VERSION or KUBE_UBUNTU_INSTALL_RUNC_VERSION
|
# if KUBE_UBUNTU_INSTALL_CONTAINERD_VERSION or KUBE_UBUNTU_INSTALL_RUNC_VERSION
|
||||||
# is set to empty then we do not override the version(s) and just
|
# is set to empty then we do not override the version(s) and just
|
||||||
# use whatever is in the default installation of containerd package
|
# use whatever is in the default installation of containerd package
|
||||||
UBUNTU_INSTALL_CONTAINERD_VERSION=${KUBE_UBUNTU_INSTALL_CONTAINERD_VERSION:-}
|
export UBUNTU_INSTALL_CONTAINERD_VERSION=${KUBE_UBUNTU_INSTALL_CONTAINERD_VERSION:-}
|
||||||
UBUNTU_INSTALL_RUNC_VERSION=${KUBE_UBUNTU_INSTALL_RUNC_VERSION:-}
|
export UBUNTU_INSTALL_RUNC_VERSION=${KUBE_UBUNTU_INSTALL_RUNC_VERSION:-}
|
||||||
|
|
||||||
# MASTER_EXTRA_METADATA is the extra instance metadata on master instance separated by commas.
|
# MASTER_EXTRA_METADATA is the extra instance metadata on master instance separated by commas.
|
||||||
MASTER_EXTRA_METADATA=${KUBE_MASTER_EXTRA_METADATA:-${KUBE_EXTRA_METADATA:-}}
|
export MASTER_EXTRA_METADATA=${KUBE_MASTER_EXTRA_METADATA:-${KUBE_EXTRA_METADATA:-}}
|
||||||
# MASTER_EXTRA_METADATA is the extra instance metadata on node instance separated by commas.
|
# MASTER_EXTRA_METADATA is the extra instance metadata on node instance separated by commas.
|
||||||
NODE_EXTRA_METADATA=${KUBE_NODE_EXTRA_METADATA:-${KUBE_EXTRA_METADATA:-}}
|
export NODE_EXTRA_METADATA=${KUBE_NODE_EXTRA_METADATA:-${KUBE_EXTRA_METADATA:-}}
|
||||||
|
|
||||||
NETWORK=${KUBE_GCE_NETWORK:-e2e-test-${USER}}
|
NETWORK=${KUBE_GCE_NETWORK:-e2e-test-${USER}}
|
||||||
if [[ "${CREATE_CUSTOM_NETWORK}" == true ]]; then
|
if [[ "${CREATE_CUSTOM_NETWORK}" = true ]]; then
|
||||||
SUBNETWORK="${SUBNETWORK:-${NETWORK}-custom-subnet}"
|
SUBNETWORK=${SUBNETWORK:-${NETWORK}-custom-subnet}
|
||||||
fi
|
fi
|
||||||
INSTANCE_PREFIX="${KUBE_GCE_INSTANCE_PREFIX:-e2e-test-${USER}}"
|
INSTANCE_PREFIX=${KUBE_GCE_INSTANCE_PREFIX:-e2e-test-${USER}}
|
||||||
CLUSTER_NAME="${CLUSTER_NAME:-${INSTANCE_PREFIX}}"
|
CLUSTER_NAME=${CLUSTER_NAME:-${INSTANCE_PREFIX}}
|
||||||
MASTER_NAME="${INSTANCE_PREFIX}-master"
|
MASTER_NAME="${INSTANCE_PREFIX}-master"
|
||||||
AGGREGATOR_MASTER_NAME="${INSTANCE_PREFIX}-aggregator"
|
export AGGREGATOR_MASTER_NAME="${INSTANCE_PREFIX}-aggregator"
|
||||||
INITIAL_ETCD_CLUSTER="${MASTER_NAME}"
|
export INITIAL_ETCD_CLUSTER=${MASTER_NAME}
|
||||||
MASTER_TAG="${INSTANCE_PREFIX}-master"
|
export MASTER_TAG="${INSTANCE_PREFIX}-master"
|
||||||
NODE_TAG="${INSTANCE_PREFIX}-minion"
|
export NODE_TAG="${INSTANCE_PREFIX}-minion"
|
||||||
|
|
||||||
CLUSTER_IP_RANGE="${CLUSTER_IP_RANGE:-$(get-cluster-ip-range)}"
|
CLUSTER_IP_RANGE=${CLUSTER_IP_RANGE:-$(get-cluster-ip-range)}
|
||||||
MASTER_IP_RANGE="${MASTER_IP_RANGE:-10.246.0.0/24}"
|
MASTER_IP_RANGE=${MASTER_IP_RANGE:-10.246.0.0/24}
|
||||||
# NODE_IP_RANGE is used when ENABLE_IP_ALIASES=true or CREATE_CUSTOM_NETWORK=true.
|
# NODE_IP_RANGE is used when ENABLE_IP_ALIASES=true or CREATE_CUSTOM_NETWORK=true.
|
||||||
# It is the primary range in the subnet and is the range used for node instance IPs.
|
# It is the primary range in the subnet and is the range used for node instance IPs.
|
||||||
NODE_IP_RANGE="$(get-node-ip-range)"
|
NODE_IP_RANGE=$(get-node-ip-range)
|
||||||
|
export NODE_IP_RANGE
|
||||||
|
|
||||||
RUNTIME_CONFIG="${KUBE_RUNTIME_CONFIG:-}"
|
export RUNTIME_CONFIG=${KUBE_RUNTIME_CONFIG:-}
|
||||||
|
|
||||||
if [[ "${KUBE_FEATURE_GATES:-}" == "AllAlpha=true" ]]; then
|
if [[ "${KUBE_FEATURE_GATES:-}" = 'AllAlpha=true' ]]; then
|
||||||
RUNTIME_CONFIG="${KUBE_RUNTIME_CONFIG:-api/all=true}"
|
RUNTIME_CONFIG=${KUBE_RUNTIME_CONFIG:-api/all=true}
|
||||||
fi
|
fi
|
||||||
|
|
||||||
# If feature gates includes AllAlpha or EndpointSlice, and EndpointSlice has not been disabled, add EndpointSlice controller to list of controllers to run.
|
# If feature gates includes AllAlpha or EndpointSlice, and EndpointSlice has not been disabled, add EndpointSlice controller to list of controllers to run.
|
||||||
if [[ (( "${KUBE_FEATURE_GATES:-}" == *"AllAlpha=true"* ) || ( "${KUBE_FEATURE_GATES:-}" == *"EndpointSlice=true"* )) && "${KUBE_FEATURE_GATES:-}" != *"EndpointSlice=false"* ]]; then
|
if [[ (( "${KUBE_FEATURE_GATES:-}" = *"AllAlpha=true"* ) || ( "${KUBE_FEATURE_GATES:-}" = *"EndpointSlice=true"* )) && "${KUBE_FEATURE_GATES:-}" != *"EndpointSlice=false"* ]]; then
|
||||||
RUN_CONTROLLERS="${RUN_CONTROLLERS:-*,endpointslice}"
|
RUN_CONTROLLERS=${RUN_CONTROLLERS:-*,endpointslice}
|
||||||
fi
|
fi
|
||||||
|
|
||||||
# Optional: set feature gates
|
# Optional: set feature gates
|
||||||
FEATURE_GATES="${KUBE_FEATURE_GATES:-}"
|
FEATURE_GATES=${KUBE_FEATURE_GATES:-}
|
||||||
|
|
||||||
TERMINATED_POD_GC_THRESHOLD=${TERMINATED_POD_GC_THRESHOLD:-100}
|
TERMINATED_POD_GC_THRESHOLD=${TERMINATED_POD_GC_THRESHOLD:-100}
|
||||||
|
|
||||||
# Extra docker options for nodes.
|
# Extra docker options for nodes.
|
||||||
EXTRA_DOCKER_OPTS="${EXTRA_DOCKER_OPTS:-}"
|
EXTRA_DOCKER_OPTS=${EXTRA_DOCKER_OPTS:-}
|
||||||
|
|
||||||
# Enable the docker debug mode.
|
# Enable the docker debug mode.
|
||||||
EXTRA_DOCKER_OPTS="${EXTRA_DOCKER_OPTS} --debug"
|
EXTRA_DOCKER_OPTS="${EXTRA_DOCKER_OPTS} --debug"
|
||||||
|
|
||||||
SERVICE_CLUSTER_IP_RANGE="10.0.0.0/16" # formerly PORTAL_NET
|
export SERVICE_CLUSTER_IP_RANGE='10.0.0.0/16' # formerly PORTAL_NET
|
||||||
|
|
||||||
# When set to true, Docker Cache is enabled by default as part of the cluster bring up.
|
# When set to true, Docker Cache is enabled by default as part of the cluster bring up.
|
||||||
ENABLE_DOCKER_REGISTRY_CACHE=true
|
export ENABLE_DOCKER_REGISTRY_CACHE=true
|
||||||
|
|
||||||
# Optional: Deploy a L7 loadbalancer controller to fulfill Ingress requests:
|
# Optional: Deploy a L7 loadbalancer controller to fulfill Ingress requests:
|
||||||
# glbc - CE L7 Load Balancer Controller
|
# glbc - CE L7 Load Balancer Controller
|
||||||
ENABLE_L7_LOADBALANCING="${KUBE_ENABLE_L7_LOADBALANCING:-glbc}"
|
export ENABLE_L7_LOADBALANCING=${KUBE_ENABLE_L7_LOADBALANCING:-glbc}
|
||||||
|
|
||||||
# Optional: Enable Metrics Server. Metrics Server should be enable everywhere,
|
# Optional: Enable Metrics Server. Metrics Server should be enable everywhere,
|
||||||
# since it's a critical component, but in the first release we need a way to disable
|
# since it's a critical component, but in the first release we need a way to disable
|
||||||
# this in case of stability issues.
|
# this in case of stability issues.
|
||||||
# TODO(piosz) remove this option once Metrics Server became a stable thing.
|
# TODO(piosz) remove this option once Metrics Server became a stable thing.
|
||||||
ENABLE_METRICS_SERVER="${KUBE_ENABLE_METRICS_SERVER:-true}"
|
export ENABLE_METRICS_SERVER=${KUBE_ENABLE_METRICS_SERVER:-true}
|
||||||
|
|
||||||
# Optional: Metadata agent to setup as part of the cluster bring up:
|
# Optional: Metadata agent to setup as part of the cluster bring up:
|
||||||
# none - No metadata agent
|
# none - No metadata agent
|
||||||
# stackdriver - Stackdriver metadata agent
|
# stackdriver - Stackdriver metadata agent
|
||||||
# Metadata agent is a daemon set that provides metadata of kubernetes objects
|
# Metadata agent is a daemon set that provides metadata of kubernetes objects
|
||||||
# running on the same node for exporting metrics and logs.
|
# running on the same node for exporting metrics and logs.
|
||||||
ENABLE_METADATA_AGENT="${KUBE_ENABLE_METADATA_AGENT:-none}"
|
export ENABLE_METADATA_AGENT=${KUBE_ENABLE_METADATA_AGENT:-none}
|
||||||
|
|
||||||
# One special node out of NUM_NODES would be created of this type if specified.
|
# One special node out of NUM_NODES would be created of this type if specified.
|
||||||
# Useful for scheduling heapster in large clusters with nodes of small size.
|
# Useful for scheduling heapster in large clusters with nodes of small size.
|
||||||
HEAPSTER_MACHINE_TYPE="${HEAPSTER_MACHINE_TYPE:-}"
|
HEAPSTER_MACHINE_TYPE=${HEAPSTER_MACHINE_TYPE:-}
|
||||||
|
|
||||||
# Optional: Additional nodes would be created if their type and number is specified.
|
# Optional: Additional nodes would be created if their type and number is specified.
|
||||||
# NUM_NODES would be lowered respectively.
|
# NUM_NODES would be lowered respectively.
|
||||||
# Useful for running cluster-level addons that needs more resources than would fit
|
# Useful for running cluster-level addons that needs more resources than would fit
|
||||||
# on small nodes, like network plugins.
|
# on small nodes, like network plugins.
|
||||||
NUM_ADDITIONAL_NODES="${NUM_ADDITIONAL_NODES:-}"
|
NUM_ADDITIONAL_NODES=${NUM_ADDITIONAL_NODES:-}
|
||||||
ADDITIONAL_MACHINE_TYPE="${ADDITIONAL_MACHINE_TYPE:-}"
|
ADDITIONAL_MACHINE_TYPE=${ADDITIONAL_MACHINE_TYPE:-}
|
||||||
|
|
||||||
# Set etcd image (e.g. k8s.gcr.io/etcd) and version (e.g. 3.4.7-0) if you need
|
# Set etcd image (e.g. k8s.gcr.io/etcd) and version (e.g. 3.4.7-0) if you need
|
||||||
# non-default version.
|
# non-default version.
|
||||||
ETCD_IMAGE="${TEST_ETCD_IMAGE:-}"
|
export ETCD_IMAGE=${TEST_ETCD_IMAGE:-}
|
||||||
ETCD_DOCKER_REPOSITORY="${TEST_ETCD_DOCKER_REPOSITORY:-}"
|
export ETCD_DOCKER_REPOSITORY=${TEST_ETCD_DOCKER_REPOSITORY:-}
|
||||||
ETCD_VERSION="${TEST_ETCD_VERSION:-}"
|
export ETCD_VERSION=${TEST_ETCD_VERSION:-}
|
||||||
|
|
||||||
# Default Log level for all components in test clusters and variables to override it in specific components.
|
# Default Log level for all components in test clusters and variables to override it in specific components.
|
||||||
TEST_CLUSTER_LOG_LEVEL="${TEST_CLUSTER_LOG_LEVEL:---v=4}"
|
TEST_CLUSTER_LOG_LEVEL=${TEST_CLUSTER_LOG_LEVEL:---v=4}
|
||||||
KUBELET_TEST_LOG_LEVEL="${KUBELET_TEST_LOG_LEVEL:-$TEST_CLUSTER_LOG_LEVEL}"
|
KUBELET_TEST_LOG_LEVEL=${KUBELET_TEST_LOG_LEVEL:-$TEST_CLUSTER_LOG_LEVEL}
|
||||||
DOCKER_TEST_LOG_LEVEL="${DOCKER_TEST_LOG_LEVEL:---log-level=info}"
|
DOCKER_TEST_LOG_LEVEL=${DOCKER_TEST_LOG_LEVEL:---log-level=info}
|
||||||
API_SERVER_TEST_LOG_LEVEL="${API_SERVER_TEST_LOG_LEVEL:-$TEST_CLUSTER_LOG_LEVEL}"
|
API_SERVER_TEST_LOG_LEVEL=${API_SERVER_TEST_LOG_LEVEL:-$TEST_CLUSTER_LOG_LEVEL}
|
||||||
CONTROLLER_MANAGER_TEST_LOG_LEVEL="${CONTROLLER_MANAGER_TEST_LOG_LEVEL:-$TEST_CLUSTER_LOG_LEVEL}"
|
CONTROLLER_MANAGER_TEST_LOG_LEVEL=${CONTROLLER_MANAGER_TEST_LOG_LEVEL:-$TEST_CLUSTER_LOG_LEVEL}
|
||||||
SCHEDULER_TEST_LOG_LEVEL="${SCHEDULER_TEST_LOG_LEVEL:-$TEST_CLUSTER_LOG_LEVEL}"
|
SCHEDULER_TEST_LOG_LEVEL=${SCHEDULER_TEST_LOG_LEVEL:-$TEST_CLUSTER_LOG_LEVEL}
|
||||||
KUBEPROXY_TEST_LOG_LEVEL="${KUBEPROXY_TEST_LOG_LEVEL:-$TEST_CLUSTER_LOG_LEVEL}"
|
KUBEPROXY_TEST_LOG_LEVEL=${KUBEPROXY_TEST_LOG_LEVEL:-$TEST_CLUSTER_LOG_LEVEL}
|
||||||
|
|
||||||
VOLUME_PLUGIN_DIR="${VOLUME_PLUGIN_DIR:-/home/kubernetes/flexvolume}"
|
VOLUME_PLUGIN_DIR=${VOLUME_PLUGIN_DIR:-/home/kubernetes/flexvolume}
|
||||||
|
|
||||||
TEST_CLUSTER_DELETE_COLLECTION_WORKERS="${TEST_CLUSTER_DELETE_COLLECTION_WORKERS:---delete-collection-workers=1}"
|
TEST_CLUSTER_DELETE_COLLECTION_WORKERS=${TEST_CLUSTER_DELETE_COLLECTION_WORKERS:---delete-collection-workers=1}
|
||||||
TEST_CLUSTER_MAX_REQUESTS_INFLIGHT="${TEST_CLUSTER_MAX_REQUESTS_INFLIGHT:-}"
|
TEST_CLUSTER_MAX_REQUESTS_INFLIGHT=${TEST_CLUSTER_MAX_REQUESTS_INFLIGHT:-}
|
||||||
TEST_CLUSTER_RESYNC_PERIOD="${TEST_CLUSTER_RESYNC_PERIOD:---min-resync-period=3m}"
|
TEST_CLUSTER_RESYNC_PERIOD=${TEST_CLUSTER_RESYNC_PERIOD:---min-resync-period=3m}
|
||||||
|
|
||||||
# ContentType used by all components to communicate with apiserver.
|
# ContentType used by all components to communicate with apiserver.
|
||||||
TEST_CLUSTER_API_CONTENT_TYPE="${TEST_CLUSTER_API_CONTENT_TYPE:-}"
|
TEST_CLUSTER_API_CONTENT_TYPE=${TEST_CLUSTER_API_CONTENT_TYPE:-}
|
||||||
|
|
||||||
KUBELET_TEST_ARGS="${KUBELET_TEST_ARGS:-} --serialize-image-pulls=false ${TEST_CLUSTER_API_CONTENT_TYPE}"
|
KUBELET_TEST_ARGS="${KUBELET_TEST_ARGS:-} --serialize-image-pulls=false ${TEST_CLUSTER_API_CONTENT_TYPE}"
|
||||||
if [[ "${NODE_OS_DISTRIBUTION}" == "gci" ]] || [[ "${NODE_OS_DISTRIBUTION}" == "ubuntu" ]] || [[ "${NODE_OS_DISTRIBUTION}" == "custom" ]]; then
|
if [[ "${NODE_OS_DISTRIBUTION}" = 'gci' ]] || [[ "${NODE_OS_DISTRIBUTION}" = 'ubuntu' ]] || [[ "${NODE_OS_DISTRIBUTION}" = 'custom' ]]; then
|
||||||
NODE_KUBELET_TEST_ARGS="${NODE_KUBELET_TEST_ARGS:-} --experimental-kernel-memcg-notification=true"
|
NODE_KUBELET_TEST_ARGS="${NODE_KUBELET_TEST_ARGS:-} --experimental-kernel-memcg-notification=true"
|
||||||
fi
|
fi
|
||||||
if [[ "${MASTER_OS_DISTRIBUTION}" == "gci" ]] || [[ "${MASTER_OS_DISTRIBUTION}" == "ubuntu" ]]; then
|
if [[ "${MASTER_OS_DISTRIBUTION}" = 'gci' ]] || [[ "${MASTER_OS_DISTRIBUTION}" = 'ubuntu' ]]; then
|
||||||
MASTER_KUBELET_TEST_ARGS="${MASTER_KUBELET_TEST_ARGS:-} --experimental-kernel-memcg-notification=true"
|
MASTER_KUBELET_TEST_ARGS="${MASTER_KUBELET_TEST_ARGS:-} --experimental-kernel-memcg-notification=true"
|
||||||
fi
|
fi
|
||||||
APISERVER_TEST_ARGS="${APISERVER_TEST_ARGS:-} --runtime-config=extensions/v1beta1,scheduling.k8s.io/v1alpha1,settings.k8s.io/v1alpha1 ${TEST_CLUSTER_DELETE_COLLECTION_WORKERS} ${TEST_CLUSTER_MAX_REQUESTS_INFLIGHT}"
|
APISERVER_TEST_ARGS="${APISERVER_TEST_ARGS:-} --runtime-config=extensions/v1beta1,scheduling.k8s.io/v1alpha1,settings.k8s.io/v1alpha1 ${TEST_CLUSTER_DELETE_COLLECTION_WORKERS} ${TEST_CLUSTER_MAX_REQUESTS_INFLIGHT}"
|
||||||
@ -228,46 +230,46 @@ CONTROLLER_MANAGER_TEST_ARGS="${CONTROLLER_MANAGER_TEST_ARGS:-} ${TEST_CLUSTER_R
|
|||||||
SCHEDULER_TEST_ARGS="${SCHEDULER_TEST_ARGS:-} ${TEST_CLUSTER_API_CONTENT_TYPE}"
|
SCHEDULER_TEST_ARGS="${SCHEDULER_TEST_ARGS:-} ${TEST_CLUSTER_API_CONTENT_TYPE}"
|
||||||
KUBEPROXY_TEST_ARGS="${KUBEPROXY_TEST_ARGS:-} ${TEST_CLUSTER_API_CONTENT_TYPE}"
|
KUBEPROXY_TEST_ARGS="${KUBEPROXY_TEST_ARGS:-} ${TEST_CLUSTER_API_CONTENT_TYPE}"
|
||||||
|
|
||||||
MASTER_NODE_LABELS="${KUBE_MASTER_NODE_LABELS:-}"
|
export MASTER_NODE_LABELS=${KUBE_MASTER_NODE_LABELS:-}
|
||||||
# NON_MASTER_NODE_LABELS are labels will only be applied on non-master nodes.
|
# NON_MASTER_NODE_LABELS are labels will only be applied on non-master nodes.
|
||||||
NON_MASTER_NODE_LABELS="${KUBE_NON_MASTER_NODE_LABELS:-}"
|
NON_MASTER_NODE_LABELS=${KUBE_NON_MASTER_NODE_LABELS:-}
|
||||||
WINDOWS_NON_MASTER_NODE_LABELS="${WINDOWS_NON_MASTER_NODE_LABELS:-}"
|
WINDOWS_NON_MASTER_NODE_LABELS=${WINDOWS_NON_MASTER_NODE_LABELS:-}
|
||||||
|
|
||||||
if [[ "${PREEMPTIBLE_MASTER}" == "true" ]]; then
|
if [[ "${PREEMPTIBLE_MASTER}" = 'true' ]]; then
|
||||||
NODE_LABELS="${NODE_LABELS},cloud.google.com/gke-preemptible=true"
|
NODE_LABELS="${NODE_LABELS},cloud.google.com/gke-preemptible=true"
|
||||||
WINDOWS_NODE_LABELS="${WINDOWS_NODE_LABELS},cloud.google.com/gke-preemptible=true"
|
WINDOWS_NODE_LABELS="${WINDOWS_NODE_LABELS},cloud.google.com/gke-preemptible=true"
|
||||||
elif [[ "${PREEMPTIBLE_NODE}" == "true" ]]; then
|
elif [[ "${PREEMPTIBLE_NODE}" = 'true' ]]; then
|
||||||
NON_MASTER_NODE_LABELS="${NON_MASTER_NODE_LABELS},cloud.google.com/gke-preemptible=true"
|
NON_MASTER_NODE_LABELS="${NON_MASTER_NODE_LABELS},cloud.google.com/gke-preemptible=true"
|
||||||
WINDOWS_NON_MASTER_NODE_LABELS="${WINDOWS_NON_MASTER_NODE_LABELS},cloud.google.com/gke-preemptible=true"
|
WINDOWS_NON_MASTER_NODE_LABELS="${WINDOWS_NON_MASTER_NODE_LABELS},cloud.google.com/gke-preemptible=true"
|
||||||
fi
|
fi
|
||||||
|
|
||||||
# Optional: Enable netd.
|
# Optional: Enable netd.
|
||||||
ENABLE_NETD="${KUBE_ENABLE_NETD:-false}"
|
ENABLE_NETD=${KUBE_ENABLE_NETD:-false}
|
||||||
CUSTOM_NETD_YAML="${KUBE_CUSTOM_NETD_YAML:-}"
|
export CUSTOM_NETD_YAML=${KUBE_CUSTOM_NETD_YAML:-}
|
||||||
CUSTOM_CALICO_NODE_DAEMONSET_YAML="${KUBE_CUSTOM_CALICO_NODE_DAEMONSET_YAML:-}"
|
export CUSTOM_CALICO_NODE_DAEMONSET_YAML=${KUBE_CUSTOM_CALICO_NODE_DAEMONSET_YAML:-}
|
||||||
CUSTOM_TYPHA_DEPLOYMENT_YAML="${KUBE_CUSTOM_TYPHA_DEPLOYMENT_YAML:-}"
|
export CUSTOM_TYPHA_DEPLOYMENT_YAML=${KUBE_CUSTOM_TYPHA_DEPLOYMENT_YAML:-}
|
||||||
|
|
||||||
# To avoid running netd on a node that is not configured appropriately,
|
# To avoid running netd on a node that is not configured appropriately,
|
||||||
# label each Node so that the DaemonSet can run the Pods only on ready Nodes.
|
# label each Node so that the DaemonSet can run the Pods only on ready Nodes.
|
||||||
# Windows nodes do not support netd.
|
# Windows nodes do not support netd.
|
||||||
if [[ ${ENABLE_NETD:-} == "true" ]]; then
|
if [[ ${ENABLE_NETD:-} = 'true' ]]; then
|
||||||
NON_MASTER_NODE_LABELS="${NON_MASTER_NODE_LABELS:+${NON_MASTER_NODE_LABELS},}cloud.google.com/gke-netd-ready=true"
|
NON_MASTER_NODE_LABELS="${NON_MASTER_NODE_LABELS:+${NON_MASTER_NODE_LABELS},}cloud.google.com/gke-netd-ready=true"
|
||||||
fi
|
fi
|
||||||
|
|
||||||
ENABLE_NODELOCAL_DNS="${KUBE_ENABLE_NODELOCAL_DNS:-false}"
|
export ENABLE_NODELOCAL_DNS=${KUBE_ENABLE_NODELOCAL_DNS:-false}
|
||||||
|
|
||||||
# To avoid running Calico on a node that is not configured appropriately,
|
# To avoid running Calico on a node that is not configured appropriately,
|
||||||
# label each Node so that the DaemonSet can run the Pods only on ready Nodes.
|
# label each Node so that the DaemonSet can run the Pods only on ready Nodes.
|
||||||
# Windows nodes do not support Calico.
|
# Windows nodes do not support Calico.
|
||||||
if [[ ${NETWORK_POLICY_PROVIDER:-} == "calico" ]]; then
|
if [[ ${NETWORK_POLICY_PROVIDER:-} = 'calico' ]]; then
|
||||||
NON_MASTER_NODE_LABELS="${NON_MASTER_NODE_LABELS:+${NON_MASTER_NODE_LABELS},}projectcalico.org/ds-ready=true"
|
NON_MASTER_NODE_LABELS="${NON_MASTER_NODE_LABELS:+${NON_MASTER_NODE_LABELS},}projectcalico.org/ds-ready=true"
|
||||||
fi
|
fi
|
||||||
|
|
||||||
# Enable metadata concealment by firewalling pod traffic to the metadata server
|
# Enable metadata concealment by firewalling pod traffic to the metadata server
|
||||||
# and run a proxy daemonset on nodes.
|
# and run a proxy daemonset on nodes.
|
||||||
ENABLE_METADATA_CONCEALMENT="${ENABLE_METADATA_CONCEALMENT:-true}" # true, false
|
ENABLE_METADATA_CONCEALMENT=${ENABLE_METADATA_CONCEALMENT:-true} # true, false
|
||||||
METADATA_CONCEALMENT_NO_FIREWALL="${METADATA_CONCEALMENT_NO_FIREWALL:-false}" # true, false
|
METADATA_CONCEALMENT_NO_FIREWALL=${METADATA_CONCEALMENT_NO_FIREWALL:-false} # true, false
|
||||||
if [[ ${ENABLE_METADATA_CONCEALMENT:-} == "true" ]]; then
|
if [[ ${ENABLE_METADATA_CONCEALMENT:-} = 'true' ]]; then
|
||||||
# Put the necessary label on the node so the daemonset gets scheduled.
|
# Put the necessary label on the node so the daemonset gets scheduled.
|
||||||
NODE_LABELS="${NODE_LABELS},cloud.google.com/metadata-proxy-ready=true"
|
NODE_LABELS="${NODE_LABELS},cloud.google.com/metadata-proxy-ready=true"
|
||||||
# Add to the provider custom variables.
|
# Add to the provider custom variables.
|
||||||
@ -275,21 +277,21 @@ if [[ ${ENABLE_METADATA_CONCEALMENT:-} == "true" ]]; then
|
|||||||
fi
|
fi
|
||||||
|
|
||||||
# Optional: Enable node logging.
|
# Optional: Enable node logging.
|
||||||
ENABLE_NODE_LOGGING="${KUBE_ENABLE_NODE_LOGGING:-true}"
|
export ENABLE_NODE_LOGGING=${KUBE_ENABLE_NODE_LOGGING:-true}
|
||||||
LOGGING_DESTINATION="${KUBE_LOGGING_DESTINATION:-gcp}" # options: elasticsearch, gcp
|
export LOGGING_DESTINATION=${KUBE_LOGGING_DESTINATION:-gcp} # options: elasticsearch, gcp
|
||||||
|
|
||||||
# Optional: When set to true, Elasticsearch and Kibana will be setup as part of the cluster bring up.
|
# Optional: When set to true, Elasticsearch and Kibana will be setup as part of the cluster bring up.
|
||||||
ENABLE_CLUSTER_LOGGING="${KUBE_ENABLE_CLUSTER_LOGGING:-true}"
|
export ENABLE_CLUSTER_LOGGING=${KUBE_ENABLE_CLUSTER_LOGGING:-true}
|
||||||
ELASTICSEARCH_LOGGING_REPLICAS=1
|
export ELASTICSEARCH_LOGGING_REPLICAS=1
|
||||||
|
|
||||||
# Optional: Don't require https for registries in our local RFC1918 network
|
# Optional: Don't require https for registries in our local RFC1918 network
|
||||||
if [[ ${KUBE_ENABLE_INSECURE_REGISTRY:-false} == "true" ]]; then
|
if [[ ${KUBE_ENABLE_INSECURE_REGISTRY:-false} = 'true' ]]; then
|
||||||
EXTRA_DOCKER_OPTS="${EXTRA_DOCKER_OPTS} --insecure-registry 10.0.0.0/8"
|
EXTRA_DOCKER_OPTS="${EXTRA_DOCKER_OPTS} --insecure-registry 10.0.0.0/8"
|
||||||
fi
|
fi
|
||||||
|
|
||||||
if [[ ! -z "${NODE_ACCELERATORS}" ]]; then
|
if [[ -n "${NODE_ACCELERATORS}" ]]; then
|
||||||
if [[ -z "${FEATURE_GATES:-}" ]]; then
|
if [[ -z "${FEATURE_GATES:-}" ]]; then
|
||||||
FEATURE_GATES="DevicePlugins=true"
|
FEATURE_GATES='DevicePlugins=true'
|
||||||
else
|
else
|
||||||
FEATURE_GATES="${FEATURE_GATES},DevicePlugins=true"
|
FEATURE_GATES="${FEATURE_GATES},DevicePlugins=true"
|
||||||
fi
|
fi
|
||||||
@ -300,45 +302,46 @@ fi
|
|||||||
|
|
||||||
# Optional: Install cluster DNS.
|
# Optional: Install cluster DNS.
|
||||||
# Set CLUSTER_DNS_CORE_DNS to 'false' to install kube-dns instead of CoreDNS.
|
# Set CLUSTER_DNS_CORE_DNS to 'false' to install kube-dns instead of CoreDNS.
|
||||||
CLUSTER_DNS_CORE_DNS="${CLUSTER_DNS_CORE_DNS:-true}"
|
CLUSTER_DNS_CORE_DNS=${CLUSTER_DNS_CORE_DNS:-true}
|
||||||
ENABLE_CLUSTER_DNS="${KUBE_ENABLE_CLUSTER_DNS:-true}"
|
export ENABLE_CLUSTER_DNS=${KUBE_ENABLE_CLUSTER_DNS:-true}
|
||||||
DNS_SERVER_IP="10.0.0.10"
|
export DNS_SERVER_IP='10.0.0.10'
|
||||||
LOCAL_DNS_IP="${KUBE_LOCAL_DNS_IP:-169.254.20.10}"
|
export LOCAL_DNS_IP=${KUBE_LOCAL_DNS_IP:-169.254.20.10}
|
||||||
DNS_DOMAIN="cluster.local"
|
export DNS_DOMAIN='cluster.local'
|
||||||
DNS_MEMORY_LIMIT="${KUBE_DNS_MEMORY_LIMIT:-170Mi}"
|
export DNS_MEMORY_LIMIT=${KUBE_DNS_MEMORY_LIMIT:-170Mi}
|
||||||
|
|
||||||
# Optional: Enable DNS horizontal autoscaler
|
# Optional: Enable DNS horizontal autoscaler
|
||||||
ENABLE_DNS_HORIZONTAL_AUTOSCALER="${KUBE_ENABLE_DNS_HORIZONTAL_AUTOSCALER:-true}"
|
export ENABLE_DNS_HORIZONTAL_AUTOSCALER=${KUBE_ENABLE_DNS_HORIZONTAL_AUTOSCALER:-true}
|
||||||
|
|
||||||
# Optional: Install Kubernetes UI
|
# Optional: Install Kubernetes UI
|
||||||
ENABLE_CLUSTER_UI="${KUBE_ENABLE_CLUSTER_UI:-true}"
|
export ENABLE_CLUSTER_UI=${KUBE_ENABLE_CLUSTER_UI:-true}
|
||||||
|
|
||||||
# Optional: Install node problem detector.
|
# Optional: Install node problem detector.
|
||||||
# none - Not run node problem detector.
|
# none - Not run node problem detector.
|
||||||
# daemonset - Run node problem detector as daemonset.
|
# daemonset - Run node problem detector as daemonset.
|
||||||
# standalone - Run node problem detector as standalone system daemon.
|
# standalone - Run node problem detector as standalone system daemon.
|
||||||
if [[ "${NODE_OS_DISTRIBUTION}" == "gci" ]]; then
|
if [[ "${NODE_OS_DISTRIBUTION}" = 'gci' ]]; then
|
||||||
# Enable standalone mode by default for gci.
|
# Enable standalone mode by default for gci.
|
||||||
ENABLE_NODE_PROBLEM_DETECTOR="${KUBE_ENABLE_NODE_PROBLEM_DETECTOR:-standalone}"
|
ENABLE_NODE_PROBLEM_DETECTOR=${KUBE_ENABLE_NODE_PROBLEM_DETECTOR:-standalone}
|
||||||
else
|
else
|
||||||
ENABLE_NODE_PROBLEM_DETECTOR="${KUBE_ENABLE_NODE_PROBLEM_DETECTOR:-daemonset}"
|
ENABLE_NODE_PROBLEM_DETECTOR=${KUBE_ENABLE_NODE_PROBLEM_DETECTOR:-daemonset}
|
||||||
fi
|
fi
|
||||||
NODE_PROBLEM_DETECTOR_VERSION="${NODE_PROBLEM_DETECTOR_VERSION:-}"
|
export ENABLE_NODE_PROBLEM_DETECTOR
|
||||||
NODE_PROBLEM_DETECTOR_TAR_HASH="${NODE_PROBLEM_DETECTOR_TAR_HASH:-}"
|
NODE_PROBLEM_DETECTOR_VERSION=${NODE_PROBLEM_DETECTOR_VERSION:-}
|
||||||
NODE_PROBLEM_DETECTOR_RELEASE_PATH="${NODE_PROBLEM_DETECTOR_RELEASE_PATH:-}"
|
NODE_PROBLEM_DETECTOR_TAR_HASH=${NODE_PROBLEM_DETECTOR_TAR_HASH:-}
|
||||||
NODE_PROBLEM_DETECTOR_CUSTOM_FLAGS="${NODE_PROBLEM_DETECTOR_CUSTOM_FLAGS:-}"
|
NODE_PROBLEM_DETECTOR_RELEASE_PATH=${NODE_PROBLEM_DETECTOR_RELEASE_PATH:-}
|
||||||
|
NODE_PROBLEM_DETECTOR_CUSTOM_FLAGS=${NODE_PROBLEM_DETECTOR_CUSTOM_FLAGS:-}
|
||||||
|
|
||||||
CNI_SHA1="${CNI_SHA1:-}"
|
CNI_SHA1=${CNI_SHA1:-}
|
||||||
CNI_TAR_PREFIX="${CNI_TAR_PREFIX:-cni-plugins-linux-amd64-}"
|
CNI_TAR_PREFIX=${CNI_TAR_PREFIX:-cni-plugins-linux-amd64-}
|
||||||
CNI_STORAGE_URL_BASE="${CNI_STORAGE_URL_BASE:-https://storage.googleapis.com/k8s-artifacts-cni/release}"
|
CNI_STORAGE_URL_BASE=${CNI_STORAGE_URL_BASE:-https://storage.googleapis.com/k8s-artifacts-cni/release}
|
||||||
|
|
||||||
# Optional: Create autoscaler for cluster's nodes.
|
# Optional: Create autoscaler for cluster's nodes.
|
||||||
ENABLE_CLUSTER_AUTOSCALER="${KUBE_ENABLE_CLUSTER_AUTOSCALER:-false}"
|
export ENABLE_CLUSTER_AUTOSCALER=${KUBE_ENABLE_CLUSTER_AUTOSCALER:-false}
|
||||||
if [[ "${ENABLE_CLUSTER_AUTOSCALER}" == "true" ]]; then
|
if [[ "${ENABLE_CLUSTER_AUTOSCALER}" = 'true' ]]; then
|
||||||
AUTOSCALER_MIN_NODES="${KUBE_AUTOSCALER_MIN_NODES:-}"
|
export AUTOSCALER_MIN_NODES=${KUBE_AUTOSCALER_MIN_NODES:-}
|
||||||
AUTOSCALER_MAX_NODES="${KUBE_AUTOSCALER_MAX_NODES:-}"
|
export AUTOSCALER_MAX_NODES=${KUBE_AUTOSCALER_MAX_NODES:-}
|
||||||
AUTOSCALER_ENABLE_SCALE_DOWN="${KUBE_AUTOSCALER_ENABLE_SCALE_DOWN:-false}"
|
export AUTOSCALER_ENABLE_SCALE_DOWN=${KUBE_AUTOSCALER_ENABLE_SCALE_DOWN:-false}
|
||||||
AUTOSCALER_EXPANDER_CONFIG="${KUBE_AUTOSCALER_EXPANDER_CONFIG:---expander=price}"
|
export AUTOSCALER_EXPANDER_CONFIG=${KUBE_AUTOSCALER_EXPANDER_CONFIG:---expander=price}
|
||||||
fi
|
fi
|
||||||
|
|
||||||
# Optional: Enable allocation of pod IPs using IP aliases.
|
# Optional: Enable allocation of pod IPs using IP aliases.
|
||||||
@ -349,20 +352,22 @@ fi
|
|||||||
# IP_ALIAS_SUBNETWORK is the subnetwork to allocate from. If empty, a
|
# IP_ALIAS_SUBNETWORK is the subnetwork to allocate from. If empty, a
|
||||||
# new subnetwork will be created for the cluster.
|
# new subnetwork will be created for the cluster.
|
||||||
ENABLE_IP_ALIASES=${KUBE_GCE_ENABLE_IP_ALIASES:-false}
|
ENABLE_IP_ALIASES=${KUBE_GCE_ENABLE_IP_ALIASES:-false}
|
||||||
NODE_IPAM_MODE=${KUBE_GCE_NODE_IPAM_MODE:-RangeAllocator}
|
export NODE_IPAM_MODE=${KUBE_GCE_NODE_IPAM_MODE:-RangeAllocator}
|
||||||
if [ ${ENABLE_IP_ALIASES} = true ]; then
|
if [ "${ENABLE_IP_ALIASES}" = true ]; then
|
||||||
# Number of Pods that can run on this node.
|
# Number of Pods that can run on this node.
|
||||||
MAX_PODS_PER_NODE=${MAX_PODS_PER_NODE:-110}
|
MAX_PODS_PER_NODE=${MAX_PODS_PER_NODE:-110}
|
||||||
# Size of ranges allocated to each node.
|
# Size of ranges allocated to each node.
|
||||||
IP_ALIAS_SIZE="/$(get-alias-range-size ${MAX_PODS_PER_NODE})"
|
IP_ALIAS_SIZE="/$(get-alias-range-size "${MAX_PODS_PER_NODE}")"
|
||||||
IP_ALIAS_SUBNETWORK=${KUBE_GCE_IP_ALIAS_SUBNETWORK:-${INSTANCE_PREFIX}-subnet-default}
|
IP_ALIAS_SUBNETWORK=${KUBE_GCE_IP_ALIAS_SUBNETWORK:-${INSTANCE_PREFIX}-subnet-default}
|
||||||
# If we're using custom network, use the subnet we already create for it as the one for ip-alias.
|
# If we're using custom network, use the subnet we already create for it as the one for ip-alias.
|
||||||
# Note that this means SUBNETWORK would override KUBE_GCE_IP_ALIAS_SUBNETWORK in case of custom network.
|
# Note that this means SUBNETWORK would override KUBE_GCE_IP_ALIAS_SUBNETWORK in case of custom network.
|
||||||
if [[ "${CREATE_CUSTOM_NETWORK}" == true ]]; then
|
if [[ "${CREATE_CUSTOM_NETWORK}" = true ]]; then
|
||||||
IP_ALIAS_SUBNETWORK="${SUBNETWORK}"
|
IP_ALIAS_SUBNETWORK=${SUBNETWORK}
|
||||||
fi
|
fi
|
||||||
|
export IP_ALIAS_SIZE
|
||||||
|
export IP_ALIAS_SUBNETWORK
|
||||||
# Reserve the services IP space to avoid being allocated for other GCP resources.
|
# Reserve the services IP space to avoid being allocated for other GCP resources.
|
||||||
SERVICE_CLUSTER_IP_SUBNETWORK=${KUBE_GCE_SERVICE_CLUSTER_IP_SUBNETWORK:-${INSTANCE_PREFIX}-subnet-services}
|
export SERVICE_CLUSTER_IP_SUBNETWORK=${KUBE_GCE_SERVICE_CLUSTER_IP_SUBNETWORK:-${INSTANCE_PREFIX}-subnet-services}
|
||||||
NODE_IPAM_MODE=${KUBE_GCE_NODE_IPAM_MODE:-CloudAllocator}
|
NODE_IPAM_MODE=${KUBE_GCE_NODE_IPAM_MODE:-CloudAllocator}
|
||||||
SECONDARY_RANGE_NAME=${SECONDARY_RANGE_NAME:-}
|
SECONDARY_RANGE_NAME=${SECONDARY_RANGE_NAME:-}
|
||||||
# Add to the provider custom variables.
|
# Add to the provider custom variables.
|
||||||
@ -371,7 +376,7 @@ if [ ${ENABLE_IP_ALIASES} = true ]; then
|
|||||||
PROVIDER_VARS="${PROVIDER_VARS:-} SECONDARY_RANGE_NAME"
|
PROVIDER_VARS="${PROVIDER_VARS:-} SECONDARY_RANGE_NAME"
|
||||||
elif [[ -n "${MAX_PODS_PER_NODE:-}" ]]; then
|
elif [[ -n "${MAX_PODS_PER_NODE:-}" ]]; then
|
||||||
# Should not have MAX_PODS_PER_NODE set for route-based clusters.
|
# Should not have MAX_PODS_PER_NODE set for route-based clusters.
|
||||||
echo -e "${color_red}Cannot set MAX_PODS_PER_NODE for route-based projects for ${PROJECT}." >&2
|
echo -e "${color_red:-}Cannot set MAX_PODS_PER_NODE for route-based projects for ${PROJECT}." >&2
|
||||||
exit 1
|
exit 1
|
||||||
fi
|
fi
|
||||||
|
|
||||||
@ -389,11 +394,11 @@ fi
|
|||||||
if [[ -n "${GCE_GLBC_IMAGE:-}" ]]; then
|
if [[ -n "${GCE_GLBC_IMAGE:-}" ]]; then
|
||||||
PROVIDER_VARS="${PROVIDER_VARS:-} GCE_GLBC_IMAGE"
|
PROVIDER_VARS="${PROVIDER_VARS:-} GCE_GLBC_IMAGE"
|
||||||
fi
|
fi
|
||||||
CUSTOM_INGRESS_YAML="${CUSTOM_INGRESS_YAML:-}"
|
CUSTOM_INGRESS_YAML=${CUSTOM_INGRESS_YAML:-}
|
||||||
|
|
||||||
if [[ -z "${KUBE_ADMISSION_CONTROL:-}" ]]; then
|
if [[ -z "${KUBE_ADMISSION_CONTROL:-}" ]]; then
|
||||||
ADMISSION_CONTROL="NamespaceLifecycle,LimitRanger,ServiceAccount,PersistentVolumeLabel,PodPreset,DefaultStorageClass,DefaultTolerationSeconds,NodeRestriction,Priority,StorageObjectInUseProtection,PersistentVolumeClaimResize,RuntimeClass"
|
ADMISSION_CONTROL='NamespaceLifecycle,LimitRanger,ServiceAccount,PersistentVolumeLabel,PodPreset,DefaultStorageClass,DefaultTolerationSeconds,NodeRestriction,Priority,StorageObjectInUseProtection,PersistentVolumeClaimResize,RuntimeClass'
|
||||||
if [[ "${ENABLE_POD_SECURITY_POLICY:-}" == "true" ]]; then
|
if [[ "${ENABLE_POD_SECURITY_POLICY:-}" = 'true' ]]; then
|
||||||
ADMISSION_CONTROL="${ADMISSION_CONTROL},PodSecurityPolicy"
|
ADMISSION_CONTROL="${ADMISSION_CONTROL},PodSecurityPolicy"
|
||||||
fi
|
fi
|
||||||
# ResourceQuota must come last, or a creation is recorded, but the pod may be forbidden.
|
# ResourceQuota must come last, or a creation is recorded, but the pod may be forbidden.
|
||||||
@ -402,7 +407,7 @@ else
|
|||||||
ADMISSION_CONTROL=${KUBE_ADMISSION_CONTROL}
|
ADMISSION_CONTROL=${KUBE_ADMISSION_CONTROL}
|
||||||
fi
|
fi
|
||||||
|
|
||||||
ENABLE_APISERVER_DYNAMIC_AUDIT="${ENABLE_APISERVER_DYNAMIC_AUDIT:-false}"
|
ENABLE_APISERVER_DYNAMIC_AUDIT=${ENABLE_APISERVER_DYNAMIC_AUDIT:-false}
|
||||||
|
|
||||||
# Optional: if set to true kube-up will automatically check for existing resources and clean them up.
|
# Optional: if set to true kube-up will automatically check for existing resources and clean them up.
|
||||||
KUBE_UP_AUTOMATIC_CLEANUP=${KUBE_UP_AUTOMATIC_CLEANUP:-false}
|
KUBE_UP_AUTOMATIC_CLEANUP=${KUBE_UP_AUTOMATIC_CLEANUP:-false}
|
||||||
@ -411,48 +416,48 @@ KUBE_UP_AUTOMATIC_CLEANUP=${KUBE_UP_AUTOMATIC_CLEANUP:-false}
|
|||||||
# so that we can use pulled kubernetes binaries, even if binaries
|
# so that we can use pulled kubernetes binaries, even if binaries
|
||||||
# are pre-installed in the image. Note that currently this logic
|
# are pre-installed in the image. Note that currently this logic
|
||||||
# is only supported in trusty or GCI.
|
# is only supported in trusty or GCI.
|
||||||
TEST_CLUSTER="${TEST_CLUSTER:-true}"
|
TEST_CLUSTER=${TEST_CLUSTER:-true}
|
||||||
|
|
||||||
# Storage backend. 'etcd2' and 'etcd3' are supported.
|
# Storage backend. 'etcd2' and 'etcd3' are supported.
|
||||||
STORAGE_BACKEND=${STORAGE_BACKEND:-}
|
STORAGE_BACKEND=${STORAGE_BACKEND:-}
|
||||||
# Storage media type: application/json and application/vnd.kubernetes.protobuf are supported.
|
# Storage media type: application/json and application/vnd.kubernetes.protobuf are supported.
|
||||||
STORAGE_MEDIA_TYPE=${STORAGE_MEDIA_TYPE:-}
|
STORAGE_MEDIA_TYPE=${STORAGE_MEDIA_TYPE:-}
|
||||||
|
|
||||||
NETWORK_PROVIDER="${NETWORK_PROVIDER:-kubenet}" # none, kubenet
|
NETWORK_PROVIDER=${NETWORK_PROVIDER:-kubenet} # none, kubenet
|
||||||
|
|
||||||
# Network Policy plugin specific settings.
|
# Network Policy plugin specific settings.
|
||||||
NETWORK_POLICY_PROVIDER="${NETWORK_POLICY_PROVIDER:-none}" # calico
|
NETWORK_POLICY_PROVIDER=${NETWORK_POLICY_PROVIDER:-none} # calico
|
||||||
|
|
||||||
NON_MASQUERADE_CIDR="0.0.0.0/0"
|
export NON_MASQUERADE_CIDR='0.0.0.0/0'
|
||||||
|
|
||||||
# How should the kubelet configure hairpin mode?
|
# How should the kubelet configure hairpin mode?
|
||||||
HAIRPIN_MODE="${HAIRPIN_MODE:-hairpin-veth}" # promiscuous-bridge, hairpin-veth, none
|
HAIRPIN_MODE=${HAIRPIN_MODE:-hairpin-veth} # promiscuous-bridge, hairpin-veth, none
|
||||||
|
|
||||||
# Optional: if set to true, kube-up will configure the cluster to run e2e tests.
|
# Optional: if set to true, kube-up will configure the cluster to run e2e tests.
|
||||||
E2E_STORAGE_TEST_ENVIRONMENT=${KUBE_E2E_STORAGE_TEST_ENVIRONMENT:-false}
|
export E2E_STORAGE_TEST_ENVIRONMENT=${KUBE_E2E_STORAGE_TEST_ENVIRONMENT:-false}
|
||||||
|
|
||||||
# Evict pods whenever compute resource availability on the nodes gets below a threshold.
|
# Evict pods whenever compute resource availability on the nodes gets below a threshold.
|
||||||
EVICTION_HARD="${EVICTION_HARD:-memory.available<250Mi,nodefs.available<10%,nodefs.inodesFree<5%}"
|
EVICTION_HARD=${EVICTION_HARD:-memory.available<250Mi,nodefs.available<10%,nodefs.inodesFree<5%}
|
||||||
|
|
||||||
# Optional: custom scheduling algorithm
|
# Optional: custom scheduling algorithm
|
||||||
SCHEDULING_ALGORITHM_PROVIDER="${SCHEDULING_ALGORITHM_PROVIDER:-}"
|
SCHEDULING_ALGORITHM_PROVIDER=${SCHEDULING_ALGORITHM_PROVIDER:-}
|
||||||
|
|
||||||
# Optional: install a default StorageClass
|
# Optional: install a default StorageClass
|
||||||
ENABLE_DEFAULT_STORAGE_CLASS="${ENABLE_DEFAULT_STORAGE_CLASS:-true}"
|
ENABLE_DEFAULT_STORAGE_CLASS=${ENABLE_DEFAULT_STORAGE_CLASS:-true}
|
||||||
|
|
||||||
# Optional: install volume snapshot CRDs
|
# Optional: install volume snapshot CRDs
|
||||||
ENABLE_VOLUME_SNAPSHOTS="${ENABLE_VOLUME_SNAPSHOTS:-true}"
|
ENABLE_VOLUME_SNAPSHOTS=${ENABLE_VOLUME_SNAPSHOTS:-true}
|
||||||
|
|
||||||
# Optional: Enable legacy ABAC policy that makes all service accounts superusers.
|
# Optional: Enable legacy ABAC policy that makes all service accounts superusers.
|
||||||
# Disabling this by default in tests ensures default RBAC policies are sufficient from 1.6+
|
# Disabling this by default in tests ensures default RBAC policies are sufficient from 1.6+
|
||||||
# Upgrade test jobs that go from a version < 1.6 to a version >= 1.6 should override this to be true.
|
# Upgrade test jobs that go from a version < 1.6 to a version >= 1.6 should override this to be true.
|
||||||
ENABLE_LEGACY_ABAC="${ENABLE_LEGACY_ABAC:-false}" # true, false
|
ENABLE_LEGACY_ABAC=${ENABLE_LEGACY_ABAC:-false} # true, false
|
||||||
|
|
||||||
# Enable a simple "AdvancedAuditing" setup for testing.
|
# Enable a simple "AdvancedAuditing" setup for testing.
|
||||||
ENABLE_APISERVER_ADVANCED_AUDIT="${ENABLE_APISERVER_ADVANCED_AUDIT:-true}" # true, false
|
ENABLE_APISERVER_ADVANCED_AUDIT=${ENABLE_APISERVER_ADVANCED_AUDIT:-true} # true, false
|
||||||
ADVANCED_AUDIT_LOG_MODE="${ADVANCED_AUDIT_LOG_MODE:-batch}" # batch, blocking
|
ADVANCED_AUDIT_LOG_MODE=${ADVANCED_AUDIT_LOG_MODE:-batch} # batch, blocking
|
||||||
|
|
||||||
ENABLE_BIG_CLUSTER_SUBNETS="${ENABLE_BIG_CLUSTER_SUBNETS:-false}"
|
ENABLE_BIG_CLUSTER_SUBNETS=${ENABLE_BIG_CLUSTER_SUBNETS:-false}
|
||||||
|
|
||||||
if [[ -n "${LOGROTATE_FILES_MAX_COUNT:-}" ]]; then
|
if [[ -n "${LOGROTATE_FILES_MAX_COUNT:-}" ]]; then
|
||||||
PROVIDER_VARS="${PROVIDER_VARS:-} LOGROTATE_FILES_MAX_COUNT"
|
PROVIDER_VARS="${PROVIDER_VARS:-} LOGROTATE_FILES_MAX_COUNT"
|
||||||
@ -471,59 +476,59 @@ fi
|
|||||||
|
|
||||||
# Fluentd requirements
|
# Fluentd requirements
|
||||||
# YAML exists to trigger a configuration refresh when changes are made.
|
# YAML exists to trigger a configuration refresh when changes are made.
|
||||||
FLUENTD_GCP_YAML_VERSION="v3.2.0"
|
export FLUENTD_GCP_YAML_VERSION='v3.2.0'
|
||||||
FLUENTD_GCP_VERSION="${FLUENTD_GCP_VERSION:-1.6.17}"
|
FLUENTD_GCP_VERSION=${FLUENTD_GCP_VERSION:-1.6.17}
|
||||||
FLUENTD_GCP_MEMORY_LIMIT="${FLUENTD_GCP_MEMORY_LIMIT:-}"
|
FLUENTD_GCP_MEMORY_LIMIT=${FLUENTD_GCP_MEMORY_LIMIT:-}
|
||||||
FLUENTD_GCP_CPU_REQUEST="${FLUENTD_GCP_CPU_REQUEST:-}"
|
FLUENTD_GCP_CPU_REQUEST=${FLUENTD_GCP_CPU_REQUEST:-}
|
||||||
FLUENTD_GCP_MEMORY_REQUEST="${FLUENTD_GCP_MEMORY_REQUEST:-}"
|
FLUENTD_GCP_MEMORY_REQUEST=${FLUENTD_GCP_MEMORY_REQUEST:-}
|
||||||
|
|
||||||
# Heapster requirements
|
# Heapster requirements
|
||||||
HEAPSTER_GCP_BASE_MEMORY="${HEAPSTER_GCP_BASE_MEMORY:-140Mi}"
|
HEAPSTER_GCP_BASE_MEMORY=${HEAPSTER_GCP_BASE_MEMORY:-140Mi}
|
||||||
HEAPSTER_GCP_MEMORY_PER_NODE="${HEAPSTER_GCP_MEMORY_PER_NODE:-4}"
|
HEAPSTER_GCP_MEMORY_PER_NODE=${HEAPSTER_GCP_MEMORY_PER_NODE:-4}
|
||||||
HEAPSTER_GCP_BASE_CPU="${HEAPSTER_GCP_BASE_CPU:-80m}"
|
HEAPSTER_GCP_BASE_CPU=${HEAPSTER_GCP_BASE_CPU:-80m}
|
||||||
HEAPSTER_GCP_CPU_PER_NODE="${HEAPSTER_GCP_CPU_PER_NODE:-0.5}"
|
HEAPSTER_GCP_CPU_PER_NODE=${HEAPSTER_GCP_CPU_PER_NODE:-0.5}
|
||||||
|
|
||||||
# Optional: custom system banner for dashboard addon
|
# Optional: custom system banner for dashboard addon
|
||||||
CUSTOM_KUBE_DASHBOARD_BANNER="${CUSTOM_KUBE_DASHBOARD_BANNER:-}"
|
CUSTOM_KUBE_DASHBOARD_BANNER=${CUSTOM_KUBE_DASHBOARD_BANNER:-}
|
||||||
|
|
||||||
# Default Stackdriver resources version exported by Fluentd-gcp addon
|
# Default Stackdriver resources version exported by Fluentd-gcp addon
|
||||||
LOGGING_STACKDRIVER_RESOURCE_TYPES="${LOGGING_STACKDRIVER_RESOURCE_TYPES:-old}"
|
LOGGING_STACKDRIVER_RESOURCE_TYPES=${LOGGING_STACKDRIVER_RESOURCE_TYPES:-old}
|
||||||
|
|
||||||
# Adding to PROVIDER_VARS, since this is GCP-specific.
|
# Adding to PROVIDER_VARS, since this is GCP-specific.
|
||||||
PROVIDER_VARS="${PROVIDER_VARS:-} FLUENTD_GCP_YAML_VERSION FLUENTD_GCP_VERSION FLUENTD_GCP_MEMORY_LIMIT FLUENTD_GCP_CPU_REQUEST FLUENTD_GCP_MEMORY_REQUEST HEAPSTER_GCP_BASE_MEMORY HEAPSTER_GCP_MEMORY_PER_NODE HEAPSTER_GCP_BASE_CPU HEAPSTER_GCP_CPU_PER_NODE CUSTOM_KUBE_DASHBOARD_BANNER LOGGING_STACKDRIVER_RESOURCE_TYPES"
|
PROVIDER_VARS="${PROVIDER_VARS:-} FLUENTD_GCP_YAML_VERSION FLUENTD_GCP_VERSION FLUENTD_GCP_MEMORY_LIMIT FLUENTD_GCP_CPU_REQUEST FLUENTD_GCP_MEMORY_REQUEST HEAPSTER_GCP_BASE_MEMORY HEAPSTER_GCP_MEMORY_PER_NODE HEAPSTER_GCP_BASE_CPU HEAPSTER_GCP_CPU_PER_NODE CUSTOM_KUBE_DASHBOARD_BANNER LOGGING_STACKDRIVER_RESOURCE_TYPES"
|
||||||
|
|
||||||
# Fluentd configuration for node-journal
|
# Fluentd configuration for node-journal
|
||||||
ENABLE_NODE_JOURNAL="${ENABLE_NODE_JOURNAL:-false}"
|
ENABLE_NODE_JOURNAL=${ENABLE_NODE_JOURNAL:-false}
|
||||||
|
|
||||||
# prometheus-to-sd configuration
|
# prometheus-to-sd configuration
|
||||||
PROMETHEUS_TO_SD_ENDPOINT="${PROMETHEUS_TO_SD_ENDPOINT:-https://monitoring.googleapis.com/}"
|
PROMETHEUS_TO_SD_ENDPOINT=${PROMETHEUS_TO_SD_ENDPOINT:-https://monitoring.googleapis.com/}
|
||||||
PROMETHEUS_TO_SD_PREFIX="${PROMETHEUS_TO_SD_PREFIX:-custom.googleapis.com}"
|
PROMETHEUS_TO_SD_PREFIX=${PROMETHEUS_TO_SD_PREFIX:-custom.googleapis.com}
|
||||||
ENABLE_PROMETHEUS_TO_SD="${ENABLE_PROMETHEUS_TO_SD:-true}"
|
ENABLE_PROMETHEUS_TO_SD=${ENABLE_PROMETHEUS_TO_SD:-true}
|
||||||
|
|
||||||
# TODO(#51292): Make kube-proxy Daemonset default and remove the configuration here.
|
# TODO(#51292): Make kube-proxy Daemonset default and remove the configuration here.
|
||||||
# Optional: [Experiment Only] Run kube-proxy as a DaemonSet if set to true, run as static pods otherwise.
|
# Optional: [Experiment Only] Run kube-proxy as a DaemonSet if set to true, run as static pods otherwise.
|
||||||
KUBE_PROXY_DAEMONSET="${KUBE_PROXY_DAEMONSET:-false}" # true, false
|
KUBE_PROXY_DAEMONSET=${KUBE_PROXY_DAEMONSET:-false} # true, false
|
||||||
|
|
||||||
# Optional: Change the kube-proxy implementation. Choices are [iptables, ipvs].
|
# Optional: Change the kube-proxy implementation. Choices are [iptables, ipvs].
|
||||||
KUBE_PROXY_MODE="${KUBE_PROXY_MODE:-iptables}"
|
KUBE_PROXY_MODE=${KUBE_PROXY_MODE:-iptables}
|
||||||
|
|
||||||
# Optional: duration of cluster signed certificates.
|
# Optional: duration of cluster signed certificates.
|
||||||
CLUSTER_SIGNING_DURATION="${CLUSTER_SIGNING_DURATION:-}"
|
CLUSTER_SIGNING_DURATION=${CLUSTER_SIGNING_DURATION:-}
|
||||||
|
|
||||||
# Optional: enable certificate rotation of the kubelet certificates.
|
# Optional: enable certificate rotation of the kubelet certificates.
|
||||||
ROTATE_CERTIFICATES="${ROTATE_CERTIFICATES:-}"
|
ROTATE_CERTIFICATES=${ROTATE_CERTIFICATES:-}
|
||||||
|
|
||||||
# The number of services that are allowed to sync concurrently. Will be passed
|
# The number of services that are allowed to sync concurrently. Will be passed
|
||||||
# into kube-controller-manager via `--concurrent-service-syncs`
|
# into kube-controller-manager via `--concurrent-service-syncs`
|
||||||
CONCURRENT_SERVICE_SYNCS="${CONCURRENT_SERVICE_SYNCS:-}"
|
CONCURRENT_SERVICE_SYNCS=${CONCURRENT_SERVICE_SYNCS:-}
|
||||||
|
|
||||||
# The value kubernetes.default.svc is only usable in Pods and should only be
|
# The value kubernetes.default.svc is only usable in Pods and should only be
|
||||||
# set for tests. DO NOT COPY THIS VALUE FOR PRODUCTION CLUSTERS.
|
# set for tests. DO NOT COPY THIS VALUE FOR PRODUCTION CLUSTERS.
|
||||||
SERVICEACCOUNT_ISSUER="https://kubernetes.default.svc"
|
export SERVICEACCOUNT_ISSUER='https://kubernetes.default.svc'
|
||||||
|
|
||||||
# Optional: Enable Node termination Handler for Preemptible and GPU VMs.
|
# Optional: Enable Node termination Handler for Preemptible and GPU VMs.
|
||||||
# https://github.com/GoogleCloudPlatform/k8s-node-termination-handler
|
# https://github.com/GoogleCloudPlatform/k8s-node-termination-handler
|
||||||
ENABLE_NODE_TERMINATION_HANDLER="${ENABLE_NODE_TERMINATION_HANDLER:-false}"
|
ENABLE_NODE_TERMINATION_HANDLER=${ENABLE_NODE_TERMINATION_HANDLER:-false}
|
||||||
# Override default Node Termination Handler Image
|
# Override default Node Termination Handler Image
|
||||||
if [[ "${NODE_TERMINATION_HANDLER_IMAGE:-}" ]]; then
|
if [[ "${NODE_TERMINATION_HANDLER_IMAGE:-}" ]]; then
|
||||||
PROVIDER_VARS="${PROVIDER_VARS:-} NODE_TERMINATION_HANDLER_IMAGE"
|
PROVIDER_VARS="${PROVIDER_VARS:-} NODE_TERMINATION_HANDLER_IMAGE"
|
||||||
@ -531,12 +536,12 @@ fi
|
|||||||
|
|
||||||
# Taint Windows nodes by default to prevent Linux workloads from being
|
# Taint Windows nodes by default to prevent Linux workloads from being
|
||||||
# scheduled onto them.
|
# scheduled onto them.
|
||||||
WINDOWS_NODE_TAINTS="${WINDOWS_NODE_TAINTS:-node.kubernetes.io/os=win1809:NoSchedule}"
|
WINDOWS_NODE_TAINTS=${WINDOWS_NODE_TAINTS:-node.kubernetes.io/os=win1809:NoSchedule}
|
||||||
|
|
||||||
# Whether to set up a private GCE cluster, i.e. a cluster where nodes have only private IPs.
|
# Whether to set up a private GCE cluster, i.e. a cluster where nodes have only private IPs.
|
||||||
GCE_PRIVATE_CLUSTER="${KUBE_GCE_PRIVATE_CLUSTER:-false}"
|
export GCE_PRIVATE_CLUSTER=${KUBE_GCE_PRIVATE_CLUSTER:-false}
|
||||||
GCE_PRIVATE_CLUSTER_PORTS_PER_VM="${KUBE_GCE_PRIVATE_CLUSTER_PORTS_PER_VM:-}"
|
export GCE_PRIVATE_CLUSTER_PORTS_PER_VM=${KUBE_GCE_PRIVATE_CLUSTER_PORTS_PER_VM:-}
|
||||||
|
|
||||||
ETCD_LISTEN_CLIENT_IP=0.0.0.0
|
export ETCD_LISTEN_CLIENT_IP=0.0.0.0
|
||||||
|
|
||||||
GCE_UPLOAD_KUBCONFIG_TO_MASTER_METADATA=true
|
export GCE_UPLOAD_KUBCONFIG_TO_MASTER_METADATA=true
|
||||||
|
@ -1,5 +1,4 @@
|
|||||||
./cluster/gce/config-default.sh
|
./cluster/gce/config-default.sh
|
||||||
./cluster/gce/config-test.sh
|
|
||||||
./cluster/gce/gci/configure-helper.sh
|
./cluster/gce/gci/configure-helper.sh
|
||||||
./cluster/gce/gci/configure.sh
|
./cluster/gce/gci/configure.sh
|
||||||
./cluster/gce/gci/master-helper.sh
|
./cluster/gce/gci/master-helper.sh
|
||||||
|
Loading…
Reference in New Issue
Block a user