mirror of
				https://github.com/k3s-io/kubernetes.git
				synced 2025-11-03 23:40:03 +00:00 
			
		
		
		
	
		
			
				
	
	
		
			1231 lines
		
	
	
		
			47 KiB
		
	
	
	
		
			Bash
		
	
	
		
			Executable File
		
	
	
	
	
			
		
		
	
	
			1231 lines
		
	
	
		
			47 KiB
		
	
	
	
		
			Bash
		
	
	
		
			Executable File
		
	
	
	
	
#!/usr/bin/env bash
 | 
						|
 | 
						|
# Copyright 2014 The Kubernetes Authors.
 | 
						|
#
 | 
						|
# Licensed under the Apache License, Version 2.0 (the "License");
 | 
						|
# you may not use this file except in compliance with the License.
 | 
						|
# You may obtain a copy of the License at
 | 
						|
#
 | 
						|
#     http://www.apache.org/licenses/LICENSE-2.0
 | 
						|
#
 | 
						|
# Unless required by applicable law or agreed to in writing, software
 | 
						|
# distributed under the License is distributed on an "AS IS" BASIS,
 | 
						|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 | 
						|
# See the License for the specific language governing permissions and
 | 
						|
# limitations under the License.
 | 
						|
 | 
						|
KUBE_ROOT=$(dirname "${BASH_SOURCE[0]}")/..
 | 
						|
 | 
						|
# This script builds and runs a local kubernetes cluster. You may need to run
 | 
						|
# this as root to allow kubelet to open docker's socket, and to write the test
 | 
						|
# CA in /var/run/kubernetes.
 | 
						|
# Usage: `hack/local-up-cluster.sh`.
 | 
						|
 | 
						|
DOCKER_OPTS=${DOCKER_OPTS:-""}
 | 
						|
export DOCKER=(docker "${DOCKER_OPTS[@]}")
 | 
						|
DOCKER_ROOT=${DOCKER_ROOT:-""}
 | 
						|
ALLOW_PRIVILEGED=${ALLOW_PRIVILEGED:-""}
 | 
						|
DENY_SECURITY_CONTEXT_ADMISSION=${DENY_SECURITY_CONTEXT_ADMISSION:-""}
 | 
						|
PSP_ADMISSION=${PSP_ADMISSION:-""}
 | 
						|
RUNTIME_CONFIG=${RUNTIME_CONFIG:-""}
 | 
						|
KUBELET_AUTHORIZATION_WEBHOOK=${KUBELET_AUTHORIZATION_WEBHOOK:-""}
 | 
						|
KUBELET_AUTHENTICATION_WEBHOOK=${KUBELET_AUTHENTICATION_WEBHOOK:-""}
 | 
						|
POD_MANIFEST_PATH=${POD_MANIFEST_PATH:-"/var/run/kubernetes/static-pods"}
 | 
						|
KUBELET_FLAGS=${KUBELET_FLAGS:-""}
 | 
						|
KUBELET_IMAGE=${KUBELET_IMAGE:-""}
 | 
						|
# many dev environments run with swap on, so we don't fail in this env
 | 
						|
FAIL_SWAP_ON=${FAIL_SWAP_ON:-"false"}
 | 
						|
# Name of the dns addon, eg: "kube-dns" or "coredns"
 | 
						|
DNS_ADDON=${DNS_ADDON:-"coredns"}
 | 
						|
CLUSTER_CIDR=${CLUSTER_CIDR:-10.1.0.0/16}
 | 
						|
SERVICE_CLUSTER_IP_RANGE=${SERVICE_CLUSTER_IP_RANGE:-10.0.0.0/24}
 | 
						|
FIRST_SERVICE_CLUSTER_IP=${FIRST_SERVICE_CLUSTER_IP:-10.0.0.1}
 | 
						|
# if enabled, must set CGROUP_ROOT
 | 
						|
CGROUPS_PER_QOS=${CGROUPS_PER_QOS:-true}
 | 
						|
# name of the cgroup driver, i.e. cgroupfs or systemd
 | 
						|
CGROUP_DRIVER=${CGROUP_DRIVER:-""}
 | 
						|
# if cgroups per qos is enabled, optionally change cgroup root
 | 
						|
CGROUP_ROOT=${CGROUP_ROOT:-""}
 | 
						|
# owner of client certs, default to current user if not specified
 | 
						|
USER=${USER:-$(whoami)}
 | 
						|
 | 
						|
# required for cni installation
 | 
						|
CNI_CONFIG_DIR=${CNI_CONFIG_DIR:-/etc/cni/net.d}
 | 
						|
CNI_PLUGINS_VERSION=${CNI_PLUGINS_VERSION:-"v1.0.1"}
 | 
						|
CNI_TARGETARCH=${CNI_TARGETARCH:-amd64}
 | 
						|
CNI_PLUGINS_TARBALL="${CNI_PLUGINS_VERSION}/cni-plugins-linux-${CNI_TARGETARCH}-${CNI_PLUGINS_VERSION}.tgz"
 | 
						|
CNI_PLUGINS_URL="https://github.com/containernetworking/plugins/releases/download/${CNI_PLUGINS_TARBALL}"
 | 
						|
CNI_PLUGINS_AMD64_SHA256SUM=${CNI_PLUGINS_AMD64_SHA256SUM:-"5238fbb2767cbf6aae736ad97a7aa29167525dcd405196dfbc064672a730d3cf"}
 | 
						|
CNI_PLUGINS_ARM64_SHA256SUM=${CNI_PLUGINS_ARM64_SHA256SUM:-"2d4528c45bdd0a8875f849a75082bc4eafe95cb61f9bcc10a6db38a031f67226"}
 | 
						|
CNI_PLUGINS_PPC64LE_SHA256SUM=${CNI_PLUGINS_PPC64LE_SHA256SUM:-"f078e33067e6daaef3a3a5010d6440f2464b7973dec3ca0b5d5be22fdcb1fd96"}
 | 
						|
CNI_PLUGINS_S390X_SHA256SUM=${CNI_PLUGINS_S390X_SHA256SUM:-"468d33e16440d9ca4395c6bb2d5b71b35ae4a4df26301e4da85ac70c5ce56822"}
 | 
						|
 | 
						|
# enables testing eviction scenarios locally.
 | 
						|
EVICTION_HARD=${EVICTION_HARD:-"memory.available<100Mi,nodefs.available<10%,nodefs.inodesFree<5%"}
 | 
						|
EVICTION_SOFT=${EVICTION_SOFT:-""}
 | 
						|
EVICTION_PRESSURE_TRANSITION_PERIOD=${EVICTION_PRESSURE_TRANSITION_PERIOD:-"1m"}
 | 
						|
 | 
						|
# This script uses docker0 (or whatever container bridge docker is currently using)
 | 
						|
# and we don't know the IP of the DNS pod to pass in as --cluster-dns.
 | 
						|
# To set this up by hand, set this flag and change DNS_SERVER_IP.
 | 
						|
# Note also that you need API_HOST (defined below) for correct DNS.
 | 
						|
KUBE_PROXY_MODE=${KUBE_PROXY_MODE:-""}
 | 
						|
ENABLE_CLUSTER_DNS=${KUBE_ENABLE_CLUSTER_DNS:-true}
 | 
						|
ENABLE_NODELOCAL_DNS=${KUBE_ENABLE_NODELOCAL_DNS:-false}
 | 
						|
DNS_SERVER_IP=${KUBE_DNS_SERVER_IP:-10.0.0.10}
 | 
						|
LOCAL_DNS_IP=${KUBE_LOCAL_DNS_IP:-169.254.20.10}
 | 
						|
DNS_MEMORY_LIMIT=${KUBE_DNS_MEMORY_LIMIT:-170Mi}
 | 
						|
DNS_DOMAIN=${KUBE_DNS_NAME:-"cluster.local"}
 | 
						|
KUBECTL=${KUBECTL:-"${KUBE_ROOT}/cluster/kubectl.sh"}
 | 
						|
WAIT_FOR_URL_API_SERVER=${WAIT_FOR_URL_API_SERVER:-60}
 | 
						|
MAX_TIME_FOR_URL_API_SERVER=${MAX_TIME_FOR_URL_API_SERVER:-1}
 | 
						|
ENABLE_DAEMON=${ENABLE_DAEMON:-false}
 | 
						|
HOSTNAME_OVERRIDE=${HOSTNAME_OVERRIDE:-"127.0.0.1"}
 | 
						|
EXTERNAL_CLOUD_PROVIDER=${EXTERNAL_CLOUD_PROVIDER:-false}
 | 
						|
EXTERNAL_CLOUD_PROVIDER_BINARY=${EXTERNAL_CLOUD_PROVIDER_BINARY:-""}
 | 
						|
EXTERNAL_CLOUD_VOLUME_PLUGIN=${EXTERNAL_CLOUD_VOLUME_PLUGIN:-""}
 | 
						|
CONFIGURE_CLOUD_ROUTES=${CONFIGURE_CLOUD_ROUTES:-true}
 | 
						|
CLOUD_CTLRMGR_FLAGS=${CLOUD_CTLRMGR_FLAGS:-""}
 | 
						|
CLOUD_PROVIDER=${CLOUD_PROVIDER:-""}
 | 
						|
CLOUD_CONFIG=${CLOUD_CONFIG:-""}
 | 
						|
KUBELET_PROVIDER_ID=${KUBELET_PROVIDER_ID:-"$(hostname)"}
 | 
						|
FEATURE_GATES=${FEATURE_GATES:-"AllAlpha=false"}
 | 
						|
STORAGE_BACKEND=${STORAGE_BACKEND:-"etcd3"}
 | 
						|
STORAGE_MEDIA_TYPE=${STORAGE_MEDIA_TYPE:-"application/vnd.kubernetes.protobuf"}
 | 
						|
# preserve etcd data. you also need to set ETCD_DIR.
 | 
						|
PRESERVE_ETCD="${PRESERVE_ETCD:-false}"
 | 
						|
 | 
						|
# enable Kubernetes-CSI snapshotter
 | 
						|
ENABLE_CSI_SNAPSHOTTER=${ENABLE_CSI_SNAPSHOTTER:-false}
 | 
						|
 | 
						|
# RBAC Mode options
 | 
						|
AUTHORIZATION_MODE=${AUTHORIZATION_MODE:-"Node,RBAC"}
 | 
						|
KUBECONFIG_TOKEN=${KUBECONFIG_TOKEN:-""}
 | 
						|
AUTH_ARGS=${AUTH_ARGS:-""}
 | 
						|
 | 
						|
# WebHook Authentication and Authorization
 | 
						|
AUTHORIZATION_WEBHOOK_CONFIG_FILE=${AUTHORIZATION_WEBHOOK_CONFIG_FILE:-""}
 | 
						|
AUTHENTICATION_WEBHOOK_CONFIG_FILE=${AUTHENTICATION_WEBHOOK_CONFIG_FILE:-""}
 | 
						|
 | 
						|
# Install a default storage class (enabled by default)
 | 
						|
DEFAULT_STORAGE_CLASS=${KUBE_DEFAULT_STORAGE_CLASS:-true}
 | 
						|
 | 
						|
# Do not run the mutation detector by default on a local cluster.
 | 
						|
# It is intended for a specific type of testing and inherently leaks memory.
 | 
						|
KUBE_CACHE_MUTATION_DETECTOR="${KUBE_CACHE_MUTATION_DETECTOR:-false}"
 | 
						|
export KUBE_CACHE_MUTATION_DETECTOR
 | 
						|
 | 
						|
# panic the server on watch decode errors since they are considered coder mistakes
 | 
						|
KUBE_PANIC_WATCH_DECODE_ERROR="${KUBE_PANIC_WATCH_DECODE_ERROR:-true}"
 | 
						|
export KUBE_PANIC_WATCH_DECODE_ERROR
 | 
						|
 | 
						|
# Default list of admission Controllers to invoke prior to persisting objects in cluster
 | 
						|
# The order defined here does not matter.
 | 
						|
ENABLE_ADMISSION_PLUGINS=${ENABLE_ADMISSION_PLUGINS:-"NamespaceLifecycle,LimitRanger,ServiceAccount,DefaultStorageClass,DefaultTolerationSeconds,Priority,MutatingAdmissionWebhook,ValidatingAdmissionWebhook,ResourceQuota,NodeRestriction"}
 | 
						|
DISABLE_ADMISSION_PLUGINS=${DISABLE_ADMISSION_PLUGINS:-""}
 | 
						|
ADMISSION_CONTROL_CONFIG_FILE=${ADMISSION_CONTROL_CONFIG_FILE:-""}
 | 
						|
 | 
						|
# START_MODE can be 'all', 'kubeletonly', 'nokubelet', 'nokubeproxy', or 'nokubelet,nokubeproxy'
 | 
						|
START_MODE=${START_MODE:-"all"}
 | 
						|
 | 
						|
# A list of controllers to enable
 | 
						|
KUBE_CONTROLLERS="${KUBE_CONTROLLERS:-"*"}"
 | 
						|
 | 
						|
# Audit policy
 | 
						|
AUDIT_POLICY_FILE=${AUDIT_POLICY_FILE:-""}
 | 
						|
 | 
						|
# sanity check for OpenStack provider
 | 
						|
if [ "${CLOUD_PROVIDER}" == "openstack" ]; then
 | 
						|
    if [ "${CLOUD_CONFIG}" == "" ]; then
 | 
						|
        echo "Missing CLOUD_CONFIG env for OpenStack provider!"
 | 
						|
        exit 1
 | 
						|
    fi
 | 
						|
    if [ ! -f "${CLOUD_CONFIG}" ]; then
 | 
						|
        echo "Cloud config ${CLOUD_CONFIG} doesn't exist"
 | 
						|
        exit 1
 | 
						|
    fi
 | 
						|
fi
 | 
						|
 | 
						|
# Stop right away if the build fails
 | 
						|
set -e
 | 
						|
 | 
						|
source "${KUBE_ROOT}/hack/lib/init.sh"
 | 
						|
kube::util::ensure-gnu-sed
 | 
						|
 | 
						|
function usage {
 | 
						|
            echo "This script starts a local kube cluster. "
 | 
						|
            echo "Example 0: hack/local-up-cluster.sh -h  (this 'help' usage description)"
 | 
						|
            echo "Example 1: hack/local-up-cluster.sh -o _output/dockerized/bin/linux/amd64/ (run from docker output)"
 | 
						|
            echo "Example 2: hack/local-up-cluster.sh -O (auto-guess the bin path for your platform)"
 | 
						|
            echo "Example 3: hack/local-up-cluster.sh (build a local copy of the source)"
 | 
						|
}
 | 
						|
 | 
						|
# This function guesses where the existing cached binary build is for the `-O`
 | 
						|
# flag
 | 
						|
function guess_built_binary_path {
 | 
						|
  local apiserver_path
 | 
						|
  apiserver_path=$(kube::util::find-binary "kube-apiserver")
 | 
						|
  if [[ -z "${apiserver_path}" ]]; then
 | 
						|
    return
 | 
						|
  fi
 | 
						|
  echo -n "$(dirname "${apiserver_path}")"
 | 
						|
}
 | 
						|
 | 
						|
### Allow user to supply the source directory.
 | 
						|
GO_OUT=${GO_OUT:-}
 | 
						|
while getopts "ho:O" OPTION
 | 
						|
do
 | 
						|
    case ${OPTION} in
 | 
						|
        o)
 | 
						|
            echo "skipping build"
 | 
						|
            GO_OUT="${OPTARG}"
 | 
						|
            echo "using source ${GO_OUT}"
 | 
						|
            ;;
 | 
						|
        O)
 | 
						|
            GO_OUT=$(guess_built_binary_path)
 | 
						|
            if [ "${GO_OUT}" == "" ]; then
 | 
						|
                echo "Could not guess the correct output directory to use."
 | 
						|
                exit 1
 | 
						|
            fi
 | 
						|
            ;;
 | 
						|
        h)
 | 
						|
            usage
 | 
						|
            exit
 | 
						|
            ;;
 | 
						|
        ?)
 | 
						|
            usage
 | 
						|
            exit
 | 
						|
            ;;
 | 
						|
    esac
 | 
						|
done
 | 
						|
 | 
						|
if [ "x${GO_OUT}" == "x" ]; then
 | 
						|
    make -C "${KUBE_ROOT}" WHAT="cmd/kubectl cmd/kube-apiserver cmd/kube-controller-manager cmd/cloud-controller-manager cmd/kubelet cmd/kube-proxy cmd/kube-scheduler"
 | 
						|
else
 | 
						|
    echo "skipped the build."
 | 
						|
fi
 | 
						|
 | 
						|
# Shut down anyway if there's an error.
 | 
						|
set +e
 | 
						|
 | 
						|
API_PORT=${API_PORT:-0}
 | 
						|
API_SECURE_PORT=${API_SECURE_PORT:-6443}
 | 
						|
 | 
						|
# WARNING: For DNS to work on most setups you should export API_HOST as the docker0 ip address,
 | 
						|
API_HOST=${API_HOST:-localhost}
 | 
						|
API_HOST_IP=${API_HOST_IP:-"127.0.0.1"}
 | 
						|
ADVERTISE_ADDRESS=${ADVERTISE_ADDRESS:-""}
 | 
						|
NODE_PORT_RANGE=${NODE_PORT_RANGE:-""}
 | 
						|
API_BIND_ADDR=${API_BIND_ADDR:-"0.0.0.0"}
 | 
						|
EXTERNAL_HOSTNAME=${EXTERNAL_HOSTNAME:-localhost}
 | 
						|
 | 
						|
KUBELET_HOST=${KUBELET_HOST:-"127.0.0.1"}
 | 
						|
KUBELET_RESOLV_CONF=${KUBELET_RESOLV_CONF:-"/etc/resolv.conf"}
 | 
						|
# By default only allow CORS for requests on localhost
 | 
						|
API_CORS_ALLOWED_ORIGINS=${API_CORS_ALLOWED_ORIGINS:-/127.0.0.1(:[0-9]+)?$,/localhost(:[0-9]+)?$}
 | 
						|
KUBELET_PORT=${KUBELET_PORT:-10250}
 | 
						|
# By default we use 0(close it) for it's insecure
 | 
						|
KUBELET_READ_ONLY_PORT=${KUBELET_READ_ONLY_PORT:-0}
 | 
						|
LOG_LEVEL=${LOG_LEVEL:-3}
 | 
						|
# Use to increase verbosity on particular files, e.g. LOG_SPEC=token_controller*=5,other_controller*=4
 | 
						|
LOG_SPEC=${LOG_SPEC:-""}
 | 
						|
LOG_DIR=${LOG_DIR:-"/tmp"}
 | 
						|
CONTAINER_RUNTIME=${CONTAINER_RUNTIME:-"remote"}
 | 
						|
CONTAINER_RUNTIME_ENDPOINT=${CONTAINER_RUNTIME_ENDPOINT:-"unix:///run/containerd/containerd.sock"}
 | 
						|
RUNTIME_REQUEST_TIMEOUT=${RUNTIME_REQUEST_TIMEOUT:-"2m"}
 | 
						|
IMAGE_SERVICE_ENDPOINT=${IMAGE_SERVICE_ENDPOINT:-""}
 | 
						|
CPU_CFS_QUOTA=${CPU_CFS_QUOTA:-true}
 | 
						|
ENABLE_HOSTPATH_PROVISIONER=${ENABLE_HOSTPATH_PROVISIONER:-"false"}
 | 
						|
CLAIM_BINDER_SYNC_PERIOD=${CLAIM_BINDER_SYNC_PERIOD:-"15s"} # current k8s default
 | 
						|
ENABLE_CONTROLLER_ATTACH_DETACH=${ENABLE_CONTROLLER_ATTACH_DETACH:-"true"} # current default
 | 
						|
# This is the default dir and filename where the apiserver will generate a self-signed cert
 | 
						|
# which should be able to be used as the CA to verify itself
 | 
						|
CERT_DIR=${CERT_DIR:-"/var/run/kubernetes"}
 | 
						|
ROOT_CA_FILE=${CERT_DIR}/server-ca.crt
 | 
						|
CLUSTER_SIGNING_CERT_FILE=${CLUSTER_SIGNING_CERT_FILE:-"${CERT_DIR}/client-ca.crt"}
 | 
						|
CLUSTER_SIGNING_KEY_FILE=${CLUSTER_SIGNING_KEY_FILE:-"${CERT_DIR}/client-ca.key"}
 | 
						|
# Reuse certs will skip generate new ca/cert files under CERT_DIR
 | 
						|
# it's useful with PRESERVE_ETCD=true because new ca will make existed service account secrets invalided
 | 
						|
REUSE_CERTS=${REUSE_CERTS:-false}
 | 
						|
 | 
						|
 | 
						|
# Ensure CERT_DIR is created for auto-generated crt/key and kubeconfig
 | 
						|
mkdir -p "${CERT_DIR}" &>/dev/null || sudo mkdir -p "${CERT_DIR}"
 | 
						|
CONTROLPLANE_SUDO=$(test -w "${CERT_DIR}" || echo "sudo -E")
 | 
						|
 | 
						|
function test_apiserver_off {
 | 
						|
    # For the common local scenario, fail fast if server is already running.
 | 
						|
    # this can happen if you run local-up-cluster.sh twice and kill etcd in between.
 | 
						|
    if [[ "${API_PORT}" -gt "0" ]]; then
 | 
						|
        if ! curl --silent -g "${API_HOST}:${API_PORT}" ; then
 | 
						|
            echo "API SERVER insecure port is free, proceeding..."
 | 
						|
        else
 | 
						|
            echo "ERROR starting API SERVER, exiting. Some process on ${API_HOST} is serving already on ${API_PORT}"
 | 
						|
            exit 1
 | 
						|
        fi
 | 
						|
    fi
 | 
						|
 | 
						|
    if ! curl --silent -k -g "${API_HOST}:${API_SECURE_PORT}" ; then
 | 
						|
        echo "API SERVER secure port is free, proceeding..."
 | 
						|
    else
 | 
						|
        echo "ERROR starting API SERVER, exiting. Some process on ${API_HOST} is serving already on ${API_SECURE_PORT}"
 | 
						|
        exit 1
 | 
						|
    fi
 | 
						|
}
 | 
						|
 | 
						|
function detect_binary {
 | 
						|
    # Detect the OS name/arch so that we can find our binary
 | 
						|
    case "$(uname -s)" in
 | 
						|
      Darwin)
 | 
						|
        host_os=darwin
 | 
						|
        ;;
 | 
						|
      Linux)
 | 
						|
        host_os=linux
 | 
						|
        ;;
 | 
						|
      *)
 | 
						|
        echo "Unsupported host OS.  Must be Linux or Mac OS X." >&2
 | 
						|
        exit 1
 | 
						|
        ;;
 | 
						|
    esac
 | 
						|
 | 
						|
    case "$(uname -m)" in
 | 
						|
      x86_64*)
 | 
						|
        host_arch=amd64
 | 
						|
        ;;
 | 
						|
      i?86_64*)
 | 
						|
        host_arch=amd64
 | 
						|
        ;;
 | 
						|
      amd64*)
 | 
						|
        host_arch=amd64
 | 
						|
        ;;
 | 
						|
      aarch64*)
 | 
						|
        host_arch=arm64
 | 
						|
        ;;
 | 
						|
      arm64*)
 | 
						|
        host_arch=arm64
 | 
						|
        ;;
 | 
						|
      arm*)
 | 
						|
        host_arch=arm
 | 
						|
        ;;
 | 
						|
      i?86*)
 | 
						|
        host_arch=x86
 | 
						|
        ;;
 | 
						|
      s390x*)
 | 
						|
        host_arch=s390x
 | 
						|
        ;;
 | 
						|
      ppc64le*)
 | 
						|
        host_arch=ppc64le
 | 
						|
        ;;
 | 
						|
      *)
 | 
						|
        echo "Unsupported host arch. Must be x86_64, 386, arm, arm64, s390x or ppc64le." >&2
 | 
						|
        exit 1
 | 
						|
        ;;
 | 
						|
    esac
 | 
						|
 | 
						|
   GO_OUT="${KUBE_ROOT}/_output/local/bin/${host_os}/${host_arch}"
 | 
						|
}
 | 
						|
 | 
						|
cleanup()
 | 
						|
{
 | 
						|
  echo "Cleaning up..."
 | 
						|
  # delete running images
 | 
						|
  # if [[ "${ENABLE_CLUSTER_DNS}" == true ]]; then
 | 
						|
  # Still need to figure why this commands throw an error: Error from server: client: etcd cluster is unavailable or misconfigured
 | 
						|
  #     ${KUBECTL} --namespace=kube-system delete service kube-dns
 | 
						|
  # And this one hang forever:
 | 
						|
  #     ${KUBECTL} --namespace=kube-system delete rc kube-dns-v10
 | 
						|
  # fi
 | 
						|
 | 
						|
  # Check if the API server is still running
 | 
						|
  [[ -n "${APISERVER_PID-}" ]] && kube::util::read-array APISERVER_PIDS < <(pgrep -P "${APISERVER_PID}" ; ps -o pid= -p "${APISERVER_PID}")
 | 
						|
  [[ -n "${APISERVER_PIDS-}" ]] && sudo kill "${APISERVER_PIDS[@]}" 2>/dev/null
 | 
						|
 | 
						|
  # Check if the controller-manager is still running
 | 
						|
  [[ -n "${CTLRMGR_PID-}" ]] && kube::util::read-array CTLRMGR_PIDS < <(pgrep -P "${CTLRMGR_PID}" ; ps -o pid= -p "${CTLRMGR_PID}")
 | 
						|
  [[ -n "${CTLRMGR_PIDS-}" ]] && sudo kill "${CTLRMGR_PIDS[@]}" 2>/dev/null
 | 
						|
 | 
						|
  # Check if the cloud-controller-manager is still running
 | 
						|
  [[ -n "${CLOUD_CTLRMGR_PID-}" ]] && kube::util::read-array CLOUD_CTLRMGR_PIDS < <(pgrep -P "${CLOUD_CTLRMGR_PID}" ; ps -o pid= -p "${CLOUD_CTLRMGR_PID}")
 | 
						|
  [[ -n "${CLOUD_CTLRMGR_PIDS-}" ]] && sudo kill "${CLOUD_CTLRMGR_PIDS[@]}" 2>/dev/null
 | 
						|
 | 
						|
  # Check if the kubelet is still running
 | 
						|
  [[ -n "${KUBELET_PID-}" ]] && kube::util::read-array KUBELET_PIDS < <(pgrep -P "${KUBELET_PID}" ; ps -o pid= -p "${KUBELET_PID}")
 | 
						|
  [[ -n "${KUBELET_PIDS-}" ]] && sudo kill "${KUBELET_PIDS[@]}" 2>/dev/null
 | 
						|
 | 
						|
  # Check if the proxy is still running
 | 
						|
  [[ -n "${PROXY_PID-}" ]] && kube::util::read-array PROXY_PIDS < <(pgrep -P "${PROXY_PID}" ; ps -o pid= -p "${PROXY_PID}")
 | 
						|
  [[ -n "${PROXY_PIDS-}" ]] && sudo kill "${PROXY_PIDS[@]}" 2>/dev/null
 | 
						|
 | 
						|
  # Check if the scheduler is still running
 | 
						|
  [[ -n "${SCHEDULER_PID-}" ]] && kube::util::read-array SCHEDULER_PIDS < <(pgrep -P "${SCHEDULER_PID}" ; ps -o pid= -p "${SCHEDULER_PID}")
 | 
						|
  [[ -n "${SCHEDULER_PIDS-}" ]] && sudo kill "${SCHEDULER_PIDS[@]}" 2>/dev/null
 | 
						|
 | 
						|
  # Check if the etcd is still running
 | 
						|
  [[ -n "${ETCD_PID-}" ]] && kube::etcd::stop
 | 
						|
  if [[ "${PRESERVE_ETCD}" == "false" ]]; then
 | 
						|
    [[ -n "${ETCD_DIR-}" ]] && kube::etcd::clean_etcd_dir
 | 
						|
  fi
 | 
						|
 | 
						|
  exit 0
 | 
						|
}
 | 
						|
 | 
						|
# Check if all processes are still running. Prints a warning once each time
 | 
						|
# a process dies unexpectedly.
 | 
						|
function healthcheck {
 | 
						|
  if [[ -n "${APISERVER_PID-}" ]] && ! sudo kill -0 "${APISERVER_PID}" 2>/dev/null; then
 | 
						|
    warning_log "API server terminated unexpectedly, see ${APISERVER_LOG}"
 | 
						|
    APISERVER_PID=
 | 
						|
  fi
 | 
						|
 | 
						|
  if [[ -n "${CTLRMGR_PID-}" ]] && ! sudo kill -0 "${CTLRMGR_PID}" 2>/dev/null; then
 | 
						|
    warning_log "kube-controller-manager terminated unexpectedly, see ${CTLRMGR_LOG}"
 | 
						|
    CTLRMGR_PID=
 | 
						|
  fi
 | 
						|
 | 
						|
  if [[ -n "${KUBELET_PID-}" ]] && ! sudo kill -0 "${KUBELET_PID}" 2>/dev/null; then
 | 
						|
    warning_log "kubelet terminated unexpectedly, see ${KUBELET_LOG}"
 | 
						|
    KUBELET_PID=
 | 
						|
  fi
 | 
						|
 | 
						|
  if [[ -n "${PROXY_PID-}" ]] && ! sudo kill -0 "${PROXY_PID}" 2>/dev/null; then
 | 
						|
    warning_log "kube-proxy terminated unexpectedly, see ${PROXY_LOG}"
 | 
						|
    PROXY_PID=
 | 
						|
  fi
 | 
						|
 | 
						|
  if [[ -n "${SCHEDULER_PID-}" ]] && ! sudo kill -0 "${SCHEDULER_PID}" 2>/dev/null; then
 | 
						|
    warning_log "scheduler terminated unexpectedly, see ${SCHEDULER_LOG}"
 | 
						|
    SCHEDULER_PID=
 | 
						|
  fi
 | 
						|
 | 
						|
  if [[ -n "${ETCD_PID-}" ]] && ! sudo kill -0 "${ETCD_PID}" 2>/dev/null; then
 | 
						|
    warning_log "etcd terminated unexpectedly"
 | 
						|
    ETCD_PID=
 | 
						|
  fi
 | 
						|
}
 | 
						|
 | 
						|
function print_color {
 | 
						|
  message=$1
 | 
						|
  prefix=${2:+$2: } # add colon only if defined
 | 
						|
  color=${3:-1}     # default is red
 | 
						|
  echo -n "$(tput bold)$(tput setaf "${color}")"
 | 
						|
  echo "${prefix}${message}"
 | 
						|
  echo -n "$(tput sgr0)"
 | 
						|
}
 | 
						|
 | 
						|
function warning_log {
 | 
						|
  print_color "$1" "W$(date "+%m%d %H:%M:%S")]" 1
 | 
						|
}
 | 
						|
 | 
						|
function start_etcd {
 | 
						|
    echo "Starting etcd"
 | 
						|
    export ETCD_LOGFILE=${LOG_DIR}/etcd.log
 | 
						|
    kube::etcd::start
 | 
						|
}
 | 
						|
 | 
						|
function set_service_accounts {
 | 
						|
    SERVICE_ACCOUNT_LOOKUP=${SERVICE_ACCOUNT_LOOKUP:-true}
 | 
						|
    SERVICE_ACCOUNT_KEY=${SERVICE_ACCOUNT_KEY:-/tmp/kube-serviceaccount.key}
 | 
						|
    # Generate ServiceAccount key if needed
 | 
						|
    if [[ ! -f "${SERVICE_ACCOUNT_KEY}" ]]; then
 | 
						|
      mkdir -p "$(dirname "${SERVICE_ACCOUNT_KEY}")"
 | 
						|
      openssl genrsa -out "${SERVICE_ACCOUNT_KEY}" 2048 2>/dev/null
 | 
						|
    fi
 | 
						|
}
 | 
						|
 | 
						|
function generate_certs {
 | 
						|
    # Create CA signers
 | 
						|
    if [[ "${ENABLE_SINGLE_CA_SIGNER:-}" = true ]]; then
 | 
						|
        kube::util::create_signing_certkey "${CONTROLPLANE_SUDO}" "${CERT_DIR}" server '"client auth","server auth"'
 | 
						|
        sudo cp "${CERT_DIR}/server-ca.key" "${CERT_DIR}/client-ca.key"
 | 
						|
        sudo cp "${CERT_DIR}/server-ca.crt" "${CERT_DIR}/client-ca.crt"
 | 
						|
        sudo cp "${CERT_DIR}/server-ca-config.json" "${CERT_DIR}/client-ca-config.json"
 | 
						|
    else
 | 
						|
        kube::util::create_signing_certkey "${CONTROLPLANE_SUDO}" "${CERT_DIR}" server '"server auth"'
 | 
						|
        kube::util::create_signing_certkey "${CONTROLPLANE_SUDO}" "${CERT_DIR}" client '"client auth"'
 | 
						|
    fi
 | 
						|
 | 
						|
    # Create auth proxy client ca
 | 
						|
    kube::util::create_signing_certkey "${CONTROLPLANE_SUDO}" "${CERT_DIR}" request-header '"client auth"'
 | 
						|
 | 
						|
    # serving cert for kube-apiserver
 | 
						|
    kube::util::create_serving_certkey "${CONTROLPLANE_SUDO}" "${CERT_DIR}" "server-ca" kube-apiserver kubernetes.default kubernetes.default.svc "localhost" "${API_HOST_IP}" "${API_HOST}" "${FIRST_SERVICE_CLUSTER_IP}"
 | 
						|
 | 
						|
    # Create client certs signed with client-ca, given id, given CN and a number of groups
 | 
						|
    kube::util::create_client_certkey "${CONTROLPLANE_SUDO}" "${CERT_DIR}" 'client-ca' controller system:kube-controller-manager
 | 
						|
    kube::util::create_client_certkey "${CONTROLPLANE_SUDO}" "${CERT_DIR}" 'client-ca' scheduler  system:kube-scheduler
 | 
						|
    kube::util::create_client_certkey "${CONTROLPLANE_SUDO}" "${CERT_DIR}" 'client-ca' admin system:admin system:masters
 | 
						|
    kube::util::create_client_certkey "${CONTROLPLANE_SUDO}" "${CERT_DIR}" 'client-ca' kube-apiserver kube-apiserver
 | 
						|
 | 
						|
    # Create matching certificates for kube-aggregator
 | 
						|
    kube::util::create_serving_certkey "${CONTROLPLANE_SUDO}" "${CERT_DIR}" "server-ca" kube-aggregator api.kube-public.svc "localhost" "${API_HOST_IP}"
 | 
						|
    kube::util::create_client_certkey "${CONTROLPLANE_SUDO}" "${CERT_DIR}" request-header-ca auth-proxy system:auth-proxy
 | 
						|
 | 
						|
    # TODO remove masters and add rolebinding
 | 
						|
    kube::util::create_client_certkey "${CONTROLPLANE_SUDO}" "${CERT_DIR}" 'client-ca' kube-aggregator system:kube-aggregator system:masters
 | 
						|
    kube::util::write_client_kubeconfig "${CONTROLPLANE_SUDO}" "${CERT_DIR}" "${ROOT_CA_FILE}" "${API_HOST}" "${API_SECURE_PORT}" kube-aggregator
 | 
						|
}
 | 
						|
 | 
						|
function generate_kubeproxy_certs {
 | 
						|
    kube::util::create_client_certkey "${CONTROLPLANE_SUDO}" "${CERT_DIR}" 'client-ca' kube-proxy system:kube-proxy system:nodes
 | 
						|
    kube::util::write_client_kubeconfig "${CONTROLPLANE_SUDO}" "${CERT_DIR}" "${ROOT_CA_FILE}" "${API_HOST}" "${API_SECURE_PORT}" kube-proxy
 | 
						|
}
 | 
						|
 | 
						|
function generate_kubelet_certs {
 | 
						|
    kube::util::create_client_certkey "${CONTROLPLANE_SUDO}" "${CERT_DIR}" 'client-ca' kubelet "system:node:${HOSTNAME_OVERRIDE}" system:nodes
 | 
						|
    kube::util::write_client_kubeconfig "${CONTROLPLANE_SUDO}" "${CERT_DIR}" "${ROOT_CA_FILE}" "${API_HOST}" "${API_SECURE_PORT}" kubelet
 | 
						|
}
 | 
						|
 | 
						|
function start_apiserver {
 | 
						|
    security_admission=""
 | 
						|
    if [[ -n "${DENY_SECURITY_CONTEXT_ADMISSION}" ]]; then
 | 
						|
      security_admission=",SecurityContextDeny"
 | 
						|
    fi
 | 
						|
    if [[ -n "${PSP_ADMISSION}" ]]; then
 | 
						|
      security_admission=",PodSecurityPolicy"
 | 
						|
    fi
 | 
						|
 | 
						|
    # Append security_admission plugin
 | 
						|
    ENABLE_ADMISSION_PLUGINS="${ENABLE_ADMISSION_PLUGINS}${security_admission}"
 | 
						|
 | 
						|
    authorizer_arg=""
 | 
						|
    if [[ -n "${AUTHORIZATION_MODE}" ]]; then
 | 
						|
      authorizer_arg="--authorization-mode=${AUTHORIZATION_MODE}"
 | 
						|
    fi
 | 
						|
    priv_arg=""
 | 
						|
    if [[ -n "${ALLOW_PRIVILEGED}" ]]; then
 | 
						|
      priv_arg="--allow-privileged=${ALLOW_PRIVILEGED}"
 | 
						|
    fi
 | 
						|
 | 
						|
    runtime_config=""
 | 
						|
    if [[ -n "${RUNTIME_CONFIG}" ]]; then
 | 
						|
      runtime_config="--runtime-config=${RUNTIME_CONFIG}"
 | 
						|
    fi
 | 
						|
 | 
						|
    # Let the API server pick a default address when API_HOST_IP
 | 
						|
    # is set to 127.0.0.1
 | 
						|
    advertise_address=""
 | 
						|
    if [[ "${API_HOST_IP}" != "127.0.0.1" ]]; then
 | 
						|
        advertise_address="--advertise-address=${API_HOST_IP}"
 | 
						|
    fi
 | 
						|
    if [[ "${ADVERTISE_ADDRESS}" != "" ]] ; then
 | 
						|
        advertise_address="--advertise-address=${ADVERTISE_ADDRESS}"
 | 
						|
    fi
 | 
						|
    node_port_range=""
 | 
						|
    if [[ "${NODE_PORT_RANGE}" != "" ]] ; then
 | 
						|
        node_port_range="--service-node-port-range=${NODE_PORT_RANGE}"
 | 
						|
    fi
 | 
						|
 | 
						|
    if [[ "${REUSE_CERTS}" != true ]]; then
 | 
						|
      # Create Certs
 | 
						|
      generate_certs
 | 
						|
    fi
 | 
						|
 | 
						|
    cloud_config_arg="--cloud-provider=${CLOUD_PROVIDER} --cloud-config=${CLOUD_CONFIG}"
 | 
						|
    if [[ "${EXTERNAL_CLOUD_PROVIDER:-}" == "true" ]]; then
 | 
						|
      cloud_config_arg="--cloud-provider=external"
 | 
						|
    fi
 | 
						|
 | 
						|
    if [[ -z "${EGRESS_SELECTOR_CONFIG_FILE:-}" ]]; then
 | 
						|
      cat <<EOF > /tmp/kube_egress_selector_configuration.yaml
 | 
						|
apiVersion: apiserver.k8s.io/v1beta1
 | 
						|
kind: EgressSelectorConfiguration
 | 
						|
egressSelections:
 | 
						|
- name: cluster
 | 
						|
  connection:
 | 
						|
    proxyProtocol: Direct
 | 
						|
- name: controlplane
 | 
						|
  connection:
 | 
						|
    proxyProtocol: Direct
 | 
						|
- name: etcd
 | 
						|
  connection:
 | 
						|
    proxyProtocol: Direct
 | 
						|
EOF
 | 
						|
      EGRESS_SELECTOR_CONFIG_FILE="/tmp/kube_egress_selector_configuration.yaml"
 | 
						|
    fi
 | 
						|
 | 
						|
    if [[ -z "${AUDIT_POLICY_FILE}" ]]; then
 | 
						|
      cat <<EOF > /tmp/kube-audit-policy-file
 | 
						|
# Log all requests at the Metadata level.
 | 
						|
apiVersion: audit.k8s.io/v1
 | 
						|
kind: Policy
 | 
						|
rules:
 | 
						|
- level: Metadata
 | 
						|
EOF
 | 
						|
      AUDIT_POLICY_FILE="/tmp/kube-audit-policy-file"
 | 
						|
    fi
 | 
						|
 | 
						|
    APISERVER_LOG=${LOG_DIR}/kube-apiserver.log
 | 
						|
    # shellcheck disable=SC2086
 | 
						|
    ${CONTROLPLANE_SUDO} "${GO_OUT}/kube-apiserver" "${authorizer_arg}" "${priv_arg}" ${runtime_config} \
 | 
						|
      ${cloud_config_arg} \
 | 
						|
      "${advertise_address}" \
 | 
						|
      "${node_port_range}" \
 | 
						|
      --v="${LOG_LEVEL}" \
 | 
						|
      --vmodule="${LOG_SPEC}" \
 | 
						|
      --audit-policy-file="${AUDIT_POLICY_FILE}" \
 | 
						|
      --audit-log-path="${LOG_DIR}/kube-apiserver-audit.log" \
 | 
						|
      --authorization-webhook-config-file="${AUTHORIZATION_WEBHOOK_CONFIG_FILE}" \
 | 
						|
      --authentication-token-webhook-config-file="${AUTHENTICATION_WEBHOOK_CONFIG_FILE}" \
 | 
						|
      --cert-dir="${CERT_DIR}" \
 | 
						|
      --egress-selector-config-file="${EGRESS_SELECTOR_CONFIG_FILE:-}" \
 | 
						|
      --client-ca-file="${CERT_DIR}/client-ca.crt" \
 | 
						|
      --kubelet-client-certificate="${CERT_DIR}/client-kube-apiserver.crt" \
 | 
						|
      --kubelet-client-key="${CERT_DIR}/client-kube-apiserver.key" \
 | 
						|
      --service-account-key-file="${SERVICE_ACCOUNT_KEY}" \
 | 
						|
      --service-account-lookup="${SERVICE_ACCOUNT_LOOKUP}" \
 | 
						|
      --service-account-issuer="https://kubernetes.default.svc" \
 | 
						|
      --service-account-jwks-uri="https://kubernetes.default.svc/openid/v1/jwks" \
 | 
						|
      --service-account-signing-key-file="${SERVICE_ACCOUNT_KEY}" \
 | 
						|
      --enable-admission-plugins="${ENABLE_ADMISSION_PLUGINS}" \
 | 
						|
      --disable-admission-plugins="${DISABLE_ADMISSION_PLUGINS}" \
 | 
						|
      --admission-control-config-file="${ADMISSION_CONTROL_CONFIG_FILE}" \
 | 
						|
      --bind-address="${API_BIND_ADDR}" \
 | 
						|
      --secure-port="${API_SECURE_PORT}" \
 | 
						|
      --tls-cert-file="${CERT_DIR}/serving-kube-apiserver.crt" \
 | 
						|
      --tls-private-key-file="${CERT_DIR}/serving-kube-apiserver.key" \
 | 
						|
      --storage-backend="${STORAGE_BACKEND}" \
 | 
						|
      --storage-media-type="${STORAGE_MEDIA_TYPE}" \
 | 
						|
      --etcd-servers="http://${ETCD_HOST}:${ETCD_PORT}" \
 | 
						|
      --service-cluster-ip-range="${SERVICE_CLUSTER_IP_RANGE}" \
 | 
						|
      --feature-gates="${FEATURE_GATES}" \
 | 
						|
      --external-hostname="${EXTERNAL_HOSTNAME}" \
 | 
						|
      --requestheader-username-headers=X-Remote-User \
 | 
						|
      --requestheader-group-headers=X-Remote-Group \
 | 
						|
      --requestheader-extra-headers-prefix=X-Remote-Extra- \
 | 
						|
      --requestheader-client-ca-file="${CERT_DIR}/request-header-ca.crt" \
 | 
						|
      --requestheader-allowed-names=system:auth-proxy \
 | 
						|
      --proxy-client-cert-file="${CERT_DIR}/client-auth-proxy.crt" \
 | 
						|
      --proxy-client-key-file="${CERT_DIR}/client-auth-proxy.key" \
 | 
						|
      --cors-allowed-origins="${API_CORS_ALLOWED_ORIGINS}" >"${APISERVER_LOG}" 2>&1 &
 | 
						|
    APISERVER_PID=$!
 | 
						|
 | 
						|
    # Wait for kube-apiserver to come up before launching the rest of the components.
 | 
						|
    echo "Waiting for apiserver to come up"
 | 
						|
    kube::util::wait_for_url "https://${API_HOST_IP}:${API_SECURE_PORT}/healthz" "apiserver: " 1 "${WAIT_FOR_URL_API_SERVER}" "${MAX_TIME_FOR_URL_API_SERVER}" \
 | 
						|
        || { echo "check apiserver logs: ${APISERVER_LOG}" ; exit 1 ; }
 | 
						|
 | 
						|
    # Create kubeconfigs for all components, using client certs
 | 
						|
    kube::util::write_client_kubeconfig "${CONTROLPLANE_SUDO}" "${CERT_DIR}" "${ROOT_CA_FILE}" "${API_HOST}" "${API_SECURE_PORT}" admin
 | 
						|
    ${CONTROLPLANE_SUDO} chown "${USER}" "${CERT_DIR}/client-admin.key" # make readable for kubectl
 | 
						|
    kube::util::write_client_kubeconfig "${CONTROLPLANE_SUDO}" "${CERT_DIR}" "${ROOT_CA_FILE}" "${API_HOST}" "${API_SECURE_PORT}" controller
 | 
						|
    kube::util::write_client_kubeconfig "${CONTROLPLANE_SUDO}" "${CERT_DIR}" "${ROOT_CA_FILE}" "${API_HOST}" "${API_SECURE_PORT}" scheduler
 | 
						|
 | 
						|
    if [[ -z "${AUTH_ARGS}" ]]; then
 | 
						|
        AUTH_ARGS="--client-key=${CERT_DIR}/client-admin.key --client-certificate=${CERT_DIR}/client-admin.crt"
 | 
						|
    fi
 | 
						|
    # Grant apiserver permission to speak to the kubelet
 | 
						|
    ${KUBECTL} --kubeconfig "${CERT_DIR}/admin.kubeconfig" create clusterrolebinding kube-apiserver-kubelet-admin --clusterrole=system:kubelet-api-admin --user=kube-apiserver
 | 
						|
 | 
						|
    # Grant kubelets permission to request client certificates
 | 
						|
    ${KUBECTL} --kubeconfig "${CERT_DIR}/admin.kubeconfig" create clusterrolebinding kubelet-csr --clusterrole=system:certificates.k8s.io:certificatesigningrequests:selfnodeclient --group=system:nodes
 | 
						|
 | 
						|
    ${CONTROLPLANE_SUDO} cp "${CERT_DIR}/admin.kubeconfig" "${CERT_DIR}/admin-kube-aggregator.kubeconfig"
 | 
						|
    ${CONTROLPLANE_SUDO} chown -R "$(whoami)" "${CERT_DIR}"
 | 
						|
    ${KUBECTL} config set-cluster local-up-cluster --kubeconfig="${CERT_DIR}/admin-kube-aggregator.kubeconfig" --server="https://${API_HOST_IP}:31090"
 | 
						|
    echo "use 'kubectl --kubeconfig=${CERT_DIR}/admin-kube-aggregator.kubeconfig' to use the aggregated API server"
 | 
						|
 | 
						|
}
 | 
						|
 | 
						|
function start_controller_manager {
 | 
						|
    cloud_config_arg=("--cloud-provider=${CLOUD_PROVIDER}" "--cloud-config=${CLOUD_CONFIG}")
 | 
						|
    cloud_config_arg+=("--configure-cloud-routes=${CONFIGURE_CLOUD_ROUTES}")
 | 
						|
    if [[ "${EXTERNAL_CLOUD_PROVIDER:-}" == "true" ]]; then
 | 
						|
      cloud_config_arg=("--cloud-provider=external")
 | 
						|
      cloud_config_arg+=("--external-cloud-volume-plugin=${EXTERNAL_CLOUD_VOLUME_PLUGIN}")
 | 
						|
      cloud_config_arg+=("--cloud-config=${CLOUD_CONFIG}")
 | 
						|
    fi
 | 
						|
 | 
						|
    CTLRMGR_LOG=${LOG_DIR}/kube-controller-manager.log
 | 
						|
    ${CONTROLPLANE_SUDO} "${GO_OUT}/kube-controller-manager" \
 | 
						|
      --v="${LOG_LEVEL}" \
 | 
						|
      --vmodule="${LOG_SPEC}" \
 | 
						|
      --service-account-private-key-file="${SERVICE_ACCOUNT_KEY}" \
 | 
						|
      --service-cluster-ip-range="${SERVICE_CLUSTER_IP_RANGE}" \
 | 
						|
      --root-ca-file="${ROOT_CA_FILE}" \
 | 
						|
      --cluster-signing-cert-file="${CLUSTER_SIGNING_CERT_FILE}" \
 | 
						|
      --cluster-signing-key-file="${CLUSTER_SIGNING_KEY_FILE}" \
 | 
						|
      --enable-hostpath-provisioner="${ENABLE_HOSTPATH_PROVISIONER}" \
 | 
						|
      --pvclaimbinder-sync-period="${CLAIM_BINDER_SYNC_PERIOD}" \
 | 
						|
      --feature-gates="${FEATURE_GATES}" \
 | 
						|
      "${cloud_config_arg[@]}" \
 | 
						|
      --authentication-kubeconfig "${CERT_DIR}"/controller.kubeconfig \
 | 
						|
      --authorization-kubeconfig "${CERT_DIR}"/controller.kubeconfig \
 | 
						|
      --kubeconfig "${CERT_DIR}"/controller.kubeconfig \
 | 
						|
      --use-service-account-credentials \
 | 
						|
      --controllers="${KUBE_CONTROLLERS}" \
 | 
						|
      --leader-elect=false \
 | 
						|
      --cert-dir="${CERT_DIR}" \
 | 
						|
      --master="https://${API_HOST}:${API_SECURE_PORT}" >"${CTLRMGR_LOG}" 2>&1 &
 | 
						|
    CTLRMGR_PID=$!
 | 
						|
}
 | 
						|
 | 
						|
function start_cloud_controller_manager {
 | 
						|
    if [ -z "${CLOUD_CONFIG}" ]; then
 | 
						|
      echo "CLOUD_CONFIG cannot be empty!"
 | 
						|
      exit 1
 | 
						|
    fi
 | 
						|
    if [ ! -f "${CLOUD_CONFIG}" ]; then
 | 
						|
      echo "Cloud config ${CLOUD_CONFIG} doesn't exist"
 | 
						|
      exit 1
 | 
						|
    fi
 | 
						|
 | 
						|
    CLOUD_CTLRMGR_LOG=${LOG_DIR}/cloud-controller-manager.log
 | 
						|
    ${CONTROLPLANE_SUDO} "${EXTERNAL_CLOUD_PROVIDER_BINARY:-"${GO_OUT}/cloud-controller-manager"}" \
 | 
						|
      "${CLOUD_CTLRMGR_FLAGS}" \
 | 
						|
      --v="${LOG_LEVEL}" \
 | 
						|
      --vmodule="${LOG_SPEC}" \
 | 
						|
      --feature-gates="${FEATURE_GATES}" \
 | 
						|
      --cloud-provider="${CLOUD_PROVIDER}" \
 | 
						|
      --cloud-config="${CLOUD_CONFIG}" \
 | 
						|
      --configure-cloud-routes="${CONFIGURE_CLOUD_ROUTES}" \
 | 
						|
      --kubeconfig "${CERT_DIR}"/controller.kubeconfig \
 | 
						|
      --use-service-account-credentials \
 | 
						|
      --leader-elect=false \
 | 
						|
      --master="https://${API_HOST}:${API_SECURE_PORT}" >"${CLOUD_CTLRMGR_LOG}" 2>&1 &
 | 
						|
    export CLOUD_CTLRMGR_PID=$!
 | 
						|
}
 | 
						|
 | 
						|
function wait_node_ready(){
 | 
						|
  # check the nodes information after kubelet daemon start
 | 
						|
  local nodes_stats="${KUBECTL} --kubeconfig '${CERT_DIR}/admin.kubeconfig' get nodes"
 | 
						|
  local node_name=$HOSTNAME_OVERRIDE
 | 
						|
  local system_node_wait_time=60
 | 
						|
  local interval_time=2
 | 
						|
  kube::util::wait_for_success "$system_node_wait_time" "$interval_time" "$nodes_stats | grep $node_name"
 | 
						|
  if [ $? == "1" ]; then
 | 
						|
    echo "time out on waiting $node_name info"
 | 
						|
    exit 1
 | 
						|
  fi
 | 
						|
}
 | 
						|
 | 
						|
function start_kubelet {
 | 
						|
    KUBELET_LOG=${LOG_DIR}/kubelet.log
 | 
						|
    mkdir -p "${POD_MANIFEST_PATH}" &>/dev/null || sudo mkdir -p "${POD_MANIFEST_PATH}"
 | 
						|
 | 
						|
    cloud_config_arg=("--cloud-provider=${CLOUD_PROVIDER}" "--cloud-config=${CLOUD_CONFIG}")
 | 
						|
    if [[ "${EXTERNAL_CLOUD_PROVIDER:-}" == "true" ]]; then
 | 
						|
       cloud_config_arg=("--cloud-provider=external")
 | 
						|
       if [[ "${CLOUD_PROVIDER:-}" == "aws" ]]; then
 | 
						|
         cloud_config_arg+=("--provider-id=$(curl http://169.254.169.254/latest/meta-data/instance-id)")
 | 
						|
       else
 | 
						|
         cloud_config_arg+=("--provider-id=${KUBELET_PROVIDER_ID}")
 | 
						|
       fi
 | 
						|
    fi
 | 
						|
 | 
						|
    mkdir -p "/var/lib/kubelet" &>/dev/null || sudo mkdir -p "/var/lib/kubelet"
 | 
						|
    container_runtime_endpoint_args=()
 | 
						|
    if [[ -n "${CONTAINER_RUNTIME_ENDPOINT}" ]]; then
 | 
						|
      container_runtime_endpoint_args=("--container-runtime-endpoint=${CONTAINER_RUNTIME_ENDPOINT}")
 | 
						|
    fi
 | 
						|
 | 
						|
    image_service_endpoint_args=()
 | 
						|
    if [[ -n "${IMAGE_SERVICE_ENDPOINT}" ]]; then
 | 
						|
      image_service_endpoint_args=("--image-service-endpoint=${IMAGE_SERVICE_ENDPOINT}")
 | 
						|
    fi
 | 
						|
 | 
						|
    # shellcheck disable=SC2206
 | 
						|
    all_kubelet_flags=(
 | 
						|
      "--v=${LOG_LEVEL}"
 | 
						|
      "--vmodule=${LOG_SPEC}"
 | 
						|
      "--container-runtime=${CONTAINER_RUNTIME}"
 | 
						|
      "--hostname-override=${HOSTNAME_OVERRIDE}"
 | 
						|
      "${cloud_config_arg[@]}"
 | 
						|
      "--bootstrap-kubeconfig=${CERT_DIR}/kubelet.kubeconfig"
 | 
						|
      "--kubeconfig=${CERT_DIR}/kubelet-rotated.kubeconfig"
 | 
						|
      ${container_runtime_endpoint_args[@]+"${container_runtime_endpoint_args[@]}"}
 | 
						|
      ${image_service_endpoint_args[@]+"${image_service_endpoint_args[@]}"}
 | 
						|
      ${KUBELET_FLAGS}
 | 
						|
    )
 | 
						|
 | 
						|
    # warn if users are running with swap allowed
 | 
						|
    if [ "${FAIL_SWAP_ON}" == "false" ]; then
 | 
						|
        echo "WARNING : The kubelet is configured to not fail even if swap is enabled; production deployments should disable swap unless testing NodeSwap feature."
 | 
						|
    fi
 | 
						|
 | 
						|
    if [[ "${REUSE_CERTS}" != true ]]; then
 | 
						|
        # clear previous dynamic certs
 | 
						|
        sudo rm -fr "/var/lib/kubelet/pki" "${CERT_DIR}/kubelet-rotated.kubeconfig"
 | 
						|
        # create new certs
 | 
						|
        generate_kubelet_certs
 | 
						|
    fi
 | 
						|
 | 
						|
    cat <<EOF > /tmp/kubelet.yaml
 | 
						|
apiVersion: kubelet.config.k8s.io/v1beta1
 | 
						|
kind: KubeletConfiguration
 | 
						|
address: "${KUBELET_HOST}"
 | 
						|
cgroupDriver: "${CGROUP_DRIVER}"
 | 
						|
cgroupRoot: "${CGROUP_ROOT}"
 | 
						|
cgroupsPerQOS: ${CGROUPS_PER_QOS}
 | 
						|
cpuCFSQuota: ${CPU_CFS_QUOTA}
 | 
						|
enableControllerAttachDetach: ${ENABLE_CONTROLLER_ATTACH_DETACH}
 | 
						|
evictionPressureTransitionPeriod: "${EVICTION_PRESSURE_TRANSITION_PERIOD}"
 | 
						|
failSwapOn: ${FAIL_SWAP_ON}
 | 
						|
port: ${KUBELET_PORT}
 | 
						|
readOnlyPort: ${KUBELET_READ_ONLY_PORT}
 | 
						|
rotateCertificates: true
 | 
						|
runtimeRequestTimeout: "${RUNTIME_REQUEST_TIMEOUT}"
 | 
						|
staticPodPath: "${POD_MANIFEST_PATH}"
 | 
						|
resolvConf: "${KUBELET_RESOLV_CONF}"
 | 
						|
EOF
 | 
						|
    {
 | 
						|
      # authentication
 | 
						|
      echo "authentication:"
 | 
						|
      echo "  webhook:"
 | 
						|
      if [[ "${KUBELET_AUTHENTICATION_WEBHOOK:-}" != "false" ]]; then
 | 
						|
        echo "    enabled: true"
 | 
						|
      else
 | 
						|
        echo "    enabled: false"
 | 
						|
      fi
 | 
						|
      echo "  x509:"
 | 
						|
      if [[ -n "${CLIENT_CA_FILE:-}" ]]; then
 | 
						|
        echo "    clientCAFile: \"${CLIENT_CA_FILE}\""
 | 
						|
      else
 | 
						|
        echo "    clientCAFile: \"${CERT_DIR}/client-ca.crt\""
 | 
						|
      fi
 | 
						|
 | 
						|
      # authorization
 | 
						|
      if [[ "${KUBELET_AUTHORIZATION_WEBHOOK:-}" != "false" ]]; then
 | 
						|
        echo "authorization:"
 | 
						|
        echo "  mode: Webhook"
 | 
						|
      fi
 | 
						|
 | 
						|
      # dns
 | 
						|
      if [[ "${ENABLE_CLUSTER_DNS}" = true ]]; then
 | 
						|
        if [[ "${ENABLE_NODELOCAL_DNS:-}" == "true" ]]; then
 | 
						|
          echo "clusterDNS: [ \"${LOCAL_DNS_IP}\" ]"
 | 
						|
        else
 | 
						|
          echo "clusterDNS: [ \"${DNS_SERVER_IP}\" ]"
 | 
						|
        fi
 | 
						|
        echo "clusterDomain: \"${DNS_DOMAIN}\""
 | 
						|
      else
 | 
						|
        # To start a private DNS server set ENABLE_CLUSTER_DNS and
 | 
						|
        # DNS_SERVER_IP/DOMAIN. This will at least provide a working
 | 
						|
        # DNS server for real world hostnames.
 | 
						|
        echo "clusterDNS: [ \"8.8.8.8\" ]"
 | 
						|
      fi
 | 
						|
 | 
						|
      # eviction
 | 
						|
      if [[ -n ${EVICTION_HARD} ]]; then
 | 
						|
        echo "evictionHard:"
 | 
						|
        parse_eviction "${EVICTION_HARD}"
 | 
						|
      fi
 | 
						|
      if [[ -n ${EVICTION_SOFT} ]]; then
 | 
						|
        echo "evictionSoft:"
 | 
						|
        parse_eviction "${EVICTION_SOFT}"
 | 
						|
      fi
 | 
						|
 | 
						|
      # feature gate
 | 
						|
      if [[ -n ${FEATURE_GATES} ]]; then
 | 
						|
        parse_feature_gates "${FEATURE_GATES}"
 | 
						|
      fi
 | 
						|
    } >>/tmp/kubelet.yaml
 | 
						|
 | 
						|
    # shellcheck disable=SC2024
 | 
						|
    sudo -E "${GO_OUT}/kubelet" "${all_kubelet_flags[@]}" \
 | 
						|
      --config=/tmp/kubelet.yaml >"${KUBELET_LOG}" 2>&1 &
 | 
						|
    KUBELET_PID=$!
 | 
						|
 | 
						|
    # Quick check that kubelet is running.
 | 
						|
    if [ -n "${KUBELET_PID}" ] && ps -p ${KUBELET_PID} > /dev/null; then
 | 
						|
      echo "kubelet ( ${KUBELET_PID} ) is running."
 | 
						|
    else
 | 
						|
      cat "${KUBELET_LOG}" ; exit 1
 | 
						|
    fi
 | 
						|
}
 | 
						|
 | 
						|
function start_kubeproxy {
 | 
						|
    PROXY_LOG=${LOG_DIR}/kube-proxy.log
 | 
						|
 | 
						|
    if [[ "${START_MODE}" != *"nokubelet"* ]]; then
 | 
						|
      # wait for kubelet collect node information
 | 
						|
      echo "wait kubelet ready"
 | 
						|
      wait_node_ready
 | 
						|
    fi
 | 
						|
 | 
						|
    cat <<EOF > /tmp/kube-proxy.yaml
 | 
						|
apiVersion: kubeproxy.config.k8s.io/v1alpha1
 | 
						|
kind: KubeProxyConfiguration
 | 
						|
clientConnection:
 | 
						|
  kubeconfig: ${CERT_DIR}/kube-proxy.kubeconfig
 | 
						|
hostnameOverride: ${HOSTNAME_OVERRIDE}
 | 
						|
mode: ${KUBE_PROXY_MODE}
 | 
						|
conntrack:
 | 
						|
# Skip setting sysctl value "net.netfilter.nf_conntrack_max"
 | 
						|
  maxPerCore: 0
 | 
						|
# Skip setting "net.netfilter.nf_conntrack_tcp_timeout_established"
 | 
						|
  tcpEstablishedTimeout: 0s
 | 
						|
# Skip setting "net.netfilter.nf_conntrack_tcp_timeout_close"
 | 
						|
  tcpCloseWaitTimeout: 0s
 | 
						|
EOF
 | 
						|
    if [[ -n ${FEATURE_GATES} ]]; then
 | 
						|
      parse_feature_gates "${FEATURE_GATES}"
 | 
						|
    fi >>/tmp/kube-proxy.yaml
 | 
						|
 | 
						|
    if [[ "${REUSE_CERTS}" != true ]]; then
 | 
						|
        generate_kubeproxy_certs
 | 
						|
    fi
 | 
						|
 | 
						|
    # shellcheck disable=SC2024
 | 
						|
    sudo "${GO_OUT}/kube-proxy" \
 | 
						|
      --v="${LOG_LEVEL}" \
 | 
						|
      --config=/tmp/kube-proxy.yaml \
 | 
						|
      --master="https://${API_HOST}:${API_SECURE_PORT}" >"${PROXY_LOG}" 2>&1 &
 | 
						|
    PROXY_PID=$!
 | 
						|
}
 | 
						|
 | 
						|
function start_kubescheduler {
 | 
						|
    SCHEDULER_LOG=${LOG_DIR}/kube-scheduler.log
 | 
						|
 | 
						|
    cat <<EOF > /tmp/kube-scheduler.yaml
 | 
						|
apiVersion: kubescheduler.config.k8s.io/v1beta2
 | 
						|
kind: KubeSchedulerConfiguration
 | 
						|
clientConnection:
 | 
						|
  kubeconfig: ${CERT_DIR}/scheduler.kubeconfig
 | 
						|
leaderElection:
 | 
						|
  leaderElect: false
 | 
						|
EOF
 | 
						|
    ${CONTROLPLANE_SUDO} "${GO_OUT}/kube-scheduler" \
 | 
						|
      --v="${LOG_LEVEL}" \
 | 
						|
      --config=/tmp/kube-scheduler.yaml \
 | 
						|
      --feature-gates="${FEATURE_GATES}" \
 | 
						|
      --authentication-kubeconfig "${CERT_DIR}"/scheduler.kubeconfig \
 | 
						|
      --authorization-kubeconfig "${CERT_DIR}"/scheduler.kubeconfig \
 | 
						|
      --master="https://${API_HOST}:${API_SECURE_PORT}" >"${SCHEDULER_LOG}" 2>&1 &
 | 
						|
    SCHEDULER_PID=$!
 | 
						|
}
 | 
						|
 | 
						|
function start_dns_addon {
 | 
						|
    if [[ "${ENABLE_CLUSTER_DNS}" = true ]]; then
 | 
						|
        cp "${KUBE_ROOT}/cluster/addons/dns/${DNS_ADDON}/${DNS_ADDON}.yaml.in" dns.yaml
 | 
						|
        ${SED} -i -e "s/dns_domain/${DNS_DOMAIN}/g" dns.yaml
 | 
						|
        ${SED} -i -e "s/dns_server/${DNS_SERVER_IP}/g" dns.yaml
 | 
						|
        ${SED} -i -e "s/dns_memory_limit/${DNS_MEMORY_LIMIT}/g" dns.yaml
 | 
						|
        # TODO update to dns role once we have one.
 | 
						|
        # use kubectl to create dns addon
 | 
						|
        if ${KUBECTL} --kubeconfig="${CERT_DIR}/admin.kubeconfig" --namespace=kube-system create -f dns.yaml ; then
 | 
						|
            echo "${DNS_ADDON} addon successfully deployed."
 | 
						|
        else
 | 
						|
		echo "Something is wrong with your DNS input"
 | 
						|
		cat dns.yaml
 | 
						|
		exit 1
 | 
						|
        fi
 | 
						|
        rm dns.yaml
 | 
						|
    fi
 | 
						|
}
 | 
						|
 | 
						|
function start_nodelocaldns {
 | 
						|
  cp "${KUBE_ROOT}/cluster/addons/dns/nodelocaldns/nodelocaldns.yaml" nodelocaldns.yaml
 | 
						|
  # eventually all the __PILLAR__ stuff will be gone, but theyre still in nodelocaldns for backward compat.
 | 
						|
  ${SED} -i -e "s/__PILLAR__DNS__DOMAIN__/${DNS_DOMAIN}/g" nodelocaldns.yaml
 | 
						|
  ${SED} -i -e "s/__PILLAR__DNS__SERVER__/${DNS_SERVER_IP}/g" nodelocaldns.yaml
 | 
						|
  ${SED} -i -e "s/__PILLAR__LOCAL__DNS__/${LOCAL_DNS_IP}/g" nodelocaldns.yaml
 | 
						|
 | 
						|
  # use kubectl to create nodelocaldns addon
 | 
						|
  ${KUBECTL} --kubeconfig="${CERT_DIR}/admin.kubeconfig" --namespace=kube-system create -f nodelocaldns.yaml
 | 
						|
  echo "NodeLocalDNS addon successfully deployed."
 | 
						|
  rm nodelocaldns.yaml
 | 
						|
}
 | 
						|
 | 
						|
function start_csi_snapshotter {
 | 
						|
    if [[ "${ENABLE_CSI_SNAPSHOTTER}" = true ]]; then
 | 
						|
        echo "Creating Kubernetes-CSI snapshotter"
 | 
						|
        ${KUBECTL} --kubeconfig="${CERT_DIR}/admin.kubeconfig" apply -f "${KUBE_ROOT}/cluster/addons/volumesnapshots/crd/snapshot.storage.k8s.io_volumesnapshots.yaml"
 | 
						|
        ${KUBECTL} --kubeconfig="${CERT_DIR}/admin.kubeconfig" apply -f "${KUBE_ROOT}/cluster/addons/volumesnapshots/crd/snapshot.storage.k8s.io_volumesnapshotclasses.yaml"
 | 
						|
        ${KUBECTL} --kubeconfig="${CERT_DIR}/admin.kubeconfig" apply -f "${KUBE_ROOT}/cluster/addons/volumesnapshots/crd/snapshot.storage.k8s.io_volumesnapshotcontents.yaml"
 | 
						|
        ${KUBECTL} --kubeconfig="${CERT_DIR}/admin.kubeconfig" apply -f "${KUBE_ROOT}/cluster/addons/volumesnapshots/volume-snapshot-controller/rbac-volume-snapshot-controller.yaml"
 | 
						|
        ${KUBECTL} --kubeconfig="${CERT_DIR}/admin.kubeconfig" apply -f "${KUBE_ROOT}/cluster/addons/volumesnapshots/volume-snapshot-controller/volume-snapshot-controller-deployment.yaml"
 | 
						|
 | 
						|
        echo "Kubernetes-CSI snapshotter successfully deployed."
 | 
						|
    fi
 | 
						|
}
 | 
						|
 | 
						|
function create_psp_policy {
 | 
						|
    echo "Create podsecuritypolicy policies for RBAC."
 | 
						|
    ${KUBECTL} --kubeconfig="${CERT_DIR}/admin.kubeconfig" create -f "${KUBE_ROOT}/examples/podsecuritypolicy/rbac/policies.yaml"
 | 
						|
    ${KUBECTL} --kubeconfig="${CERT_DIR}/admin.kubeconfig" create -f "${KUBE_ROOT}/examples/podsecuritypolicy/rbac/roles.yaml"
 | 
						|
    ${KUBECTL} --kubeconfig="${CERT_DIR}/admin.kubeconfig" create -f "${KUBE_ROOT}/examples/podsecuritypolicy/rbac/bindings.yaml"
 | 
						|
}
 | 
						|
 | 
						|
function create_storage_class {
 | 
						|
    if [ -z "${CLOUD_PROVIDER}" ]; then
 | 
						|
        CLASS_FILE=${KUBE_ROOT}/cluster/addons/storage-class/local/default.yaml
 | 
						|
    else
 | 
						|
        CLASS_FILE=${KUBE_ROOT}/cluster/addons/storage-class/${CLOUD_PROVIDER}/default.yaml
 | 
						|
    fi
 | 
						|
 | 
						|
    if [ -e "${CLASS_FILE}" ]; then
 | 
						|
        echo "Create default storage class for ${CLOUD_PROVIDER}"
 | 
						|
        ${KUBECTL} --kubeconfig="${CERT_DIR}/admin.kubeconfig" create -f "${CLASS_FILE}"
 | 
						|
    else
 | 
						|
        echo "No storage class available for ${CLOUD_PROVIDER}."
 | 
						|
    fi
 | 
						|
}
 | 
						|
 | 
						|
function print_success {
 | 
						|
if [[ "${START_MODE}" != "kubeletonly" ]]; then
 | 
						|
  if [[ "${ENABLE_DAEMON}" = false ]]; then
 | 
						|
    echo "Local Kubernetes cluster is running. Press Ctrl-C to shut it down."
 | 
						|
  else
 | 
						|
    echo "Local Kubernetes cluster is running."
 | 
						|
  fi
 | 
						|
  cat <<EOF
 | 
						|
 | 
						|
Logs:
 | 
						|
  ${APISERVER_LOG:-}
 | 
						|
  ${CTLRMGR_LOG:-}
 | 
						|
  ${CLOUD_CTLRMGR_LOG:-}
 | 
						|
  ${PROXY_LOG:-}
 | 
						|
  ${SCHEDULER_LOG:-}
 | 
						|
EOF
 | 
						|
fi
 | 
						|
 | 
						|
if [[ "${START_MODE}" == "all" ]]; then
 | 
						|
  echo "  ${KUBELET_LOG}"
 | 
						|
elif [[ "${START_MODE}" == *"nokubelet"* ]]; then
 | 
						|
  echo
 | 
						|
  echo "No kubelet was started because you set START_MODE=nokubelet"
 | 
						|
  echo "Run this script again with START_MODE=kubeletonly to run a kubelet"
 | 
						|
fi
 | 
						|
 | 
						|
if [[ "${START_MODE}" != "kubeletonly" ]]; then
 | 
						|
  echo
 | 
						|
  if [[ "${ENABLE_DAEMON}" = false ]]; then
 | 
						|
    echo "To start using your cluster, you can open up another terminal/tab and run:"
 | 
						|
  else
 | 
						|
    echo "To start using your cluster, run:"
 | 
						|
  fi
 | 
						|
  cat <<EOF
 | 
						|
 | 
						|
  export KUBECONFIG=${CERT_DIR}/admin.kubeconfig
 | 
						|
  cluster/kubectl.sh
 | 
						|
 | 
						|
Alternatively, you can write to the default kubeconfig:
 | 
						|
 | 
						|
  export KUBERNETES_PROVIDER=local
 | 
						|
 | 
						|
  cluster/kubectl.sh config set-cluster local --server=https://${API_HOST}:${API_SECURE_PORT} --certificate-authority=${ROOT_CA_FILE}
 | 
						|
  cluster/kubectl.sh config set-credentials myself ${AUTH_ARGS}
 | 
						|
  cluster/kubectl.sh config set-context local --cluster=local --user=myself
 | 
						|
  cluster/kubectl.sh config use-context local
 | 
						|
  cluster/kubectl.sh
 | 
						|
EOF
 | 
						|
else
 | 
						|
  cat <<EOF
 | 
						|
The kubelet was started.
 | 
						|
 | 
						|
Logs:
 | 
						|
  ${KUBELET_LOG}
 | 
						|
EOF
 | 
						|
fi
 | 
						|
}
 | 
						|
 | 
						|
function parse_feature_gates {
 | 
						|
  echo "featureGates:"
 | 
						|
  # Convert from foo=true,bar=false to
 | 
						|
  #   foo: true
 | 
						|
  #   bar: false
 | 
						|
  for gate in $(echo "$1" | tr ',' ' '); do
 | 
						|
    echo "${gate}" | ${SED} -e 's/\(.*\)=\(.*\)/  \1: \2/'
 | 
						|
  done
 | 
						|
}
 | 
						|
 | 
						|
function parse_eviction {
 | 
						|
  # Convert from memory.available<100Mi,nodefs.available<10%,nodefs.inodesFree<5% to
 | 
						|
  #   memory.available: "100Mi"
 | 
						|
  #   nodefs.available: "10%"
 | 
						|
  #   nodefs.inodesFree: "5%"
 | 
						|
  for eviction in $(echo "$1" | tr ',' ' '); do
 | 
						|
    echo "${eviction}" | ${SED} -e 's/</: \"/' | ${SED} -e 's/^/  /' | ${SED} -e 's/$/\"/'
 | 
						|
  done
 | 
						|
}
 | 
						|
 | 
						|
function install_cni {
 | 
						|
  echo "Installing CNI plugin binaries ..." \
 | 
						|
    && curl -sSL --retry 5 --output /tmp/cni."${CNI_TARGETARCH}".tgz "${CNI_PLUGINS_URL}" \
 | 
						|
    && echo "${CNI_PLUGINS_AMD64_SHA256SUM}  /tmp/cni.amd64.tgz" | tee /tmp/cni.sha256 \
 | 
						|
    && sha256sum --ignore-missing -c /tmp/cni.sha256 \
 | 
						|
    && rm -f /tmp/cni.sha256 \
 | 
						|
    && sudo mkdir -p /opt/cni/bin \
 | 
						|
    && sudo tar -C /opt/cni/bin -xzvf /tmp/cni."${CNI_TARGETARCH}".tgz \
 | 
						|
    && rm -rf /tmp/cni."${CNI_TARGETARCH}".tgz \
 | 
						|
    && sudo find /opt/cni/bin -type f -not \( \
 | 
						|
         -iname host-local \
 | 
						|
         -o -iname bridge \
 | 
						|
         -o -iname portmap \
 | 
						|
         -o -iname loopback \
 | 
						|
      \) \
 | 
						|
      -delete
 | 
						|
 | 
						|
  # containerd 1.4.12 installed by docker in kubekins supports CNI version 0.4.0
 | 
						|
  echo "Configuring cni"
 | 
						|
  sudo mkdir -p "$CNI_CONFIG_DIR"
 | 
						|
  cat << EOF | sudo tee "$CNI_CONFIG_DIR"/10-containerd-net.conflist
 | 
						|
{
 | 
						|
  "cniVersion": "0.4.0",
 | 
						|
  "name": "containerd-net",
 | 
						|
  "plugins": [
 | 
						|
    {
 | 
						|
      "type": "bridge",
 | 
						|
      "bridge": "cni0",
 | 
						|
      "isGateway": true,
 | 
						|
      "ipMasq": true,
 | 
						|
      "promiscMode": true,
 | 
						|
      "ipam": {
 | 
						|
        "type": "host-local",
 | 
						|
        "ranges": [
 | 
						|
          [{
 | 
						|
            "subnet": "10.88.0.0/16"
 | 
						|
          }],
 | 
						|
          [{
 | 
						|
            "subnet": "2001:4860:4860::/64"
 | 
						|
          }]
 | 
						|
        ],
 | 
						|
        "routes": [
 | 
						|
          { "dst": "0.0.0.0/0" },
 | 
						|
          { "dst": "::/0" }
 | 
						|
        ]
 | 
						|
      }
 | 
						|
    },
 | 
						|
    {
 | 
						|
      "type": "portmap",
 | 
						|
      "capabilities": {"portMappings": true}
 | 
						|
    }
 | 
						|
  ]
 | 
						|
}
 | 
						|
EOF
 | 
						|
}
 | 
						|
 | 
						|
function install_cni_if_needed {
 | 
						|
  echo "Checking CNI Installation at /opt/cni/bin"
 | 
						|
  if ! command -v /opt/cni/bin/loopback &> /dev/null ; then
 | 
						|
    echo "CNI Installation not found at /opt/cni/bin"
 | 
						|
    install_cni
 | 
						|
  fi
 | 
						|
}
 | 
						|
 | 
						|
# If we are running in the CI, we need a few more things before we can start
 | 
						|
if [[ "${KUBETEST_IN_DOCKER:-}" == "true" ]]; then
 | 
						|
  echo "Preparing to test ..."
 | 
						|
  "${KUBE_ROOT}"/hack/install-etcd.sh
 | 
						|
  export PATH="${KUBE_ROOT}/third_party/etcd:${PATH}"
 | 
						|
  KUBE_FASTBUILD=true make ginkgo cross
 | 
						|
 | 
						|
  apt-get update && apt-get install -y sudo
 | 
						|
  apt-get remove -y systemd
 | 
						|
 | 
						|
  # configure shared mounts to prevent failure in DIND scenarios
 | 
						|
  mount --make-rshared /
 | 
						|
 | 
						|
  # kubekins has a special directory for docker root
 | 
						|
  DOCKER_ROOT="/docker-graph"
 | 
						|
 | 
						|
  # to use docker installed containerd as kubelet container runtime
 | 
						|
  # we need to enable cri and install cni
 | 
						|
  # install cni for docker in docker
 | 
						|
  install_cni 
 | 
						|
 | 
						|
  # enable cri for docker in docker
 | 
						|
  echo "enable cri"
 | 
						|
  echo "DOCKER_OPTS=\"\${DOCKER_OPTS} --cri-containerd\"" >> /etc/default/docker
 | 
						|
 | 
						|
  echo "restarting docker"
 | 
						|
  service docker restart
 | 
						|
fi
 | 
						|
 | 
						|
# validate that etcd is: not running, in path, and has minimum required version.
 | 
						|
if [[ "${START_MODE}" != "kubeletonly" ]]; then
 | 
						|
  kube::etcd::validate
 | 
						|
fi
 | 
						|
 | 
						|
if [[ "${START_MODE}" != "kubeletonly" ]]; then
 | 
						|
  test_apiserver_off
 | 
						|
fi
 | 
						|
 | 
						|
kube::util::test_openssl_installed
 | 
						|
kube::util::ensure-cfssl
 | 
						|
 | 
						|
### IF the user didn't supply an output/ for the build... Then we detect.
 | 
						|
if [ "${GO_OUT}" == "" ]; then
 | 
						|
  detect_binary
 | 
						|
fi
 | 
						|
echo "Detected host and ready to start services.  Doing some housekeeping first..."
 | 
						|
echo "Using GO_OUT ${GO_OUT}"
 | 
						|
export KUBELET_CIDFILE=/tmp/kubelet.cid
 | 
						|
if [[ "${ENABLE_DAEMON}" = false ]]; then
 | 
						|
  trap cleanup EXIT
 | 
						|
fi
 | 
						|
 | 
						|
echo "Starting services now!"
 | 
						|
if [[ "${START_MODE}" != "kubeletonly" ]]; then
 | 
						|
  start_etcd
 | 
						|
  set_service_accounts
 | 
						|
  start_apiserver
 | 
						|
  start_controller_manager
 | 
						|
  if [[ "${EXTERNAL_CLOUD_PROVIDER:-}" == "true" ]]; then
 | 
						|
    start_cloud_controller_manager
 | 
						|
  fi
 | 
						|
  start_kubescheduler
 | 
						|
  start_dns_addon
 | 
						|
  if [[ "${ENABLE_NODELOCAL_DNS:-}" == "true" ]]; then
 | 
						|
    start_nodelocaldns
 | 
						|
  fi
 | 
						|
  start_csi_snapshotter
 | 
						|
fi
 | 
						|
 | 
						|
if [[ "${START_MODE}" != *"nokubelet"* ]]; then
 | 
						|
  ## TODO remove this check if/when kubelet is supported on darwin
 | 
						|
  # Detect the OS name/arch and display appropriate error.
 | 
						|
    case "$(uname -s)" in
 | 
						|
      Darwin)
 | 
						|
        print_color "kubelet is not currently supported in darwin, kubelet aborted."
 | 
						|
        KUBELET_LOG=""
 | 
						|
        ;;
 | 
						|
      Linux)
 | 
						|
        install_cni_if_needed
 | 
						|
        start_kubelet
 | 
						|
        ;;
 | 
						|
      *)
 | 
						|
        print_color "Unsupported host OS.  Must be Linux or Mac OS X, kubelet aborted."
 | 
						|
        ;;
 | 
						|
    esac
 | 
						|
fi
 | 
						|
 | 
						|
if [[ "${START_MODE}" != "kubeletonly" ]]; then
 | 
						|
  if [[ "${START_MODE}" != *"nokubeproxy"* ]]; then
 | 
						|
    ## TODO remove this check if/when kubelet is supported on darwin
 | 
						|
    # Detect the OS name/arch and display appropriate error.
 | 
						|
    case "$(uname -s)" in
 | 
						|
      Darwin)
 | 
						|
        print_color "kubelet is not currently supported in darwin, kube-proxy aborted."
 | 
						|
        ;;
 | 
						|
      Linux)
 | 
						|
        start_kubeproxy
 | 
						|
        ;;
 | 
						|
      *)
 | 
						|
        print_color "Unsupported host OS.  Must be Linux or Mac OS X, kube-proxy aborted."
 | 
						|
        ;;
 | 
						|
    esac
 | 
						|
  fi
 | 
						|
fi
 | 
						|
 | 
						|
if [[ -n "${PSP_ADMISSION}" && "${AUTHORIZATION_MODE}" = *RBAC* ]]; then
 | 
						|
  create_psp_policy
 | 
						|
fi
 | 
						|
 | 
						|
if [[ "${DEFAULT_STORAGE_CLASS}" = "true" ]]; then
 | 
						|
  create_storage_class
 | 
						|
fi
 | 
						|
 | 
						|
print_success
 | 
						|
 | 
						|
if [[ "${ENABLE_DAEMON}" = false ]]; then
 | 
						|
  while true; do sleep 1; healthcheck; done
 | 
						|
fi
 | 
						|
 | 
						|
if [[ "${KUBETEST_IN_DOCKER:-}" == "true" ]]; then
 | 
						|
  cluster/kubectl.sh config set-cluster local --server=https://localhost:6443 --certificate-authority=/var/run/kubernetes/server-ca.crt
 | 
						|
  cluster/kubectl.sh config set-credentials myself --client-key=/var/run/kubernetes/client-admin.key --client-certificate=/var/run/kubernetes/client-admin.crt
 | 
						|
  cluster/kubectl.sh config set-context local --cluster=local --user=myself
 | 
						|
  cluster/kubectl.sh config use-context local
 | 
						|
fi
 |