mirror of
				https://github.com/k3s-io/kubernetes.git
				synced 2025-11-03 23:40:03 +00:00 
			
		
		
		
	
		
			
				
	
	
		
			461 lines
		
	
	
		
			19 KiB
		
	
	
	
		
			Bash
		
	
	
	
	
	
			
		
		
	
	
			461 lines
		
	
	
		
			19 KiB
		
	
	
	
		
			Bash
		
	
	
	
	
	
# Copyright 2014 The Kubernetes Authors.
 | 
						|
#
 | 
						|
# Licensed under the Apache License, Version 2.0 (the "License");
 | 
						|
# you may not use this file except in compliance with the License.
 | 
						|
# You may obtain a copy of the License at
 | 
						|
#
 | 
						|
#     http://www.apache.org/licenses/LICENSE-2.0
 | 
						|
#
 | 
						|
# Unless required by applicable law or agreed to in writing, software
 | 
						|
# distributed under the License is distributed on an "AS IS" BASIS,
 | 
						|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 | 
						|
# See the License for the specific language governing permissions and
 | 
						|
# limitations under the License.
 | 
						|
 | 
						|
# required:
 | 
						|
# KUBE_ROOT: path of the root of the Kubernetes repository
 | 
						|
 | 
						|
: "${KUBE_ROOT?Must set KUBE_ROOT env var}"
 | 
						|
 | 
						|
# Provides the $KUBERNETES_PROVIDER, kubeconfig-federation-context()
 | 
						|
# and detect-project function
 | 
						|
source "${KUBE_ROOT}/cluster/kube-util.sh"
 | 
						|
 | 
						|
# kubefed configuration
 | 
						|
FEDERATION_NAME="${FEDERATION_NAME:-e2e-federation}"
 | 
						|
FEDERATION_NAMESPACE=${FEDERATION_NAMESPACE:-federation-system}
 | 
						|
FEDERATION_KUBE_CONTEXT="${FEDERATION_KUBE_CONTEXT:-${FEDERATION_NAME}}"
 | 
						|
HOST_CLUSTER_ZONE="${FEDERATION_HOST_CLUSTER_ZONE:-}"
 | 
						|
# If $HOST_CLUSTER_ZONE isn't specified, arbitrarily choose
 | 
						|
# last zone as the host cluster zone.
 | 
						|
if [[ -z "${HOST_CLUSTER_ZONE}" ]]; then
 | 
						|
  E2E_ZONES_ARR=(${E2E_ZONES:-})
 | 
						|
  if [[ ${#E2E_ZONES_ARR[@]} > 0 ]]; then
 | 
						|
    HOST_CLUSTER_ZONE=${E2E_ZONES_ARR[-1]}
 | 
						|
  fi
 | 
						|
fi
 | 
						|
 | 
						|
HOST_CLUSTER_CONTEXT="${FEDERATION_HOST_CLUSTER_CONTEXT:-}"
 | 
						|
if [[ -z "${HOST_CLUSTER_CONTEXT}" ]]; then
 | 
						|
  # Sets ${CLUSTER_CONTEXT}
 | 
						|
  if [[ -z "${HOST_CLUSTER_ZONE:-}" ]]; then
 | 
						|
    echo "At least one of FEDERATION_HOST_CLUSTER_CONTEXT, FEDERATION_HOST_CLUSTER_ZONE or E2E_ZONES is required."
 | 
						|
    exit 1
 | 
						|
  fi
 | 
						|
  kubeconfig-federation-context "${HOST_CLUSTER_ZONE:-}"
 | 
						|
  HOST_CLUSTER_CONTEXT="${CLUSTER_CONTEXT}"
 | 
						|
fi
 | 
						|
 | 
						|
# kube-dns configuration.
 | 
						|
KUBEDNS_CONFIGMAP_NAME="kube-dns"
 | 
						|
KUBEDNS_CONFIGMAP_NAMESPACE="kube-system"
 | 
						|
KUBEDNS_FEDERATION_FLAG="federations"
 | 
						|
 | 
						|
function federation_cluster_contexts() {
 | 
						|
  local -r contexts=$("${KUBE_ROOT}/cluster/kubectl.sh" config get-contexts -o name)
 | 
						|
  federation_contexts=()
 | 
						|
  for context in ${contexts}; do
 | 
						|
    # Skip federation context
 | 
						|
    if [[ "${context}" == "${FEDERATION_NAME}" ]]; then
 | 
						|
      continue
 | 
						|
    fi
 | 
						|
    # Skip contexts not beginning with "federation"
 | 
						|
    if [[ "${context}" != federation* ]]; then
 | 
						|
      continue
 | 
						|
    fi
 | 
						|
    federation_contexts+=("${context}")
 | 
						|
  done
 | 
						|
  echo ${federation_contexts[@]}
 | 
						|
}
 | 
						|
 | 
						|
 | 
						|
#-----------------------------------------------------------------#
 | 
						|
# NOTE:                                                           #
 | 
						|
# Everything below this line is deprecated. It will be removed    #
 | 
						|
# once we have sufficient confidence in kubefed based testing.    #
 | 
						|
#-----------------------------------------------------------------#
 | 
						|
 | 
						|
# optional override
 | 
						|
# FEDERATION_IMAGE_REPO_BASE: repo which federated images are tagged under (default gcr.io/google_containers)
 | 
						|
# FEDERATION_NAMESPACE: name of the namespace will created for the federated components in the underlying cluster.
 | 
						|
# KUBE_PLATFORM
 | 
						|
# KUBE_ARCH
 | 
						|
# KUBE_BUILD_STAGE
 | 
						|
 | 
						|
source "${KUBE_ROOT}/cluster/common.sh"
 | 
						|
 | 
						|
host_kubectl="${KUBE_ROOT}/cluster/kubectl.sh --namespace=${FEDERATION_NAMESPACE}"
 | 
						|
 | 
						|
# If $FEDERATION_PUSH_REPO_BASE isn't set, then set the GCR registry name
 | 
						|
# based on the detected project name for gce and gke providers.
 | 
						|
FEDERATION_PUSH_REPO_BASE=${FEDERATION_PUSH_REPO_BASE:-}
 | 
						|
if [[ -z "${FEDERATION_PUSH_REPO_BASE}" ]]; then
 | 
						|
    if [[ "${KUBERNETES_PROVIDER}" == "gke" || "${KUBERNETES_PROVIDER}" == "gce" ]]; then
 | 
						|
        # Populates $PROJECT
 | 
						|
        detect-project
 | 
						|
        if [[ ${PROJECT} == *':'* ]]; then
 | 
						|
            echo "${PROJECT} contains ':' and can not be used as FEDERATION_PUSH_REPO_BASE. Please set FEDERATION_PUSH_REPO_BASE explicitly."
 | 
						|
            exit 1
 | 
						|
        fi
 | 
						|
        FEDERATION_PUSH_REPO_BASE=gcr.io/${PROJECT}
 | 
						|
    else
 | 
						|
        echo "Must set FEDERATION_PUSH_REPO_BASE env var"
 | 
						|
        exit 1
 | 
						|
    fi
 | 
						|
fi
 | 
						|
 | 
						|
FEDERATION_IMAGE_REPO_BASE=${FEDERATION_IMAGE_REPO_BASE:-'gcr.io/google_containers'}
 | 
						|
 | 
						|
KUBE_PLATFORM=${KUBE_PLATFORM:-linux}
 | 
						|
KUBE_ARCH=${KUBE_ARCH:-amd64}
 | 
						|
KUBE_BUILD_STAGE=${KUBE_BUILD_STAGE:-release-stage}
 | 
						|
 | 
						|
source "${KUBE_ROOT}/cluster/common.sh"
 | 
						|
 | 
						|
host_kubectl="${KUBE_ROOT}/cluster/kubectl.sh --namespace=${FEDERATION_NAMESPACE}"
 | 
						|
 | 
						|
# required:
 | 
						|
# FEDERATION_PUSH_REPO_BASE: repo to which federated container images will be pushed
 | 
						|
# FEDERATION_IMAGE_TAG: reference and pull all federated images with this tag.
 | 
						|
function create-federation-api-objects {
 | 
						|
(
 | 
						|
    : "${FEDERATION_PUSH_REPO_BASE?Must set FEDERATION_PUSH_REPO_BASE env var}"
 | 
						|
    : "${FEDERATION_IMAGE_TAG?Must set FEDERATION_IMAGE_TAG env var}"
 | 
						|
 | 
						|
    export FEDERATION_APISERVER_DEPLOYMENT_NAME="federation-apiserver"
 | 
						|
    export FEDERATION_APISERVER_IMAGE_REPO="${FEDERATION_PUSH_REPO_BASE}/hyperkube-amd64"
 | 
						|
    export FEDERATION_APISERVER_IMAGE_TAG="${FEDERATION_IMAGE_TAG}"
 | 
						|
 | 
						|
    export FEDERATION_CONTROLLER_MANAGER_DEPLOYMENT_NAME="federation-controller-manager"
 | 
						|
    export FEDERATION_CONTROLLER_MANAGER_IMAGE_REPO="${FEDERATION_PUSH_REPO_BASE}/hyperkube-amd64"
 | 
						|
    export FEDERATION_CONTROLLER_MANAGER_IMAGE_TAG="${FEDERATION_IMAGE_TAG}"
 | 
						|
 | 
						|
    if [[ -z "${FEDERATION_DNS_PROVIDER:-}" ]]; then
 | 
						|
      # Set the appropriate value based on cloud provider.
 | 
						|
      if [[ "$KUBERNETES_PROVIDER" == "gce" || "${KUBERNETES_PROVIDER}" == "gke" ]]; then
 | 
						|
        echo "setting dns provider to google-clouddns"
 | 
						|
        export FEDERATION_DNS_PROVIDER="google-clouddns"
 | 
						|
      elif [[ "${KUBERNETES_PROVIDER}" == "aws" ]]; then
 | 
						|
        echo "setting dns provider to aws-route53"
 | 
						|
        export FEDERATION_DNS_PROVIDER="aws-route53"
 | 
						|
      else
 | 
						|
        echo "Must set FEDERATION_DNS_PROVIDER env var"
 | 
						|
        exit 1
 | 
						|
      fi
 | 
						|
    fi
 | 
						|
 | 
						|
    export FEDERATION_SERVICE_CIDR=${FEDERATION_SERVICE_CIDR:-"10.10.0.0/24"}
 | 
						|
 | 
						|
    #Only used for providers that require a nodeport service (vagrant for now)
 | 
						|
    #We will use loadbalancer services where we can
 | 
						|
    export FEDERATION_API_NODEPORT=32111
 | 
						|
    export FEDERATION_NAMESPACE
 | 
						|
    export FEDERATION_NAME="${FEDERATION_NAME:-federation}"
 | 
						|
    export DNS_ZONE_NAME="${DNS_ZONE_NAME:-federation.example.}"  # See https://tools.ietf.org/html/rfc2606
 | 
						|
 | 
						|
    template="go run ${KUBE_ROOT}/federation/cluster/template.go"
 | 
						|
 | 
						|
    FEDERATION_KUBECONFIG_PATH="${KUBE_ROOT}/federation/cluster/kubeconfig"
 | 
						|
 | 
						|
    federation_kubectl="${KUBE_ROOT}/cluster/kubectl.sh --context=federation-cluster --namespace=default"
 | 
						|
 | 
						|
    manifests_root="${KUBE_ROOT}/federation/manifests/"
 | 
						|
 | 
						|
    $template "${manifests_root}/federation-ns.yaml" | $host_kubectl apply -f -
 | 
						|
 | 
						|
    cleanup-federation-api-objects
 | 
						|
 | 
						|
    export FEDERATION_API_HOST=""
 | 
						|
    export KUBE_MASTER_IP=""
 | 
						|
    export IS_DNS_NAME="false"
 | 
						|
    if [[ "$KUBERNETES_PROVIDER" == "vagrant" ]];then
 | 
						|
	# The vagrant approach is to use a nodeport service, and point kubectl at one of the nodes
 | 
						|
	$template "${manifests_root}/federation-apiserver-nodeport-service.yaml" | $host_kubectl create -f -
 | 
						|
	node_addresses=`$host_kubectl get nodes -o=jsonpath='{.items[*].status.addresses[?(@.type=="InternalIP")].address}'`
 | 
						|
	FEDERATION_API_HOST=`printf "$node_addresses" | cut -d " " -f1`
 | 
						|
	KUBE_MASTER_IP="${FEDERATION_API_HOST}:${FEDERATION_API_NODEPORT}"
 | 
						|
    elif [[ "$KUBERNETES_PROVIDER" == "gce" || "$KUBERNETES_PROVIDER" == "gke" || "$KUBERNETES_PROVIDER" == "aws" ]];then
 | 
						|
 | 
						|
	# Any providers where ingress is a DNS name should tick this box.
 | 
						|
	# TODO(chom): attempt to do this automatically
 | 
						|
	if [[ "$KUBERNETES_PROVIDER" == "aws" ]];then
 | 
						|
	    IS_DNS_NAME="true"
 | 
						|
	fi
 | 
						|
	# any capable providers should use a loadbalancer service
 | 
						|
	# we check for ingress.ip and ingress.hostname, so should work for any loadbalancer-providing provider
 | 
						|
	# allows 30x5 = 150 seconds for loadbalancer creation
 | 
						|
	$template "${manifests_root}/federation-apiserver-lb-service.yaml" | $host_kubectl create -f -
 | 
						|
	for i in {1..30};do
 | 
						|
	    echo "attempting to get federation-apiserver loadbalancer hostname ($i / 30)"
 | 
						|
	    LB_STATUS=`${host_kubectl} get -o=jsonpath svc/${FEDERATION_APISERVER_DEPLOYMENT_NAME} --template '{.status.loadBalancer}'`
 | 
						|
	    # Check if ingress field has been set in load balancer status.
 | 
						|
	    if [[ "${LB_STATUS}" != *"ingress"* ]]; then
 | 
						|
	        echo "Waiting for load balancer status to be set"
 | 
						|
	        sleep 5
 | 
						|
	        continue
 | 
						|
	    fi
 | 
						|
	    for field in ip hostname;do
 | 
						|
		FEDERATION_API_HOST=`${host_kubectl} get -o=jsonpath svc/${FEDERATION_APISERVER_DEPLOYMENT_NAME} --template '{.status.loadBalancer.ingress[*].'"${field}}"`
 | 
						|
		if [[ ! -z "${FEDERATION_API_HOST// }" ]];then
 | 
						|
		    break 2
 | 
						|
		fi
 | 
						|
	    done
 | 
						|
	    if [[ $i -eq 30 ]];then
 | 
						|
		echo "Could not find ingress hostname for federation-apiserver loadbalancer service"
 | 
						|
		exit 1
 | 
						|
	    fi
 | 
						|
	    sleep 5
 | 
						|
	done
 | 
						|
	KUBE_MASTER_IP="${FEDERATION_API_HOST}:443"
 | 
						|
    else
 | 
						|
	echo "provider ${KUBERNETES_PROVIDER} is not (yet) supported for e2e testing"
 | 
						|
	exit 1
 | 
						|
    fi
 | 
						|
    echo "Found federation-apiserver host at $FEDERATION_API_HOST"
 | 
						|
 | 
						|
    FEDERATION_API_TOKEN="$(dd if=/dev/urandom bs=128 count=1 2>/dev/null | base64 | tr -d "=+/" | dd bs=32 count=1 2>/dev/null)"
 | 
						|
    export FEDERATION_API_KNOWN_TOKENS="${FEDERATION_API_TOKEN},admin,admin"
 | 
						|
    gen-kube-basicauth
 | 
						|
    export FEDERATION_API_BASIC_AUTH="${KUBE_PASSWORD},${KUBE_USER},admin"
 | 
						|
 | 
						|
    # Create a kubeconfig with credentials for federation-apiserver. We will
 | 
						|
    # then use this kubeconfig to create a secret which the federation
 | 
						|
    # controller manager can use to talk to the federation-apiserver.
 | 
						|
    # Note that the file name should be "kubeconfig" so that the secret key gets the same name.
 | 
						|
    KUBECONFIG_DIR=$(dirname ${KUBECONFIG:-$DEFAULT_KUBECONFIG})
 | 
						|
    CONTEXT=${FEDERATION_KUBE_CONTEXT} \
 | 
						|
	   KUBE_BEARER_TOKEN="$FEDERATION_API_TOKEN" \
 | 
						|
           KUBE_USER="${KUBE_USER}" \
 | 
						|
           KUBE_PASSWORD="${KUBE_PASSWORD}" \
 | 
						|
           KUBECONFIG="${KUBECONFIG_DIR}/federation/federation-apiserver/kubeconfig" \
 | 
						|
	   create-kubeconfig
 | 
						|
 | 
						|
    # Create secret with federation-apiserver's kubeconfig
 | 
						|
    $host_kubectl create secret generic federation-apiserver-kubeconfig --from-file="${KUBECONFIG_DIR}/federation/federation-apiserver/kubeconfig" --namespace="${FEDERATION_NAMESPACE}"
 | 
						|
 | 
						|
    # Create secrets with all the kubernetes-apiserver's kubeconfigs.
 | 
						|
    # Note: This is used only by the test setup (where kubernetes clusters are
 | 
						|
    # brought up with FEDERATION=true). Users are expected to create this secret
 | 
						|
    # themselves.
 | 
						|
    for dir in ${KUBECONFIG_DIR}/federation/kubernetes-apiserver/*; do
 | 
						|
      # We create a secret with the same name as the directory name (which is
 | 
						|
      # same as cluster name in kubeconfig).
 | 
						|
      # Massage the name so that it is valid (should not contain "_" and max 253
 | 
						|
      # chars)
 | 
						|
      name=$(basename $dir)
 | 
						|
      name=$(echo "$name" | sed -e "s/_/-/g")  # Replace "_" by "-"
 | 
						|
      name=${name:0:252}
 | 
						|
      echo "Creating secret with name: $name"
 | 
						|
      $host_kubectl create secret generic ${name} --from-file="${dir}/kubeconfig" --namespace="${FEDERATION_NAMESPACE}"
 | 
						|
    done
 | 
						|
 | 
						|
    # Create server certificates.
 | 
						|
    kube::util::ensure-temp-dir
 | 
						|
    echo "Creating federation apiserver certs for federation api host: ${FEDERATION_API_HOST} ( is this a dns name?: ${IS_DNS_NAME} )"
 | 
						|
    MASTER_NAME="federation-apiserver" create-federation-apiserver-certs ${FEDERATION_API_HOST}
 | 
						|
    export FEDERATION_APISERVER_CA_CERT_BASE64="${FEDERATION_APISERVER_CA_CERT_BASE64}"
 | 
						|
    export FEDERATION_APISERVER_CERT_BASE64="${FEDERATION_APISERVER_CERT_BASE64}"
 | 
						|
    export FEDERATION_APISERVER_KEY_BASE64="${FEDERATION_APISERVER_KEY_BASE64}"
 | 
						|
 | 
						|
    # Enable the NamespaceLifecycle admission control by default.
 | 
						|
    export FEDERATION_ADMISSION_CONTROL="${FEDERATION_ADMISSION_CONTROL:-NamespaceLifecycle}"
 | 
						|
 | 
						|
    for file in federation-etcd-pvc.yaml federation-apiserver-{deployment,secrets}.yaml federation-controller-manager-deployment.yaml; do
 | 
						|
      echo "Creating manifest: ${file}"
 | 
						|
      $template "${manifests_root}/${file}"
 | 
						|
      $template "${manifests_root}/${file}" | $host_kubectl create -f -
 | 
						|
    done
 | 
						|
 | 
						|
    # Update the users kubeconfig to include federation-apiserver credentials.
 | 
						|
    CONTEXT=${FEDERATION_KUBE_CONTEXT} \
 | 
						|
	   KUBE_BEARER_TOKEN="${FEDERATION_API_TOKEN}" \
 | 
						|
           KUBE_USER="${KUBE_USER}" \
 | 
						|
           KUBE_PASSWORD="${KUBE_PASSWORD}" \
 | 
						|
	   SECONDARY_KUBECONFIG=true \
 | 
						|
	   create-kubeconfig
 | 
						|
 | 
						|
    # Don't finish provisioning until federation-apiserver pod is running
 | 
						|
    for i in {1..30};do
 | 
						|
	#TODO(colhom): in the future this needs to scale out for N pods. This assumes just one pod
 | 
						|
	phase="$($host_kubectl get -o=jsonpath pods -lapp=federated-cluster,module=federation-apiserver --template '{.items[*].status.phase}')"
 | 
						|
	echo "Waiting for federation-apiserver to be running...(phase= $phase)"
 | 
						|
	if [[ "$phase" == "Running" ]];then
 | 
						|
	    echo "federation-apiserver pod is running!"
 | 
						|
	    break
 | 
						|
	fi
 | 
						|
 | 
						|
	if [[ $i -eq 30 ]];then
 | 
						|
	    echo "federation-apiserver pod is not running! giving up."
 | 
						|
	    exit 1
 | 
						|
	fi
 | 
						|
 | 
						|
	sleep 4
 | 
						|
    done
 | 
						|
 | 
						|
    # Verify that federation-controller-manager pod is running.
 | 
						|
    for i in {1..30};do
 | 
						|
	#TODO(colhom): in the future this needs to scale out for N pods. This assumes just one pod
 | 
						|
	phase="$($host_kubectl get -o=jsonpath pods -lapp=federated-cluster,module=federation-controller-manager --template '{.items[*].status.phase}')"
 | 
						|
	echo "Waiting for federation-controller-manager to be running...(phase= $phase)"
 | 
						|
	if [[ "$phase" == "Running" ]];then
 | 
						|
	    echo "federation-controller-manager pod is running!"
 | 
						|
	    break
 | 
						|
	fi
 | 
						|
 | 
						|
	if [[ $i -eq 30 ]];then
 | 
						|
	    echo "federation-controller-manager pod is not running! giving up."
 | 
						|
	    exit 1
 | 
						|
	fi
 | 
						|
 | 
						|
	sleep 4
 | 
						|
    done
 | 
						|
)
 | 
						|
}
 | 
						|
 | 
						|
# Creates the required certificates for federation apiserver.
 | 
						|
# $1: The public IP or DNS name for the master.
 | 
						|
#
 | 
						|
# Assumed vars
 | 
						|
#   KUBE_TEMP
 | 
						|
#   MASTER_NAME
 | 
						|
#   IS_DNS_NAME=true|false
 | 
						|
function create-federation-apiserver-certs {
 | 
						|
  local primary_cn
 | 
						|
  local sans
 | 
						|
 | 
						|
  if [[ "${IS_DNS_NAME:-}" == "true" ]];then
 | 
						|
    primary_cn="$(printf "${1}" | sha1sum | tr " -" " ")"
 | 
						|
    sans="DNS:${1},DNS:${MASTER_NAME}"
 | 
						|
  else
 | 
						|
    primary_cn="${1}"
 | 
						|
    sans="IP:${1},DNS:${MASTER_NAME}"
 | 
						|
  fi
 | 
						|
 | 
						|
  echo "Generating certs for alternate-names: ${sans}"
 | 
						|
 | 
						|
  local kube_temp="${KUBE_TEMP}/federation"
 | 
						|
  mkdir -p "${kube_temp}"
 | 
						|
  KUBE_TEMP="${kube_temp}" PRIMARY_CN="${primary_cn}" SANS="${sans}" generate-certs
 | 
						|
 | 
						|
  local cert_dir="${kube_temp}/easy-rsa-master/easyrsa3"
 | 
						|
  # By default, linux wraps base64 output every 76 cols, so we use 'tr -d' to remove whitespaces.
 | 
						|
  # Note 'base64 -w0' doesn't work on Mac OS X, which has different flags.
 | 
						|
  FEDERATION_APISERVER_CA_CERT_BASE64=$(cat "${cert_dir}/pki/ca.crt" | base64 | tr -d '\r\n')
 | 
						|
  FEDERATION_APISERVER_CERT_BASE64=$(cat "${cert_dir}/pki/issued/${MASTER_NAME}.crt" | base64 | tr -d '\r\n')
 | 
						|
  FEDERATION_APISERVER_KEY_BASE64=$(cat "${cert_dir}/pki/private/${MASTER_NAME}.key" | base64 | tr -d '\r\n')
 | 
						|
}
 | 
						|
 | 
						|
 | 
						|
# Required
 | 
						|
# FEDERATION_PUSH_REPO_BASE: the docker repo where federated images will be pushed
 | 
						|
# FEDERATION_IMAGE_TAG: the tag of the image to be pushed
 | 
						|
function push-federation-images {
 | 
						|
    : "${FEDERATION_PUSH_REPO_BASE?Must set FEDERATION_PUSH_REPO_BASE env var}"
 | 
						|
    : "${FEDERATION_IMAGE_TAG?Must set FEDERATION_IMAGE_TAG env var}"
 | 
						|
 | 
						|
    source "${KUBE_ROOT}/build/common.sh"
 | 
						|
    source "${KUBE_ROOT}/hack/lib/util.sh"
 | 
						|
 | 
						|
    local FEDERATION_BINARIES=${FEDERATION_BINARIES:-"hyperkube-amd64"}
 | 
						|
 | 
						|
    local bin_dir="${KUBE_ROOT}/_output/${KUBE_BUILD_STAGE}/server/${KUBE_PLATFORM}-${KUBE_ARCH}/kubernetes/server/bin"
 | 
						|
 | 
						|
    if [[ ! -d "${bin_dir}" ]];then
 | 
						|
        echo "${bin_dir} does not exist! Run make quick-release or make release"
 | 
						|
        exit 1
 | 
						|
    fi
 | 
						|
 | 
						|
    for binary in ${FEDERATION_BINARIES}; do
 | 
						|
        local bin_path="${bin_dir}/${binary}"
 | 
						|
 | 
						|
        if [[ ! -f "${bin_path}" ]]; then
 | 
						|
            echo "${bin_path} does not exist!"
 | 
						|
            exit 1
 | 
						|
        fi
 | 
						|
 | 
						|
        local docker_build_path="${bin_path}.dockerbuild"
 | 
						|
        local docker_file_path="${docker_build_path}/Dockerfile"
 | 
						|
 | 
						|
        rm -rf ${docker_build_path}
 | 
						|
        mkdir -p ${docker_build_path}
 | 
						|
 | 
						|
        ln "${bin_path}" "${docker_build_path}/${binary}"
 | 
						|
        printf " FROM debian:jessie \n ADD ${binary} /usr/local/bin/${binary}\n" > ${docker_file_path}
 | 
						|
 | 
						|
        local docker_image_tag="${FEDERATION_PUSH_REPO_BASE}/${binary}:${FEDERATION_IMAGE_TAG}"
 | 
						|
 | 
						|
        # Build the docker image on-the-fly.
 | 
						|
        #
 | 
						|
        # NOTE: This is only a temporary fix until the proposal in issue
 | 
						|
        # https://github.com/kubernetes/kubernetes/issues/28630 is implemented.
 | 
						|
        # Also, the new turn up mechanism completely obviates this step.
 | 
						|
        #
 | 
						|
        # TODO(madhusudancs): Remove this code when the new turn up mechanism work
 | 
						|
        # is merged.
 | 
						|
        kube::log::status "Building docker image ${docker_image_tag} from the binary"
 | 
						|
        docker build --pull -q -t "${docker_image_tag}" ${docker_build_path} >/dev/null
 | 
						|
 | 
						|
        rm -rf ${docker_build_path}
 | 
						|
 | 
						|
        kube::log::status "Pushing ${docker_image_tag}"
 | 
						|
        if [[ "${FEDERATION_PUSH_REPO_BASE}" == "gcr.io/"* ]]; then
 | 
						|
            echo " -> GCR repository detected. Using gcloud"
 | 
						|
            gcloud docker -- push "${docker_image_tag}"
 | 
						|
        else
 | 
						|
            docker push "${docker_image_tag}"
 | 
						|
        fi
 | 
						|
 | 
						|
        kube::log::status "Deleting docker image ${docker_image_tag}"
 | 
						|
        docker rmi "${docker_image_tag}" 2>/dev/null || true
 | 
						|
    done
 | 
						|
}
 | 
						|
 | 
						|
function cleanup-federation-api-objects {
 | 
						|
  # This is a cleanup function. We cannot stop on errors here. So disable
 | 
						|
  # errexit in this function.
 | 
						|
  set +o errexit
 | 
						|
 | 
						|
  echo "Cleaning Federation control plane objects"
 | 
						|
  # Delete all resources with the federated-cluster label.
 | 
						|
  $host_kubectl delete pods,svc,rc,deployment,secret -lapp=federated-cluster
 | 
						|
 | 
						|
  # Delete all PVs bound to PVCs in FEDERATION_NAMESPACE
 | 
						|
  pvs=$($host_kubectl get pvc --namespace=${FEDERATION_NAMESPACE} -o jsonpath='{.items[*].spec.volumeName}')
 | 
						|
  while $host_kubectl delete pv ${pvs} >/dev/null 2>&1; do
 | 
						|
    sleep 2
 | 
						|
  done
 | 
						|
 | 
						|
  # Delete all resources in FEDERATION_NAMESPACE.
 | 
						|
  $host_kubectl delete pvc,pods,svc,rc,deployment,secret --namespace=${FEDERATION_NAMESPACE} --all
 | 
						|
 | 
						|
  # This is a big hammer. We get rid of federation-system namespace from
 | 
						|
  # all the clusters
 | 
						|
  for context in $(federation_cluster_contexts); do
 | 
						|
    (
 | 
						|
      local -r role="federation-controller-manager:${FEDERATION_NAME}-${context}-${HOST_CLUSTER_CONTEXT}"
 | 
						|
      kube::log::status "Removing namespace \"${FEDERATION_NAMESPACE}\", cluster role \"${role}\" and cluster role binding \"${role}\" from \"${context}\""
 | 
						|
      # Try deleting until the namespace is completely gone.
 | 
						|
      while $host_kubectl --context="${context}" delete namespace "${FEDERATION_NAMESPACE}" >/dev/null 2>&1; do
 | 
						|
        # It is usually slower to remove a namespace because it involves
 | 
						|
        # performing a cascading deletion of all the resources in the
 | 
						|
        # namespace. So we sleep a little longer than other resources
 | 
						|
        # before retrying
 | 
						|
        sleep 5
 | 
						|
      done
 | 
						|
      kube::log::status "Removed namespace \"${FEDERATION_NAMESPACE}\" from \"${context}\""
 | 
						|
 | 
						|
      while $host_kubectl --context="${context}" delete clusterrole "${role}" >/dev/null 2>&1; do
 | 
						|
        sleep 2
 | 
						|
      done
 | 
						|
      kube::log::status "Removed cluster role \"${role}\" from \"${context}\""
 | 
						|
 | 
						|
      while $host_kubectl --context="${context}" delete clusterrolebinding "${role}" >/dev/null 2>&1; do
 | 
						|
        sleep 2
 | 
						|
      done
 | 
						|
      kube::log::status "Removed cluster role binding \"${role}\" from \"${context}\""
 | 
						|
    ) &
 | 
						|
  done
 | 
						|
  wait
 | 
						|
  set -o errexit
 | 
						|
}
 |