mirror of
https://github.com/k3s-io/kubernetes.git
synced 2025-10-22 06:59:03 +00:00
For historical reasons InitConfiguration is used almost everywhere in kubeadm as a carrier of various configuration components such as ClusterConfiguration, local API server endpoint, node registration settings, etc. Since v1alpha2, InitConfiguration is meant to be used solely as a way to supply the kubeadm init configuration from a config file. Its usage outside of this context is caused by technical dept, it's clunky and requires hacks to fetch a working InitConfiguration from the cluster (as it's not stored in the config map in its entirety). This change is a small step towards removing all unnecessary usages of InitConfiguration. It reduces its usage by replacing it in some places with some of the following: - ClusterConfiguration only. - APIEndpoint (as local API server endpoint). - NodeRegistrationOptions only. - Some combinations of the above types, or if single fields from them are used, only those field. Signed-off-by: Rostislav M. Georgiev <rostislavg@vmware.com>
337 lines
15 KiB
Go
337 lines
15 KiB
Go
/*
|
|
Copyright 2016 The Kubernetes Authors.
|
|
|
|
Licensed under the Apache License, Version 2.0 (the "License");
|
|
you may not use this file except in compliance with the License.
|
|
You may obtain a copy of the License at
|
|
|
|
http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
Unless required by applicable law or agreed to in writing, software
|
|
distributed under the License is distributed on an "AS IS" BASIS,
|
|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
See the License for the specific language governing permissions and
|
|
limitations under the License.
|
|
*/
|
|
|
|
package controlplane
|
|
|
|
import (
|
|
"fmt"
|
|
"net"
|
|
"os"
|
|
"path/filepath"
|
|
"strconv"
|
|
"strings"
|
|
|
|
"github.com/pkg/errors"
|
|
"k8s.io/api/core/v1"
|
|
"k8s.io/apimachinery/pkg/util/intstr"
|
|
"k8s.io/apimachinery/pkg/util/version"
|
|
"k8s.io/klog"
|
|
kubeadmapi "k8s.io/kubernetes/cmd/kubeadm/app/apis/kubeadm"
|
|
kubeadmconstants "k8s.io/kubernetes/cmd/kubeadm/app/constants"
|
|
"k8s.io/kubernetes/cmd/kubeadm/app/images"
|
|
certphase "k8s.io/kubernetes/cmd/kubeadm/app/phases/certs"
|
|
kubeadmutil "k8s.io/kubernetes/cmd/kubeadm/app/util"
|
|
staticpodutil "k8s.io/kubernetes/cmd/kubeadm/app/util/staticpod"
|
|
authzmodes "k8s.io/kubernetes/pkg/kubeapiserver/authorizer/modes"
|
|
)
|
|
|
|
// CreateInitStaticPodManifestFiles will write all static pod manifest files needed to bring up the control plane.
|
|
func CreateInitStaticPodManifestFiles(manifestDir string, cfg *kubeadmapi.InitConfiguration) error {
|
|
klog.V(1).Infoln("[control-plane] creating static Pod files")
|
|
return CreateStaticPodFiles(manifestDir, &cfg.ClusterConfiguration, &cfg.LocalAPIEndpoint, kubeadmconstants.KubeAPIServer, kubeadmconstants.KubeControllerManager, kubeadmconstants.KubeScheduler)
|
|
}
|
|
|
|
// GetStaticPodSpecs returns all staticPodSpecs actualized to the context of the current configuration
|
|
// NB. this methods holds the information about how kubeadm creates static pod manifests.
|
|
func GetStaticPodSpecs(cfg *kubeadmapi.ClusterConfiguration, endpoint *kubeadmapi.APIEndpoint, k8sVersion *version.Version) map[string]v1.Pod {
|
|
// Get the required hostpath mounts
|
|
mounts := getHostPathVolumesForTheControlPlane(cfg)
|
|
|
|
// Prepare static pod specs
|
|
staticPodSpecs := map[string]v1.Pod{
|
|
kubeadmconstants.KubeAPIServer: staticpodutil.ComponentPod(v1.Container{
|
|
Name: kubeadmconstants.KubeAPIServer,
|
|
Image: images.GetKubernetesImage(kubeadmconstants.KubeAPIServer, cfg),
|
|
ImagePullPolicy: v1.PullIfNotPresent,
|
|
Command: getAPIServerCommand(cfg, endpoint),
|
|
VolumeMounts: staticpodutil.VolumeMountMapToSlice(mounts.GetVolumeMounts(kubeadmconstants.KubeAPIServer)),
|
|
LivenessProbe: livenessProbe(staticpodutil.GetAPIServerProbeAddress(endpoint), int(endpoint.BindPort), v1.URISchemeHTTPS),
|
|
Resources: staticpodutil.ComponentResources("250m"),
|
|
Env: getProxyEnvVars(),
|
|
}, mounts.GetVolumes(kubeadmconstants.KubeAPIServer)),
|
|
kubeadmconstants.KubeControllerManager: staticpodutil.ComponentPod(v1.Container{
|
|
Name: kubeadmconstants.KubeControllerManager,
|
|
Image: images.GetKubernetesImage(kubeadmconstants.KubeControllerManager, cfg),
|
|
ImagePullPolicy: v1.PullIfNotPresent,
|
|
Command: getControllerManagerCommand(cfg, k8sVersion),
|
|
VolumeMounts: staticpodutil.VolumeMountMapToSlice(mounts.GetVolumeMounts(kubeadmconstants.KubeControllerManager)),
|
|
LivenessProbe: livenessProbe(staticpodutil.GetControllerManagerProbeAddress(cfg), 10252, v1.URISchemeHTTP),
|
|
Resources: staticpodutil.ComponentResources("200m"),
|
|
Env: getProxyEnvVars(),
|
|
}, mounts.GetVolumes(kubeadmconstants.KubeControllerManager)),
|
|
kubeadmconstants.KubeScheduler: staticpodutil.ComponentPod(v1.Container{
|
|
Name: kubeadmconstants.KubeScheduler,
|
|
Image: images.GetKubernetesImage(kubeadmconstants.KubeScheduler, cfg),
|
|
ImagePullPolicy: v1.PullIfNotPresent,
|
|
Command: getSchedulerCommand(cfg),
|
|
VolumeMounts: staticpodutil.VolumeMountMapToSlice(mounts.GetVolumeMounts(kubeadmconstants.KubeScheduler)),
|
|
LivenessProbe: livenessProbe(staticpodutil.GetSchedulerProbeAddress(cfg), 10251, v1.URISchemeHTTP),
|
|
Resources: staticpodutil.ComponentResources("100m"),
|
|
Env: getProxyEnvVars(),
|
|
}, mounts.GetVolumes(kubeadmconstants.KubeScheduler)),
|
|
}
|
|
return staticPodSpecs
|
|
}
|
|
|
|
func livenessProbe(host string, port int, scheme v1.URIScheme) *v1.Probe {
|
|
return &v1.Probe{
|
|
Handler: v1.Handler{
|
|
HTTPGet: &v1.HTTPGetAction{
|
|
Host: host,
|
|
Path: "/healthz",
|
|
Port: intstr.FromInt(port),
|
|
Scheme: scheme,
|
|
},
|
|
},
|
|
InitialDelaySeconds: 15,
|
|
TimeoutSeconds: 15,
|
|
FailureThreshold: 8,
|
|
}
|
|
}
|
|
|
|
// CreateStaticPodFiles creates all the requested static pod files.
|
|
func CreateStaticPodFiles(manifestDir string, cfg *kubeadmapi.ClusterConfiguration, endpoint *kubeadmapi.APIEndpoint, componentNames ...string) error {
|
|
// TODO: Move the "pkg/util/version".Version object into the internal API instead of always parsing the string
|
|
k8sVersion, err := version.ParseSemantic(cfg.KubernetesVersion)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
// gets the StaticPodSpecs, actualized for the current ClusterConfiguration
|
|
klog.V(1).Infoln("[control-plane] getting StaticPodSpecs")
|
|
specs := GetStaticPodSpecs(cfg, endpoint, k8sVersion)
|
|
|
|
// creates required static pod specs
|
|
for _, componentName := range componentNames {
|
|
// retrives the StaticPodSpec for given component
|
|
spec, exists := specs[componentName]
|
|
if !exists {
|
|
return errors.Errorf("couldn't retrive StaticPodSpec for %q", componentName)
|
|
}
|
|
|
|
// writes the StaticPodSpec to disk
|
|
if err := staticpodutil.WriteStaticPodToDisk(componentName, manifestDir, spec); err != nil {
|
|
return errors.Wrapf(err, "failed to create static pod manifest file for %q", componentName)
|
|
}
|
|
|
|
klog.V(1).Infof("[control-plane] wrote static Pod manifest for component %q to %q\n", componentName, kubeadmconstants.GetStaticPodFilepath(componentName, manifestDir))
|
|
}
|
|
|
|
return nil
|
|
}
|
|
|
|
// getAPIServerCommand builds the right API server command from the given config object and version
|
|
func getAPIServerCommand(cfg *kubeadmapi.ClusterConfiguration, localAPIEndpoint *kubeadmapi.APIEndpoint) []string {
|
|
defaultArguments := map[string]string{
|
|
"advertise-address": localAPIEndpoint.AdvertiseAddress,
|
|
"enable-admission-plugins": "NodeRestriction",
|
|
"service-cluster-ip-range": cfg.Networking.ServiceSubnet,
|
|
"service-account-key-file": filepath.Join(cfg.CertificatesDir, kubeadmconstants.ServiceAccountPublicKeyName),
|
|
"client-ca-file": filepath.Join(cfg.CertificatesDir, kubeadmconstants.CACertName),
|
|
"tls-cert-file": filepath.Join(cfg.CertificatesDir, kubeadmconstants.APIServerCertName),
|
|
"tls-private-key-file": filepath.Join(cfg.CertificatesDir, kubeadmconstants.APIServerKeyName),
|
|
"kubelet-client-certificate": filepath.Join(cfg.CertificatesDir, kubeadmconstants.APIServerKubeletClientCertName),
|
|
"kubelet-client-key": filepath.Join(cfg.CertificatesDir, kubeadmconstants.APIServerKubeletClientKeyName),
|
|
"enable-bootstrap-token-auth": "true",
|
|
"secure-port": fmt.Sprintf("%d", localAPIEndpoint.BindPort),
|
|
"allow-privileged": "true",
|
|
"kubelet-preferred-address-types": "InternalIP,ExternalIP,Hostname",
|
|
// add options to configure the front proxy. Without the generated client cert, this will never be useable
|
|
// so add it unconditionally with recommended values
|
|
"requestheader-username-headers": "X-Remote-User",
|
|
"requestheader-group-headers": "X-Remote-Group",
|
|
"requestheader-extra-headers-prefix": "X-Remote-Extra-",
|
|
"requestheader-client-ca-file": filepath.Join(cfg.CertificatesDir, kubeadmconstants.FrontProxyCACertName),
|
|
"requestheader-allowed-names": "front-proxy-client",
|
|
"proxy-client-cert-file": filepath.Join(cfg.CertificatesDir, kubeadmconstants.FrontProxyClientCertName),
|
|
"proxy-client-key-file": filepath.Join(cfg.CertificatesDir, kubeadmconstants.FrontProxyClientKeyName),
|
|
}
|
|
|
|
command := []string{"kube-apiserver"}
|
|
|
|
// If the user set endpoints for an external etcd cluster
|
|
if cfg.Etcd.External != nil {
|
|
defaultArguments["etcd-servers"] = strings.Join(cfg.Etcd.External.Endpoints, ",")
|
|
|
|
// Use any user supplied etcd certificates
|
|
if cfg.Etcd.External.CAFile != "" {
|
|
defaultArguments["etcd-cafile"] = cfg.Etcd.External.CAFile
|
|
}
|
|
if cfg.Etcd.External.CertFile != "" && cfg.Etcd.External.KeyFile != "" {
|
|
defaultArguments["etcd-certfile"] = cfg.Etcd.External.CertFile
|
|
defaultArguments["etcd-keyfile"] = cfg.Etcd.External.KeyFile
|
|
}
|
|
} else {
|
|
// Default to etcd static pod on localhost
|
|
defaultArguments["etcd-servers"] = fmt.Sprintf("https://127.0.0.1:%d", kubeadmconstants.EtcdListenClientPort)
|
|
defaultArguments["etcd-cafile"] = filepath.Join(cfg.CertificatesDir, kubeadmconstants.EtcdCACertName)
|
|
defaultArguments["etcd-certfile"] = filepath.Join(cfg.CertificatesDir, kubeadmconstants.APIServerEtcdClientCertName)
|
|
defaultArguments["etcd-keyfile"] = filepath.Join(cfg.CertificatesDir, kubeadmconstants.APIServerEtcdClientKeyName)
|
|
|
|
// Apply user configurations for local etcd
|
|
if cfg.Etcd.Local != nil {
|
|
if value, ok := cfg.Etcd.Local.ExtraArgs["advertise-client-urls"]; ok {
|
|
defaultArguments["etcd-servers"] = value
|
|
}
|
|
}
|
|
}
|
|
|
|
if cfg.APIServer.ExtraArgs == nil {
|
|
cfg.APIServer.ExtraArgs = map[string]string{}
|
|
}
|
|
cfg.APIServer.ExtraArgs["authorization-mode"] = getAuthzModes(cfg.APIServer.ExtraArgs["authorization-mode"])
|
|
command = append(command, kubeadmutil.BuildArgumentListFromMap(defaultArguments, cfg.APIServer.ExtraArgs)...)
|
|
|
|
return command
|
|
}
|
|
|
|
// getAuthzModes gets the authorization-related parameters to the api server
|
|
// Node,RBAC should be fixed in this order at the beginning
|
|
// AlwaysAllow and AlwaysDeny is ignored as they are only for testing
|
|
func getAuthzModes(authzModeExtraArgs string) string {
|
|
modes := []string{
|
|
authzmodes.ModeNode,
|
|
authzmodes.ModeRBAC,
|
|
}
|
|
if strings.Contains(authzModeExtraArgs, authzmodes.ModeABAC) {
|
|
modes = append(modes, authzmodes.ModeABAC)
|
|
}
|
|
if strings.Contains(authzModeExtraArgs, authzmodes.ModeWebhook) {
|
|
modes = append(modes, authzmodes.ModeWebhook)
|
|
}
|
|
return strings.Join(modes, ",")
|
|
}
|
|
|
|
// calcNodeCidrSize determines the size of the subnets used on each node, based
|
|
// on the pod subnet provided. For IPv4, we assume that the pod subnet will
|
|
// be /16 and use /24. If the pod subnet cannot be parsed, the IPv4 value will
|
|
// be used (/24).
|
|
//
|
|
// For IPv6, the algorithm will do two three. First, the node CIDR will be set
|
|
// to a multiple of 8, using the available bits for easier readability by user.
|
|
// Second, the number of nodes will be 512 to 64K to attempt to maximize the
|
|
// number of nodes (see NOTE below). Third, pod networks of /113 and larger will
|
|
// be rejected, as the amount of bits available is too small.
|
|
//
|
|
// A special case is when the pod network size is /112, where /120 will be used,
|
|
// only allowing 256 nodes and 256 pods.
|
|
//
|
|
// If the pod network size is /113 or larger, the node CIDR will be set to the same
|
|
// size and this will be rejected later in validation.
|
|
//
|
|
// NOTE: Currently, the design allows a maximum of 64K nodes. This algorithm splits
|
|
// the available bits to maximize the number used for nodes, but still have the node
|
|
// CIDR be a multiple of eight.
|
|
//
|
|
func calcNodeCidrSize(podSubnet string) string {
|
|
maskSize := "24"
|
|
if ip, podCidr, err := net.ParseCIDR(podSubnet); err == nil {
|
|
if ip.To4() == nil {
|
|
var nodeCidrSize int
|
|
podNetSize, totalBits := podCidr.Mask.Size()
|
|
switch {
|
|
case podNetSize == 112:
|
|
// Special case, allows 256 nodes, 256 pods/node
|
|
nodeCidrSize = 120
|
|
case podNetSize < 112:
|
|
// Use multiple of 8 for node CIDR, with 512 to 64K nodes
|
|
nodeCidrSize = totalBits - ((totalBits-podNetSize-1)/8-1)*8
|
|
default:
|
|
// Not enough bits, will fail later, when validate
|
|
nodeCidrSize = podNetSize
|
|
}
|
|
maskSize = strconv.Itoa(nodeCidrSize)
|
|
}
|
|
}
|
|
return maskSize
|
|
}
|
|
|
|
// getControllerManagerCommand builds the right controller manager command from the given config object and version
|
|
func getControllerManagerCommand(cfg *kubeadmapi.ClusterConfiguration, k8sVersion *version.Version) []string {
|
|
|
|
kubeconfigFile := filepath.Join(kubeadmconstants.KubernetesDir, kubeadmconstants.ControllerManagerKubeConfigFileName)
|
|
caFile := filepath.Join(cfg.CertificatesDir, kubeadmconstants.CACertName)
|
|
|
|
defaultArguments := map[string]string{
|
|
"bind-address": "127.0.0.1",
|
|
"leader-elect": "true",
|
|
"kubeconfig": kubeconfigFile,
|
|
"authentication-kubeconfig": kubeconfigFile,
|
|
"authorization-kubeconfig": kubeconfigFile,
|
|
"client-ca-file": caFile,
|
|
"requestheader-client-ca-file": filepath.Join(cfg.CertificatesDir, kubeadmconstants.FrontProxyCACertName),
|
|
"root-ca-file": caFile,
|
|
"service-account-private-key-file": filepath.Join(cfg.CertificatesDir, kubeadmconstants.ServiceAccountPrivateKeyName),
|
|
"cluster-signing-cert-file": caFile,
|
|
"cluster-signing-key-file": filepath.Join(cfg.CertificatesDir, kubeadmconstants.CAKeyName),
|
|
"use-service-account-credentials": "true",
|
|
"controllers": "*,bootstrapsigner,tokencleaner",
|
|
}
|
|
|
|
// If using external CA, pass empty string to controller manager instead of ca.key/ca.crt path,
|
|
// so that the csrsigning controller fails to start
|
|
if res, _ := certphase.UsingExternalCA(cfg); res {
|
|
defaultArguments["cluster-signing-key-file"] = ""
|
|
defaultArguments["cluster-signing-cert-file"] = ""
|
|
}
|
|
|
|
// Let the controller-manager allocate Node CIDRs for the Pod network.
|
|
// Each node will get a subspace of the address CIDR provided with --pod-network-cidr.
|
|
if cfg.Networking.PodSubnet != "" {
|
|
maskSize := calcNodeCidrSize(cfg.Networking.PodSubnet)
|
|
defaultArguments["allocate-node-cidrs"] = "true"
|
|
defaultArguments["cluster-cidr"] = cfg.Networking.PodSubnet
|
|
defaultArguments["node-cidr-mask-size"] = maskSize
|
|
}
|
|
|
|
command := []string{"kube-controller-manager"}
|
|
command = append(command, kubeadmutil.BuildArgumentListFromMap(defaultArguments, cfg.ControllerManager.ExtraArgs)...)
|
|
|
|
return command
|
|
}
|
|
|
|
// getSchedulerCommand builds the right scheduler command from the given config object and version
|
|
func getSchedulerCommand(cfg *kubeadmapi.ClusterConfiguration) []string {
|
|
defaultArguments := map[string]string{
|
|
"bind-address": "127.0.0.1",
|
|
"leader-elect": "true",
|
|
"kubeconfig": filepath.Join(kubeadmconstants.KubernetesDir, kubeadmconstants.SchedulerKubeConfigFileName),
|
|
}
|
|
|
|
command := []string{"kube-scheduler"}
|
|
command = append(command, kubeadmutil.BuildArgumentListFromMap(defaultArguments, cfg.Scheduler.ExtraArgs)...)
|
|
return command
|
|
}
|
|
|
|
// getProxyEnvVars builds a list of environment variables to use in the control plane containers in order to use the right proxy
|
|
func getProxyEnvVars() []v1.EnvVar {
|
|
envs := []v1.EnvVar{}
|
|
for _, env := range os.Environ() {
|
|
pos := strings.Index(env, "=")
|
|
if pos == -1 {
|
|
// malformed environment variable, skip it.
|
|
continue
|
|
}
|
|
name := env[:pos]
|
|
value := env[pos+1:]
|
|
if strings.HasSuffix(strings.ToLower(name), "_proxy") && value != "" {
|
|
envVar := v1.EnvVar{Name: name, Value: value}
|
|
envs = append(envs, envVar)
|
|
}
|
|
}
|
|
return envs
|
|
}
|