diff --git a/cluster/aws/coreos/node.yaml b/cluster/aws/coreos/node.yaml index 286abdea1ac..c85aeda552c 100644 --- a/cluster/aws/coreos/node.yaml +++ b/cluster/aws/coreos/node.yaml @@ -65,12 +65,12 @@ coreos: ExecStartPre=/usr/bin/mkdir -p /etc/kubernetes/manifests ExecStartPre=/run/setup-auth.sh ExecStart=/opt/kubernetes/server/bin/kubelet \ - --api_servers=https://${MASTER_IP} \ + --api-servers=https://${MASTER_IP} \ --config=/etc/kubernetes/manifests \ - --allow_privileged=False \ + --allow-privileged=False \ --v=2 \ - --cluster_dns=10.0.0.10 \ - --cluster_domain=${DNS_DOMAIN} \ + --cluster-dns=10.0.0.10 \ + --cluster-domain=${DNS_DOMAIN} \ --logtostderr=true \ --hostname-override=${HOSTNAME_OVERRIDE} \ --container-runtime=${KUBERNETES_CONTAINER_RUNTIME} diff --git a/cluster/gce/coreos/node.yaml b/cluster/gce/coreos/node.yaml index 955f990d527..0ec352965d5 100644 --- a/cluster/gce/coreos/node.yaml +++ b/cluster/gce/coreos/node.yaml @@ -142,12 +142,12 @@ coreos: ExecStartPre=/usr/bin/mkdir -p /etc/kubernetes/manifests ExecStartPre=/run/setup-auth.sh ExecStart=/opt/kubernetes/server/bin/kubelet \ - --api_servers=https://${KUBERNETES_MASTER_NAME}.c.${PROJECT_ID}.internal \ + --api-servers=https://${KUBERNETES_MASTER_NAME}.c.${PROJECT_ID}.internal \ --config=/etc/kubernetes/manifests \ - --allow_privileged=False \ + --allow-privileged=False \ --v=2 \ - --cluster_dns=10.0.0.10 \ - --cluster_domain=cluster.local \ + --cluster-dns=10.0.0.10 \ + --cluster-domain=cluster.local \ --logtostderr=true \ --container-runtime=${KUBERNETES_CONTAINER_RUNTIME} Restart=always diff --git a/cluster/images/hyperkube/master-multi.json b/cluster/images/hyperkube/master-multi.json index a3c2ccff950..d6fb3574f6f 100644 --- a/cluster/images/hyperkube/master-multi.json +++ b/cluster/images/hyperkube/master-multi.json @@ -23,8 +23,8 @@ "apiserver", "--portal-net=10.0.0.1/24", "--address=0.0.0.0", - "--etcd_servers=http://127.0.0.1:4001", - "--cluster_name=kubernetes", + "--etcd-servers=http://127.0.0.1:4001", + "--cluster-name=kubernetes", "--v=2" ] }, diff --git a/cluster/images/hyperkube/master.json b/cluster/images/hyperkube/master.json index f1e39c9a72b..8b935d14951 100644 --- a/cluster/images/hyperkube/master.json +++ b/cluster/images/hyperkube/master.json @@ -23,8 +23,8 @@ "apiserver", "--portal-net=10.0.0.1/24", "--address=127.0.0.1", - "--etcd_servers=http://127.0.0.1:4001", - "--cluster_name=kubernetes", + "--etcd-servers=http://127.0.0.1:4001", + "--cluster-name=kubernetes", "--v=2" ] }, diff --git a/cluster/juju/charms/trusty/kubernetes-master/files/apiserver.upstart.tmpl b/cluster/juju/charms/trusty/kubernetes-master/files/apiserver.upstart.tmpl index 3f0109eb83a..28d9799978c 100644 --- a/cluster/juju/charms/trusty/kubernetes-master/files/apiserver.upstart.tmpl +++ b/cluster/juju/charms/trusty/kubernetes-master/files/apiserver.upstart.tmpl @@ -9,7 +9,7 @@ kill timeout 30 # wait 30s between SIGTERM and SIGKILL. exec /usr/local/bin/apiserver \ --address=%(api_bind_address)s \ - --etcd_servers=%(etcd_servers)s \ + --etcd-servers=%(etcd_servers)s \ --logtostderr=true \ --service-cluster-ip-range=10.244.240.0/20 diff --git a/cluster/juju/charms/trusty/kubernetes/files/kubelet.upstart.tmpl b/cluster/juju/charms/trusty/kubernetes/files/kubelet.upstart.tmpl index 0aa6cb72349..5b8dc39d50f 100644 --- a/cluster/juju/charms/trusty/kubernetes/files/kubelet.upstart.tmpl +++ b/cluster/juju/charms/trusty/kubernetes/files/kubelet.upstart.tmpl @@ -9,7 +9,7 @@ kill timeout 60 # wait 60s between SIGTERM and SIGKILL. exec /usr/local/bin/kubelet \ --address=%(kubelet_bind_addr)s \ - --api_servers=%(kubeapi_server)s \ - --hostname_override=%(kubelet_bind_addr)s \ - --cadvisor_port=4193 \ + --api-servers=%(kubeapi_server)s \ + --hostname-override=%(kubelet_bind_addr)s \ + --cadvisor-port=4193 \ --logtostderr=true diff --git a/cluster/libvirt-coreos/user_data_master.yml b/cluster/libvirt-coreos/user_data_master.yml index 2e5f0764d4f..cd48387b278 100644 --- a/cluster/libvirt-coreos/user_data_master.yml +++ b/cluster/libvirt-coreos/user_data_master.yml @@ -16,8 +16,8 @@ coreos: ExecStart=/opt/kubernetes/bin/kube-apiserver \ --address=0.0.0.0 \ --port=8080 \ - --etcd_servers=http://127.0.0.1:4001 \ - --kubelet_port=10250 \ + --etcd-servers=http://127.0.0.1:4001 \ + --kubelet-port=10250 \ --service-cluster-ip-range=${SERVICE_CLUSTER_IP_RANGE} Restart=always RestartSec=2 diff --git a/cluster/libvirt-coreos/user_data_minion.yml b/cluster/libvirt-coreos/user_data_minion.yml index 3a843415910..5c9c5c9d8fe 100644 --- a/cluster/libvirt-coreos/user_data_minion.yml +++ b/cluster/libvirt-coreos/user_data_minion.yml @@ -15,10 +15,10 @@ coreos: [Service] ExecStart=/opt/kubernetes/bin/kubelet \ --address=0.0.0.0 \ - --hostname_override=${MINION_IPS[$i]} \ - --api_servers=http://${MASTER_IP}:8080 \ - $( [[ "$ENABLE_CLUSTER_DNS" == "true" ]] && echo "--cluster_dns=${DNS_SERVER_IP}" ) \ - $( [[ "$ENABLE_CLUSTER_DNS" == "true" ]] && echo "--cluster_domain=${DNS_DOMAIN}" ) \ + --hostname-override=${MINION_IPS[$i]} \ + --api-servers=http://${MASTER_IP}:8080 \ + $( [[ "$ENABLE_CLUSTER_DNS" == "true" ]] && echo "--cluster-dns=${DNS_SERVER_IP}" ) \ + $( [[ "$ENABLE_CLUSTER_DNS" == "true" ]] && echo "--cluster-domain=${DNS_DOMAIN}" ) \ --config=/opt/kubernetes/manifests Restart=always RestartSec=2 diff --git a/cluster/rackspace/cloud-config/master-cloud-config.yaml b/cluster/rackspace/cloud-config/master-cloud-config.yaml index a55a132805c..7d9b0e8bf4d 100644 --- a/cluster/rackspace/cloud-config/master-cloud-config.yaml +++ b/cluster/rackspace/cloud-config/master-cloud-config.yaml @@ -87,9 +87,9 @@ coreos: ExecStartPre=/usr/bin/mkdir -p /var/lib/kube-apiserver ExecStart=/opt/bin/kube-apiserver \ --address=127.0.0.1 \ - --cloud_provider=rackspace \ - --cloud_config=/etc/cloud.conf \ - --etcd_servers=http://127.0.0.1:4001 \ + --cloud-provider=rackspace \ + --cloud-config=/etc/cloud.conf \ + --etcd-servers=http://127.0.0.1:4001 \ --logtostderr=true \ --port=8080 \ --service-cluster-ip-range=SERVICE_CLUSTER_IP_RANGE \ @@ -122,8 +122,8 @@ coreos: [Service] ExecStartPre=/usr/bin/ln -sf /opt/kubernetes/server/bin/kube-controller-manager /opt/bin/kube-controller-manager ExecStart=/opt/bin/kube-controller-manager \ - --cloud_provider=rackspace \ - --cloud_config=/etc/cloud.conf \ + --cloud-provider=rackspace \ + --cloud-config=/etc/cloud.conf \ --logtostderr=true \ --master=127.0.0.1:8080 \ --v=2 diff --git a/cluster/rackspace/cloud-config/minion-cloud-config.yaml b/cluster/rackspace/cloud-config/minion-cloud-config.yaml index d1bab0db3f0..0f59551195b 100644 --- a/cluster/rackspace/cloud-config/minion-cloud-config.yaml +++ b/cluster/rackspace/cloud-config/minion-cloud-config.yaml @@ -150,11 +150,11 @@ coreos: ExecStartPre=/usr/bin/ln -sf /opt/kubernetes/server/bin/kubelet /opt/bin/kubelet ExecStart=/opt/bin/kubelet \ --address=$private_ipv4 \ - --api_servers=${FIRST_APISERVER_URL} \ - --cluster_dns=DNS_SERVER_IP \ - --cluster_domain=DNS_DOMAIN \ + --api-servers=${FIRST_APISERVER_URL} \ + --cluster-dns=DNS_SERVER_IP \ + --cluster-domain=DNS_DOMAIN \ --healthz-bind-address=$private_ipv4 \ - --hostname_override=$private_ipv4 \ + --hostname-override=$private_ipv4 \ --logtostderr=true \ --v=2 Restart=always @@ -176,7 +176,7 @@ coreos: ExecStartPre=/run/config-kube-proxy.sh ExecStartPre=/usr/bin/ln -sf /opt/kubernetes/server/bin/kube-proxy /opt/bin/kube-proxy ExecStart=/opt/bin/kube-proxy \ - --bind_address=$private_ipv4 \ + --bind-address=$private_ipv4 \ --kubeconfig=/var/lib/kube-proxy/kubeconfig \ --logtostderr=true \ --master=${FIRST_APISERVER_URL} diff --git a/cluster/saltbase/salt/kube-apiserver/kube-apiserver.manifest b/cluster/saltbase/salt/kube-apiserver/kube-apiserver.manifest index 5f66068102a..0afab9c79c0 100644 --- a/cluster/saltbase/salt/kube-apiserver/kube-apiserver.manifest +++ b/cluster/saltbase/salt/kube-apiserver/kube-apiserver.manifest @@ -9,10 +9,10 @@ {% set cloud_config_volume = "" -%} {% if grains.cloud is defined -%} - {% set cloud_provider = "--cloud_provider=" + grains.cloud -%} + {% set cloud_provider = "--cloud-provider=" + grains.cloud -%} {% if grains.cloud in [ 'aws', 'gce' ] and grains.cloud_config is defined -%} - {% set cloud_config = "--cloud_config=" + grains.cloud_config -%} + {% set cloud_config = "--cloud-config=" + grains.cloud_config -%} {% set cloud_config_mount = "{\"name\": \"cloudconfigmount\",\"mountPath\": \"" + grains.cloud_config + "\", \"readOnly\": true}," -%} {% set cloud_config_volume = "{\"name\": \"cloudconfigmount\",\"hostPath\": {\"path\": \"" + grains.cloud_config + "\"}}," -%} {% endif -%} @@ -35,7 +35,7 @@ {% set cluster_name = "" -%} {% if pillar['instance_prefix'] is defined -%} - {% set cluster_name = "--cluster_name=" + pillar['instance_prefix'] -%} + {% set cluster_name = "--cluster-name=" + pillar['instance_prefix'] -%} {% endif -%} {% set bind_address = "" -%} @@ -43,24 +43,24 @@ {% set bind_address = "--bind-address=" + grains.publicAddressOverride -%} {% endif -%} -{% set etcd_servers = "--etcd_servers=http://127.0.0.1:4001" -%} +{% set etcd_servers = "--etcd-servers=http://127.0.0.1:4001" -%} {% set service_cluster_ip_range = "" -%} {% if pillar['service_cluster_ip_range'] is defined -%} {% set service_cluster_ip_range = "--service-cluster-ip-range=" + pillar['service_cluster_ip_range'] -%} {% endif -%} -{% set cert_file = "--tls_cert_file=/srv/kubernetes/server.cert" -%} -{% set key_file = "--tls_private_key_file=/srv/kubernetes/server.key" -%} +{% set cert_file = "--tls-cert-file=/srv/kubernetes/server.cert" -%} +{% set key_file = "--tls-private-key-file=/srv/kubernetes/server.key" -%} {% set client_ca_file = "" -%} {% set secure_port = "6443" -%} {% if grains['cloud'] is defined and grains.cloud in [ 'aws', 'gce', 'vagrant' ] %} {% set secure_port = "443" -%} - {% set client_ca_file = "--client_ca_file=/srv/kubernetes/ca.crt" -%} + {% set client_ca_file = "--client-ca-file=/srv/kubernetes/ca.crt" -%} {% endif -%} -{% set token_auth_file = "--token_auth_file=/dev/null" -%} +{% set token_auth_file = "--token-auth-file=/dev/null" -%} {% set basic_auth_file = "" -%} {% set min_request_timeout = "" -%} @@ -70,26 +70,26 @@ {% if grains.cloud is defined -%} {% if grains.cloud in [ 'aws', 'gce', 'vagrant' ] -%} - {% set token_auth_file = "--token_auth_file=/srv/kubernetes/known_tokens.csv" -%} + {% set token_auth_file = "--token-auth-file=/srv/kubernetes/known_tokens.csv" -%} {% endif -%} {% endif -%} {% if grains['cloud'] is defined and grains.cloud in [ 'aws', 'gce', 'vagrant'] %} - {% set basic_auth_file = "--basic_auth_file=/srv/kubernetes/basic_auth.csv" -%} + {% set basic_auth_file = "--basic-auth-file=/srv/kubernetes/basic_auth.csv" -%} {% endif -%} {% set admission_control = "" -%} {% if pillar['admission_control'] is defined -%} - {% set admission_control = "--admission_control=" + pillar['admission_control'] -%} + {% set admission_control = "--admission-control=" + pillar['admission_control'] -%} {% endif -%} {% set runtime_config = "" -%} {% if grains.runtime_config is defined -%} - {% set runtime_config = "--runtime_config=" + grains.runtime_config -%} + {% set runtime_config = "--runtime-config=" + grains.runtime_config -%} {% endif -%} {% set params = address + " " + etcd_servers + " " + cloud_provider + " " + cloud_config + " " + runtime_config + " " + admission_control + " " + service_cluster_ip_range + " " + client_ca_file + " " + basic_auth_file + " " + min_request_timeout -%} -{% set params = params + " " + cluster_name + " " + cert_file + " " + key_file + " --secure_port=" + secure_port + " " + token_auth_file + " " + bind_address + " " + pillar['log_level'] + " " + advertise_address + " " + proxy_ssh_options -%} +{% set params = params + " " + cluster_name + " " + cert_file + " " + key_file + " --secure-port=" + secure_port + " " + token_auth_file + " " + bind_address + " " + pillar['log_level'] + " " + advertise_address + " " + proxy_ssh_options -%} # test_args has to be kept at the end, so they'll overwrite any prior configuration {% if pillar['apiserver_test_args'] is defined -%} @@ -117,7 +117,7 @@ "command": [ "/bin/sh", "-c", - "/usr/local/bin/kube-apiserver {{params}} --allow_privileged={{pillar['allow_privileged']}} 1>>/var/log/kube-apiserver.log 2>&1" + "/usr/local/bin/kube-apiserver {{params}} --allow-privileged={{pillar['allow_privileged']}} 1>>/var/log/kube-apiserver.log 2>&1" ], "livenessProbe": { "httpGet": { diff --git a/cluster/saltbase/salt/kube-controller-manager/kube-controller-manager.manifest b/cluster/saltbase/salt/kube-controller-manager/kube-controller-manager.manifest index 1725030fe06..7c0e214213a 100644 --- a/cluster/saltbase/salt/kube-controller-manager/kube-controller-manager.manifest +++ b/cluster/saltbase/salt/kube-controller-manager/kube-controller-manager.manifest @@ -3,7 +3,7 @@ {% set allocate_node_cidrs = "" -%} {% if pillar['instance_prefix'] is defined -%} - {% set cluster_name = "--cluster_name=" + pillar['instance_prefix'] -%} + {% set cluster_name = "--cluster-name=" + pillar['instance_prefix'] -%} {% endif -%} {% if pillar['cluster_cidr'] is defined and pillar['cluster_cidr'] != "" -%} {% set cluster_cidr = "--cluster-cidr=" + pillar['cluster_cidr'] -%} @@ -18,11 +18,11 @@ {% set cloud_config_volume = "" -%} {% if grains.cloud is defined -%} - {% set cloud_provider = "--cloud_provider=" + grains.cloud -%} - {% set service_account_key = " --service_account_private_key_file=/srv/kubernetes/server.key " -%} + {% set cloud_provider = "--cloud-provider=" + grains.cloud -%} + {% set service_account_key = " --service-account-private-key-file=/srv/kubernetes/server.key " -%} {% if grains.cloud in [ 'aws', 'gce' ] and grains.cloud_config is defined -%} - {% set cloud_config = "--cloud_config=" + grains.cloud_config -%} + {% set cloud_config = "--cloud-config=" + grains.cloud_config -%} {% set cloud_config_mount = "{\"name\": \"cloudconfigmount\",\"mountPath\": \"" + grains.cloud_config + "\", \"readOnly\": true}," -%} {% set cloud_config_volume = "{\"name\": \"cloudconfigmount\",\"hostPath\": {\"path\": \"" + grains.cloud_config + "\"}}," -%} {% endif -%} @@ -31,7 +31,7 @@ {% set root_ca_file = "" -%} {% if grains['cloud'] is defined and grains.cloud in [ 'aws', 'gce', 'vagrant' ] %} - {% set root_ca_file = "--root_ca_file=/srv/kubernetes/ca.crt" -%} + {% set root_ca_file = "--root-ca-file=/srv/kubernetes/ca.crt" -%} {% endif -%} {% set params = "--master=127.0.0.1:8080" + " " + cluster_name + " " + cluster_cidr + " " + allocate_node_cidrs + " " + cloud_provider + " " + cloud_config + service_account_key + pillar['log_level'] + " " + root_ca_file -%} diff --git a/cluster/saltbase/salt/kubelet/default b/cluster/saltbase/salt/kubelet/default index 48fddbf8e5a..6a9dcdc7d21 100644 --- a/cluster/saltbase/salt/kubelet/default +++ b/cluster/saltbase/salt/kubelet/default @@ -4,15 +4,15 @@ {% endif -%} {% if grains.api_servers is defined -%} - {% set api_servers = "--api_servers=https://" + grains.api_servers -%} + {% set api_servers = "--api-servers=https://" + grains.api_servers -%} {% elif grains.apiservers is defined -%} # TODO(remove after 0.16.0): Deprecated form - {% set api_servers = "--api_servers=https://" + grains.apiservers -%} + {% set api_servers = "--api-servers=https://" + grains.apiservers -%} {% elif grains['roles'][0] == 'kubernetes-master' -%} {% set master_ipv4 = salt['grains.get']('fqdn_ip4')[0] -%} - {% set api_servers = "--api_servers=https://" + master_ipv4 -%} + {% set api_servers = "--api-servers=https://" + master_ipv4 -%} {% else -%} {% set ips = salt['mine.get']('roles:kubernetes-master', 'network.ip_addrs', 'grain').values() -%} - {% set api_servers = "--api_servers=https://" + ips[0][0] -%} + {% set api_servers = "--api-servers=https://" + ips[0][0] -%} {% endif -%} # TODO: remove nginx for other cloud providers. @@ -44,7 +44,7 @@ {% set cloud_provider = "" -%} {% if grains.cloud is defined -%} - {% set cloud_provider = "--cloud_provider=" + grains.cloud -%} + {% set cloud_provider = "--cloud-provider=" + grains.cloud -%} {% endif -%} {% set config = "--config=/etc/kubernetes/manifests" -%} @@ -56,24 +56,24 @@ {% set hostname_override = "" -%} {% if grains.hostname_override is defined -%} - {% set hostname_override = " --hostname_override=" + grains.hostname_override -%} + {% set hostname_override = " --hostname-override=" + grains.hostname_override -%} {% endif -%} {% set cluster_dns = "" %} {% set cluster_domain = "" %} {% if pillar.get('enable_cluster_dns', '').lower() == 'true' %} - {% set cluster_dns = "--cluster_dns=" + pillar['dns_server'] %} - {% set cluster_domain = "--cluster_domain=" + pillar['dns_domain'] %} + {% set cluster_dns = "--cluster-dns=" + pillar['dns_server'] %} + {% set cluster_domain = "--cluster-domain=" + pillar['dns_domain'] %} {% endif %} {% set docker_root = "" -%} {% if grains.docker_root is defined -%} - {% set docker_root = " --docker_root=" + grains.docker_root -%} + {% set docker_root = " --docker-root=" + grains.docker_root -%} {% endif -%} {% set kubelet_root = "" -%} {% if grains.kubelet_root is defined -%} - {% set kubelet_root = " --root_dir=" + grains.kubelet_root -%} + {% set kubelet_root = " --root-dir=" + grains.kubelet_root -%} {% endif -%} {% set configure_cbr0 = "" -%} @@ -86,10 +86,10 @@ {% set cgroup_root = "" -%} {% if grains['os_family'] == 'Debian' -%} {% set system_container = "--system-container=/system" -%} - {% set cgroup_root = "--cgroup_root=/" -%} + {% set cgroup_root = "--cgroup-root=/" -%} {% endif -%} {% if grains['oscodename'] == 'vivid' -%} - {% set cgroup_root = "--cgroup_root=docker" -%} + {% set cgroup_root = "--cgroup-root=docker" -%} {% endif -%} {% set pod_cidr = "" %} @@ -103,4 +103,4 @@ {% endif -%} # test_args has to be kept at the end, so they'll overwrite any prior configuration -DAEMON_ARGS="{{daemon_args}} {{api_servers_with_port}} {{debugging_handlers}} {{hostname_override}} {{cloud_provider}} {{config}} {{manifest_url}} --allow_privileged={{pillar['allow_privileged']}} {{pillar['log_level']}} {{cluster_dns}} {{cluster_domain}} {{docker_root}} {{kubelet_root}} {{configure_cbr0}} {{cgroup_root}} {{system_container}} {{pod_cidr}} {{test_args}}" +DAEMON_ARGS="{{daemon_args}} {{api_servers_with_port}} {{debugging_handlers}} {{hostname_override}} {{cloud_provider}} {{config}} {{manifest_url}} --allow-privileged={{pillar['allow_privileged']}} {{pillar['log_level']}} {{cluster_dns}} {{cluster_domain}} {{docker_root}} {{kubelet_root}} {{configure_cbr0}} {{cgroup_root}} {{system_container}} {{pod_cidr}} {{test_args}}" diff --git a/cmd/kubelet/app/server.go b/cmd/kubelet/app/server.go index 5dd01372ff2..6bb9d4a952a 100644 --- a/cmd/kubelet/app/server.go +++ b/cmd/kubelet/app/server.go @@ -215,7 +215,7 @@ func (s *KubeletServer) AddFlags(fs *pflag.FlagSet) { fs.BoolVar(&s.AllowPrivileged, "allow-privileged", s.AllowPrivileged, "If true, allow containers to request privileged mode. [default=false]") fs.StringVar(&s.HostNetworkSources, "host-network-sources", s.HostNetworkSources, "Comma-separated list of sources from which the Kubelet allows pods to use of host network. For all sources use \"*\" [default=\"file\"]") fs.Float64Var(&s.RegistryPullQPS, "registry-qps", s.RegistryPullQPS, "If > 0, limit registry pull QPS to this value. If 0, unlimited. [default=0.0]") - fs.IntVar(&s.RegistryBurst, "registry-burst", s.RegistryBurst, "Maximum size of a bursty pulls, temporarily allows pulls to burst to this number, while still not exceeding registry_qps. Only used if --registry-qps > 0") + fs.IntVar(&s.RegistryBurst, "registry-burst", s.RegistryBurst, "Maximum size of a bursty pulls, temporarily allows pulls to burst to this number, while still not exceeding registry-qps. Only used if --registry-qps > 0") fs.BoolVar(&s.RunOnce, "runonce", s.RunOnce, "If true, exit after spawning pods from local manifests or remote urls. Exclusive with --api-servers, and --enable-server") fs.BoolVar(&s.EnableDebuggingHandlers, "enable-debugging-handlers", s.EnableDebuggingHandlers, "Enables server endpoints for log collection and local running of containers and commands") fs.DurationVar(&s.MinimumGCAge, "minimum-container-ttl-duration", s.MinimumGCAge, "Minimum age for a finished container before it is garbage collected. Examples: '300ms', '10s' or '2h45m'") @@ -227,7 +227,7 @@ func (s *KubeletServer) AddFlags(fs *pflag.FlagSet) { fs.UintVar(&s.CadvisorPort, "cadvisor-port", s.CadvisorPort, "The port of the localhost cAdvisor endpoint") fs.IntVar(&s.HealthzPort, "healthz-port", s.HealthzPort, "The port of the localhost healthz endpoint") fs.Var(&s.HealthzBindAddress, "healthz-bind-address", "The IP address for the healthz server to serve on, defaulting to 127.0.0.1 (set to 0.0.0.0 for all interfaces)") - fs.IntVar(&s.OOMScoreAdj, "oom-score-adj", s.OOMScoreAdj, "The oom_score_adj value for kubelet process. Values must be within the range [-1000, 1000]") + fs.IntVar(&s.OOMScoreAdj, "oom-score-adj", s.OOMScoreAdj, "The oom-score-adj value for kubelet process. Values must be within the range [-1000, 1000]") fs.StringSliceVar(&s.APIServerList, "api-servers", []string{}, "List of Kubernetes API servers for publishing events, and reading pods and services. (ip:port), comma separated.") fs.BoolVar(&s.RegisterNode, "register-node", s.RegisterNode, "Register the node with the apiserver (defaults to true if --api-server is set)") fs.StringVar(&s.ClusterDomain, "cluster-domain", s.ClusterDomain, "Domain for this cluster. If set, kubelet will configure all containers to search this domain in addition to the host's search domains") @@ -243,8 +243,8 @@ func (s *KubeletServer) AddFlags(fs *pflag.FlagSet) { fs.StringVar(&s.CloudProvider, "cloud-provider", s.CloudProvider, "The provider for cloud services. Empty string for no provider.") fs.StringVar(&s.CloudConfigFile, "cloud-config", s.CloudConfigFile, "The path to the cloud provider configuration file. Empty string for no configuration file.") fs.StringVar(&s.ResourceContainer, "resource-container", s.ResourceContainer, "Absolute name of the resource-only container to create and run the Kubelet in (Default: /kubelet).") - fs.StringVar(&s.CgroupRoot, "cgroup_root", s.CgroupRoot, "Optional root cgroup to use for pods. This is handled by the container runtime on a best effort basis. Default: '', which means use the container runtime default.") - fs.StringVar(&s.ContainerRuntime, "container_runtime", s.ContainerRuntime, "The container runtime to use. Possible values: 'docker', 'rkt'. Default: 'docker'.") + fs.StringVar(&s.CgroupRoot, "cgroup-root", s.CgroupRoot, "Optional root cgroup to use for pods. This is handled by the container runtime on a best effort basis. Default: '', which means use the container runtime default.") + fs.StringVar(&s.ContainerRuntime, "container-runtime", s.ContainerRuntime, "The container runtime to use. Possible values: 'docker', 'rkt'. Default: 'docker'.") fs.StringVar(&s.SystemContainer, "system-container", s.SystemContainer, "Optional resource-only container in which to place all non-kernel processes that are not already in a container. Empty for no container. Rolling back the flag requires a reboot. (Default: \"\").") fs.BoolVar(&s.ConfigureCBR0, "configure-cbr0", s.ConfigureCBR0, "If true, kubelet will configure cbr0 based on Node.Spec.PodCIDR.") fs.IntVar(&s.MaxPods, "max-pods", 40, "Number of Pods that can run on this Kubelet.") diff --git a/contrib/ansible/roles/kubernetes/files/make-ca-cert.sh b/contrib/ansible/roles/kubernetes/files/make-ca-cert.sh index fec89bd7fac..81df11e15be 100755 --- a/contrib/ansible/roles/kubernetes/files/make-ca-cert.sh +++ b/contrib/ansible/roles/kubernetes/files/make-ca-cert.sh @@ -21,7 +21,7 @@ set -o pipefail # Caller should set in the ev: # MASTER_IP - this may be an ip or things like "_use_gce_external_ip_" # MASTER_NAME - DNS name for the master -# DNS_DOMAIN - which will be passed to minions in --cluster_domain +# DNS_DOMAIN - which will be passed to minions in --cluster-domain # SERVICE_CLUSTER_IP_RANGE - where all service IPs are allocated # Also the following will be respected diff --git a/contrib/ansible/roles/kubernetes/templates/config.j2 b/contrib/ansible/roles/kubernetes/templates/config.j2 index cf83970267e..aadc92979f8 100644 --- a/contrib/ansible/roles/kubernetes/templates/config.j2 +++ b/contrib/ansible/roles/kubernetes/templates/config.j2 @@ -17,7 +17,7 @@ KUBE_LOGTOSTDERR="--logtostderr=true" KUBE_LOG_LEVEL="--v=0" # Should this cluster be allowed to run privileged docker containers -KUBE_ALLOW_PRIV="--allow_privileged=true" +KUBE_ALLOW_PRIV="--allow-privileged=true" # How the replication controller, scheduler, and proxy KUBE_MASTER="--master=https://{{ groups['masters'][0] }}:{{ kube_master_api_port }}" diff --git a/contrib/ansible/roles/master/templates/apiserver.j2 b/contrib/ansible/roles/master/templates/apiserver.j2 index eb42ef09427..aff516ca150 100644 --- a/contrib/ansible/roles/master/templates/apiserver.j2 +++ b/contrib/ansible/roles/master/templates/apiserver.j2 @@ -11,16 +11,16 @@ KUBE_API_ADDRESS="--insecure-bind-address=127.0.0.1" KUBE_API_PORT="--secure-port={{ kube_master_api_port }}" # Port nodes listen on -# KUBELET_PORT="--kubelet_port=10250" +# KUBELET_PORT="--kubelet-port=10250" # Address range to use for services KUBE_SERVICE_ADDRESSES="--service-cluster-ip-range={{ kube_service_addresses }}" # Location of the etcd cluster -KUBE_ETCD_SERVERS="--etcd_servers={% for node in groups['etcd'] %}http://{{ node }}:2379{% if not loop.last %},{% endif %}{% endfor %}" +KUBE_ETCD_SERVERS="--etcd-servers={% for node in groups['etcd'] %}http://{{ node }}:2379{% if not loop.last %},{% endif %}{% endfor %}" # default admission control policies -KUBE_ADMISSION_CONTROL="--admission_control=NamespaceLifecycle,NamespaceExists,LimitRanger,SecurityContextDeny,ServiceAccount,ResourceQuota" +KUBE_ADMISSION_CONTROL="--admission-control=NamespaceLifecycle,NamespaceExists,LimitRanger,SecurityContextDeny,ServiceAccount,ResourceQuota" # Add your own! -KUBE_API_ARGS="--tls_cert_file={{ kube_cert_dir }}/server.crt --tls_private_key_file={{ kube_cert_dir }}/server.key --client_ca_file={{ kube_cert_dir }}/ca.crt --token_auth_file={{ kube_token_dir }}/known_tokens.csv --service_account_key_file={{ kube_cert_dir }}/server.crt" +KUBE_API_ARGS="--tls-cert-file={{ kube_cert_dir }}/server.crt --tls-private-key-file={{ kube_cert_dir }}/server.key --client-ca-file={{ kube_cert_dir }}/ca.crt --token-auth-file={{ kube_token_dir }}/known_tokens.csv --service-account-key-file={{ kube_cert_dir }}/server.crt" diff --git a/contrib/ansible/roles/master/templates/controller-manager.j2 b/contrib/ansible/roles/master/templates/controller-manager.j2 index 389a60fa0e3..0dde136fbd5 100644 --- a/contrib/ansible/roles/master/templates/controller-manager.j2 +++ b/contrib/ansible/roles/master/templates/controller-manager.j2 @@ -4,4 +4,4 @@ # defaults from config and apiserver should be adequate # Add your own! -KUBE_CONTROLLER_MANAGER_ARGS="--kubeconfig={{ kube_config_dir }}/controller-manager.kubeconfig --service_account_private_key_file={{ kube_cert_dir }}/server.key --root_ca_file={{ kube_cert_dir }}/ca.crt" +KUBE_CONTROLLER_MANAGER_ARGS="--kubeconfig={{ kube_config_dir }}/controller-manager.kubeconfig --service-account-private-key-file={{ kube_cert_dir }}/server.key --root-ca-file={{ kube_cert_dir }}/ca.crt" diff --git a/contrib/ansible/roles/node/templates/kubelet.j2 b/contrib/ansible/roles/node/templates/kubelet.j2 index 3e9029f6839..283b3e514c2 100644 --- a/contrib/ansible/roles/node/templates/kubelet.j2 +++ b/contrib/ansible/roles/node/templates/kubelet.j2 @@ -8,14 +8,14 @@ KUBELET_ADDRESS="--address=0.0.0.0" # KUBELET_PORT="--port=10250" # You may leave this blank to use the actual hostname -KUBELET_HOSTNAME="--hostname_override={{ inventory_hostname }}" +KUBELET_HOSTNAME="--hostname-override={{ inventory_hostname }}" # location of the api-server -KUBELET_API_SERVER="--api_servers=https://{{ groups['masters'][0] }}:{{ kube_master_api_port }}" +KUBELET_API_SERVER="--api-servers=https://{{ groups['masters'][0] }}:{{ kube_master_api_port }}" # Add your own! {% if dns_setup %} -KUBELET_ARGS="--cluster_dns={{ dns_server }} --cluster_domain={{ dns_domain }} --kubeconfig={{ kube_config_dir}}/kubelet.kubeconfig --config={{ kube_manifest_dir }}" +KUBELET_ARGS="--cluster-dns={{ dns_server }} --cluster-domain={{ dns_domain }} --kubeconfig={{ kube_config_dir}}/kubelet.kubeconfig --config={{ kube_manifest_dir }}" {% else %} KUBELET_ARGS="--kubeconfig={{ kube_config_dir }}/kubelet.kubeconfig --config={{ kube_manifest_dir }}" {% endif %} diff --git a/contrib/init/systemd/environ/apiserver b/contrib/init/systemd/environ/apiserver index 3a2da7d1aff..223b99203c7 100644 --- a/contrib/init/systemd/environ/apiserver +++ b/contrib/init/systemd/environ/apiserver @@ -11,16 +11,16 @@ KUBE_API_ADDRESS="--insecure-bind-address=127.0.0.1" # KUBE_API_PORT="--port=8080" # Port minions listen on -# KUBELET_PORT="--kubelet_port=10250" +# KUBELET_PORT="--kubelet-port=10250" # Comma separated list of nodes in the etcd cluster -KUBE_ETCD_SERVERS="--etcd_servers=http://127.0.0.1:4001" +KUBE_ETCD_SERVERS="--etcd-servers=http://127.0.0.1:4001" # Address range to use for services KUBE_SERVICE_ADDRESSES="--service-cluster-ip-range=10.254.0.0/16" # default admission control policies -KUBE_ADMISSION_CONTROL="--admission_control=NamespaceLifecycle,NamespaceExists,LimitRanger,SecurityContextDeny,ServiceAccount,ResourceQuota" +KUBE_ADMISSION_CONTROL="--admission-control=NamespaceLifecycle,NamespaceExists,LimitRanger,SecurityContextDeny,ServiceAccount,ResourceQuota" # Add your own! KUBE_API_ARGS="" diff --git a/contrib/init/systemd/environ/config b/contrib/init/systemd/environ/config index b90a756a531..8c0a28493fd 100644 --- a/contrib/init/systemd/environ/config +++ b/contrib/init/systemd/environ/config @@ -16,7 +16,7 @@ KUBE_LOGTOSTDERR="--logtostderr=true" KUBE_LOG_LEVEL="--v=0" # Should this cluster be allowed to run privileged docker containers -KUBE_ALLOW_PRIV="--allow_privileged=false" +KUBE_ALLOW_PRIV="--allow-privileged=false" # How the controller-manager, scheduler, and proxy find the apiserver KUBE_MASTER="--master=http://127.0.0.1:8080" diff --git a/contrib/init/systemd/environ/kubelet b/contrib/init/systemd/environ/kubelet index 64a7b25c454..db93a5eeb42 100644 --- a/contrib/init/systemd/environ/kubelet +++ b/contrib/init/systemd/environ/kubelet @@ -8,10 +8,10 @@ KUBELET_ADDRESS="--address=127.0.0.1" # KUBELET_PORT="--port=10250" # You may leave this blank to use the actual hostname -KUBELET_HOSTNAME="--hostname_override=127.0.0.1" +KUBELET_HOSTNAME="--hostname-override=127.0.0.1" # location of the api-server -KUBELET_API_SERVER="--api_servers=http://127.0.0.1:8080" +KUBELET_API_SERVER="--api-servers=http://127.0.0.1:8080" # Add your own! KUBELET_ARGS="" diff --git a/docs/admin/high-availability/default-kubelet b/docs/admin/high-availability/default-kubelet index d581ec5e70d..41ee5301510 100644 --- a/docs/admin/high-availability/default-kubelet +++ b/docs/admin/high-availability/default-kubelet @@ -3,6 +3,6 @@ MASTER_IP= # This should be the internal service IP address reserved for DNS DNS_IP= -DAEMON_ARGS="$DAEMON_ARGS --api_servers=https://${MASTER_IP} --enable-debugging-handlers=true --cloud_provider= -gce --config=/etc/kubernetes/manifests --allow_privileged=False --v=2 --cluster_dns=${DNS_IP} --cluster_domain=c -luster.local --configure-cbr0=true --cgroup_root=/ --system-container=/system " \ No newline at end of file +DAEMON_ARGS="$DAEMON_ARGS --api-servers=https://${MASTER_IP} --enable-debugging-handlers=true --cloud-provider= +gce --config=/etc/kubernetes/manifests --allow-privileged=False --v=2 --cluster-dns=${DNS_IP} --cluster-domain=c +luster.local --configure-cbr0=true --cgroup-root=/ --system-container=/system " diff --git a/docs/admin/high-availability/kube-apiserver.yaml b/docs/admin/high-availability/kube-apiserver.yaml index 3bfe65f9620..f5081925aa3 100644 --- a/docs/admin/high-availability/kube-apiserver.yaml +++ b/docs/admin/high-availability/kube-apiserver.yaml @@ -10,13 +10,13 @@ spec: command: - /bin/sh - -c - - /usr/local/bin/kube-apiserver --address=127.0.0.1 --etcd_servers=http://127.0.0.1:4001 - --cloud_provider=gce --admission_control=NamespaceLifecycle,NamespaceExists,LimitRanger,SecurityContextDeny,ServiceAccount,ResourceQuota - --service-cluster-ip-range=10.0.0.0/16 --client_ca_file=/srv/kubernetes/ca.crt - --basic_auth_file=/srv/kubernetes/basic_auth.csv --cluster_name=e2e-test-bburns - --tls_cert_file=/srv/kubernetes/server.cert --tls_private_key_file=/srv/kubernetes/server.key - --secure_port=443 --token_auth_file=/srv/kubernetes/known_tokens.csv --v=2 - --allow_privileged=False 1>>/var/log/kube-apiserver.log 2>&1 + - /usr/local/bin/kube-apiserver --address=127.0.0.1 --etcd-servers=http://127.0.0.1:4001 + --cloud-provider=gce --admission-control=NamespaceLifecycle,NamespaceExists,LimitRanger,SecurityContextDeny,ServiceAccount,ResourceQuota + --service-cluster-ip-range=10.0.0.0/16 --client-ca-file=/srv/kubernetes/ca.crt + --basic-auth-file=/srv/kubernetes/basic_auth.csv --cluster-name=e2e-test-bburns + --tls-cert-file=/srv/kubernetes/server.cert --tls-private-key-file=/srv/kubernetes/server.key + --secure-port=443 --token-auth-file=/srv/kubernetes/known_tokens.csv --v=2 + --allow-privileged=False 1>>/var/log/kube-apiserver.log 2>&1 ports: - containerPort: 443 hostPort: 443 diff --git a/docs/admin/high-availability/kube-controller-manager.yaml b/docs/admin/high-availability/kube-controller-manager.yaml index 09acc340504..736181fc7c2 100644 --- a/docs/admin/high-availability/kube-controller-manager.yaml +++ b/docs/admin/high-availability/kube-controller-manager.yaml @@ -7,8 +7,8 @@ spec: - command: - /bin/sh - -c - - /usr/local/bin/kube-controller-manager --master=127.0.0.1:8080 --cluster_name=e2e-test-bburns - --cluster-cidr=10.245.0.0/16 --allocate-node-cidrs=true --cloud_provider=gce --service_account_private_key_file=/srv/kubernetes/server.key + - /usr/local/bin/kube-controller-manager --master=127.0.0.1:8080 --cluster-name=e2e-test-bburns + --cluster-cidr=10.245.0.0/16 --allocate-node-cidrs=true --cloud-provider=gce --service-account-private-key-file=/srv/kubernetes/server.key --v=2 1>>/var/log/kube-controller-manager.log 2>&1 image: gcr.io/google_containers/kube-controller-manager:fda24638d51a48baa13c35337fcd4793 livenessProbe: diff --git a/docs/getting-started-guides/coreos/azure/cloud_config_templates/kubernetes-cluster-main-nodes-template.yml b/docs/getting-started-guides/coreos/azure/cloud_config_templates/kubernetes-cluster-main-nodes-template.yml index fcf2c6c573d..1e4305133a2 100644 --- a/docs/getting-started-guides/coreos/azure/cloud_config_templates/kubernetes-cluster-main-nodes-template.yml +++ b/docs/getting-started-guides/coreos/azure/cloud_config_templates/kubernetes-cluster-main-nodes-template.yml @@ -288,11 +288,11 @@ coreos: ExecStart=/opt/kubernetes/server/bin/kubelet \ --address=0.0.0.0 \ --port=10250 \ - --hostname_override=%H \ - --api_servers=http://kube-00:8080 \ + --hostname-override=%H \ + --api-servers=http://kube-00:8080 \ --logtostderr=true \ - --cluster_dns=10.1.0.3 \ - --cluster_domain=kube.local \ + --cluster-dns=10.1.0.3 \ + --cluster-domain=kube.local \ --config=/etc/kubernetes/manifests/ Restart=always RestartSec=10 diff --git a/docs/getting-started-guides/coreos/azure/lib/deployment_logic/kubernetes.js b/docs/getting-started-guides/coreos/azure/lib/deployment_logic/kubernetes.js index dfeca7eb628..2002b43a53e 100644 --- a/docs/getting-started-guides/coreos/azure/lib/deployment_logic/kubernetes.js +++ b/docs/getting-started-guides/coreos/azure/lib/deployment_logic/kubernetes.js @@ -38,7 +38,7 @@ etcd_initial_cluster_conf_kube = function (conf) { 'name': 'kube-apiserver.service', 'drop-ins': [{ 'name': '50-etcd-initial-cluster.conf', - 'content': _.template("[Service]\nEnvironment=ETCD_SERVERS=--etcd_servers=<%= nodes.join(',') %>\n")(data), + 'content': _.template("[Service]\nEnvironment=ETCD_SERVERS=--etcd-servers=<%= nodes.join(',') %>\n")(data), }], }; }; diff --git a/docs/getting-started-guides/coreos/cloud-configs/master.yaml b/docs/getting-started-guides/coreos/cloud-configs/master.yaml index af4ee5c30ff..195d7b8773d 100644 --- a/docs/getting-started-guides/coreos/cloud-configs/master.yaml +++ b/docs/getting-started-guides/coreos/cloud-configs/master.yaml @@ -87,18 +87,18 @@ coreos: ExecStartPre=/usr/bin/chmod +x /opt/bin/kube-apiserver ExecStartPre=/opt/bin/wupiao 127.0.0.1:2379/v2/machines ExecStart=/opt/bin/kube-apiserver \ - --service_account_key_file=/opt/bin/kube-serviceaccount.key \ - --service_account_lookup=false \ - --admission_control=NamespaceLifecycle,NamespaceAutoProvision,LimitRanger,SecurityContextDeny,ServiceAccount,ResourceQuota \ - --runtime_config=api/v1 \ - --allow_privileged=true \ - --insecure_bind_address=0.0.0.0 \ - --insecure_port=8080 \ - --kubelet_https=true \ - --secure_port=6443 \ + --service-account-key-file=/opt/bin/kube-serviceaccount.key \ + --service-account-lookup=false \ + --admission-control=NamespaceLifecycle,NamespaceAutoProvision,LimitRanger,SecurityContextDeny,ServiceAccount,ResourceQuota \ + --runtime-config=api/v1 \ + --allow-privileged=true \ + --insecure-bind-address=0.0.0.0 \ + --insecure-port=8080 \ + --kubelet-https=true \ + --secure-port=6443 \ --service-cluster-ip-range=10.100.0.0/16 \ - --etcd_servers=http://127.0.0.1:2379 \ - --public_address_override=${DEFAULT_IPV4} \ + --etcd-servers=http://127.0.0.1:2379 \ + --public-address-override=${DEFAULT_IPV4} \ --logtostderr=true Restart=always RestartSec=10 @@ -115,7 +115,7 @@ coreos: ExecStartPre=/usr/bin/curl -L -o /opt/bin/kube-controller-manager -z /opt/bin/kube-controller-manager https://storage.googleapis.com/kubernetes-release/release/v1.0.1/bin/linux/amd64/kube-controller-manager ExecStartPre=/usr/bin/chmod +x /opt/bin/kube-controller-manager ExecStart=/opt/bin/kube-controller-manager \ - --service_account_private_key_file=/opt/bin/kube-serviceaccount.key \ + --service-account-private-key-file=/opt/bin/kube-serviceaccount.key \ --master=127.0.0.1:8080 \ --logtostderr=true Restart=always diff --git a/docs/getting-started-guides/coreos/cloud-configs/node.yaml b/docs/getting-started-guides/coreos/cloud-configs/node.yaml index 58a33812980..1b4dc2e48cd 100644 --- a/docs/getting-started-guides/coreos/cloud-configs/node.yaml +++ b/docs/getting-started-guides/coreos/cloud-configs/node.yaml @@ -84,13 +84,13 @@ coreos: ExecStart=/opt/bin/kubelet \ --address=0.0.0.0 \ --port=10250 \ - --hostname_override=${DEFAULT_IPV4} \ - --api_servers=:8080 \ - --allow_privileged=true \ + --hostname-override=${DEFAULT_IPV4} \ + --api-servers=:8080 \ + --allow-privileged=true \ --logtostderr=true \ - --cadvisor_port=4194 \ - --healthz_bind_address=0.0.0.0 \ - --healthz_port=10248 + --cadvisor-port=4194 \ + --healthz-bind-address=0.0.0.0 \ + --healthz-port=10248 Restart=always RestartSec=10 update: diff --git a/examples/high-availability/etc_kubernetes_kubelet b/examples/high-availability/etc_kubernetes_kubelet index 6c963f416c7..26d32a2a970 100644 --- a/examples/high-availability/etc_kubernetes_kubelet +++ b/examples/high-availability/etc_kubernetes_kubelet @@ -9,13 +9,13 @@ KUBELET_ADDRESS="" # KUBELET_PORT="--port=10250" # You may leave this blank to use the actual hostname -# KUBELET_HOSTNAME="--hostname_override=0.0.0." +# KUBELET_HOSTNAME="--hostname-override=0.0.0." # location of the api-server -KUBELET_API_SERVER="--api_servers=http://0.0.0.0:8080,kube1.ha:8080,kube0.ha:8080 " -# --cert-dir="/var/run/kubernetes": The directory where the TLS certs are located (by default /var/run/kubernetes). If --tls_cert_file and --tls_private_key_file are provided, this flag will be ignored. -# --tls-cert-file="": File containing x509 Certificate for HTTPS. (CA cert, if any, concatenated after server cert). If --tls_cert_file and --tls_private_key_file are not provided, a self-signed certificate and key are generated for the public address and saved to the directory passed to --cert_dir. -# --tls-private-key-file="": File containing x509 private key matching --tls_cert_file. +KUBELET_API_SERVER="--api-servers=http://0.0.0.0:8080,kube1.ha:8080,kube0.ha:8080 " +# --cert-dir="/var/run/kubernetes": The directory where the TLS certs are located (by default /var/run/kubernetes). If --tls-cert-file and --tls-private-key-file are provided, this flag will be ignored. +# --tls-cert-file="": File containing x509 Certificate for HTTPS. (CA cert, if any, concatenated after server cert). If --tls-cert-file and --tls-private-key-file are not provided, a self-signed certificate and key are generated for the public address and saved to the directory passed to --cert-dir. +# --tls-private-key-file="": File containing x509 private key matching --tls-cert-file. # We modify kubelet args to do verbose logs + read from manifests dir. KUBELET_ARGS="--tls-cert-file=/vagrant/apiserver.crt --tls-private-key-file=/vagrant/apiserver.key --register-node=true --v=5 --config=/etc/kubernetes/manifests --kubeconfig=/vagrant/kubeconfig" diff --git a/examples/high-availability/kube-apiserver.manifest b/examples/high-availability/kube-apiserver.manifest index 99f0f6f7c46..2e9488c3c62 100644 --- a/examples/high-availability/kube-apiserver.manifest +++ b/examples/high-availability/kube-apiserver.manifest @@ -14,7 +14,7 @@ "command": [ "/bin/sh", "-c", - "/usr/local/bin/kube-apiserver --address=0.0.0.0 --etcd_servers=http://kube0.ha:2379 --service-cluster-ip-range=10.0.0.0/16 --v=4 --allow_privileged=True 1>>/var/log/kube-apiserver.log 2>&1" + "/usr/local/bin/kube-apiserver --address=0.0.0.0 --etcd-servers=http://kube0.ha:2379 --service-cluster-ip-range=10.0.0.0/16 --v=4 --allow-privileged=True 1>>/var/log/kube-apiserver.log 2>&1" ], "livenessProbe": { "httpGet": { diff --git a/examples/high-availability/kube-controller-manager.manifest b/examples/high-availability/kube-controller-manager.manifest index 4d358da5d32..209c8821513 100644 --- a/examples/high-availability/kube-controller-manager.manifest +++ b/examples/high-availability/kube-controller-manager.manifest @@ -14,7 +14,7 @@ "command": [ "/bin/sh", "-c", - "/usr/local/bin/kube-controller-manager --master=192.168.4.102:8080 --service_account_private_key_file=/srv/kubernetes/server.key --v=4 1>>/var/log/kube-controller-manager.log 2>&1" + "/usr/local/bin/kube-controller-manager --master=192.168.4.102:8080 --service-account-private-key-file=/srv/kubernetes/server.key --v=4 1>>/var/log/kube-controller-manager.log 2>&1" ], "livenessProbe": { "httpGet": { diff --git a/hack/local-up-cluster.sh b/hack/local-up-cluster.sh index 49387258373..7f1adcc7cba 100755 --- a/hack/local-up-cluster.sh +++ b/hack/local-up-cluster.sh @@ -213,15 +213,15 @@ function start_apiserver { APISERVER_LOG=/tmp/kube-apiserver.log sudo -E "${GO_OUT}/kube-apiserver" ${priv_arg}\ --v=${LOG_LEVEL} \ - --cert_dir="${CERT_DIR}" \ - --service_account_key_file="${SERVICE_ACCOUNT_KEY}" \ - --service_account_lookup="${SERVICE_ACCOUNT_LOOKUP}" \ - --admission_control="${ADMISSION_CONTROL}" \ + --cert-dir="${CERT_DIR}" \ + --service-account-key-file="${SERVICE_ACCOUNT_KEY}" \ + --service-account-lookup="${SERVICE_ACCOUNT_LOOKUP}" \ + --admission-control="${ADMISSION_CONTROL}" \ --address="${API_HOST}" \ --port="${API_PORT}" \ - --etcd_servers="http://127.0.0.1:4001" \ + --etcd-servers="http://127.0.0.1:4001" \ --service-cluster-ip-range="10.0.0.0/24" \ - --cors_allowed_origins="${API_CORS_ALLOWED_ORIGINS}" >"${APISERVER_LOG}" 2>&1 & + --cors-allowed-origins="${API_CORS_ALLOWED_ORIGINS}" >"${APISERVER_LOG}" 2>&1 & APISERVER_PID=$! # Wait for kube-apiserver to come up before launching the rest of the components. @@ -233,8 +233,8 @@ function start_controller_manager { CTLRMGR_LOG=/tmp/kube-controller-manager.log sudo -E "${GO_OUT}/kube-controller-manager" \ --v=${LOG_LEVEL} \ - --service_account_private_key_file="${SERVICE_ACCOUNT_KEY}" \ - --root_ca_file="${ROOT_CA_FILE}" \ + --service-account-private-key-file="${SERVICE_ACCOUNT_KEY}" \ + --root-ca-file="${ROOT_CA_FILE}" \ --master="${API_HOST}:${API_PORT}" >"${CTLRMGR_LOG}" 2>&1 & CTLRMGR_PID=$! } @@ -244,11 +244,11 @@ function start_kubelet { if [[ -z "${DOCKERIZE_KUBELET}" ]]; then sudo -E "${GO_OUT}/kubelet" ${priv_arg}\ --v=${LOG_LEVEL} \ - --chaos_chance="${CHAOS_CHANCE}" \ - --container_runtime="${CONTAINER_RUNTIME}" \ - --hostname_override="127.0.0.1" \ + --chaos-chance="${CHAOS_CHANCE}" \ + --container-runtime="${CONTAINER_RUNTIME}" \ + --hostname-override="127.0.0.1" \ --address="127.0.0.1" \ - --api_servers="${API_HOST}:${API_PORT}" \ + --api-servers="${API_HOST}:${API_PORT}" \ --port="$KUBELET_PORT" >"${KUBELET_LOG}" 2>&1 & KUBELET_PID=$! else diff --git a/hack/test-cmd.sh b/hack/test-cmd.sh index 1efea1f6e3b..5e7bd483053 100755 --- a/hack/test-cmd.sh +++ b/hack/test-cmd.sh @@ -99,29 +99,29 @@ kube::log::status "Running kubectl with no options" kube::log::status "Starting kubelet in masterless mode" "${KUBE_OUTPUT_HOSTBIN}/kubelet" \ - --really_crash_for_testing=true \ - --root_dir=/tmp/kubelet.$$ \ - --cert_dir="${TMPDIR:-/tmp/}" \ - --docker_endpoint="fake://" \ - --hostname_override="127.0.0.1" \ + --really-crash-for-testing=true \ + --root-dir=/tmp/kubelet.$$ \ + --cert-dir="${TMPDIR:-/tmp/}" \ + --docker-endpoint="fake://" \ + --hostname-override="127.0.0.1" \ --address="127.0.0.1" \ --port="$KUBELET_PORT" \ - --healthz_port="${KUBELET_HEALTHZ_PORT}" 1>&2 & + --healthz-port="${KUBELET_HEALTHZ_PORT}" 1>&2 & KUBELET_PID=$! kube::util::wait_for_url "http://127.0.0.1:${KUBELET_HEALTHZ_PORT}/healthz" "kubelet(masterless)" kill ${KUBELET_PID} 1>&2 2>/dev/null kube::log::status "Starting kubelet in masterful mode" "${KUBE_OUTPUT_HOSTBIN}/kubelet" \ - --really_crash_for_testing=true \ - --root_dir=/tmp/kubelet.$$ \ - --cert_dir="${TMPDIR:-/tmp/}" \ - --docker_endpoint="fake://" \ - --hostname_override="127.0.0.1" \ + --really-crash-for-testing=true \ + --root-dir=/tmp/kubelet.$$ \ + --cert-dir="${TMPDIR:-/tmp/}" \ + --docker-endpoint="fake://" \ + --hostname-override="127.0.0.1" \ --address="127.0.0.1" \ - --api_servers="${API_HOST}:${API_PORT}" \ + --api-servers="${API_HOST}:${API_PORT}" \ --port="$KUBELET_PORT" \ - --healthz_port="${KUBELET_HEALTHZ_PORT}" 1>&2 & + --healthz-port="${KUBELET_HEALTHZ_PORT}" 1>&2 & KUBELET_PID=$! kube::util::wait_for_url "http://127.0.0.1:${KUBELET_HEALTHZ_PORT}/healthz" "kubelet" @@ -130,13 +130,13 @@ kube::util::wait_for_url "http://127.0.0.1:${KUBELET_HEALTHZ_PORT}/healthz" "kub kube::log::status "Starting kube-apiserver" KUBE_API_VERSIONS="v1" "${KUBE_OUTPUT_HOSTBIN}/kube-apiserver" \ --address="127.0.0.1" \ - --public_address_override="127.0.0.1" \ + --public-address-override="127.0.0.1" \ --port="${API_PORT}" \ - --etcd_servers="http://${ETCD_HOST}:${ETCD_PORT}" \ - --public_address_override="127.0.0.1" \ - --kubelet_port=${KUBELET_PORT} \ - --runtime_config=api/v1 \ - --cert_dir="${TMPDIR:-/tmp/}" \ + --etcd-servers="http://${ETCD_HOST}:${ETCD_PORT}" \ + --public-address-override="127.0.0.1" \ + --kubelet-port=${KUBELET_PORT} \ + --runtime-config=api/v1 \ + --cert-dir="${TMPDIR:-/tmp/}" \ --service-cluster-ip-range="10.0.0.0/24" 1>&2 & APISERVER_PID=$! diff --git a/hack/test-update-storage-objects.sh b/hack/test-update-storage-objects.sh index fcd58c622df..31a8cde237f 100755 --- a/hack/test-update-storage-objects.sh +++ b/hack/test-update-storage-objects.sh @@ -46,13 +46,13 @@ function startApiServer() { KUBE_API_VERSIONS="${KUBE_API_VERSIONS}" \ "${KUBE_OUTPUT_HOSTBIN}/kube-apiserver" \ --address="127.0.0.1" \ - --public_address_override="127.0.0.1" \ + --public-address-override="127.0.0.1" \ --port="${API_PORT}" \ - --etcd_servers="http://${ETCD_HOST}:${ETCD_PORT}" \ - --public_address_override="127.0.0.1" \ - --kubelet_port=${KUBELET_PORT} \ - --runtime_config="${RUNTIME_CONFIG}" \ - --cert_dir="${TMPDIR:-/tmp/}" \ + --etcd-servers="http://${ETCD_HOST}:${ETCD_PORT}" \ + --public-address-override="127.0.0.1" \ + --kubelet-port=${KUBELET_PORT} \ + --runtime-config="${RUNTIME_CONFIG}" \ + --cert-dir="${TMPDIR:-/tmp/}" \ --service-cluster-ip-range="10.0.0.0/24" 1>&2 & APISERVER_PID=$! diff --git a/hack/update-swagger-spec.sh b/hack/update-swagger-spec.sh index 3ec4303d544..18a7e1c2271 100755 --- a/hack/update-swagger-spec.sh +++ b/hack/update-swagger-spec.sh @@ -51,12 +51,12 @@ KUBELET_PORT=${KUBELET_PORT:-10250} kube::log::status "Starting kube-apiserver" KUBE_API_VERSIONS="v1" "${KUBE_OUTPUT_HOSTBIN}/kube-apiserver" \ --address="127.0.0.1" \ - --public_address_override="127.0.0.1" \ + --public-address-override="127.0.0.1" \ --port="${API_PORT}" \ - --etcd_servers="http://${ETCD_HOST}:${ETCD_PORT}" \ - --public_address_override="127.0.0.1" \ - --kubelet_port=${KUBELET_PORT} \ - --runtime_config=api/v1 \ + --etcd-servers="http://${ETCD_HOST}:${ETCD_PORT}" \ + --public-address-override="127.0.0.1" \ + --kubelet-port=${KUBELET_PORT} \ + --runtime-config=api/v1 \ --service-cluster-ip-range="10.0.0.0/24" >/dev/null 2>&1 & APISERVER_PID=$!