diff --git a/vendor.conf b/vendor.conf index 96f6e71d..dcd2256b 100644 --- a/vendor.conf +++ b/vendor.conf @@ -30,4 +30,4 @@ github.com/mattn/go-colorable efa589957cd060542a26d2dd7832fd6a6c6c3ad github.com/mattn/go-isatty 6ca4dbf54d38eea1a992b3c722a76a5d1c4cb25c github.com/rancher/norman f5744043a6fb81330ee78e4f7a0f04d0ef65c9f1 github.com/rancher/types 45caae4c520880a0164cd220ef2baf9ae96047b1 -github.com/rancher/kontainer-driver-metadata 821544f898e7bcbc0a475d2f0afccc6c272f691c +github.com/rancher/kontainer-driver-metadata ec99e4af8c8d678e457a6f315353101979300c52 diff --git a/vendor/github.com/rancher/kontainer-driver-metadata/rke/k8s_docker_info.go b/vendor/github.com/rancher/kontainer-driver-metadata/rke/k8s_docker_info.go index 94db1379..f704bc2a 100644 --- a/vendor/github.com/rancher/kontainer-driver-metadata/rke/k8s_docker_info.go +++ b/vendor/github.com/rancher/kontainer-driver-metadata/rke/k8s_docker_info.go @@ -9,5 +9,6 @@ func loadK8sVersionDockerInfo() map[string][]string { "1.12": {"1.11.x", "1.12.x", "1.13.x", "17.03.x", "17.06.x", "17.09.x", "18.06.x", "18.09.x", "19.03.x"}, "1.13": {"1.11.x", "1.12.x", "1.13.x", "17.03.x", "17.06.x", "17.09.x", "18.06.x", "18.09.x", "19.03.x"}, "1.14": {"1.13.x", "17.03.x", "17.06.x", "17.09.x", "18.06.x", "18.09.x", "19.03.x"}, - "1.15": {"1.13.x", "17.03.x", "17.06.x", "17.09.x", "18.06.x", "18.09.x", "19.03.x"}} + "1.15": {"1.13.x", "17.03.x", "17.06.x", "17.09.x", "18.06.x", "18.09.x", "19.03.x"}, + "1.16": {"1.13.x", "17.03.x", "17.06.x", "17.09.x", "18.06.x", "18.09.x", "19.03.x"}} } diff --git a/vendor/github.com/rancher/kontainer-driver-metadata/rke/k8s_rke_system_images.go b/vendor/github.com/rancher/kontainer-driver-metadata/rke/k8s_rke_system_images.go index bfcfda6f..8e526ab5 100644 --- a/vendor/github.com/rancher/kontainer-driver-metadata/rke/k8s_rke_system_images.go +++ b/vendor/github.com/rancher/kontainer-driver-metadata/rke/k8s_rke_system_images.go @@ -944,9 +944,11 @@ func loadK8sRKESystemImages() map[string]v3.RKESystemImages { CoreDNS: m("coredns/coredns:1.2.6"), CoreDNSAutoscaler: m("gcr.io/google_containers/cluster-proportional-autoscaler:1.0.0"), }, - "v1.13.8-rancher1-1": { + // Enabled in Rancher v2.2.7 + // v1.13.9-rancher1-1 is the backported version made available due to older rke-tools (see line #1321) + "v1.13.9-rancher1-2": { Etcd: m("quay.io/coreos/etcd:v3.2.24-rancher1"), - Kubernetes: m("rancher/hyperkube:v1.13.8-rancher1"), + Kubernetes: m("rancher/hyperkube:v1.13.9-rancher1"), Alpine: m("rancher/rke-tools:v0.1.40"), NginxProxy: m("rancher/rke-tools:v0.1.40"), CertDownloader: m("rancher/rke-tools:v0.1.40"), @@ -1058,9 +1060,10 @@ func loadK8sRKESystemImages() map[string]v3.RKESystemImages { CoreDNS: m("coredns/coredns:1.3.1"), CoreDNSAutoscaler: m("gcr.io/google_containers/cluster-proportional-autoscaler:1.3.0"), }, - "v1.14.4-rancher1-1": { + // Enabled in Rancher v2.2.7 + "v1.14.5-rancher1-1": { Etcd: m("quay.io/coreos/etcd:v3.3.10-rancher1"), - Kubernetes: m("rancher/hyperkube:v1.14.4-rancher1"), + Kubernetes: m("rancher/hyperkube:v1.14.5-rancher1"), Alpine: m("rancher/rke-tools:v0.1.40"), NginxProxy: m("rancher/rke-tools:v0.1.40"), CertDownloader: m("rancher/rke-tools:v0.1.40"), @@ -1116,9 +1119,9 @@ func loadK8sRKESystemImages() map[string]v3.RKESystemImages { CoreDNS: m("coredns/coredns:1.3.1"), CoreDNSAutoscaler: m("gcr.io/google_containers/cluster-proportional-autoscaler:1.3.0"), }, - "v1.15.0-rancher1-2": { + "v1.15.2-rancher1-1": { Etcd: m("quay.io/coreos/etcd:v3.3.10-rancher1"), - Kubernetes: m("rancher/hyperkube:v1.15.0-rancher1"), + Kubernetes: m("rancher/hyperkube:v1.15.2-rancher1"), Alpine: m("rancher/rke-tools:v0.1.40"), NginxProxy: m("rancher/rke-tools:v0.1.40"), CertDownloader: m("rancher/rke-tools:v0.1.40"), @@ -1313,5 +1316,33 @@ func loadK8sRKESystemImages() map[string]v3.RKESystemImages { IngressBackend: m("k8s.gcr.io/defaultbackend:1.4"), MetricsServer: m("gcr.io/google_containers/metrics-server-amd64:v0.3.1"), }, + // k8s version from 2.1.x release with old rke-tools to allow upgrade from 2.1.x clusters + // without all clusters being restarted + "v1.13.9-rancher1-1": { + Etcd: m("quay.io/coreos/etcd:v3.2.24"), + Kubernetes: m("rancher/hyperkube:v1.13.9-rancher1"), + Alpine: m("rancher/rke-tools:v0.1.16"), + NginxProxy: m("rancher/rke-tools:v0.1.16"), + CertDownloader: m("rancher/rke-tools:v0.1.16"), + KubernetesServicesSidecar: m("rancher/rke-tools:v0.1.16"), + KubeDNS: m("gcr.io/google_containers/k8s-dns-kube-dns-amd64:1.15.0"), + DNSmasq: m("gcr.io/google_containers/k8s-dns-dnsmasq-nanny-amd64:1.15.0"), + KubeDNSSidecar: m("gcr.io/google_containers/k8s-dns-sidecar-amd64:1.15.0"), + KubeDNSAutoscaler: m("gcr.io/google_containers/cluster-proportional-autoscaler-amd64:1.0.0"), + Flannel: m("quay.io/coreos/flannel:v0.10.0"), + FlannelCNI: m("quay.io/coreos/flannel-cni:v0.3.0"), + CalicoNode: m("quay.io/calico/node:v3.4.0"), + CalicoCNI: m("quay.io/calico/cni:v3.4.0"), + CalicoCtl: m("quay.io/calico/ctl:v2.0.0"), + CanalNode: m("quay.io/calico/node:v3.4.0"), + CanalCNI: m("quay.io/calico/cni:v3.4.0"), + CanalFlannel: m("quay.io/coreos/flannel:v0.10.0"), + WeaveNode: m("weaveworks/weave-kube:2.5.0"), + WeaveCNI: m("weaveworks/weave-npc:2.5.0"), + PodInfraContainer: m("gcr.io/google_containers/pause-amd64:3.1"), + Ingress: m("rancher/nginx-ingress-controller:0.21.0-rancher1"), + IngressBackend: m("k8s.gcr.io/defaultbackend:1.4"), + MetricsServer: m("gcr.io/google_containers/metrics-server-amd64:v0.3.1"), + }, } } diff --git a/vendor/github.com/rancher/kontainer-driver-metadata/rke/k8s_version_info.go b/vendor/github.com/rancher/kontainer-driver-metadata/rke/k8s_version_info.go index 8c2a1476..117b1d5d 100644 --- a/vendor/github.com/rancher/kontainer-driver-metadata/rke/k8s_version_info.go +++ b/vendor/github.com/rancher/kontainer-driver-metadata/rke/k8s_version_info.go @@ -4,17 +4,17 @@ import v3 "github.com/rancher/types/apis/management.cattle.io/v3" func loadRancherDefaultK8sVersions() map[string]string { return map[string]string{ - "2.3": "v1.15.0-rancher1-2", + "2.3": "v1.15.2-rancher1-1", // rancher will use default if its version is absent - "default": "v1.15.0-rancher1-2", + "default": "v1.15.2-rancher1-1", } } func loadRKEDefaultK8sVersions() map[string]string { return map[string]string{ - "0.3": "v1.15.0-rancher1-2", + "0.3": "v1.15.2-rancher1-1", // rke will use default if its version is absent - "default": "v1.15.0-rancher1-2", + "default": "v1.15.2-rancher1-1", } } diff --git a/vendor/github.com/rancher/kontainer-driver-metadata/rke/k8s_windows_defaults.go b/vendor/github.com/rancher/kontainer-driver-metadata/rke/k8s_windows_defaults.go index 9054f9d3..584bc19d 100644 --- a/vendor/github.com/rancher/kontainer-driver-metadata/rke/k8s_windows_defaults.go +++ b/vendor/github.com/rancher/kontainer-driver-metadata/rke/k8s_windows_defaults.go @@ -292,74 +292,6 @@ func loadK8sVersionWindowsSystemimages() map[string]v3.WindowsSystemImages { FlannelCNIBinaries: m("rancher/flannel-cni:v0.0.1-nanoserver-1803"), KubeletPause: m("rancher/kubelet-pause:v0.0.1-nanoserver-1803"), }, - "v1.12.10-rancher1-1": { - NginxProxy: m("rancher/nginx-proxy:v0.0.1-nanoserver-1803"), - KubernetesBinaries: m("rancher/hyperkube:v1.12.10-nanoserver-1803"), - FlannelCNIBinaries: m("rancher/flannel-cni:v0.0.1-nanoserver-1803"), - KubeletPause: m("rancher/kubelet-pause:v0.0.1-nanoserver-1803"), - }, - "v1.13.1-rancher1-2": { - NginxProxy: m("rancher/nginx-proxy:v0.0.1-nanoserver-1803"), - KubernetesBinaries: m("rancher/hyperkube:v1.13.1-nanoserver-1803"), - FlannelCNIBinaries: m("rancher/flannel-cni:v0.0.1-nanoserver-1803"), - KubeletPause: m("rancher/kubelet-pause:v0.0.1-nanoserver-1803"), - }, - "v1.13.4-rancher1-1": { - NginxProxy: m("rancher/nginx-proxy:v0.0.1-nanoserver-1803"), - KubernetesBinaries: m("rancher/hyperkube:v1.13.4-nanoserver-1803"), - FlannelCNIBinaries: m("rancher/flannel-cni:v0.0.1-nanoserver-1803"), - CalicoCNIBinaries: m("rancher/calico-cni:v0.0.1-nanoserver-1803"), - CanalCNIBinaries: m("rancher/canal-cni:v0.0.1-nanoserver-1803"), - KubeletPause: m("rancher/kubelet-pause:v0.0.1-nanoserver-1803"), - }, - "v1.13.4-rancher1-2": { - NginxProxy: m("rancher/nginx-proxy:v0.0.1-nanoserver-1803"), - KubernetesBinaries: m("rancher/hyperkube:v1.13.4-nanoserver-1803"), - FlannelCNIBinaries: m("rancher/flannel-cni:v0.0.1-nanoserver-1803"), - CalicoCNIBinaries: m("rancher/calico-cni:v0.0.1-nanoserver-1803"), - CanalCNIBinaries: m("rancher/canal-cni:v0.0.1-nanoserver-1803"), - KubeletPause: m("rancher/kubelet-pause:v0.0.1-nanoserver-1803"), - }, - "v1.13.5-rancher1-1": { - NginxProxy: m("rancher/nginx-proxy:v0.0.1-nanoserver-1803"), - KubernetesBinaries: m("rancher/hyperkube:v1.13.5-nanoserver-1803"), - FlannelCNIBinaries: m("rancher/flannel-cni:v0.0.1-nanoserver-1803"), - CalicoCNIBinaries: m("rancher/calico-cni:v0.0.1-nanoserver-1803"), - CanalCNIBinaries: m("rancher/canal-cni:v0.0.1-nanoserver-1803"), - KubeletPause: m("rancher/kubelet-pause:v0.0.1-nanoserver-1803"), - }, - "v1.13.5-rancher1-2": { - NginxProxy: m("rancher/nginx-proxy:v0.0.1-nanoserver-1803"), - KubernetesBinaries: m("rancher/hyperkube:v1.13.5-nanoserver-1803"), - FlannelCNIBinaries: m("rancher/flannel-cni:v0.0.1-nanoserver-1803"), - CalicoCNIBinaries: m("rancher/calico-cni:v0.0.1-nanoserver-1803"), - CanalCNIBinaries: m("rancher/canal-cni:v0.0.1-nanoserver-1803"), - KubeletPause: m("rancher/kubelet-pause:v0.0.1-nanoserver-1803"), - }, - "v1.13.5-rancher1-3": { - NginxProxy: m("rancher/nginx-proxy:v0.0.1-nanoserver-1803"), - KubernetesBinaries: m("rancher/hyperkube:v1.13.5-nanoserver-1803"), - FlannelCNIBinaries: m("rancher/flannel-cni:v0.0.1-nanoserver-1803"), - CalicoCNIBinaries: m("rancher/calico-cni:v0.0.1-nanoserver-1803"), - CanalCNIBinaries: m("rancher/canal-cni:v0.0.1-nanoserver-1803"), - KubeletPause: m("rancher/kubelet-pause:v0.0.1-nanoserver-1803"), - }, - "v1.13.7-rancher1-1": { - NginxProxy: m("rancher/nginx-proxy:v0.0.1-nanoserver-1803"), - KubernetesBinaries: m("rancher/hyperkube:v1.13.7-nanoserver-1803"), - FlannelCNIBinaries: m("rancher/flannel-cni:v0.0.1-nanoserver-1803"), - CalicoCNIBinaries: m("rancher/calico-cni:v0.0.1-nanoserver-1803"), - CanalCNIBinaries: m("rancher/canal-cni:v0.0.1-nanoserver-1803"), - KubeletPause: m("rancher/kubelet-pause:v0.0.1-nanoserver-1803"), - }, - "v1.13.8-rancher1-1": { - NginxProxy: m("rancher/nginx-proxy:v0.0.1-nanoserver-1803"), - KubernetesBinaries: m("rancher/hyperkube:v1.13.8-nanoserver-1803"), - FlannelCNIBinaries: m("rancher/flannel-cni:v0.0.1-nanoserver-1803"), - CalicoCNIBinaries: m("rancher/calico-cni:v0.0.1-nanoserver-1803"), - CanalCNIBinaries: m("rancher/canal-cni:v0.0.1-nanoserver-1803"), - KubeletPause: m("rancher/kubelet-pause:v0.0.1-nanoserver-1803"), - }, "v1.14.1-rancher1-1": { NginxProxy: m("rancher/nginx-proxy:v0.0.1-nanoserver-1803"), KubernetesBinaries: m("rancher/hyperkube:v1.14.1-nanoserver-1803"), @@ -380,23 +312,11 @@ func loadK8sVersionWindowsSystemimages() map[string]v3.WindowsSystemImages { FlannelCNIBinaries: m("rancher/flannel-cni:v0.3.0-rancher4"), KubeletPause: m("rancher/kubelet-pause:v0.1.2"), }, - "v1.14.4-rancher1-1": { - // NginxProxy image is replaced by host running nginx, fixed rancher#16074 - KubernetesBinaries: m("rancher/hyperkube:v1.14.4-rancher1"), - FlannelCNIBinaries: m("rancher/flannel-cni:v0.3.0-rancher4"), - KubeletPause: m("rancher/kubelet-pause:v0.1.2"), - }, "v1.15.0-rancher1-1": { // NginxProxy image is replaced by host running nginx, fixed rancher#16074 KubernetesBinaries: m("rancher/hyperkube:v1.15.0-rancher1"), FlannelCNIBinaries: m("rancher/flannel-cni:v0.3.0-rancher4"), KubeletPause: m("rancher/kubelet-pause:v0.1.2"), }, - "v1.15.0-rancher1-2": { - // NginxProxy image is replaced by host running nginx, fixed rancher#16074 - KubernetesBinaries: m("rancher/hyperkube:v1.15.0-rancher1"), - FlannelCNIBinaries: m("rancher/flannel-cni:v0.3.0-rancher4"), - KubeletPause: m("rancher/kubelet-pause:v0.1.2"), - }, } } diff --git a/vendor/github.com/rancher/kontainer-driver-metadata/rke/templates/weave.go b/vendor/github.com/rancher/kontainer-driver-metadata/rke/templates/weave.go index a2ce5ce6..eee05391 100644 --- a/vendor/github.com/rancher/kontainer-driver-metadata/rke/templates/weave.go +++ b/vendor/github.com/rancher/kontainer-driver-metadata/rke/templates/weave.go @@ -22,6 +22,7 @@ items: template: metadata: annotations: + scheduler.alpha.kubernetes.io/critical-pod: '' scheduler.alpha.kubernetes.io/tolerations: >- [{"key":"dedicated","operator":"Equal","value":"master","effect":"NoSchedule"}] labels: