mirror of
https://github.com/rancher/rke.git
synced 2025-08-14 13:03:06 +00:00
Update vendor
This commit is contained in:
parent
8a219b5c50
commit
3fd4c2eb91
@ -28,4 +28,4 @@ github.com/matttproud/golang_protobuf_extensions c12348ce28de40eed0136aa2b644d0e
|
||||
github.com/mattn/go-colorable efa589957cd060542a26d2dd7832fd6a6c6c3ade
|
||||
github.com/mattn/go-isatty 6ca4dbf54d38eea1a992b3c722a76a5d1c4cb25c
|
||||
github.com/rancher/norman 0557aa4ff31a3a0f007dcb1b684894f23cda390c
|
||||
github.com/rancher/types 58191660b8b37fe11a25061f38e6b0763479dbf6
|
||||
github.com/rancher/types a3d51c1bcc9025f7baabf8e2014b74deaab4264c
|
||||
|
77
vendor/github.com/rancher/types/apis/management.cattle.io/v3/k8s_defaults.go
generated
vendored
77
vendor/github.com/rancher/types/apis/management.cattle.io/v3/k8s_defaults.go
generated
vendored
@ -9,7 +9,7 @@ import (
|
||||
)
|
||||
|
||||
const (
|
||||
DefaultK8s = "v1.13.4-rancher1-1"
|
||||
DefaultK8s = "v1.13.4-rancher1-2"
|
||||
)
|
||||
|
||||
var (
|
||||
@ -29,7 +29,7 @@ var (
|
||||
// Account for the fact that rke-tools image is different cross branches, and it requires using
|
||||
// different k8s version tag
|
||||
"v1.12.6-rancher1-2",
|
||||
"v1.13.4-rancher1-1",
|
||||
"v1.13.4-rancher1-2",
|
||||
}
|
||||
|
||||
// K8sVersionToRKESystemImages is dynamically populated on init() with the latest versions
|
||||
@ -817,18 +817,18 @@ var (
|
||||
CoreDNSAutoscaler: m("gcr.io/google_containers/cluster-proportional-autoscaler-amd64:1.0.0"),
|
||||
},
|
||||
"v1.12.6-rancher1-2": {
|
||||
Etcd: m("quay.io/coreos/etcd:v3.2.24"),
|
||||
Etcd: m("quay.io/coreos/etcd:v3.2.24-rancher1"),
|
||||
Kubernetes: m("rancher/hyperkube:v1.12.6-rancher1"),
|
||||
Alpine: m("rancher/rke-tools:v0.1.27"),
|
||||
NginxProxy: m("rancher/rke-tools:v0.1.27"),
|
||||
CertDownloader: m("rancher/rke-tools:v0.1.27"),
|
||||
KubernetesServicesSidecar: m("rancher/rke-tools:v0.1.27"),
|
||||
KubeDNS: m("gcr.io/google_containers/k8s-dns-kube-dns-amd64:1.14.13"),
|
||||
DNSmasq: m("gcr.io/google_containers/k8s-dns-dnsmasq-nanny-amd64:1.14.13"),
|
||||
KubeDNSSidecar: m("gcr.io/google_containers/k8s-dns-sidecar-amd64:1.14.13"),
|
||||
KubeDNSAutoscaler: m("gcr.io/google_containers/cluster-proportional-autoscaler-amd64:1.0.0"),
|
||||
Flannel: m("quay.io/coreos/flannel:v0.10.0"),
|
||||
FlannelCNI: m("quay.io/coreos/flannel-cni:v0.3.0"),
|
||||
KubeDNS: m("gcr.io/google_containers/k8s-dns-kube-dns:1.14.13"),
|
||||
DNSmasq: m("gcr.io/google_containers/k8s-dns-dnsmasq-nanny:1.14.13"),
|
||||
KubeDNSSidecar: m("gcr.io/google_containers/k8s-dns-sidecar:1.14.13"),
|
||||
KubeDNSAutoscaler: m("gcr.io/google_containers/cluster-proportional-autoscaler:1.0.0"),
|
||||
Flannel: m("quay.io/coreos/flannel:v0.10.0-rancher1"),
|
||||
FlannelCNI: m("rancher/flannel-cni:v0.3.0-rancher1"),
|
||||
CalicoNode: m("quay.io/calico/node:v3.1.3"),
|
||||
CalicoCNI: m("quay.io/calico/cni:v3.1.3"),
|
||||
CalicoCtl: m("quay.io/calico/ctl:v2.0.0"),
|
||||
@ -837,12 +837,12 @@ var (
|
||||
CanalFlannel: m("quay.io/coreos/flannel:v0.10.0"),
|
||||
WeaveNode: m("weaveworks/weave-kube:2.5.0"),
|
||||
WeaveCNI: m("weaveworks/weave-npc:2.5.0"),
|
||||
PodInfraContainer: m("gcr.io/google_containers/pause-amd64:3.1"),
|
||||
Ingress: m("rancher/nginx-ingress-controller:0.21.0-rancher1"),
|
||||
IngressBackend: m("k8s.gcr.io/defaultbackend:1.4"),
|
||||
MetricsServer: m("gcr.io/google_containers/metrics-server-amd64:v0.3.1"),
|
||||
PodInfraContainer: m("gcr.io/google_containers/pause:3.1"),
|
||||
Ingress: m("rancher/nginx-ingress-controller:0.21.0-rancher3"),
|
||||
IngressBackend: m("k8s.gcr.io/defaultbackend:1.4-rancher1"),
|
||||
MetricsServer: m("gcr.io/google_containers/metrics-server:v0.3.1"),
|
||||
CoreDNS: m("coredns/coredns:1.2.2"),
|
||||
CoreDNSAutoscaler: m("gcr.io/google_containers/cluster-proportional-autoscaler-amd64:1.0.0"),
|
||||
CoreDNSAutoscaler: m("gcr.io/google_containers/cluster-proportional-autoscaler:1.0.0"),
|
||||
},
|
||||
"v1.13.1-rancher1-2": {
|
||||
Etcd: m("quay.io/coreos/etcd:v3.2.24"),
|
||||
@ -872,7 +872,36 @@ var (
|
||||
CoreDNS: m("coredns/coredns:1.2.6"),
|
||||
CoreDNSAutoscaler: m("gcr.io/google_containers/cluster-proportional-autoscaler-amd64:1.0.0"),
|
||||
},
|
||||
// Add a new version with old rke-tools to allow upgrade from 2.1.x clusters
|
||||
"v1.13.4-rancher1-2": {
|
||||
Etcd: m("quay.io/coreos/etcd:v3.2.24-rancher1"),
|
||||
Kubernetes: m("rancher/hyperkube:v1.13.4-rancher1"),
|
||||
Alpine: m("rancher/rke-tools:v0.1.27"),
|
||||
NginxProxy: m("rancher/rke-tools:v0.1.27"),
|
||||
CertDownloader: m("rancher/rke-tools:v0.1.27"),
|
||||
KubernetesServicesSidecar: m("rancher/rke-tools:v0.1.27"),
|
||||
KubeDNS: m("gcr.io/google_containers/k8s-dns-kube-dns:1.15.0"),
|
||||
DNSmasq: m("gcr.io/google_containers/k8s-dns-dnsmasq-nanny:1.15.0"),
|
||||
KubeDNSSidecar: m("gcr.io/google_containers/k8s-dns-sidecar:1.15.0"),
|
||||
KubeDNSAutoscaler: m("gcr.io/google_containers/cluster-proportional-autoscaler:1.0.0"),
|
||||
Flannel: m("quay.io/coreos/flannel:v0.10.0-rancher1"),
|
||||
FlannelCNI: m("rancher/flannel-cni:v0.3.0-rancher1"),
|
||||
CalicoNode: m("quay.io/calico/node:v3.4.0"),
|
||||
CalicoCNI: m("quay.io/calico/cni:v3.4.0"),
|
||||
CalicoCtl: m("quay.io/calico/ctl:v2.0.0"),
|
||||
CanalNode: m("quay.io/calico/node:v3.4.0"),
|
||||
CanalCNI: m("quay.io/calico/cni:v3.4.0"),
|
||||
CanalFlannel: m("quay.io/coreos/flannel:v0.10.0"),
|
||||
WeaveNode: m("weaveworks/weave-kube:2.5.0"),
|
||||
WeaveCNI: m("weaveworks/weave-npc:2.5.0"),
|
||||
PodInfraContainer: m("gcr.io/google_containers/pause:3.1"),
|
||||
Ingress: m("rancher/nginx-ingress-controller:0.21.0-rancher3"),
|
||||
IngressBackend: m("k8s.gcr.io/defaultbackend:1.4-rancher1"),
|
||||
MetricsServer: m("gcr.io/google_containers/metrics-server:v0.3.1"),
|
||||
CoreDNS: m("coredns/coredns:1.2.6"),
|
||||
CoreDNSAutoscaler: m("gcr.io/google_containers/cluster-proportional-autoscaler:1.0.0"),
|
||||
},
|
||||
// k8s version from 2.1.x release with old rke-tools to allow upgrade from 2.1.x clusters
|
||||
// without all clusters being restarted
|
||||
"v1.12.5-rancher1-1": {
|
||||
Etcd: m("quay.io/coreos/etcd:v3.2.24"),
|
||||
Kubernetes: m("rancher/hyperkube:v1.12.5-rancher1"),
|
||||
@ -899,7 +928,8 @@ var (
|
||||
IngressBackend: m("k8s.gcr.io/defaultbackend:1.4"),
|
||||
MetricsServer: m("gcr.io/google_containers/metrics-server-amd64:v0.3.1"),
|
||||
},
|
||||
// Add a new version with old rke-tools to allow upgrade from 2.1.x clusters
|
||||
// k8s version from 2.1.x release with old rke-tools to allow upgrade from 2.1.x clusters
|
||||
// without all clusters being restarted
|
||||
"v1.12.6-rancher1-1": {
|
||||
Etcd: m("quay.io/coreos/etcd:v3.2.24"),
|
||||
Kubernetes: m("rancher/hyperkube:v1.12.6-rancher1"),
|
||||
@ -926,7 +956,8 @@ var (
|
||||
IngressBackend: m("k8s.gcr.io/defaultbackend:1.4"),
|
||||
MetricsServer: m("gcr.io/google_containers/metrics-server-amd64:v0.3.1"),
|
||||
},
|
||||
// Add a new version with old rke-tools to allow upgrade from 2.1.x clusters
|
||||
// k8s version from 2.1.x release with old rke-tools to allow upgrade from 2.1.x clusters
|
||||
// without all clusters being restarted
|
||||
"v1.13.1-rancher1-1": {
|
||||
Etcd: m("quay.io/coreos/etcd:v3.2.24"),
|
||||
Kubernetes: m("rancher/hyperkube:v1.13.1-rancher1"),
|
||||
@ -953,13 +984,15 @@ var (
|
||||
IngressBackend: m("k8s.gcr.io/defaultbackend:1.4"),
|
||||
MetricsServer: m("gcr.io/google_containers/metrics-server-amd64:v0.3.1"),
|
||||
},
|
||||
// k8s version from 2.1.x release with old rke-tools to allow upgrade from 2.1.x clusters
|
||||
// without all clusters being restarted
|
||||
"v1.13.4-rancher1-1": {
|
||||
Etcd: m("quay.io/coreos/etcd:v3.2.24"),
|
||||
Kubernetes: m("rancher/hyperkube:v1.13.4-rancher1"),
|
||||
Alpine: m("rancher/rke-tools:v0.1.27"),
|
||||
NginxProxy: m("rancher/rke-tools:v0.1.27"),
|
||||
CertDownloader: m("rancher/rke-tools:v0.1.27"),
|
||||
KubernetesServicesSidecar: m("rancher/rke-tools:v0.1.27"),
|
||||
Alpine: m("rancher/rke-tools:v0.1.16"),
|
||||
NginxProxy: m("rancher/rke-tools:v0.1.16"),
|
||||
CertDownloader: m("rancher/rke-tools:v0.1.16"),
|
||||
KubernetesServicesSidecar: m("rancher/rke-tools:v0.1.16"),
|
||||
KubeDNS: m("gcr.io/google_containers/k8s-dns-kube-dns-amd64:1.15.0"),
|
||||
DNSmasq: m("gcr.io/google_containers/k8s-dns-dnsmasq-nanny-amd64:1.15.0"),
|
||||
KubeDNSSidecar: m("gcr.io/google_containers/k8s-dns-sidecar-amd64:1.15.0"),
|
||||
@ -978,8 +1011,6 @@ var (
|
||||
Ingress: m("rancher/nginx-ingress-controller:0.21.0-rancher1"),
|
||||
IngressBackend: m("k8s.gcr.io/defaultbackend:1.4"),
|
||||
MetricsServer: m("gcr.io/google_containers/metrics-server-amd64:v0.3.1"),
|
||||
CoreDNS: m("coredns/coredns:1.2.6"),
|
||||
CoreDNSAutoscaler: m("gcr.io/google_containers/cluster-proportional-autoscaler-amd64:1.0.0"),
|
||||
},
|
||||
}
|
||||
)
|
||||
|
8
vendor/github.com/rancher/types/apis/management.cattle.io/v3/k8s_windows_default.go
generated
vendored
8
vendor/github.com/rancher/types/apis/management.cattle.io/v3/k8s_windows_default.go
generated
vendored
@ -247,6 +247,14 @@ var (
|
||||
CanalCNIBinaries: m("rancher/canal-cni:v0.0.1-nanoserver-1803"),
|
||||
KubeletPause: m("rancher/kubelet-pause:v0.0.1-nanoserver-1803"),
|
||||
},
|
||||
"v1.13.4-rancher1-2": {
|
||||
NginxProxy: m("rancher/nginx-proxy:v0.0.1-nanoserver-1803"),
|
||||
KubernetesBinaries: m("rancher/hyperkube:v1.13.4-nanoserver-1803"),
|
||||
FlannelCNIBinaries: m("rancher/flannel-cni:v0.0.1-nanoserver-1803"),
|
||||
CalicoCNIBinaries: m("rancher/calico-cni:v0.0.1-nanoserver-1803"),
|
||||
CanalCNIBinaries: m("rancher/canal-cni:v0.0.1-nanoserver-1803"),
|
||||
KubeletPause: m("rancher/kubelet-pause:v0.0.1-nanoserver-1803"),
|
||||
},
|
||||
}
|
||||
)
|
||||
|
||||
|
Loading…
Reference in New Issue
Block a user