From 8614be3c714ddeafe4698b88ad2870d1a27af3ec Mon Sep 17 00:00:00 2001 From: Girish Kalele Date: Wed, 29 Jun 2016 15:07:43 -0700 Subject: [PATCH] Bump kubedns version to 1.6 with latest skynetservices/skydns code Built kube-dns for all architectures and pushed containers to gcr.io --- build/kube-dns/Makefile | 4 ++-- cluster/saltbase/salt/kube-dns/skydns-rc.yaml.base | 10 +++++----- cluster/saltbase/salt/kube-dns/skydns-rc.yaml.in | 10 +++++----- cluster/saltbase/salt/kube-dns/skydns-rc.yaml.sed | 10 +++++----- 4 files changed, 17 insertions(+), 17 deletions(-) diff --git a/build/kube-dns/Makefile b/build/kube-dns/Makefile index 0ea431cb2d4..896c0deb799 100644 --- a/build/kube-dns/Makefile +++ b/build/kube-dns/Makefile @@ -17,12 +17,12 @@ # If you update this image please bump the tag value before pushing. # # Usage: -# [ARCH=amd64] [TAG=1.5] [REGISTRY=gcr.io/google_containers] [BASEIMAGE=busybox] make (container|push) +# [ARCH=amd64] [TAG=1.6] [REGISTRY=gcr.io/google_containers] [BASEIMAGE=busybox] make (container|push) # Default registry, arch and tag. This can be overwritten by arguments to make PLATFORM?=linux ARCH?=amd64 -TAG?=1.5 +TAG?=1.6 REGISTRY?=gcr.io/google_containers GOLANG_VERSION=1.6 diff --git a/cluster/saltbase/salt/kube-dns/skydns-rc.yaml.base b/cluster/saltbase/salt/kube-dns/skydns-rc.yaml.base index caf97c7088f..5b662089884 100644 --- a/cluster/saltbase/salt/kube-dns/skydns-rc.yaml.base +++ b/cluster/saltbase/salt/kube-dns/skydns-rc.yaml.base @@ -21,27 +21,27 @@ apiVersion: v1 kind: ReplicationController metadata: - name: kube-dns-v17 + name: kube-dns-v18 namespace: kube-system labels: k8s-app: kube-dns - version: v17 + version: v18 kubernetes.io/cluster-service: "true" spec: replicas: __PILLAR__DNS__REPLICAS__ selector: k8s-app: kube-dns - version: v17 + version: v18 template: metadata: labels: k8s-app: kube-dns - version: v17 + version: v18 kubernetes.io/cluster-service: "true" spec: containers: - name: kubedns - image: gcr.io/google_containers/kubedns-amd64:1.5 + image: gcr.io/google_containers/kubedns-amd64:1.6 resources: # TODO: Set memory limits when we've profiled the container for large # clusters, then set request = limit to keep this container in diff --git a/cluster/saltbase/salt/kube-dns/skydns-rc.yaml.in b/cluster/saltbase/salt/kube-dns/skydns-rc.yaml.in index 138d3d9853f..4fe4cd7ffa4 100644 --- a/cluster/saltbase/salt/kube-dns/skydns-rc.yaml.in +++ b/cluster/saltbase/salt/kube-dns/skydns-rc.yaml.in @@ -21,27 +21,27 @@ apiVersion: v1 kind: ReplicationController metadata: - name: kube-dns-v17 + name: kube-dns-v18 namespace: kube-system labels: k8s-app: kube-dns - version: v17 + version: v18 kubernetes.io/cluster-service: "true" spec: replicas: {{ pillar['dns_replicas'] }} selector: k8s-app: kube-dns - version: v17 + version: v18 template: metadata: labels: k8s-app: kube-dns - version: v17 + version: v18 kubernetes.io/cluster-service: "true" spec: containers: - name: kubedns - image: gcr.io/google_containers/kubedns-amd64:1.5 + image: gcr.io/google_containers/kubedns-amd64:1.6 resources: # TODO: Set memory limits when we've profiled the container for large # clusters, then set request = limit to keep this container in diff --git a/cluster/saltbase/salt/kube-dns/skydns-rc.yaml.sed b/cluster/saltbase/salt/kube-dns/skydns-rc.yaml.sed index 960a6c5e3d5..fa4f67c6c1f 100644 --- a/cluster/saltbase/salt/kube-dns/skydns-rc.yaml.sed +++ b/cluster/saltbase/salt/kube-dns/skydns-rc.yaml.sed @@ -21,27 +21,27 @@ apiVersion: v1 kind: ReplicationController metadata: - name: kube-dns-v17 + name: kube-dns-v18 namespace: kube-system labels: k8s-app: kube-dns - version: v17 + version: v18 kubernetes.io/cluster-service: "true" spec: replicas: $DNS_REPLICAS selector: k8s-app: kube-dns - version: v17 + version: v18 template: metadata: labels: k8s-app: kube-dns - version: v17 + version: v18 kubernetes.io/cluster-service: "true" spec: containers: - name: kubedns - image: gcr.io/google_containers/kubedns-amd64:1.5 + image: gcr.io/google_containers/kubedns-amd64:1.6 resources: # TODO: Set memory limits when we've profiled the container for large # clusters, then set request = limit to keep this container in