Merge pull request #2559 from ijc/kubernetes

kubernetes: bump to v1.8.0
This commit is contained in:
Ian Campbell 2017-10-03 16:43:19 +01:00 committed by GitHub
commit 77f0c75b9d
10 changed files with 33 additions and 26 deletions

View File

@ -29,7 +29,7 @@ network.yaml: $(NETWORK).yaml
ln -nf $< $@
weave-%.yaml:
curl -L -o $@ https://cloud.weave.works/k8s/v1.7/net?v=$*
curl -L -o $@ https://cloud.weave.works/k8s/v1.8/net?v=$*
clean:
rm -f -r \

View File

@ -1,6 +1,6 @@
services:
- name: cri-containerd
image: linuxkitprojects/cri-containerd:3ccc80f3257647a02c4a78fb0f0557805601c958
image: linuxkitprojects/cri-containerd:2e3f5fcb87f92e1176637943b496910b9d979fb5
files:
- path: /etc/kubelet.sh.conf
contents: |

View File

@ -48,4 +48,4 @@ FROM scratch
WORKDIR /
ENTRYPOINT ["cri-containerd", "-v", "2", "--alsologtostderr", "--network-bin-dir", "/var/lib/cni/opt/bin", "--network-conf-dir", "/var/lib/cni/etc/net.d"]
COPY --from=build /out /
LABEL org.mobyproject.config='{"binds": ["/etc/resolv.conf:/etc/resolv.conf", "/run:/run:rshared,rbind", "/dev:/dev", "/tmp:/tmp", "/var:/var:rshared,rbind", "/var/lib/kubeadm:/etc/kubernetes", "/var/lib/cni/etc:/etc/cni:rshared,rbind", "/var/lib/cni/opt:/opt/cni:rshared,rbind", "/run/containerd/containerd.sock:/run/containerd/containerd.sock"], "mounts": [{"type": "cgroup", "options": ["rw","nosuid","noexec","nodev","relatime"]}], "capabilities": ["all"], "rootfsPropagation": "shared", "pid": "host", "runtime": {"mkdir": ["/var/lib/kubeadm", "/var/lib/cni/etc/net.d", "/var/lib/cni/opt"]}}'
LABEL org.mobyproject.config='{"binds": ["/etc/resolv.conf:/etc/resolv.conf", "/run:/run:rshared,rbind", "/dev:/dev", "/tmp:/tmp", "/var:/var:rshared,rbind", "/var/lib/kubeadm:/etc/kubernetes", "/var/lib/cni/etc:/etc/cni:rshared,rbind", "/var/lib/cni/opt:/opt/cni:rshared,rbind", "/run/containerd/containerd.sock:/run/containerd/containerd.sock", "/var/lib/kubelet-plugins:/usr/libexec/kubernetes/kubelet-plugins:rshared,rbind"], "mounts": [{"type": "cgroup", "options": ["rw","nosuid","noexec","nodev","relatime"]}], "capabilities": ["all"], "rootfsPropagation": "shared", "pid": "host", "runtime": {"mkdir": ["/var/lib/kubeadm", "/var/lib/cni/etc/net.d", "/var/lib/cni/opt", "/var/lib/kubelet-plugins"]}}'

View File

@ -1,3 +1,3 @@
services:
- name: kubernetes-image-cache-control-plane
image: linuxkitprojects/kubernetes-image-cache-control-plane:465c8781efd6664c9a744eb78bf5b90d6ce5db3e
image: linuxkitprojects/kubernetes-image-cache-control-plane:2c8a47f1dfc9f4b00771b27b5f8b74065c132a3a

View File

@ -16,12 +16,13 @@ services:
- /var/lib/kubeadm:/etc/kubernetes
- /var/lib/cni/etc:/etc/cni:rshared,rbind
- /var/lib/cni/opt:/opt/cni:rshared,rbind
- /var/lib/kubelet-plugins:/usr/libexec/kubernetes/kubelet-plugins:rshared,rbind
rootfsPropagation: shared
command: ["/usr/local/bin/docker-init", "/usr/local/bin/dockerd"]
runtime:
mkdir: ["/var/lib/kubeadm", "/var/lib/cni/etc", "/var/lib/cni/opt"]
mkdir: ["/var/lib/kubeadm", "/var/lib/cni/etc", "/var/lib/cni/opt", "/var/lib/kubelet-plugins"]
- name: kubernetes-image-cache-common
image: linuxkitprojects/kubernetes-image-cache-common:465c8781efd6664c9a744eb78bf5b90d6ce5db3e
image: linuxkitprojects/kubernetes-image-cache-common:2c8a47f1dfc9f4b00771b27b5f8b74065c132a3a
files:
- path: /etc/kubelet.sh.conf
contents: ""

View File

@ -1,7 +1,7 @@
#!/bin/sh
repo=gcr.io/google_containers
kube_version=v1.7.6
kube_dns_version=1.14.4
kube_version=v1.8.0
kube_dns_version=1.14.5
pause_version=3.0
etcd_version=3.0.17

View File

@ -1,13 +1,13 @@
# autogenerated by mkversions
COMMON_IMAGES := \
kube-proxy-amd64\:v1.7.6@sha256\:1509f2fc8a60501d604d21d983ed6f5d0ea40ccdd7cc6ba6c994389ef7db16d8 \
k8s-dns-sidecar-amd64\:1.14.4@sha256\:97074c951046e37d3cbb98b82ae85ed15704a290cce66a8314e7f846404edde9 \
k8s-dns-kube-dns-amd64\:1.14.4@sha256\:40790881bbe9ef4ae4ff7fe8b892498eecb7fe6dcc22661402f271e03f7de344 \
k8s-dns-dnsmasq-nanny-amd64\:1.14.4@sha256\:aeeb994acbc505eabc7415187cd9edb38cbb5364dc1c2fc748154576464b3dc2 \
kube-proxy-amd64\:v1.8.0@sha256\:3da0e03a49404119fbf104037b7ac24d871842bf565a4364390a7d6ea60f48a0 \
k8s-dns-sidecar-amd64\:1.14.5@sha256\:9aab42bf6a2a068b797fe7d91a5d8d915b10dbbc3d6f2b10492848debfba6044 \
k8s-dns-kube-dns-amd64\:1.14.5@sha256\:1a3fc069de481ae690188f6f1ba4664b5cc7760af37120f70c86505c79eea61d \
k8s-dns-dnsmasq-nanny-amd64\:1.14.5@sha256\:46b933bb70270c8a02fa6b6f87d440f6f1fce1a5a2a719e164f83f7b109f7544 \
pause-amd64\:3.0@sha256\:163ac025575b775d1c0f9bf0bdd0f086883171eb475b5068e7defa4ca9e76516
CONTROL_PLANE_IMAGES := \
kube-apiserver-amd64\:v1.7.6@sha256\:f3a208d30314a89952cf613e5ee671f9d2ed7b197cd6c5d91bebfe02571d7e1b \
kube-controller-manager-amd64\:v1.7.6@sha256\:42a42e8d39fd68de7c1db6844f909bfa6bff89019ecef86e6c542354cf8ab9fb \
kube-scheduler-amd64\:v1.7.6@sha256\:334a38ac844be07599f74876f6c923271bbd0aab48a43e7ca1ad4942e9ebdabd \
kube-apiserver-amd64\:v1.8.0@sha256\:e7377096f0b88b0fcc5dce1c56aed002f999f095a30676c68b8f686a6bb1e943 \
kube-controller-manager-amd64\:v1.8.0@sha256\:9b1ae1007d3d0f4272e7ad5e8a6d34a1f49c2e3a33339773864aed525a667d26 \
kube-scheduler-amd64\:v1.8.0@sha256\:79b5b697b1bc1023b5f18a792ea7020c4ad8403d861169b1b420d3fe673b9fe4 \
etcd-amd64\:3.0.17@sha256\:d83d3545e06fb035db8512e33bd44afb55dea007a3abd7b17742d3ac6d235940

View File

@ -36,12 +36,14 @@ services:
- name: sshd
image: linuxkit/sshd:d313eea3d9d7fbcbc927d06a6700325725db2a82
- name: kubelet
image: linuxkitprojects/kubernetes:807de925610de344864cadd21b9860342a5c2bda
image: linuxkitprojects/kubernetes:98d03686d3665b935dcd68da192f79c4cb618ec7
files:
- path: etc/linuxkit.yml
metadata: yaml
- path: /etc/kubernetes
symlink: "/var/lib/kubeadm"
- path: /usr/libexec/kubernetes/kubelet-plugins
symlink: "/var/lib/kubelet-plugins"
- path: /etc/kubeadm/kube-system.init/50-network.yaml
source: network.yaml
- path: /etc/sysctl.d/01-kubernetes.conf

View File

@ -1,6 +1,6 @@
FROM linuxkit/alpine:28254e4530703db4caa6b0199a025c30a987dfa1 AS build
ENV kubernetes_version v1.7.6
ENV kubernetes_version v1.8.0
ENV cni_version v0.6.0
RUN apk add -U --no-cache \
@ -75,4 +75,4 @@ WORKDIR /
ENTRYPOINT ["/usr/bin/kubelet.sh"]
COPY --from=build /out /
ENV KUBECONFIG "/etc/kubernetes/admin.conf"
LABEL org.mobyproject.config='{"binds": ["/dev:/dev", "/etc/resolv.conf:/etc/resolv.conf", "/run:/run:rshared,rbind", "/var:/var:rshared,rbind", "/var/lib/kubeadm:/etc/kubernetes", "/etc/kubelet.sh.conf:/etc/kubelet.sh.conf", "/etc/kubeadm:/etc/kubeadm"], "mounts": [{"type": "cgroup", "options": ["rw","nosuid","noexec","nodev","relatime"]}], "capabilities": ["all"], "rootfsPropagation": "shared", "pid": "host", "runtime": {"mkdir": ["/var/lib/kubeadm", "/var/lib/cni/etc", "/var/lib/cni/opt"], "mounts": [{"type": "bind", "source": "/var/lib/cni/opt", "destination": "/opt/cni", "options": ["rw", "bind"]}, {"type": "bind", "source": "/var/lib/cni/etc", "destination": "/etc/cni", "options": ["rw", "bind"]}]}}'
LABEL org.mobyproject.config='{"binds": ["/dev:/dev", "/etc/resolv.conf:/etc/resolv.conf", "/run:/run:rshared,rbind", "/var:/var:rshared,rbind", "/var/lib/kubeadm:/etc/kubernetes", "/etc/kubelet.sh.conf:/etc/kubelet.sh.conf", "/etc/kubeadm:/etc/kubeadm", "/var/lib/kubelet-plugins:/usr/libexec/kubernetes/kubelet-plugins:rshared,rbind"], "mounts": [{"type": "cgroup", "options": ["rw","nosuid","noexec","nodev","relatime"]}], "capabilities": ["all"], "rootfsPropagation": "shared", "pid": "host", "runtime": {"mkdir": ["/var/lib/kubeadm", "/var/lib/cni/etc", "/var/lib/cni/opt", "/var/lib/kubelet-plugins"], "mounts": [{"type": "bind", "source": "/var/lib/cni/opt", "destination": "/opt/cni", "options": ["rw", "bind"]}, {"type": "bind", "source": "/var/lib/cni/etc", "destination": "/etc/cni", "options": ["rw", "bind"]}]}}'

View File

@ -11,9 +11,9 @@ if [ -e /etc/kubelet.sh.conf ] ; then
. /etc/kubelet.sh.conf
fi
conf=/var/lib/kubeadm/kubelet.conf
await=/etc/kubernetes/kubelet.conf
if [ -f "${conf}" ] ; then
if [ -f "/etc/kubernetes/kubelet.conf" ] ; then
echo "kubelet.sh: kubelet already configured"
elif [ -e /var/config/kubeadm/init ] ; then
echo "kubelet.sh: init cluster with metadata \"$(cat /var/config/kubeadm/init)\""
@ -23,27 +23,31 @@ elif [ -e /var/config/kubeadm/init ] ; then
elif [ -e /var/config/kubeadm/join ] ; then
echo "kubelet.sh: joining cluster with metadata \"$(cat /var/config/kubeadm/join)\""
kubeadm join --skip-preflight-checks $(cat /var/config/kubeadm/join)
await=/etc/kubernetes/bootstrap-kubelet.conf
elif [ -e /var/config/userdata ] ; then
echo "kubelet.sh: joining cluster with metadata \"$(cat /var/config/userdata)\""
kubeadm join --skip-preflight-checks $(cat /var/config/userdata)
await=/etc/kubernetes/bootstrap-kubelet.conf
fi
echo "kubelet.sh: waiting for ${conf}"
echo "kubelet.sh: waiting for ${await}"
# TODO(ijc) is there a race between kubeadm creating this file and
# finishing the write where we might be able to fall through and
# start kubelet with an incomplete configuration file? I've tried
# to provoke such a race without success. An explicit
# synchronisation barrier or changing kubeadm to write
# kubelet.conf atomically might be good in any case.
until [ -f "${conf}" ] ; do
until [ -f "${await}" ] ; do
sleep 1
done
echo "kubelet.sh: ${conf} has arrived" 2>&1
echo "kubelet.sh: ${await} has arrived" 2>&1
exec kubelet --kubeconfig=${conf} \
--require-kubeconfig=true \
--pod-manifest-path=/var/lib/kubeadm/manifests \
mkdir -p /etc/kubernetes/manifests
exec kubelet --kubeconfig=/etc/kubernetes/kubelet.conf \
--bootstrap-kubeconfig=/etc/kubernetes/bootstrap-kubelet.conf \
--pod-manifest-path=/etc/kubernetes/manifests \
--allow-privileged=true \
--cluster-dns=10.96.0.10 \
--cluster-domain=cluster.local \