mirror of
https://github.com/k3s-io/kubernetes.git
synced 2025-08-04 09:49:50 +00:00
Merge pull request #67690 from feiskyer/iptables-cross
Automatic merge from submit-queue. If you want to cherry-pick this change to another branch, please follow the instructions here: https://github.com/kubernetes/community/blob/master/contributors/devel/cherry-picks.md. Kubelet: only sync iptables on linux **What this PR does / why we need it**: Iptables is only supported on Linux, kubelet should only sync NAT rules on Linux. Without this PR, Kubelet on Windows would logs following errors on each `syncNetworkUtil()`: ``` kubelet.err.log:4692:E0711 22:03:42.103939 2872 kubelet_network.go:102] Failed to ensure that nat chain KUBE-MARK-DROP exists: error creating chain "KUBE-MARK-DROP": executable file ``` **Which issue(s) this PR fixes** *(optional, in `fixes #<issue number>(, fixes #<issue_number>, ...)` format, will close the issue(s) when PR gets merged)*: Fixes #65713 **Special notes for your reviewer**: **Release note**: ```release-note Kubelet now only sync iptables on Linux. ```
This commit is contained in:
commit
4bc9e94fee
@ -14,6 +14,8 @@ go_library(
|
|||||||
"kubelet.go",
|
"kubelet.go",
|
||||||
"kubelet_getters.go",
|
"kubelet_getters.go",
|
||||||
"kubelet_network.go",
|
"kubelet_network.go",
|
||||||
|
"kubelet_network_linux.go",
|
||||||
|
"kubelet_network_others.go",
|
||||||
"kubelet_node_status.go",
|
"kubelet_node_status.go",
|
||||||
"kubelet_pods.go",
|
"kubelet_pods.go",
|
||||||
"kubelet_resources.go",
|
"kubelet_resources.go",
|
||||||
@ -155,6 +157,7 @@ go_test(
|
|||||||
"kubelet_getters_test.go",
|
"kubelet_getters_test.go",
|
||||||
"kubelet_network_test.go",
|
"kubelet_network_test.go",
|
||||||
"kubelet_node_status_test.go",
|
"kubelet_node_status_test.go",
|
||||||
|
"kubelet_pods_linux_test.go",
|
||||||
"kubelet_pods_test.go",
|
"kubelet_pods_test.go",
|
||||||
"kubelet_pods_windows_test.go",
|
"kubelet_pods_windows_test.go",
|
||||||
"kubelet_resources_test.go",
|
"kubelet_resources_test.go",
|
||||||
@ -172,7 +175,6 @@ go_test(
|
|||||||
"//pkg/capabilities:go_default_library",
|
"//pkg/capabilities:go_default_library",
|
||||||
"//pkg/features:go_default_library",
|
"//pkg/features:go_default_library",
|
||||||
"//pkg/kubelet/apis:go_default_library",
|
"//pkg/kubelet/apis:go_default_library",
|
||||||
"//pkg/kubelet/apis/cri/runtime/v1alpha2:go_default_library",
|
|
||||||
"//pkg/kubelet/cadvisor/testing:go_default_library",
|
"//pkg/kubelet/cadvisor/testing:go_default_library",
|
||||||
"//pkg/kubelet/cm:go_default_library",
|
"//pkg/kubelet/cm:go_default_library",
|
||||||
"//pkg/kubelet/config:go_default_library",
|
"//pkg/kubelet/config:go_default_library",
|
||||||
@ -242,7 +244,12 @@ go_test(
|
|||||||
"//vendor/github.com/google/cadvisor/info/v2:go_default_library",
|
"//vendor/github.com/google/cadvisor/info/v2:go_default_library",
|
||||||
"//vendor/github.com/stretchr/testify/assert:go_default_library",
|
"//vendor/github.com/stretchr/testify/assert:go_default_library",
|
||||||
"//vendor/github.com/stretchr/testify/require:go_default_library",
|
"//vendor/github.com/stretchr/testify/require:go_default_library",
|
||||||
|
] + select({
|
||||||
|
"@io_bazel_rules_go//go/platform:linux": [
|
||||||
|
"//pkg/kubelet/apis/cri/runtime/v1alpha2:go_default_library",
|
||||||
],
|
],
|
||||||
|
"//conditions:default": [],
|
||||||
|
}),
|
||||||
)
|
)
|
||||||
|
|
||||||
filegroup(
|
filegroup(
|
||||||
|
@ -77,91 +77,6 @@ func (kl *Kubelet) updatePodCIDR(cidr string) error {
|
|||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
// syncNetworkUtil ensures the network utility are present on host.
|
|
||||||
// Network util includes:
|
|
||||||
// 1. In nat table, KUBE-MARK-DROP rule to mark connections for dropping
|
|
||||||
// Marked connection will be drop on INPUT/OUTPUT Chain in filter table
|
|
||||||
// 2. In nat table, KUBE-MARK-MASQ rule to mark connections for SNAT
|
|
||||||
// Marked connection will get SNAT on POSTROUTING Chain in nat table
|
|
||||||
func (kl *Kubelet) syncNetworkUtil() {
|
|
||||||
if kl.iptablesMasqueradeBit < 0 || kl.iptablesMasqueradeBit > 31 {
|
|
||||||
glog.Errorf("invalid iptables-masquerade-bit %v not in [0, 31]", kl.iptablesMasqueradeBit)
|
|
||||||
return
|
|
||||||
}
|
|
||||||
|
|
||||||
if kl.iptablesDropBit < 0 || kl.iptablesDropBit > 31 {
|
|
||||||
glog.Errorf("invalid iptables-drop-bit %v not in [0, 31]", kl.iptablesDropBit)
|
|
||||||
return
|
|
||||||
}
|
|
||||||
|
|
||||||
if kl.iptablesDropBit == kl.iptablesMasqueradeBit {
|
|
||||||
glog.Errorf("iptables-masquerade-bit %v and iptables-drop-bit %v must be different", kl.iptablesMasqueradeBit, kl.iptablesDropBit)
|
|
||||||
return
|
|
||||||
}
|
|
||||||
|
|
||||||
// Setup KUBE-MARK-DROP rules
|
|
||||||
dropMark := getIPTablesMark(kl.iptablesDropBit)
|
|
||||||
if _, err := kl.iptClient.EnsureChain(utiliptables.TableNAT, KubeMarkDropChain); err != nil {
|
|
||||||
glog.Errorf("Failed to ensure that %s chain %s exists: %v", utiliptables.TableNAT, KubeMarkDropChain, err)
|
|
||||||
return
|
|
||||||
}
|
|
||||||
if _, err := kl.iptClient.EnsureRule(utiliptables.Append, utiliptables.TableNAT, KubeMarkDropChain, "-j", "MARK", "--set-xmark", dropMark); err != nil {
|
|
||||||
glog.Errorf("Failed to ensure marking rule for %v: %v", KubeMarkDropChain, err)
|
|
||||||
return
|
|
||||||
}
|
|
||||||
if _, err := kl.iptClient.EnsureChain(utiliptables.TableFilter, KubeFirewallChain); err != nil {
|
|
||||||
glog.Errorf("Failed to ensure that %s chain %s exists: %v", utiliptables.TableFilter, KubeFirewallChain, err)
|
|
||||||
return
|
|
||||||
}
|
|
||||||
if _, err := kl.iptClient.EnsureRule(utiliptables.Append, utiliptables.TableFilter, KubeFirewallChain,
|
|
||||||
"-m", "comment", "--comment", "kubernetes firewall for dropping marked packets",
|
|
||||||
"-m", "mark", "--mark", dropMark,
|
|
||||||
"-j", "DROP"); err != nil {
|
|
||||||
glog.Errorf("Failed to ensure rule to drop packet marked by %v in %v chain %v: %v", KubeMarkDropChain, utiliptables.TableFilter, KubeFirewallChain, err)
|
|
||||||
return
|
|
||||||
}
|
|
||||||
if _, err := kl.iptClient.EnsureRule(utiliptables.Prepend, utiliptables.TableFilter, utiliptables.ChainOutput, "-j", string(KubeFirewallChain)); err != nil {
|
|
||||||
glog.Errorf("Failed to ensure that %s chain %s jumps to %s: %v", utiliptables.TableFilter, utiliptables.ChainOutput, KubeFirewallChain, err)
|
|
||||||
return
|
|
||||||
}
|
|
||||||
if _, err := kl.iptClient.EnsureRule(utiliptables.Prepend, utiliptables.TableFilter, utiliptables.ChainInput, "-j", string(KubeFirewallChain)); err != nil {
|
|
||||||
glog.Errorf("Failed to ensure that %s chain %s jumps to %s: %v", utiliptables.TableFilter, utiliptables.ChainInput, KubeFirewallChain, err)
|
|
||||||
return
|
|
||||||
}
|
|
||||||
|
|
||||||
// Setup KUBE-MARK-MASQ rules
|
|
||||||
masqueradeMark := getIPTablesMark(kl.iptablesMasqueradeBit)
|
|
||||||
if _, err := kl.iptClient.EnsureChain(utiliptables.TableNAT, KubeMarkMasqChain); err != nil {
|
|
||||||
glog.Errorf("Failed to ensure that %s chain %s exists: %v", utiliptables.TableNAT, KubeMarkMasqChain, err)
|
|
||||||
return
|
|
||||||
}
|
|
||||||
if _, err := kl.iptClient.EnsureChain(utiliptables.TableNAT, KubePostroutingChain); err != nil {
|
|
||||||
glog.Errorf("Failed to ensure that %s chain %s exists: %v", utiliptables.TableNAT, KubePostroutingChain, err)
|
|
||||||
return
|
|
||||||
}
|
|
||||||
if _, err := kl.iptClient.EnsureRule(utiliptables.Append, utiliptables.TableNAT, KubeMarkMasqChain, "-j", "MARK", "--set-xmark", masqueradeMark); err != nil {
|
|
||||||
glog.Errorf("Failed to ensure marking rule for %v: %v", KubeMarkMasqChain, err)
|
|
||||||
return
|
|
||||||
}
|
|
||||||
if _, err := kl.iptClient.EnsureRule(utiliptables.Prepend, utiliptables.TableNAT, utiliptables.ChainPostrouting,
|
|
||||||
"-m", "comment", "--comment", "kubernetes postrouting rules", "-j", string(KubePostroutingChain)); err != nil {
|
|
||||||
glog.Errorf("Failed to ensure that %s chain %s jumps to %s: %v", utiliptables.TableNAT, utiliptables.ChainPostrouting, KubePostroutingChain, err)
|
|
||||||
return
|
|
||||||
}
|
|
||||||
if _, err := kl.iptClient.EnsureRule(utiliptables.Append, utiliptables.TableNAT, KubePostroutingChain,
|
|
||||||
"-m", "comment", "--comment", "kubernetes service traffic requiring SNAT",
|
|
||||||
"-m", "mark", "--mark", masqueradeMark, "-j", "MASQUERADE"); err != nil {
|
|
||||||
glog.Errorf("Failed to ensure SNAT rule for packets marked by %v in %v chain %v: %v", KubeMarkMasqChain, utiliptables.TableNAT, KubePostroutingChain, err)
|
|
||||||
return
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// getIPTablesMark returns the fwmark given the bit
|
|
||||||
func getIPTablesMark(bit int) string {
|
|
||||||
value := 1 << uint(bit)
|
|
||||||
return fmt.Sprintf("%#08x/%#08x", value, value)
|
|
||||||
}
|
|
||||||
|
|
||||||
// GetPodDNS returns DNS settings for the pod.
|
// GetPodDNS returns DNS settings for the pod.
|
||||||
// This function is defined in kubecontainer.RuntimeHelper interface so we
|
// This function is defined in kubecontainer.RuntimeHelper interface so we
|
||||||
// have to implement it.
|
// have to implement it.
|
||||||
|
111
pkg/kubelet/kubelet_network_linux.go
Normal file
111
pkg/kubelet/kubelet_network_linux.go
Normal file
@ -0,0 +1,111 @@
|
|||||||
|
// +build linux
|
||||||
|
|
||||||
|
/*
|
||||||
|
Copyright 2018 The Kubernetes Authors.
|
||||||
|
|
||||||
|
Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
you may not use this file except in compliance with the License.
|
||||||
|
You may obtain a copy of the License at
|
||||||
|
|
||||||
|
http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
|
||||||
|
Unless required by applicable law or agreed to in writing, software
|
||||||
|
distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
See the License for the specific language governing permissions and
|
||||||
|
limitations under the License.
|
||||||
|
*/
|
||||||
|
|
||||||
|
package kubelet
|
||||||
|
|
||||||
|
import (
|
||||||
|
"fmt"
|
||||||
|
|
||||||
|
"github.com/golang/glog"
|
||||||
|
utiliptables "k8s.io/kubernetes/pkg/util/iptables"
|
||||||
|
)
|
||||||
|
|
||||||
|
// syncNetworkUtil ensures the network utility are present on host.
|
||||||
|
// Network util includes:
|
||||||
|
// 1. In nat table, KUBE-MARK-DROP rule to mark connections for dropping
|
||||||
|
// Marked connection will be drop on INPUT/OUTPUT Chain in filter table
|
||||||
|
// 2. In nat table, KUBE-MARK-MASQ rule to mark connections for SNAT
|
||||||
|
// Marked connection will get SNAT on POSTROUTING Chain in nat table
|
||||||
|
func (kl *Kubelet) syncNetworkUtil() {
|
||||||
|
if kl.iptablesMasqueradeBit < 0 || kl.iptablesMasqueradeBit > 31 {
|
||||||
|
glog.Errorf("invalid iptables-masquerade-bit %v not in [0, 31]", kl.iptablesMasqueradeBit)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
if kl.iptablesDropBit < 0 || kl.iptablesDropBit > 31 {
|
||||||
|
glog.Errorf("invalid iptables-drop-bit %v not in [0, 31]", kl.iptablesDropBit)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
if kl.iptablesDropBit == kl.iptablesMasqueradeBit {
|
||||||
|
glog.Errorf("iptables-masquerade-bit %v and iptables-drop-bit %v must be different", kl.iptablesMasqueradeBit, kl.iptablesDropBit)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
// Setup KUBE-MARK-DROP rules
|
||||||
|
dropMark := getIPTablesMark(kl.iptablesDropBit)
|
||||||
|
if _, err := kl.iptClient.EnsureChain(utiliptables.TableNAT, KubeMarkDropChain); err != nil {
|
||||||
|
glog.Errorf("Failed to ensure that %s chain %s exists: %v", utiliptables.TableNAT, KubeMarkDropChain, err)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
if _, err := kl.iptClient.EnsureRule(utiliptables.Append, utiliptables.TableNAT, KubeMarkDropChain, "-j", "MARK", "--set-xmark", dropMark); err != nil {
|
||||||
|
glog.Errorf("Failed to ensure marking rule for %v: %v", KubeMarkDropChain, err)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
if _, err := kl.iptClient.EnsureChain(utiliptables.TableFilter, KubeFirewallChain); err != nil {
|
||||||
|
glog.Errorf("Failed to ensure that %s chain %s exists: %v", utiliptables.TableFilter, KubeFirewallChain, err)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
if _, err := kl.iptClient.EnsureRule(utiliptables.Append, utiliptables.TableFilter, KubeFirewallChain,
|
||||||
|
"-m", "comment", "--comment", "kubernetes firewall for dropping marked packets",
|
||||||
|
"-m", "mark", "--mark", dropMark,
|
||||||
|
"-j", "DROP"); err != nil {
|
||||||
|
glog.Errorf("Failed to ensure rule to drop packet marked by %v in %v chain %v: %v", KubeMarkDropChain, utiliptables.TableFilter, KubeFirewallChain, err)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
if _, err := kl.iptClient.EnsureRule(utiliptables.Prepend, utiliptables.TableFilter, utiliptables.ChainOutput, "-j", string(KubeFirewallChain)); err != nil {
|
||||||
|
glog.Errorf("Failed to ensure that %s chain %s jumps to %s: %v", utiliptables.TableFilter, utiliptables.ChainOutput, KubeFirewallChain, err)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
if _, err := kl.iptClient.EnsureRule(utiliptables.Prepend, utiliptables.TableFilter, utiliptables.ChainInput, "-j", string(KubeFirewallChain)); err != nil {
|
||||||
|
glog.Errorf("Failed to ensure that %s chain %s jumps to %s: %v", utiliptables.TableFilter, utiliptables.ChainInput, KubeFirewallChain, err)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
// Setup KUBE-MARK-MASQ rules
|
||||||
|
masqueradeMark := getIPTablesMark(kl.iptablesMasqueradeBit)
|
||||||
|
if _, err := kl.iptClient.EnsureChain(utiliptables.TableNAT, KubeMarkMasqChain); err != nil {
|
||||||
|
glog.Errorf("Failed to ensure that %s chain %s exists: %v", utiliptables.TableNAT, KubeMarkMasqChain, err)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
if _, err := kl.iptClient.EnsureChain(utiliptables.TableNAT, KubePostroutingChain); err != nil {
|
||||||
|
glog.Errorf("Failed to ensure that %s chain %s exists: %v", utiliptables.TableNAT, KubePostroutingChain, err)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
if _, err := kl.iptClient.EnsureRule(utiliptables.Append, utiliptables.TableNAT, KubeMarkMasqChain, "-j", "MARK", "--set-xmark", masqueradeMark); err != nil {
|
||||||
|
glog.Errorf("Failed to ensure marking rule for %v: %v", KubeMarkMasqChain, err)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
if _, err := kl.iptClient.EnsureRule(utiliptables.Prepend, utiliptables.TableNAT, utiliptables.ChainPostrouting,
|
||||||
|
"-m", "comment", "--comment", "kubernetes postrouting rules", "-j", string(KubePostroutingChain)); err != nil {
|
||||||
|
glog.Errorf("Failed to ensure that %s chain %s jumps to %s: %v", utiliptables.TableNAT, utiliptables.ChainPostrouting, KubePostroutingChain, err)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
if _, err := kl.iptClient.EnsureRule(utiliptables.Append, utiliptables.TableNAT, KubePostroutingChain,
|
||||||
|
"-m", "comment", "--comment", "kubernetes service traffic requiring SNAT",
|
||||||
|
"-m", "mark", "--mark", masqueradeMark, "-j", "MASQUERADE"); err != nil {
|
||||||
|
glog.Errorf("Failed to ensure SNAT rule for packets marked by %v in %v chain %v: %v", KubeMarkMasqChain, utiliptables.TableNAT, KubePostroutingChain, err)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// getIPTablesMark returns the fwmark given the bit
|
||||||
|
func getIPTablesMark(bit int) string {
|
||||||
|
value := 1 << uint(bit)
|
||||||
|
return fmt.Sprintf("%#08x/%#08x", value, value)
|
||||||
|
}
|
22
pkg/kubelet/kubelet_network_others.go
Normal file
22
pkg/kubelet/kubelet_network_others.go
Normal file
@ -0,0 +1,22 @@
|
|||||||
|
// +build !linux
|
||||||
|
|
||||||
|
/*
|
||||||
|
Copyright 2018 The Kubernetes Authors.
|
||||||
|
|
||||||
|
Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
you may not use this file except in compliance with the License.
|
||||||
|
You may obtain a copy of the License at
|
||||||
|
|
||||||
|
http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
|
||||||
|
Unless required by applicable law or agreed to in writing, software
|
||||||
|
distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
See the License for the specific language governing permissions and
|
||||||
|
limitations under the License.
|
||||||
|
*/
|
||||||
|
|
||||||
|
package kubelet
|
||||||
|
|
||||||
|
// Do nothing.
|
||||||
|
func (kl *Kubelet) syncNetworkUtil() {}
|
@ -1,3 +1,5 @@
|
|||||||
|
// +build linux
|
||||||
|
|
||||||
/*
|
/*
|
||||||
Copyright 2016 The Kubernetes Authors.
|
Copyright 2016 The Kubernetes Authors.
|
||||||
|
|
||||||
|
426
pkg/kubelet/kubelet_pods_linux_test.go
Normal file
426
pkg/kubelet/kubelet_pods_linux_test.go
Normal file
@ -0,0 +1,426 @@
|
|||||||
|
// +build linux
|
||||||
|
|
||||||
|
/*
|
||||||
|
Copyright 2018 The Kubernetes Authors.
|
||||||
|
|
||||||
|
Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
you may not use this file except in compliance with the License.
|
||||||
|
You may obtain a copy of the License at
|
||||||
|
|
||||||
|
http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
|
||||||
|
Unless required by applicable law or agreed to in writing, software
|
||||||
|
distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
See the License for the specific language governing permissions and
|
||||||
|
limitations under the License.
|
||||||
|
*/
|
||||||
|
|
||||||
|
package kubelet
|
||||||
|
|
||||||
|
import (
|
||||||
|
"testing"
|
||||||
|
|
||||||
|
"github.com/stretchr/testify/assert"
|
||||||
|
"k8s.io/api/core/v1"
|
||||||
|
utilfeature "k8s.io/apiserver/pkg/util/feature"
|
||||||
|
|
||||||
|
_ "k8s.io/kubernetes/pkg/apis/core/install"
|
||||||
|
runtimeapi "k8s.io/kubernetes/pkg/kubelet/apis/cri/runtime/v1alpha2"
|
||||||
|
kubecontainer "k8s.io/kubernetes/pkg/kubelet/container"
|
||||||
|
"k8s.io/kubernetes/pkg/util/mount"
|
||||||
|
volumetest "k8s.io/kubernetes/pkg/volume/testing"
|
||||||
|
)
|
||||||
|
|
||||||
|
func TestMakeMounts(t *testing.T) {
|
||||||
|
bTrue := true
|
||||||
|
propagationHostToContainer := v1.MountPropagationHostToContainer
|
||||||
|
propagationBidirectional := v1.MountPropagationBidirectional
|
||||||
|
propagationNone := v1.MountPropagationNone
|
||||||
|
|
||||||
|
testCases := map[string]struct {
|
||||||
|
container v1.Container
|
||||||
|
podVolumes kubecontainer.VolumeMap
|
||||||
|
expectErr bool
|
||||||
|
expectedErrMsg string
|
||||||
|
expectedMounts []kubecontainer.Mount
|
||||||
|
}{
|
||||||
|
"valid mounts in unprivileged container": {
|
||||||
|
podVolumes: kubecontainer.VolumeMap{
|
||||||
|
"disk": kubecontainer.VolumeInfo{Mounter: &stubVolume{path: "/mnt/disk"}},
|
||||||
|
"disk4": kubecontainer.VolumeInfo{Mounter: &stubVolume{path: "/mnt/host"}},
|
||||||
|
"disk5": kubecontainer.VolumeInfo{Mounter: &stubVolume{path: "/var/lib/kubelet/podID/volumes/empty/disk5"}},
|
||||||
|
},
|
||||||
|
container: v1.Container{
|
||||||
|
Name: "container1",
|
||||||
|
VolumeMounts: []v1.VolumeMount{
|
||||||
|
{
|
||||||
|
MountPath: "/etc/hosts",
|
||||||
|
Name: "disk",
|
||||||
|
ReadOnly: false,
|
||||||
|
MountPropagation: &propagationHostToContainer,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
MountPath: "/mnt/path3",
|
||||||
|
Name: "disk",
|
||||||
|
ReadOnly: true,
|
||||||
|
MountPropagation: &propagationNone,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
MountPath: "/mnt/path4",
|
||||||
|
Name: "disk4",
|
||||||
|
ReadOnly: false,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
MountPath: "/mnt/path5",
|
||||||
|
Name: "disk5",
|
||||||
|
ReadOnly: false,
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
expectedMounts: []kubecontainer.Mount{
|
||||||
|
{
|
||||||
|
Name: "disk",
|
||||||
|
ContainerPath: "/etc/hosts",
|
||||||
|
HostPath: "/mnt/disk",
|
||||||
|
ReadOnly: false,
|
||||||
|
SELinuxRelabel: false,
|
||||||
|
Propagation: runtimeapi.MountPropagation_PROPAGATION_HOST_TO_CONTAINER,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Name: "disk",
|
||||||
|
ContainerPath: "/mnt/path3",
|
||||||
|
HostPath: "/mnt/disk",
|
||||||
|
ReadOnly: true,
|
||||||
|
SELinuxRelabel: false,
|
||||||
|
Propagation: runtimeapi.MountPropagation_PROPAGATION_PRIVATE,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Name: "disk4",
|
||||||
|
ContainerPath: "/mnt/path4",
|
||||||
|
HostPath: "/mnt/host",
|
||||||
|
ReadOnly: false,
|
||||||
|
SELinuxRelabel: false,
|
||||||
|
Propagation: runtimeapi.MountPropagation_PROPAGATION_PRIVATE,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Name: "disk5",
|
||||||
|
ContainerPath: "/mnt/path5",
|
||||||
|
HostPath: "/var/lib/kubelet/podID/volumes/empty/disk5",
|
||||||
|
ReadOnly: false,
|
||||||
|
SELinuxRelabel: false,
|
||||||
|
Propagation: runtimeapi.MountPropagation_PROPAGATION_PRIVATE,
|
||||||
|
},
|
||||||
|
},
|
||||||
|
expectErr: false,
|
||||||
|
},
|
||||||
|
"valid mounts in privileged container": {
|
||||||
|
podVolumes: kubecontainer.VolumeMap{
|
||||||
|
"disk": kubecontainer.VolumeInfo{Mounter: &stubVolume{path: "/mnt/disk"}},
|
||||||
|
"disk4": kubecontainer.VolumeInfo{Mounter: &stubVolume{path: "/mnt/host"}},
|
||||||
|
"disk5": kubecontainer.VolumeInfo{Mounter: &stubVolume{path: "/var/lib/kubelet/podID/volumes/empty/disk5"}},
|
||||||
|
},
|
||||||
|
container: v1.Container{
|
||||||
|
Name: "container1",
|
||||||
|
VolumeMounts: []v1.VolumeMount{
|
||||||
|
{
|
||||||
|
MountPath: "/etc/hosts",
|
||||||
|
Name: "disk",
|
||||||
|
ReadOnly: false,
|
||||||
|
MountPropagation: &propagationBidirectional,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
MountPath: "/mnt/path3",
|
||||||
|
Name: "disk",
|
||||||
|
ReadOnly: true,
|
||||||
|
MountPropagation: &propagationHostToContainer,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
MountPath: "/mnt/path4",
|
||||||
|
Name: "disk4",
|
||||||
|
ReadOnly: false,
|
||||||
|
},
|
||||||
|
},
|
||||||
|
SecurityContext: &v1.SecurityContext{
|
||||||
|
Privileged: &bTrue,
|
||||||
|
},
|
||||||
|
},
|
||||||
|
expectedMounts: []kubecontainer.Mount{
|
||||||
|
{
|
||||||
|
Name: "disk",
|
||||||
|
ContainerPath: "/etc/hosts",
|
||||||
|
HostPath: "/mnt/disk",
|
||||||
|
ReadOnly: false,
|
||||||
|
SELinuxRelabel: false,
|
||||||
|
Propagation: runtimeapi.MountPropagation_PROPAGATION_BIDIRECTIONAL,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Name: "disk",
|
||||||
|
ContainerPath: "/mnt/path3",
|
||||||
|
HostPath: "/mnt/disk",
|
||||||
|
ReadOnly: true,
|
||||||
|
SELinuxRelabel: false,
|
||||||
|
Propagation: runtimeapi.MountPropagation_PROPAGATION_HOST_TO_CONTAINER,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Name: "disk4",
|
||||||
|
ContainerPath: "/mnt/path4",
|
||||||
|
HostPath: "/mnt/host",
|
||||||
|
ReadOnly: false,
|
||||||
|
SELinuxRelabel: false,
|
||||||
|
Propagation: runtimeapi.MountPropagation_PROPAGATION_PRIVATE,
|
||||||
|
},
|
||||||
|
},
|
||||||
|
expectErr: false,
|
||||||
|
},
|
||||||
|
"invalid absolute SubPath": {
|
||||||
|
podVolumes: kubecontainer.VolumeMap{
|
||||||
|
"disk": kubecontainer.VolumeInfo{Mounter: &stubVolume{path: "/mnt/disk"}},
|
||||||
|
},
|
||||||
|
container: v1.Container{
|
||||||
|
VolumeMounts: []v1.VolumeMount{
|
||||||
|
{
|
||||||
|
MountPath: "/mnt/path3",
|
||||||
|
SubPath: "/must/not/be/absolute",
|
||||||
|
Name: "disk",
|
||||||
|
ReadOnly: true,
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
expectErr: true,
|
||||||
|
expectedErrMsg: "error SubPath `/must/not/be/absolute` must not be an absolute path",
|
||||||
|
},
|
||||||
|
"invalid SubPath with backsteps": {
|
||||||
|
podVolumes: kubecontainer.VolumeMap{
|
||||||
|
"disk": kubecontainer.VolumeInfo{Mounter: &stubVolume{path: "/mnt/disk"}},
|
||||||
|
},
|
||||||
|
container: v1.Container{
|
||||||
|
VolumeMounts: []v1.VolumeMount{
|
||||||
|
{
|
||||||
|
MountPath: "/mnt/path3",
|
||||||
|
SubPath: "no/backsteps/../allowed",
|
||||||
|
Name: "disk",
|
||||||
|
ReadOnly: true,
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
expectErr: true,
|
||||||
|
expectedErrMsg: "unable to provision SubPath `no/backsteps/../allowed`: must not contain '..'",
|
||||||
|
},
|
||||||
|
"volume doesn't exist": {
|
||||||
|
podVolumes: kubecontainer.VolumeMap{},
|
||||||
|
container: v1.Container{
|
||||||
|
VolumeMounts: []v1.VolumeMount{
|
||||||
|
{
|
||||||
|
MountPath: "/mnt/path3",
|
||||||
|
Name: "disk",
|
||||||
|
ReadOnly: true,
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
expectErr: true,
|
||||||
|
expectedErrMsg: "cannot find volume \"disk\" to mount into container \"\"",
|
||||||
|
},
|
||||||
|
"volume mounter is nil": {
|
||||||
|
podVolumes: kubecontainer.VolumeMap{
|
||||||
|
"disk": kubecontainer.VolumeInfo{},
|
||||||
|
},
|
||||||
|
container: v1.Container{
|
||||||
|
VolumeMounts: []v1.VolumeMount{
|
||||||
|
{
|
||||||
|
MountPath: "/mnt/path3",
|
||||||
|
Name: "disk",
|
||||||
|
ReadOnly: true,
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
expectErr: true,
|
||||||
|
expectedErrMsg: "cannot find volume \"disk\" to mount into container \"\"",
|
||||||
|
},
|
||||||
|
}
|
||||||
|
|
||||||
|
for name, tc := range testCases {
|
||||||
|
t.Run(name, func(t *testing.T) {
|
||||||
|
fm := &mount.FakeMounter{}
|
||||||
|
pod := v1.Pod{
|
||||||
|
Spec: v1.PodSpec{
|
||||||
|
HostNetwork: true,
|
||||||
|
},
|
||||||
|
}
|
||||||
|
// test makeMounts with enabled mount propagation
|
||||||
|
err := utilfeature.DefaultFeatureGate.Set("MountPropagation=true")
|
||||||
|
if err != nil {
|
||||||
|
t.Errorf("Failed to enable feature gate for MountPropagation: %v", err)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
mounts, _, err := makeMounts(&pod, "/pod", &tc.container, "fakepodname", "", "", tc.podVolumes, fm, nil)
|
||||||
|
|
||||||
|
// validate only the error if we expect an error
|
||||||
|
if tc.expectErr {
|
||||||
|
if err == nil || err.Error() != tc.expectedErrMsg {
|
||||||
|
t.Fatalf("expected error message `%s` but got `%v`", tc.expectedErrMsg, err)
|
||||||
|
}
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
// otherwise validate the mounts
|
||||||
|
if err != nil {
|
||||||
|
t.Fatal(err)
|
||||||
|
}
|
||||||
|
|
||||||
|
assert.Equal(t, tc.expectedMounts, mounts, "mounts of container %+v", tc.container)
|
||||||
|
|
||||||
|
// test makeMounts with disabled mount propagation
|
||||||
|
err = utilfeature.DefaultFeatureGate.Set("MountPropagation=false")
|
||||||
|
if err != nil {
|
||||||
|
t.Errorf("Failed to enable feature gate for MountPropagation: %v", err)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
mounts, _, err = makeMounts(&pod, "/pod", &tc.container, "fakepodname", "", "", tc.podVolumes, fm, nil)
|
||||||
|
if !tc.expectErr {
|
||||||
|
expectedPrivateMounts := []kubecontainer.Mount{}
|
||||||
|
for _, mount := range tc.expectedMounts {
|
||||||
|
// all mounts are expected to be private when mount
|
||||||
|
// propagation is disabled
|
||||||
|
mount.Propagation = runtimeapi.MountPropagation_PROPAGATION_PRIVATE
|
||||||
|
expectedPrivateMounts = append(expectedPrivateMounts, mount)
|
||||||
|
}
|
||||||
|
assert.Equal(t, expectedPrivateMounts, mounts, "mounts of container %+v", tc.container)
|
||||||
|
}
|
||||||
|
})
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestMakeBlockVolumes(t *testing.T) {
|
||||||
|
testKubelet := newTestKubelet(t, false /* controllerAttachDetachEnabled */)
|
||||||
|
defer testKubelet.Cleanup()
|
||||||
|
kubelet := testKubelet.kubelet
|
||||||
|
testCases := map[string]struct {
|
||||||
|
container v1.Container
|
||||||
|
podVolumes kubecontainer.VolumeMap
|
||||||
|
expectErr bool
|
||||||
|
expectedErrMsg string
|
||||||
|
expectedDevices []kubecontainer.DeviceInfo
|
||||||
|
}{
|
||||||
|
"valid volumeDevices in container": {
|
||||||
|
podVolumes: kubecontainer.VolumeMap{
|
||||||
|
"disk1": kubecontainer.VolumeInfo{BlockVolumeMapper: &stubBlockVolume{dirPath: "/dev/", volName: "sda"}},
|
||||||
|
"disk2": kubecontainer.VolumeInfo{BlockVolumeMapper: &stubBlockVolume{dirPath: "/dev/disk/by-path/", volName: "diskPath"}, ReadOnly: true},
|
||||||
|
"disk3": kubecontainer.VolumeInfo{BlockVolumeMapper: &stubBlockVolume{dirPath: "/dev/disk/by-id/", volName: "diskUuid"}},
|
||||||
|
"disk4": kubecontainer.VolumeInfo{BlockVolumeMapper: &stubBlockVolume{dirPath: "/var/lib/", volName: "rawdisk"}, ReadOnly: true},
|
||||||
|
},
|
||||||
|
container: v1.Container{
|
||||||
|
Name: "container1",
|
||||||
|
VolumeDevices: []v1.VolumeDevice{
|
||||||
|
{
|
||||||
|
DevicePath: "/dev/sda",
|
||||||
|
Name: "disk1",
|
||||||
|
},
|
||||||
|
{
|
||||||
|
DevicePath: "/dev/xvda",
|
||||||
|
Name: "disk2",
|
||||||
|
},
|
||||||
|
{
|
||||||
|
DevicePath: "/dev/xvdb",
|
||||||
|
Name: "disk3",
|
||||||
|
},
|
||||||
|
{
|
||||||
|
DevicePath: "/mnt/rawdisk",
|
||||||
|
Name: "disk4",
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
expectedDevices: []kubecontainer.DeviceInfo{
|
||||||
|
{
|
||||||
|
PathInContainer: "/dev/sda",
|
||||||
|
PathOnHost: "/dev/sda",
|
||||||
|
Permissions: "mrw",
|
||||||
|
},
|
||||||
|
{
|
||||||
|
PathInContainer: "/dev/xvda",
|
||||||
|
PathOnHost: "/dev/disk/by-path/diskPath",
|
||||||
|
Permissions: "r",
|
||||||
|
},
|
||||||
|
{
|
||||||
|
PathInContainer: "/dev/xvdb",
|
||||||
|
PathOnHost: "/dev/disk/by-id/diskUuid",
|
||||||
|
Permissions: "mrw",
|
||||||
|
},
|
||||||
|
{
|
||||||
|
PathInContainer: "/mnt/rawdisk",
|
||||||
|
PathOnHost: "/var/lib/rawdisk",
|
||||||
|
Permissions: "r",
|
||||||
|
},
|
||||||
|
},
|
||||||
|
expectErr: false,
|
||||||
|
},
|
||||||
|
"invalid absolute Path": {
|
||||||
|
podVolumes: kubecontainer.VolumeMap{
|
||||||
|
"disk": kubecontainer.VolumeInfo{BlockVolumeMapper: &stubBlockVolume{dirPath: "/dev/", volName: "sda"}},
|
||||||
|
},
|
||||||
|
container: v1.Container{
|
||||||
|
VolumeDevices: []v1.VolumeDevice{
|
||||||
|
{
|
||||||
|
DevicePath: "must/be/absolute",
|
||||||
|
Name: "disk",
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
expectErr: true,
|
||||||
|
expectedErrMsg: "error DevicePath `must/be/absolute` must be an absolute path",
|
||||||
|
},
|
||||||
|
"volume doesn't exist": {
|
||||||
|
podVolumes: kubecontainer.VolumeMap{},
|
||||||
|
container: v1.Container{
|
||||||
|
VolumeDevices: []v1.VolumeDevice{
|
||||||
|
{
|
||||||
|
DevicePath: "/dev/sdaa",
|
||||||
|
Name: "disk",
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
expectErr: true,
|
||||||
|
expectedErrMsg: "cannot find volume \"disk\" to pass into container \"\"",
|
||||||
|
},
|
||||||
|
"volume BlockVolumeMapper is nil": {
|
||||||
|
podVolumes: kubecontainer.VolumeMap{
|
||||||
|
"disk": kubecontainer.VolumeInfo{},
|
||||||
|
},
|
||||||
|
container: v1.Container{
|
||||||
|
VolumeDevices: []v1.VolumeDevice{
|
||||||
|
{
|
||||||
|
DevicePath: "/dev/sdzz",
|
||||||
|
Name: "disk",
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
expectErr: true,
|
||||||
|
expectedErrMsg: "cannot find volume \"disk\" to pass into container \"\"",
|
||||||
|
},
|
||||||
|
}
|
||||||
|
|
||||||
|
for name, tc := range testCases {
|
||||||
|
t.Run(name, func(t *testing.T) {
|
||||||
|
pod := v1.Pod{
|
||||||
|
Spec: v1.PodSpec{
|
||||||
|
HostNetwork: true,
|
||||||
|
},
|
||||||
|
}
|
||||||
|
blkutil := volumetest.NewBlockVolumePathHandler()
|
||||||
|
blkVolumes, err := kubelet.makeBlockVolumes(&pod, &tc.container, tc.podVolumes, blkutil)
|
||||||
|
// validate only the error if we expect an error
|
||||||
|
if tc.expectErr {
|
||||||
|
if err == nil || err.Error() != tc.expectedErrMsg {
|
||||||
|
t.Fatalf("expected error message `%s` but got `%v`", tc.expectedErrMsg, err)
|
||||||
|
}
|
||||||
|
return
|
||||||
|
}
|
||||||
|
// otherwise validate the devices
|
||||||
|
if err != nil {
|
||||||
|
t.Fatal(err)
|
||||||
|
}
|
||||||
|
assert.Equal(t, tc.expectedDevices, blkVolumes, "devices of container %+v", tc.container)
|
||||||
|
})
|
||||||
|
}
|
||||||
|
}
|
@ -41,275 +41,13 @@ import (
|
|||||||
// to "v1"?
|
// to "v1"?
|
||||||
|
|
||||||
_ "k8s.io/kubernetes/pkg/apis/core/install"
|
_ "k8s.io/kubernetes/pkg/apis/core/install"
|
||||||
runtimeapi "k8s.io/kubernetes/pkg/kubelet/apis/cri/runtime/v1alpha2"
|
|
||||||
kubecontainer "k8s.io/kubernetes/pkg/kubelet/container"
|
kubecontainer "k8s.io/kubernetes/pkg/kubelet/container"
|
||||||
containertest "k8s.io/kubernetes/pkg/kubelet/container/testing"
|
containertest "k8s.io/kubernetes/pkg/kubelet/container/testing"
|
||||||
"k8s.io/kubernetes/pkg/kubelet/server/portforward"
|
"k8s.io/kubernetes/pkg/kubelet/server/portforward"
|
||||||
"k8s.io/kubernetes/pkg/kubelet/server/remotecommand"
|
"k8s.io/kubernetes/pkg/kubelet/server/remotecommand"
|
||||||
"k8s.io/kubernetes/pkg/util/mount"
|
"k8s.io/kubernetes/pkg/util/mount"
|
||||||
volumetest "k8s.io/kubernetes/pkg/volume/testing"
|
|
||||||
)
|
)
|
||||||
|
|
||||||
func TestMakeMounts(t *testing.T) {
|
|
||||||
bTrue := true
|
|
||||||
propagationHostToContainer := v1.MountPropagationHostToContainer
|
|
||||||
propagationBidirectional := v1.MountPropagationBidirectional
|
|
||||||
propagationNone := v1.MountPropagationNone
|
|
||||||
|
|
||||||
testCases := map[string]struct {
|
|
||||||
container v1.Container
|
|
||||||
podVolumes kubecontainer.VolumeMap
|
|
||||||
expectErr bool
|
|
||||||
expectedErrMsg string
|
|
||||||
expectedMounts []kubecontainer.Mount
|
|
||||||
}{
|
|
||||||
"valid mounts in unprivileged container": {
|
|
||||||
podVolumes: kubecontainer.VolumeMap{
|
|
||||||
"disk": kubecontainer.VolumeInfo{Mounter: &stubVolume{path: "/mnt/disk"}},
|
|
||||||
"disk4": kubecontainer.VolumeInfo{Mounter: &stubVolume{path: "/mnt/host"}},
|
|
||||||
"disk5": kubecontainer.VolumeInfo{Mounter: &stubVolume{path: "/var/lib/kubelet/podID/volumes/empty/disk5"}},
|
|
||||||
},
|
|
||||||
container: v1.Container{
|
|
||||||
Name: "container1",
|
|
||||||
VolumeMounts: []v1.VolumeMount{
|
|
||||||
{
|
|
||||||
MountPath: "/etc/hosts",
|
|
||||||
Name: "disk",
|
|
||||||
ReadOnly: false,
|
|
||||||
MountPropagation: &propagationHostToContainer,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
MountPath: "/mnt/path3",
|
|
||||||
Name: "disk",
|
|
||||||
ReadOnly: true,
|
|
||||||
MountPropagation: &propagationNone,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
MountPath: "/mnt/path4",
|
|
||||||
Name: "disk4",
|
|
||||||
ReadOnly: false,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
MountPath: "/mnt/path5",
|
|
||||||
Name: "disk5",
|
|
||||||
ReadOnly: false,
|
|
||||||
},
|
|
||||||
},
|
|
||||||
},
|
|
||||||
expectedMounts: []kubecontainer.Mount{
|
|
||||||
{
|
|
||||||
Name: "disk",
|
|
||||||
ContainerPath: "/etc/hosts",
|
|
||||||
HostPath: "/mnt/disk",
|
|
||||||
ReadOnly: false,
|
|
||||||
SELinuxRelabel: false,
|
|
||||||
Propagation: runtimeapi.MountPropagation_PROPAGATION_HOST_TO_CONTAINER,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
Name: "disk",
|
|
||||||
ContainerPath: "/mnt/path3",
|
|
||||||
HostPath: "/mnt/disk",
|
|
||||||
ReadOnly: true,
|
|
||||||
SELinuxRelabel: false,
|
|
||||||
Propagation: runtimeapi.MountPropagation_PROPAGATION_PRIVATE,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
Name: "disk4",
|
|
||||||
ContainerPath: "/mnt/path4",
|
|
||||||
HostPath: "/mnt/host",
|
|
||||||
ReadOnly: false,
|
|
||||||
SELinuxRelabel: false,
|
|
||||||
Propagation: runtimeapi.MountPropagation_PROPAGATION_PRIVATE,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
Name: "disk5",
|
|
||||||
ContainerPath: "/mnt/path5",
|
|
||||||
HostPath: "/var/lib/kubelet/podID/volumes/empty/disk5",
|
|
||||||
ReadOnly: false,
|
|
||||||
SELinuxRelabel: false,
|
|
||||||
Propagation: runtimeapi.MountPropagation_PROPAGATION_PRIVATE,
|
|
||||||
},
|
|
||||||
},
|
|
||||||
expectErr: false,
|
|
||||||
},
|
|
||||||
"valid mounts in privileged container": {
|
|
||||||
podVolumes: kubecontainer.VolumeMap{
|
|
||||||
"disk": kubecontainer.VolumeInfo{Mounter: &stubVolume{path: "/mnt/disk"}},
|
|
||||||
"disk4": kubecontainer.VolumeInfo{Mounter: &stubVolume{path: "/mnt/host"}},
|
|
||||||
"disk5": kubecontainer.VolumeInfo{Mounter: &stubVolume{path: "/var/lib/kubelet/podID/volumes/empty/disk5"}},
|
|
||||||
},
|
|
||||||
container: v1.Container{
|
|
||||||
Name: "container1",
|
|
||||||
VolumeMounts: []v1.VolumeMount{
|
|
||||||
{
|
|
||||||
MountPath: "/etc/hosts",
|
|
||||||
Name: "disk",
|
|
||||||
ReadOnly: false,
|
|
||||||
MountPropagation: &propagationBidirectional,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
MountPath: "/mnt/path3",
|
|
||||||
Name: "disk",
|
|
||||||
ReadOnly: true,
|
|
||||||
MountPropagation: &propagationHostToContainer,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
MountPath: "/mnt/path4",
|
|
||||||
Name: "disk4",
|
|
||||||
ReadOnly: false,
|
|
||||||
},
|
|
||||||
},
|
|
||||||
SecurityContext: &v1.SecurityContext{
|
|
||||||
Privileged: &bTrue,
|
|
||||||
},
|
|
||||||
},
|
|
||||||
expectedMounts: []kubecontainer.Mount{
|
|
||||||
{
|
|
||||||
Name: "disk",
|
|
||||||
ContainerPath: "/etc/hosts",
|
|
||||||
HostPath: "/mnt/disk",
|
|
||||||
ReadOnly: false,
|
|
||||||
SELinuxRelabel: false,
|
|
||||||
Propagation: runtimeapi.MountPropagation_PROPAGATION_BIDIRECTIONAL,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
Name: "disk",
|
|
||||||
ContainerPath: "/mnt/path3",
|
|
||||||
HostPath: "/mnt/disk",
|
|
||||||
ReadOnly: true,
|
|
||||||
SELinuxRelabel: false,
|
|
||||||
Propagation: runtimeapi.MountPropagation_PROPAGATION_HOST_TO_CONTAINER,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
Name: "disk4",
|
|
||||||
ContainerPath: "/mnt/path4",
|
|
||||||
HostPath: "/mnt/host",
|
|
||||||
ReadOnly: false,
|
|
||||||
SELinuxRelabel: false,
|
|
||||||
Propagation: runtimeapi.MountPropagation_PROPAGATION_PRIVATE,
|
|
||||||
},
|
|
||||||
},
|
|
||||||
expectErr: false,
|
|
||||||
},
|
|
||||||
"invalid absolute SubPath": {
|
|
||||||
podVolumes: kubecontainer.VolumeMap{
|
|
||||||
"disk": kubecontainer.VolumeInfo{Mounter: &stubVolume{path: "/mnt/disk"}},
|
|
||||||
},
|
|
||||||
container: v1.Container{
|
|
||||||
VolumeMounts: []v1.VolumeMount{
|
|
||||||
{
|
|
||||||
MountPath: "/mnt/path3",
|
|
||||||
SubPath: "/must/not/be/absolute",
|
|
||||||
Name: "disk",
|
|
||||||
ReadOnly: true,
|
|
||||||
},
|
|
||||||
},
|
|
||||||
},
|
|
||||||
expectErr: true,
|
|
||||||
expectedErrMsg: "error SubPath `/must/not/be/absolute` must not be an absolute path",
|
|
||||||
},
|
|
||||||
"invalid SubPath with backsteps": {
|
|
||||||
podVolumes: kubecontainer.VolumeMap{
|
|
||||||
"disk": kubecontainer.VolumeInfo{Mounter: &stubVolume{path: "/mnt/disk"}},
|
|
||||||
},
|
|
||||||
container: v1.Container{
|
|
||||||
VolumeMounts: []v1.VolumeMount{
|
|
||||||
{
|
|
||||||
MountPath: "/mnt/path3",
|
|
||||||
SubPath: "no/backsteps/../allowed",
|
|
||||||
Name: "disk",
|
|
||||||
ReadOnly: true,
|
|
||||||
},
|
|
||||||
},
|
|
||||||
},
|
|
||||||
expectErr: true,
|
|
||||||
expectedErrMsg: "unable to provision SubPath `no/backsteps/../allowed`: must not contain '..'",
|
|
||||||
},
|
|
||||||
"volume doesn't exist": {
|
|
||||||
podVolumes: kubecontainer.VolumeMap{},
|
|
||||||
container: v1.Container{
|
|
||||||
VolumeMounts: []v1.VolumeMount{
|
|
||||||
{
|
|
||||||
MountPath: "/mnt/path3",
|
|
||||||
Name: "disk",
|
|
||||||
ReadOnly: true,
|
|
||||||
},
|
|
||||||
},
|
|
||||||
},
|
|
||||||
expectErr: true,
|
|
||||||
expectedErrMsg: "cannot find volume \"disk\" to mount into container \"\"",
|
|
||||||
},
|
|
||||||
"volume mounter is nil": {
|
|
||||||
podVolumes: kubecontainer.VolumeMap{
|
|
||||||
"disk": kubecontainer.VolumeInfo{},
|
|
||||||
},
|
|
||||||
container: v1.Container{
|
|
||||||
VolumeMounts: []v1.VolumeMount{
|
|
||||||
{
|
|
||||||
MountPath: "/mnt/path3",
|
|
||||||
Name: "disk",
|
|
||||||
ReadOnly: true,
|
|
||||||
},
|
|
||||||
},
|
|
||||||
},
|
|
||||||
expectErr: true,
|
|
||||||
expectedErrMsg: "cannot find volume \"disk\" to mount into container \"\"",
|
|
||||||
},
|
|
||||||
}
|
|
||||||
|
|
||||||
for name, tc := range testCases {
|
|
||||||
t.Run(name, func(t *testing.T) {
|
|
||||||
fm := &mount.FakeMounter{}
|
|
||||||
pod := v1.Pod{
|
|
||||||
Spec: v1.PodSpec{
|
|
||||||
HostNetwork: true,
|
|
||||||
},
|
|
||||||
}
|
|
||||||
// test makeMounts with enabled mount propagation
|
|
||||||
err := utilfeature.DefaultFeatureGate.Set("MountPropagation=true")
|
|
||||||
if err != nil {
|
|
||||||
t.Errorf("Failed to enable feature gate for MountPropagation: %v", err)
|
|
||||||
return
|
|
||||||
}
|
|
||||||
|
|
||||||
mounts, _, err := makeMounts(&pod, "/pod", &tc.container, "fakepodname", "", "", tc.podVolumes, fm, nil)
|
|
||||||
|
|
||||||
// validate only the error if we expect an error
|
|
||||||
if tc.expectErr {
|
|
||||||
if err == nil || err.Error() != tc.expectedErrMsg {
|
|
||||||
t.Fatalf("expected error message `%s` but got `%v`", tc.expectedErrMsg, err)
|
|
||||||
}
|
|
||||||
return
|
|
||||||
}
|
|
||||||
|
|
||||||
// otherwise validate the mounts
|
|
||||||
if err != nil {
|
|
||||||
t.Fatal(err)
|
|
||||||
}
|
|
||||||
|
|
||||||
assert.Equal(t, tc.expectedMounts, mounts, "mounts of container %+v", tc.container)
|
|
||||||
|
|
||||||
// test makeMounts with disabled mount propagation
|
|
||||||
err = utilfeature.DefaultFeatureGate.Set("MountPropagation=false")
|
|
||||||
if err != nil {
|
|
||||||
t.Errorf("Failed to enable feature gate for MountPropagation: %v", err)
|
|
||||||
return
|
|
||||||
}
|
|
||||||
mounts, _, err = makeMounts(&pod, "/pod", &tc.container, "fakepodname", "", "", tc.podVolumes, fm, nil)
|
|
||||||
if !tc.expectErr {
|
|
||||||
expectedPrivateMounts := []kubecontainer.Mount{}
|
|
||||||
for _, mount := range tc.expectedMounts {
|
|
||||||
// all mounts are expected to be private when mount
|
|
||||||
// propagation is disabled
|
|
||||||
mount.Propagation = runtimeapi.MountPropagation_PROPAGATION_PRIVATE
|
|
||||||
expectedPrivateMounts = append(expectedPrivateMounts, mount)
|
|
||||||
}
|
|
||||||
assert.Equal(t, expectedPrivateMounts, mounts, "mounts of container %+v", tc.container)
|
|
||||||
}
|
|
||||||
})
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
func TestDisabledSubpath(t *testing.T) {
|
func TestDisabledSubpath(t *testing.T) {
|
||||||
fm := &mount.FakeMounter{}
|
fm := &mount.FakeMounter{}
|
||||||
pod := v1.Pod{
|
pod := v1.Pod{
|
||||||
@ -366,139 +104,6 @@ func TestDisabledSubpath(t *testing.T) {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestMakeBlockVolumes(t *testing.T) {
|
|
||||||
testKubelet := newTestKubelet(t, false /* controllerAttachDetachEnabled */)
|
|
||||||
defer testKubelet.Cleanup()
|
|
||||||
kubelet := testKubelet.kubelet
|
|
||||||
testCases := map[string]struct {
|
|
||||||
container v1.Container
|
|
||||||
podVolumes kubecontainer.VolumeMap
|
|
||||||
expectErr bool
|
|
||||||
expectedErrMsg string
|
|
||||||
expectedDevices []kubecontainer.DeviceInfo
|
|
||||||
}{
|
|
||||||
"valid volumeDevices in container": {
|
|
||||||
podVolumes: kubecontainer.VolumeMap{
|
|
||||||
"disk1": kubecontainer.VolumeInfo{BlockVolumeMapper: &stubBlockVolume{dirPath: "/dev/", volName: "sda"}},
|
|
||||||
"disk2": kubecontainer.VolumeInfo{BlockVolumeMapper: &stubBlockVolume{dirPath: "/dev/disk/by-path/", volName: "diskPath"}, ReadOnly: true},
|
|
||||||
"disk3": kubecontainer.VolumeInfo{BlockVolumeMapper: &stubBlockVolume{dirPath: "/dev/disk/by-id/", volName: "diskUuid"}},
|
|
||||||
"disk4": kubecontainer.VolumeInfo{BlockVolumeMapper: &stubBlockVolume{dirPath: "/var/lib/", volName: "rawdisk"}, ReadOnly: true},
|
|
||||||
},
|
|
||||||
container: v1.Container{
|
|
||||||
Name: "container1",
|
|
||||||
VolumeDevices: []v1.VolumeDevice{
|
|
||||||
{
|
|
||||||
DevicePath: "/dev/sda",
|
|
||||||
Name: "disk1",
|
|
||||||
},
|
|
||||||
{
|
|
||||||
DevicePath: "/dev/xvda",
|
|
||||||
Name: "disk2",
|
|
||||||
},
|
|
||||||
{
|
|
||||||
DevicePath: "/dev/xvdb",
|
|
||||||
Name: "disk3",
|
|
||||||
},
|
|
||||||
{
|
|
||||||
DevicePath: "/mnt/rawdisk",
|
|
||||||
Name: "disk4",
|
|
||||||
},
|
|
||||||
},
|
|
||||||
},
|
|
||||||
expectedDevices: []kubecontainer.DeviceInfo{
|
|
||||||
{
|
|
||||||
PathInContainer: "/dev/sda",
|
|
||||||
PathOnHost: "/dev/sda",
|
|
||||||
Permissions: "mrw",
|
|
||||||
},
|
|
||||||
{
|
|
||||||
PathInContainer: "/dev/xvda",
|
|
||||||
PathOnHost: "/dev/disk/by-path/diskPath",
|
|
||||||
Permissions: "r",
|
|
||||||
},
|
|
||||||
{
|
|
||||||
PathInContainer: "/dev/xvdb",
|
|
||||||
PathOnHost: "/dev/disk/by-id/diskUuid",
|
|
||||||
Permissions: "mrw",
|
|
||||||
},
|
|
||||||
{
|
|
||||||
PathInContainer: "/mnt/rawdisk",
|
|
||||||
PathOnHost: "/var/lib/rawdisk",
|
|
||||||
Permissions: "r",
|
|
||||||
},
|
|
||||||
},
|
|
||||||
expectErr: false,
|
|
||||||
},
|
|
||||||
"invalid absolute Path": {
|
|
||||||
podVolumes: kubecontainer.VolumeMap{
|
|
||||||
"disk": kubecontainer.VolumeInfo{BlockVolumeMapper: &stubBlockVolume{dirPath: "/dev/", volName: "sda"}},
|
|
||||||
},
|
|
||||||
container: v1.Container{
|
|
||||||
VolumeDevices: []v1.VolumeDevice{
|
|
||||||
{
|
|
||||||
DevicePath: "must/be/absolute",
|
|
||||||
Name: "disk",
|
|
||||||
},
|
|
||||||
},
|
|
||||||
},
|
|
||||||
expectErr: true,
|
|
||||||
expectedErrMsg: "error DevicePath `must/be/absolute` must be an absolute path",
|
|
||||||
},
|
|
||||||
"volume doesn't exist": {
|
|
||||||
podVolumes: kubecontainer.VolumeMap{},
|
|
||||||
container: v1.Container{
|
|
||||||
VolumeDevices: []v1.VolumeDevice{
|
|
||||||
{
|
|
||||||
DevicePath: "/dev/sdaa",
|
|
||||||
Name: "disk",
|
|
||||||
},
|
|
||||||
},
|
|
||||||
},
|
|
||||||
expectErr: true,
|
|
||||||
expectedErrMsg: "cannot find volume \"disk\" to pass into container \"\"",
|
|
||||||
},
|
|
||||||
"volume BlockVolumeMapper is nil": {
|
|
||||||
podVolumes: kubecontainer.VolumeMap{
|
|
||||||
"disk": kubecontainer.VolumeInfo{},
|
|
||||||
},
|
|
||||||
container: v1.Container{
|
|
||||||
VolumeDevices: []v1.VolumeDevice{
|
|
||||||
{
|
|
||||||
DevicePath: "/dev/sdzz",
|
|
||||||
Name: "disk",
|
|
||||||
},
|
|
||||||
},
|
|
||||||
},
|
|
||||||
expectErr: true,
|
|
||||||
expectedErrMsg: "cannot find volume \"disk\" to pass into container \"\"",
|
|
||||||
},
|
|
||||||
}
|
|
||||||
|
|
||||||
for name, tc := range testCases {
|
|
||||||
t.Run(name, func(t *testing.T) {
|
|
||||||
pod := v1.Pod{
|
|
||||||
Spec: v1.PodSpec{
|
|
||||||
HostNetwork: true,
|
|
||||||
},
|
|
||||||
}
|
|
||||||
blkutil := volumetest.NewBlockVolumePathHandler()
|
|
||||||
blkVolumes, err := kubelet.makeBlockVolumes(&pod, &tc.container, tc.podVolumes, blkutil)
|
|
||||||
// validate only the error if we expect an error
|
|
||||||
if tc.expectErr {
|
|
||||||
if err == nil || err.Error() != tc.expectedErrMsg {
|
|
||||||
t.Fatalf("expected error message `%s` but got `%v`", tc.expectedErrMsg, err)
|
|
||||||
}
|
|
||||||
return
|
|
||||||
}
|
|
||||||
// otherwise validate the devices
|
|
||||||
if err != nil {
|
|
||||||
t.Fatal(err)
|
|
||||||
}
|
|
||||||
assert.Equal(t, tc.expectedDevices, blkVolumes, "devices of container %+v", tc.container)
|
|
||||||
})
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
func TestNodeHostsFileContent(t *testing.T) {
|
func TestNodeHostsFileContent(t *testing.T) {
|
||||||
testCases := []struct {
|
testCases := []struct {
|
||||||
hostsFileName string
|
hostsFileName string
|
||||||
|
@ -180,7 +180,7 @@ func newTestKubeletWithImageList(
|
|||||||
kubelet.nodeName = types.NodeName(testKubeletHostname)
|
kubelet.nodeName = types.NodeName(testKubeletHostname)
|
||||||
kubelet.runtimeState = newRuntimeState(maxWaitForContainerRuntime)
|
kubelet.runtimeState = newRuntimeState(maxWaitForContainerRuntime)
|
||||||
kubelet.runtimeState.setNetworkState(nil)
|
kubelet.runtimeState.setNetworkState(nil)
|
||||||
if tempDir, err := ioutil.TempDir("/tmp", "kubelet_test."); err != nil {
|
if tempDir, err := ioutil.TempDir("", "kubelet_test."); err != nil {
|
||||||
t.Fatalf("can't make a temp rootdir: %v", err)
|
t.Fatalf("can't make a temp rootdir: %v", err)
|
||||||
} else {
|
} else {
|
||||||
kubelet.rootDirectory = tempDir
|
kubelet.rootDirectory = tempDir
|
||||||
|
Loading…
Reference in New Issue
Block a user