mirror of
https://github.com/k3s-io/kubernetes.git
synced 2025-08-03 01:06:27 +00:00
Merge pull request #124690 from mowangdk/test/add_e2e_test_for_volume_health
chore: Add e2e test for NodeGetVolumeStats
This commit is contained in:
commit
b2799bbda7
@ -395,6 +395,19 @@ var (
|
|||||||
// TODO: document the feature (owning SIG, when to use this feature for a test)
|
// TODO: document the feature (owning SIG, when to use this feature for a test)
|
||||||
VolumeSourceXFS = framework.WithFeature(framework.ValidFeatures.Add("VolumeSourceXFS"))
|
VolumeSourceXFS = framework.WithFeature(framework.ValidFeatures.Add("VolumeSourceXFS"))
|
||||||
|
|
||||||
|
// Ownerd by SIG Storage
|
||||||
|
// kep: https://kep.k8s.io/1432
|
||||||
|
// test-infra jobs:
|
||||||
|
// - pull-kubernetes-e2e-storage-kind-alpha-features (need manual trigger)
|
||||||
|
// - ci-kubernetes-e2e-storage-kind-alpha-features
|
||||||
|
// When this label is added to a test, it means that the cluster must be created
|
||||||
|
// with the feature-gate "CSIVolumeHealth=true".
|
||||||
|
//
|
||||||
|
// Once the feature is stable, this label should be removed and these tests will
|
||||||
|
// be run by default on any cluster. The test-infra job also should be updated to
|
||||||
|
// not focus on this feature anymore.
|
||||||
|
CSIVolumeHealth = framework.WithFeature(framework.ValidFeatures.Add("CSIVolumeHealth"))
|
||||||
|
|
||||||
// TODO: document the feature (owning SIG, when to use this feature for a test)
|
// TODO: document the feature (owning SIG, when to use this feature for a test)
|
||||||
Vsphere = framework.WithFeature(framework.ValidFeatures.Add("vsphere"))
|
Vsphere = framework.WithFeature(framework.ValidFeatures.Add("vsphere"))
|
||||||
|
|
||||||
|
@ -56,6 +56,8 @@ const (
|
|||||||
csiPodUnschedulableTimeout = 5 * time.Minute
|
csiPodUnschedulableTimeout = 5 * time.Minute
|
||||||
csiResizeWaitPeriod = 5 * time.Minute
|
csiResizeWaitPeriod = 5 * time.Minute
|
||||||
csiVolumeAttachmentTimeout = 7 * time.Minute
|
csiVolumeAttachmentTimeout = 7 * time.Minute
|
||||||
|
// how long to wait for GetVolumeStats
|
||||||
|
csiNodeVolumeStatWaitPeriod = 2 * time.Minute
|
||||||
// how long to wait for Resizing Condition on PVC to appear
|
// how long to wait for Resizing Condition on PVC to appear
|
||||||
csiResizingConditionWait = 2 * time.Minute
|
csiResizingConditionWait = 2 * time.Minute
|
||||||
|
|
||||||
@ -96,6 +98,7 @@ type testParameters struct {
|
|||||||
disableResizingOnDriver bool
|
disableResizingOnDriver bool
|
||||||
enableSnapshot bool
|
enableSnapshot bool
|
||||||
enableVolumeMountGroup bool // enable the VOLUME_MOUNT_GROUP node capability in the CSI mock driver.
|
enableVolumeMountGroup bool // enable the VOLUME_MOUNT_GROUP node capability in the CSI mock driver.
|
||||||
|
enableNodeVolumeCondition bool
|
||||||
hooks *drivers.Hooks
|
hooks *drivers.Hooks
|
||||||
tokenRequests []storagev1.TokenRequest
|
tokenRequests []storagev1.TokenRequest
|
||||||
requiresRepublish *bool
|
requiresRepublish *bool
|
||||||
@ -168,6 +171,7 @@ func (m *mockDriverSetup) init(ctx context.Context, tp testParameters) {
|
|||||||
DisableAttach: tp.disableAttach,
|
DisableAttach: tp.disableAttach,
|
||||||
EnableResizing: tp.enableResizing,
|
EnableResizing: tp.enableResizing,
|
||||||
EnableNodeExpansion: tp.enableNodeExpansion,
|
EnableNodeExpansion: tp.enableNodeExpansion,
|
||||||
|
EnableNodeVolumeCondition: tp.enableNodeVolumeCondition,
|
||||||
EnableSnapshot: tp.enableSnapshot,
|
EnableSnapshot: tp.enableSnapshot,
|
||||||
EnableVolumeMountGroup: tp.enableVolumeMountGroup,
|
EnableVolumeMountGroup: tp.enableVolumeMountGroup,
|
||||||
TokenRequests: tp.tokenRequests,
|
TokenRequests: tp.tokenRequests,
|
||||||
|
199
test/e2e/storage/csimock/csi_node_volume_health.go
Normal file
199
test/e2e/storage/csimock/csi_node_volume_health.go
Normal file
@ -0,0 +1,199 @@
|
|||||||
|
/*
|
||||||
|
Copyright 2024 The Kubernetes Authors.
|
||||||
|
|
||||||
|
Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
you may not use this file except in compliance with the License.
|
||||||
|
You may obtain a copy of the License at
|
||||||
|
|
||||||
|
http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
|
||||||
|
Unless required by applicable law or agreed to in writing, software
|
||||||
|
distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
See the License for the specific language governing permissions and
|
||||||
|
limitations under the License.
|
||||||
|
*/
|
||||||
|
|
||||||
|
package csimock
|
||||||
|
|
||||||
|
import (
|
||||||
|
"context"
|
||||||
|
"strings"
|
||||||
|
"time"
|
||||||
|
|
||||||
|
"google.golang.org/grpc/codes"
|
||||||
|
|
||||||
|
csipbv1 "github.com/container-storage-interface/spec/lib/go/csi"
|
||||||
|
v1 "k8s.io/api/core/v1"
|
||||||
|
metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
|
||||||
|
"k8s.io/apimachinery/pkg/util/wait"
|
||||||
|
"k8s.io/kubernetes/pkg/features"
|
||||||
|
kubeletmetrics "k8s.io/kubernetes/pkg/kubelet/metrics"
|
||||||
|
"k8s.io/kubernetes/test/e2e/feature"
|
||||||
|
"k8s.io/kubernetes/test/e2e/framework"
|
||||||
|
e2emetrics "k8s.io/kubernetes/test/e2e/framework/metrics"
|
||||||
|
e2epod "k8s.io/kubernetes/test/e2e/framework/pod"
|
||||||
|
e2epv "k8s.io/kubernetes/test/e2e/framework/pv"
|
||||||
|
"k8s.io/kubernetes/test/e2e/storage/drivers"
|
||||||
|
"k8s.io/kubernetes/test/e2e/storage/utils"
|
||||||
|
admissionapi "k8s.io/pod-security-admission/api"
|
||||||
|
|
||||||
|
"github.com/onsi/ginkgo/v2"
|
||||||
|
)
|
||||||
|
|
||||||
|
var _ = utils.SIGDescribe("CSI Mock Node Volume Health", feature.CSIVolumeHealth, framework.WithFeatureGate(features.CSIVolumeHealth), func() {
|
||||||
|
f := framework.NewDefaultFramework("csi-mock-node-volume-health")
|
||||||
|
f.NamespacePodSecurityLevel = admissionapi.LevelPrivileged
|
||||||
|
m := newMockDriverSetup(f)
|
||||||
|
|
||||||
|
f.Context("CSI Mock Node Volume Health", f.WithSlow(), func() {
|
||||||
|
trackedCalls := []string{
|
||||||
|
"NodeGetVolumeStats",
|
||||||
|
}
|
||||||
|
tests := []struct {
|
||||||
|
name string
|
||||||
|
expectedCalls []csiCall
|
||||||
|
nodeVolumeConditionRequired bool
|
||||||
|
nodeAbnormalVolumeCondition bool
|
||||||
|
}{
|
||||||
|
{
|
||||||
|
name: "return normal volume stats",
|
||||||
|
expectedCalls: []csiCall{
|
||||||
|
{
|
||||||
|
expectedMethod: "NodeGetVolumeStats",
|
||||||
|
expectedError: codes.OK,
|
||||||
|
},
|
||||||
|
},
|
||||||
|
nodeVolumeConditionRequired: true,
|
||||||
|
nodeAbnormalVolumeCondition: false,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "return normal volume stats without volume condition",
|
||||||
|
expectedCalls: []csiCall{
|
||||||
|
{
|
||||||
|
expectedMethod: "NodeGetVolumeStats",
|
||||||
|
expectedError: codes.OK,
|
||||||
|
},
|
||||||
|
},
|
||||||
|
nodeVolumeConditionRequired: false,
|
||||||
|
nodeAbnormalVolumeCondition: false,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "return normal volume stats with abnormal volume condition",
|
||||||
|
expectedCalls: []csiCall{
|
||||||
|
{
|
||||||
|
expectedMethod: "NodeGetVolumeStats",
|
||||||
|
expectedError: codes.OK,
|
||||||
|
},
|
||||||
|
},
|
||||||
|
nodeVolumeConditionRequired: true,
|
||||||
|
nodeAbnormalVolumeCondition: true,
|
||||||
|
},
|
||||||
|
}
|
||||||
|
for _, t := range tests {
|
||||||
|
test := t
|
||||||
|
ginkgo.It(test.name, func(ctx context.Context) {
|
||||||
|
m.init(ctx, testParameters{
|
||||||
|
registerDriver: true,
|
||||||
|
enableNodeVolumeCondition: test.nodeVolumeConditionRequired,
|
||||||
|
hooks: createGetVolumeStatsHook(test.nodeAbnormalVolumeCondition),
|
||||||
|
})
|
||||||
|
ginkgo.DeferCleanup(m.cleanup)
|
||||||
|
_, claim, pod := m.createPod(ctx, pvcReference)
|
||||||
|
if pod == nil {
|
||||||
|
return
|
||||||
|
}
|
||||||
|
// Wait for PVC to get bound to make sure the CSI driver is fully started.
|
||||||
|
err := e2epv.WaitForPersistentVolumeClaimPhase(ctx, v1.ClaimBound, f.ClientSet, f.Namespace.Name, claim.Name, time.Second, framework.ClaimProvisionTimeout)
|
||||||
|
framework.ExpectNoError(err, "while waiting for PVC to get provisioned")
|
||||||
|
|
||||||
|
ginkgo.By("Waiting for pod to be running")
|
||||||
|
err = e2epod.WaitForPodNameRunningInNamespace(ctx, m.cs, pod.Name, pod.Namespace)
|
||||||
|
framework.ExpectNoError(err, "wait for running pod")
|
||||||
|
ginkgo.By("Waiting for all remaining expected CSI calls")
|
||||||
|
err = wait.PollUntilContextTimeout(ctx, time.Second, csiNodeVolumeStatWaitPeriod, true, func(c context.Context) (done bool, err error) {
|
||||||
|
var index int
|
||||||
|
_, index, err = compareCSICalls(ctx, trackedCalls, test.expectedCalls, m.driver.GetCalls)
|
||||||
|
if err != nil {
|
||||||
|
return true, err
|
||||||
|
}
|
||||||
|
if index == 0 {
|
||||||
|
// No CSI call received yet
|
||||||
|
return false, nil
|
||||||
|
}
|
||||||
|
if len(test.expectedCalls) == index {
|
||||||
|
// all calls received
|
||||||
|
return true, nil
|
||||||
|
}
|
||||||
|
return false, nil
|
||||||
|
})
|
||||||
|
framework.ExpectNoError(err, "while waiting for all CSI calls")
|
||||||
|
// try to use ```csi.NewMetricsCsi(pv.handler).GetMetrics()``` to get metrics from csimock driver but failed.
|
||||||
|
// the mocked csidriver register doesn't regist itself to normal csidriver.
|
||||||
|
if test.nodeVolumeConditionRequired {
|
||||||
|
pod, err := f.ClientSet.CoreV1().Pods(pod.Namespace).Get(ctx, pod.Name, metav1.GetOptions{})
|
||||||
|
framework.ExpectNoError(err, "get running pod")
|
||||||
|
grabber, err := e2emetrics.NewMetricsGrabber(ctx, f.ClientSet, nil, f.ClientConfig(), true, false, false, false, false, false)
|
||||||
|
framework.ExpectNoError(err, "creating the metrics grabber")
|
||||||
|
waitErr := wait.PollUntilContextTimeout(ctx, 30*time.Second, csiNodeVolumeStatWaitPeriod, true, func(ctx context.Context) (bool, error) {
|
||||||
|
framework.Logf("Grabbing Kubelet metrics")
|
||||||
|
// Grab kubelet metrics from the node the pod was scheduled on
|
||||||
|
var err error
|
||||||
|
kubeMetrics, err := grabber.GrabFromKubelet(ctx, pod.Spec.NodeName)
|
||||||
|
if err != nil {
|
||||||
|
framework.Logf("Error fetching kubelet metrics err: %v", err)
|
||||||
|
return false, err
|
||||||
|
}
|
||||||
|
if !findVolumeConditionMetrics(f.Namespace.Name, claim.Name, kubeMetrics, test.nodeAbnormalVolumeCondition) {
|
||||||
|
return false, nil
|
||||||
|
}
|
||||||
|
return true, nil
|
||||||
|
})
|
||||||
|
framework.ExpectNoError(waitErr, "call metrics should not have any error")
|
||||||
|
}
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
||||||
|
})
|
||||||
|
})
|
||||||
|
|
||||||
|
func findVolumeConditionMetrics(pvcNamespace, pvcName string, kubeMetrics e2emetrics.KubeletMetrics, nodeAbnormalVolumeCondition bool) bool {
|
||||||
|
|
||||||
|
found := false
|
||||||
|
framework.Logf("Looking for sample tagged with namespace `%s`, PVC `%s`", pvcNamespace, pvcName)
|
||||||
|
for key, value := range kubeMetrics {
|
||||||
|
for _, sample := range value {
|
||||||
|
framework.Logf("Found sample %++v with key: %s", sample, key)
|
||||||
|
samplePVC, ok := sample.Metric["persistentvolumeclaim"]
|
||||||
|
if !ok {
|
||||||
|
break
|
||||||
|
}
|
||||||
|
sampleNS, ok := sample.Metric["namespace"]
|
||||||
|
if !ok {
|
||||||
|
break
|
||||||
|
}
|
||||||
|
|
||||||
|
if string(samplePVC) == pvcName && string(sampleNS) == pvcNamespace && strings.Contains(key, kubeletmetrics.VolumeStatsHealthStatusAbnormalKey) {
|
||||||
|
if (nodeAbnormalVolumeCondition && sample.Value.String() == "1") || (!nodeAbnormalVolumeCondition && sample.Value.String() == "0") {
|
||||||
|
found = true
|
||||||
|
break
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return found
|
||||||
|
}
|
||||||
|
|
||||||
|
func createGetVolumeStatsHook(abnormalVolumeCondition bool) *drivers.Hooks {
|
||||||
|
return &drivers.Hooks{
|
||||||
|
Pre: func(ctx context.Context, fullMethod string, request interface{}) (reply interface{}, err error) {
|
||||||
|
if req, ok := request.(*csipbv1.NodeGetVolumeStatsRequest); ok {
|
||||||
|
if abnormalVolumeCondition {
|
||||||
|
req.VolumePath = "/tmp/csi/health/abnormal"
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return nil, nil
|
||||||
|
},
|
||||||
|
}
|
||||||
|
|
||||||
|
}
|
@ -359,13 +359,6 @@ func (s *service) NodeGetCapabilities(
|
|||||||
},
|
},
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
{
|
|
||||||
Type: &csi.NodeServiceCapability_Rpc{
|
|
||||||
Rpc: &csi.NodeServiceCapability_RPC{
|
|
||||||
Type: csi.NodeServiceCapability_RPC_VOLUME_CONDITION,
|
|
||||||
},
|
|
||||||
},
|
|
||||||
},
|
|
||||||
}
|
}
|
||||||
if s.config.NodeExpansionRequired {
|
if s.config.NodeExpansionRequired {
|
||||||
capabilities = append(capabilities, &csi.NodeServiceCapability{
|
capabilities = append(capabilities, &csi.NodeServiceCapability{
|
||||||
@ -377,6 +370,16 @@ func (s *service) NodeGetCapabilities(
|
|||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
|
if s.config.NodeVolumeConditionRequired {
|
||||||
|
capabilities = append(capabilities, &csi.NodeServiceCapability{
|
||||||
|
Type: &csi.NodeServiceCapability_Rpc{
|
||||||
|
Rpc: &csi.NodeServiceCapability_RPC{
|
||||||
|
Type: csi.NodeServiceCapability_RPC_VOLUME_CONDITION,
|
||||||
|
},
|
||||||
|
},
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
||||||
if s.config.VolumeMountGroupRequired {
|
if s.config.VolumeMountGroupRequired {
|
||||||
capabilities = append(capabilities, &csi.NodeServiceCapability{
|
capabilities = append(capabilities, &csi.NodeServiceCapability{
|
||||||
Type: &csi.NodeServiceCapability_Rpc{
|
Type: &csi.NodeServiceCapability_Rpc{
|
||||||
@ -417,7 +420,10 @@ func (s *service) NodeGetVolumeStats(ctx context.Context,
|
|||||||
req *csi.NodeGetVolumeStatsRequest) (*csi.NodeGetVolumeStatsResponse, error) {
|
req *csi.NodeGetVolumeStatsRequest) (*csi.NodeGetVolumeStatsResponse, error) {
|
||||||
|
|
||||||
resp := &csi.NodeGetVolumeStatsResponse{
|
resp := &csi.NodeGetVolumeStatsResponse{
|
||||||
VolumeCondition: &csi.VolumeCondition{},
|
VolumeCondition: &csi.VolumeCondition{
|
||||||
|
Abnormal: false,
|
||||||
|
Message: "volume is normal",
|
||||||
|
},
|
||||||
}
|
}
|
||||||
|
|
||||||
if len(req.GetVolumeId()) == 0 {
|
if len(req.GetVolumeId()) == 0 {
|
||||||
@ -437,6 +443,19 @@ func (s *service) NodeGetVolumeStats(ctx context.Context,
|
|||||||
|
|
||||||
nodeMntPathKey := path.Join(s.nodeID, req.VolumePath)
|
nodeMntPathKey := path.Join(s.nodeID, req.VolumePath)
|
||||||
|
|
||||||
|
resp.Usage = []*csi.VolumeUsage{
|
||||||
|
{
|
||||||
|
Total: v.GetCapacityBytes(),
|
||||||
|
Unit: csi.VolumeUsage_BYTES,
|
||||||
|
},
|
||||||
|
}
|
||||||
|
|
||||||
|
if req.GetVolumePath() == "/tmp/csi/health/abnormal" {
|
||||||
|
resp.VolumeCondition.Abnormal = true
|
||||||
|
resp.VolumeCondition.Message = "volume is abnormal"
|
||||||
|
return resp, nil
|
||||||
|
}
|
||||||
|
|
||||||
_, exists := v.VolumeContext[nodeMntPathKey]
|
_, exists := v.VolumeContext[nodeMntPathKey]
|
||||||
if !exists {
|
if !exists {
|
||||||
msg := fmt.Sprintf("volume %q doest not exist on the specified path %q", req.VolumeId, req.VolumePath)
|
msg := fmt.Sprintf("volume %q doest not exist on the specified path %q", req.VolumeId, req.VolumePath)
|
||||||
@ -449,12 +468,5 @@ func (s *service) NodeGetVolumeStats(ctx context.Context,
|
|||||||
return nil, status.Errorf(hookVal, hookMsg)
|
return nil, status.Errorf(hookVal, hookMsg)
|
||||||
}
|
}
|
||||||
|
|
||||||
resp.Usage = []*csi.VolumeUsage{
|
|
||||||
{
|
|
||||||
Total: v.GetCapacityBytes(),
|
|
||||||
Unit: csi.VolumeUsage_BYTES,
|
|
||||||
},
|
|
||||||
}
|
|
||||||
|
|
||||||
return resp, nil
|
return resp, nil
|
||||||
}
|
}
|
||||||
|
@ -51,16 +51,17 @@ var Manifest = map[string]string{
|
|||||||
}
|
}
|
||||||
|
|
||||||
type Config struct {
|
type Config struct {
|
||||||
DisableAttach bool
|
DisableAttach bool
|
||||||
DriverName string
|
DriverName string
|
||||||
AttachLimit int64
|
AttachLimit int64
|
||||||
NodeExpansionRequired bool
|
NodeExpansionRequired bool
|
||||||
VolumeMountGroupRequired bool
|
NodeVolumeConditionRequired bool
|
||||||
DisableControllerExpansion bool
|
VolumeMountGroupRequired bool
|
||||||
DisableOnlineExpansion bool
|
DisableControllerExpansion bool
|
||||||
PermissiveTargetPath bool
|
DisableOnlineExpansion bool
|
||||||
EnableTopology bool
|
PermissiveTargetPath bool
|
||||||
IO DirIO
|
EnableTopology bool
|
||||||
|
IO DirIO
|
||||||
}
|
}
|
||||||
|
|
||||||
// DirIO is an abstraction over direct os calls.
|
// DirIO is an abstraction over direct os calls.
|
||||||
|
@ -344,6 +344,7 @@ type mockCSIDriver struct {
|
|||||||
requiresRepublish *bool
|
requiresRepublish *bool
|
||||||
fsGroupPolicy *storagev1.FSGroupPolicy
|
fsGroupPolicy *storagev1.FSGroupPolicy
|
||||||
enableVolumeMountGroup bool
|
enableVolumeMountGroup bool
|
||||||
|
enableNodeVolumeCondition bool
|
||||||
embedded bool
|
embedded bool
|
||||||
calls MockCSICalls
|
calls MockCSICalls
|
||||||
embeddedCSIDriver *mockdriver.CSIDriver
|
embeddedCSIDriver *mockdriver.CSIDriver
|
||||||
@ -393,6 +394,7 @@ type CSIMockDriverOpts struct {
|
|||||||
EnableNodeExpansion bool
|
EnableNodeExpansion bool
|
||||||
EnableSnapshot bool
|
EnableSnapshot bool
|
||||||
EnableVolumeMountGroup bool
|
EnableVolumeMountGroup bool
|
||||||
|
EnableNodeVolumeCondition bool
|
||||||
TokenRequests []storagev1.TokenRequest
|
TokenRequests []storagev1.TokenRequest
|
||||||
RequiresRepublish *bool
|
RequiresRepublish *bool
|
||||||
FSGroupPolicy *storagev1.FSGroupPolicy
|
FSGroupPolicy *storagev1.FSGroupPolicy
|
||||||
@ -547,6 +549,7 @@ func InitMockCSIDriver(driverOpts CSIMockDriverOpts) MockCSITestDriver {
|
|||||||
attachable: !driverOpts.DisableAttach,
|
attachable: !driverOpts.DisableAttach,
|
||||||
attachLimit: driverOpts.AttachLimit,
|
attachLimit: driverOpts.AttachLimit,
|
||||||
enableNodeExpansion: driverOpts.EnableNodeExpansion,
|
enableNodeExpansion: driverOpts.EnableNodeExpansion,
|
||||||
|
enableNodeVolumeCondition: driverOpts.EnableNodeVolumeCondition,
|
||||||
tokenRequests: driverOpts.TokenRequests,
|
tokenRequests: driverOpts.TokenRequests,
|
||||||
requiresRepublish: driverOpts.RequiresRepublish,
|
requiresRepublish: driverOpts.RequiresRepublish,
|
||||||
fsGroupPolicy: driverOpts.FSGroupPolicy,
|
fsGroupPolicy: driverOpts.FSGroupPolicy,
|
||||||
@ -621,12 +624,13 @@ func (m *mockCSIDriver) PrepareTest(ctx context.Context, f *framework.Framework)
|
|||||||
// for cleanup callbacks.
|
// for cleanup callbacks.
|
||||||
ctx, cancel := context.WithCancel(context.Background())
|
ctx, cancel := context.WithCancel(context.Background())
|
||||||
serviceConfig := mockservice.Config{
|
serviceConfig := mockservice.Config{
|
||||||
DisableAttach: !m.attachable,
|
DisableAttach: !m.attachable,
|
||||||
DriverName: "csi-mock-" + f.UniqueName,
|
DriverName: "csi-mock-" + f.UniqueName,
|
||||||
AttachLimit: int64(m.attachLimit),
|
AttachLimit: int64(m.attachLimit),
|
||||||
NodeExpansionRequired: m.enableNodeExpansion,
|
NodeExpansionRequired: m.enableNodeExpansion,
|
||||||
VolumeMountGroupRequired: m.enableVolumeMountGroup,
|
NodeVolumeConditionRequired: m.enableNodeVolumeCondition,
|
||||||
EnableTopology: m.enableTopology,
|
VolumeMountGroupRequired: m.enableVolumeMountGroup,
|
||||||
|
EnableTopology: m.enableTopology,
|
||||||
IO: proxy.PodDirIO{
|
IO: proxy.PodDirIO{
|
||||||
F: f,
|
F: f,
|
||||||
Namespace: m.driverNamespace.Name,
|
Namespace: m.driverNamespace.Name,
|
||||||
|
Loading…
Reference in New Issue
Block a user