mirror of
				https://github.com/k3s-io/kubernetes.git
				synced 2025-10-31 22:01:06 +00:00 
			
		
		
		
	
		
			
				
	
	
		
			267 lines
		
	
	
		
			7.6 KiB
		
	
	
	
		
			Go
		
	
	
	
	
	
			
		
		
	
	
			267 lines
		
	
	
		
			7.6 KiB
		
	
	
	
		
			Go
		
	
	
	
	
	
| /*
 | |
| Copyright 2015 The Kubernetes Authors.
 | |
| 
 | |
| Licensed under the Apache License, Version 2.0 (the "License");
 | |
| you may not use this file except in compliance with the License.
 | |
| You may obtain a copy of the License at
 | |
| 
 | |
|     http://www.apache.org/licenses/LICENSE-2.0
 | |
| 
 | |
| Unless required by applicable law or agreed to in writing, software
 | |
| distributed under the License is distributed on an "AS IS" BASIS,
 | |
| WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 | |
| See the License for the specific language governing permissions and
 | |
| limitations under the License.
 | |
| */
 | |
| 
 | |
| package cinder
 | |
| 
 | |
| import (
 | |
| 	"errors"
 | |
| 	"fmt"
 | |
| 	"io/ioutil"
 | |
| 	"os"
 | |
| 	"strings"
 | |
| 	"time"
 | |
| 
 | |
| 	"github.com/golang/glog"
 | |
| 	metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
 | |
| 
 | |
| 	"k8s.io/api/core/v1"
 | |
| 	"k8s.io/apimachinery/pkg/util/sets"
 | |
| 	clientset "k8s.io/client-go/kubernetes"
 | |
| 	kubeletapis "k8s.io/kubernetes/pkg/kubelet/apis"
 | |
| 	"k8s.io/kubernetes/pkg/volume"
 | |
| 	volutil "k8s.io/kubernetes/pkg/volume/util"
 | |
| 	"k8s.io/utils/exec"
 | |
| )
 | |
| 
 | |
| // DiskUtil has utility/helper methods
 | |
| type DiskUtil struct{}
 | |
| 
 | |
| // AttachDisk attaches a disk specified by a volume.CinderPersistenDisk to the current kubelet.
 | |
| // Mounts the disk to its global path.
 | |
| func (util *DiskUtil) AttachDisk(b *cinderVolumeMounter, globalPDPath string) error {
 | |
| 	options := []string{}
 | |
| 	if b.readOnly {
 | |
| 		options = append(options, "ro")
 | |
| 	}
 | |
| 	cloud, err := b.plugin.getCloudProvider()
 | |
| 	if err != nil {
 | |
| 		return err
 | |
| 	}
 | |
| 	instanceid, err := cloud.InstanceID()
 | |
| 	if err != nil {
 | |
| 		return err
 | |
| 	}
 | |
| 	diskid, err := cloud.AttachDisk(instanceid, b.pdName)
 | |
| 	if err != nil {
 | |
| 		return err
 | |
| 	}
 | |
| 
 | |
| 	var devicePath string
 | |
| 	numTries := 0
 | |
| 	for {
 | |
| 		devicePath = cloud.GetDevicePath(diskid)
 | |
| 		probeAttachedVolume()
 | |
| 
 | |
| 		_, err := os.Stat(devicePath)
 | |
| 		if err == nil {
 | |
| 			break
 | |
| 		}
 | |
| 		if err != nil && !os.IsNotExist(err) {
 | |
| 			return err
 | |
| 		}
 | |
| 		numTries++
 | |
| 		if numTries == 10 {
 | |
| 			return errors.New("Could not attach disk: Timeout after 60s")
 | |
| 		}
 | |
| 		time.Sleep(time.Second * 6)
 | |
| 	}
 | |
| 	notmnt, err := b.mounter.IsLikelyNotMountPoint(globalPDPath)
 | |
| 	if err != nil {
 | |
| 		if os.IsNotExist(err) {
 | |
| 			if err := os.MkdirAll(globalPDPath, 0750); err != nil {
 | |
| 				return err
 | |
| 			}
 | |
| 			notmnt = true
 | |
| 		} else {
 | |
| 			return err
 | |
| 		}
 | |
| 	}
 | |
| 	if notmnt {
 | |
| 		err = b.blockDeviceMounter.FormatAndMount(devicePath, globalPDPath, b.fsType, options)
 | |
| 		if err != nil {
 | |
| 			os.Remove(globalPDPath)
 | |
| 			return err
 | |
| 		}
 | |
| 		glog.V(2).Infof("Safe mount successful: %q\n", devicePath)
 | |
| 	}
 | |
| 	return nil
 | |
| }
 | |
| 
 | |
| // DetachDisk unmounts the device and detaches the disk from the kubelet's host machine.
 | |
| func (util *DiskUtil) DetachDisk(cd *cinderVolumeUnmounter) error {
 | |
| 	globalPDPath := makeGlobalPDName(cd.plugin.host, cd.pdName)
 | |
| 	if err := cd.mounter.Unmount(globalPDPath); err != nil {
 | |
| 		return err
 | |
| 	}
 | |
| 	if err := os.Remove(globalPDPath); err != nil {
 | |
| 		return err
 | |
| 	}
 | |
| 	glog.V(2).Infof("Successfully unmounted main device: %s\n", globalPDPath)
 | |
| 
 | |
| 	cloud, err := cd.plugin.getCloudProvider()
 | |
| 	if err != nil {
 | |
| 		return err
 | |
| 	}
 | |
| 	instanceid, err := cloud.InstanceID()
 | |
| 	if err != nil {
 | |
| 		return err
 | |
| 	}
 | |
| 	if err = cloud.DetachDisk(instanceid, cd.pdName); err != nil {
 | |
| 		return err
 | |
| 	}
 | |
| 	glog.V(2).Infof("Successfully detached cinder volume %s", cd.pdName)
 | |
| 	return nil
 | |
| }
 | |
| 
 | |
| // DeleteVolume uses the cloud entrypoint to delete specified volume
 | |
| func (util *DiskUtil) DeleteVolume(cd *cinderVolumeDeleter) error {
 | |
| 	cloud, err := cd.plugin.getCloudProvider()
 | |
| 	if err != nil {
 | |
| 		return err
 | |
| 	}
 | |
| 
 | |
| 	if err = cloud.DeleteVolume(cd.pdName); err != nil {
 | |
| 		// OpenStack cloud provider returns volume.tryAgainError when necessary,
 | |
| 		// no handling needed here.
 | |
| 		glog.V(2).Infof("Error deleting cinder volume %s: %v", cd.pdName, err)
 | |
| 		return err
 | |
| 	}
 | |
| 	glog.V(2).Infof("Successfully deleted cinder volume %s", cd.pdName)
 | |
| 	return nil
 | |
| }
 | |
| 
 | |
| func getZonesFromNodes(kubeClient clientset.Interface) (sets.String, error) {
 | |
| 	// TODO: caching, currently it is overkill because it calls this function
 | |
| 	// only when it creates dynamic PV
 | |
| 	zones := make(sets.String)
 | |
| 	nodes, err := kubeClient.CoreV1().Nodes().List(metav1.ListOptions{})
 | |
| 	if err != nil {
 | |
| 		glog.V(2).Infof("Error listing nodes")
 | |
| 		return zones, err
 | |
| 	}
 | |
| 	for _, node := range nodes.Items {
 | |
| 		if zone, ok := node.Labels[kubeletapis.LabelZoneFailureDomain]; ok {
 | |
| 			zones.Insert(zone)
 | |
| 		}
 | |
| 	}
 | |
| 	glog.V(4).Infof("zones found: %v", zones)
 | |
| 	return zones, nil
 | |
| }
 | |
| 
 | |
| // CreateVolume uses the cloud provider entrypoint for creating a volume
 | |
| func (util *DiskUtil) CreateVolume(c *cinderVolumeProvisioner) (volumeID string, volumeSizeGB int, volumeLabels map[string]string, fstype string, err error) {
 | |
| 	cloud, err := c.plugin.getCloudProvider()
 | |
| 	if err != nil {
 | |
| 		return "", 0, nil, "", err
 | |
| 	}
 | |
| 
 | |
| 	capacity := c.options.PVC.Spec.Resources.Requests[v1.ResourceName(v1.ResourceStorage)]
 | |
| 	// Cinder works with gigabytes, convert to GiB with rounding up
 | |
| 	volSizeGiB, err := volutil.RoundUpToGiBInt(capacity)
 | |
| 	if err != nil {
 | |
| 		return "", 0, nil, "", err
 | |
| 	}
 | |
| 
 | |
| 	name := volutil.GenerateVolumeName(c.options.ClusterName, c.options.PVName, 255) // Cinder volume name can have up to 255 characters
 | |
| 	vtype := ""
 | |
| 	availability := ""
 | |
| 	// Apply ProvisionerParameters (case-insensitive). We leave validation of
 | |
| 	// the values to the cloud provider.
 | |
| 	for k, v := range c.options.Parameters {
 | |
| 		switch strings.ToLower(k) {
 | |
| 		case "type":
 | |
| 			vtype = v
 | |
| 		case "availability":
 | |
| 			availability = v
 | |
| 		case volume.VolumeParameterFSType:
 | |
| 			fstype = v
 | |
| 		default:
 | |
| 			return "", 0, nil, "", fmt.Errorf("invalid option %q for volume plugin %s", k, c.plugin.GetPluginName())
 | |
| 		}
 | |
| 	}
 | |
| 	// TODO: implement PVC.Selector parsing
 | |
| 	if c.options.PVC.Spec.Selector != nil {
 | |
| 		return "", 0, nil, "", fmt.Errorf("claim.Spec.Selector is not supported for dynamic provisioning on Cinder")
 | |
| 	}
 | |
| 
 | |
| 	if availability == "" {
 | |
| 		// No zone specified, choose one randomly in the same region
 | |
| 		zones, err := getZonesFromNodes(c.plugin.host.GetKubeClient())
 | |
| 		if err != nil {
 | |
| 			glog.V(2).Infof("error getting zone information: %v", err)
 | |
| 			return "", 0, nil, "", err
 | |
| 		}
 | |
| 		// if we did not get any zones, lets leave it blank and gophercloud will
 | |
| 		// use zone "nova" as default
 | |
| 		if len(zones) > 0 {
 | |
| 			availability = volutil.ChooseZoneForVolume(zones, c.options.PVC.Name)
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	volumeID, volumeAZ, volumeRegion, IgnoreVolumeAZ, err := cloud.CreateVolume(name, volSizeGiB, vtype, availability, c.options.CloudTags)
 | |
| 	if err != nil {
 | |
| 		glog.V(2).Infof("Error creating cinder volume: %v", err)
 | |
| 		return "", 0, nil, "", err
 | |
| 	}
 | |
| 	glog.V(2).Infof("Successfully created cinder volume %s", volumeID)
 | |
| 
 | |
| 	// these are needed that pod is spawning to same AZ
 | |
| 	volumeLabels = make(map[string]string)
 | |
| 	if IgnoreVolumeAZ == false {
 | |
| 		volumeLabels[kubeletapis.LabelZoneFailureDomain] = volumeAZ
 | |
| 		volumeLabels[kubeletapis.LabelZoneRegion] = volumeRegion
 | |
| 	}
 | |
| 	return volumeID, volSizeGiB, volumeLabels, fstype, nil
 | |
| }
 | |
| 
 | |
| func probeAttachedVolume() error {
 | |
| 	// rescan scsi bus
 | |
| 	scsiHostRescan()
 | |
| 
 | |
| 	executor := exec.New()
 | |
| 
 | |
| 	// udevadm settle waits for udevd to process the device creation
 | |
| 	// events for all hardware devices, thus ensuring that any device
 | |
| 	// nodes have been created successfully before proceeding.
 | |
| 	argsSettle := []string{"settle"}
 | |
| 	cmdSettle := executor.Command("udevadm", argsSettle...)
 | |
| 	_, errSettle := cmdSettle.CombinedOutput()
 | |
| 	if errSettle != nil {
 | |
| 		glog.Errorf("error running udevadm settle %v\n", errSettle)
 | |
| 	}
 | |
| 
 | |
| 	args := []string{"trigger"}
 | |
| 	cmd := executor.Command("udevadm", args...)
 | |
| 	_, err := cmd.CombinedOutput()
 | |
| 	if err != nil {
 | |
| 		glog.Errorf("error running udevadm trigger %v\n", err)
 | |
| 		return err
 | |
| 	}
 | |
| 	glog.V(4).Infof("Successfully probed all attachments")
 | |
| 	return nil
 | |
| }
 | |
| 
 | |
| func scsiHostRescan() {
 | |
| 	scsiPath := "/sys/class/scsi_host/"
 | |
| 	if dirs, err := ioutil.ReadDir(scsiPath); err == nil {
 | |
| 		for _, f := range dirs {
 | |
| 			name := scsiPath + f.Name() + "/scan"
 | |
| 			data := []byte("- - -")
 | |
| 			ioutil.WriteFile(name, data, 0666)
 | |
| 		}
 | |
| 	}
 | |
| }
 |