mirror of
https://github.com/k3s-io/kubernetes.git
synced 2025-07-24 20:24:09 +00:00
Merge pull request #93011 from andyzhangx/list-disks
fix: azure disk resize error if source does not exist
This commit is contained in:
commit
4093df78ea
@ -248,7 +248,7 @@ type Cloud struct {
|
|||||||
|
|
||||||
ResourceRequestBackoff wait.Backoff
|
ResourceRequestBackoff wait.Backoff
|
||||||
metadata *InstanceMetadataService
|
metadata *InstanceMetadataService
|
||||||
vmSet VMSet
|
VMSet VMSet
|
||||||
|
|
||||||
// ipv6DualStack allows overriding for unit testing. It's normally initialized from featuregates
|
// ipv6DualStack allows overriding for unit testing. It's normally initialized from featuregates
|
||||||
ipv6DualStackEnabled bool
|
ipv6DualStackEnabled bool
|
||||||
@ -491,12 +491,12 @@ func (az *Cloud) InitializeCloudFromConfig(config *Config, fromSecret bool) erro
|
|||||||
}
|
}
|
||||||
|
|
||||||
if strings.EqualFold(vmTypeVMSS, az.Config.VMType) {
|
if strings.EqualFold(vmTypeVMSS, az.Config.VMType) {
|
||||||
az.vmSet, err = newScaleSet(az)
|
az.VMSet, err = newScaleSet(az)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
} else {
|
} else {
|
||||||
az.vmSet = newAvailabilitySet(az)
|
az.VMSet = newAvailabilitySet(az)
|
||||||
}
|
}
|
||||||
|
|
||||||
az.vmCache, err = az.newVMCache()
|
az.vmCache, err = az.newVMCache()
|
||||||
|
@ -111,7 +111,7 @@ func (az *Cloud) getPrivateIPsForMachineWithRetry(nodeName types.NodeName) ([]st
|
|||||||
var privateIPs []string
|
var privateIPs []string
|
||||||
err := wait.ExponentialBackoff(az.RequestBackoff(), func() (bool, error) {
|
err := wait.ExponentialBackoff(az.RequestBackoff(), func() (bool, error) {
|
||||||
var retryErr error
|
var retryErr error
|
||||||
privateIPs, retryErr = az.vmSet.GetPrivateIPsByNodeName(string(nodeName))
|
privateIPs, retryErr = az.VMSet.GetPrivateIPsByNodeName(string(nodeName))
|
||||||
if retryErr != nil {
|
if retryErr != nil {
|
||||||
// won't retry since the instance doesn't exist on Azure.
|
// won't retry since the instance doesn't exist on Azure.
|
||||||
if retryErr == cloudprovider.InstanceNotFound {
|
if retryErr == cloudprovider.InstanceNotFound {
|
||||||
@ -135,7 +135,7 @@ func (az *Cloud) GetIPForMachineWithRetry(name types.NodeName) (string, string,
|
|||||||
var ip, publicIP string
|
var ip, publicIP string
|
||||||
err := wait.ExponentialBackoff(az.RequestBackoff(), func() (bool, error) {
|
err := wait.ExponentialBackoff(az.RequestBackoff(), func() (bool, error) {
|
||||||
var retryErr error
|
var retryErr error
|
||||||
ip, publicIP, retryErr = az.vmSet.GetIPByNodeName(string(name))
|
ip, publicIP, retryErr = az.VMSet.GetIPByNodeName(string(name))
|
||||||
if retryErr != nil {
|
if retryErr != nil {
|
||||||
klog.Errorf("GetIPForMachineWithRetry(%s): backoff failure, will retry,err=%v", name, retryErr)
|
klog.Errorf("GetIPForMachineWithRetry(%s): backoff failure, will retry,err=%v", name, retryErr)
|
||||||
return false, nil
|
return false, nil
|
||||||
|
@ -90,15 +90,15 @@ type controllerCommon struct {
|
|||||||
|
|
||||||
// getNodeVMSet gets the VMSet interface based on config.VMType and the real virtual machine type.
|
// getNodeVMSet gets the VMSet interface based on config.VMType and the real virtual machine type.
|
||||||
func (c *controllerCommon) getNodeVMSet(nodeName types.NodeName, crt azcache.AzureCacheReadType) (VMSet, error) {
|
func (c *controllerCommon) getNodeVMSet(nodeName types.NodeName, crt azcache.AzureCacheReadType) (VMSet, error) {
|
||||||
// 1. vmType is standard, return cloud.vmSet directly.
|
// 1. vmType is standard, return cloud.VMSet directly.
|
||||||
if c.cloud.VMType == vmTypeStandard {
|
if c.cloud.VMType == vmTypeStandard {
|
||||||
return c.cloud.vmSet, nil
|
return c.cloud.VMSet, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
// 2. vmType is Virtual Machine Scale Set (vmss), convert vmSet to scaleSet.
|
// 2. vmType is Virtual Machine Scale Set (vmss), convert vmSet to scaleSet.
|
||||||
ss, ok := c.cloud.vmSet.(*scaleSet)
|
ss, ok := c.cloud.VMSet.(*scaleSet)
|
||||||
if !ok {
|
if !ok {
|
||||||
return nil, fmt.Errorf("error of converting vmSet (%q) to scaleSet with vmType %q", c.cloud.vmSet, c.cloud.VMType)
|
return nil, fmt.Errorf("error of converting vmSet (%q) to scaleSet with vmType %q", c.cloud.VMSet, c.cloud.VMType)
|
||||||
}
|
}
|
||||||
|
|
||||||
// 3. If the node is managed by availability set, then return ss.availabilitySet.
|
// 3. If the node is managed by availability set, then return ss.availabilitySet.
|
||||||
|
@ -222,7 +222,7 @@ func TestCommonAttachDiskWithVMSS(t *testing.T) {
|
|||||||
}
|
}
|
||||||
ss, err := newScaleSet(testCloud)
|
ss, err := newScaleSet(testCloud)
|
||||||
assert.Nil(t, err)
|
assert.Nil(t, err)
|
||||||
testCloud.vmSet = ss
|
testCloud.VMSet = ss
|
||||||
}
|
}
|
||||||
|
|
||||||
common := &controllerCommon{
|
common := &controllerCommon{
|
||||||
|
@ -79,7 +79,7 @@ func TestStandardAttachDisk(t *testing.T) {
|
|||||||
|
|
||||||
for i, test := range testCases {
|
for i, test := range testCases {
|
||||||
testCloud := GetTestCloud(ctrl)
|
testCloud := GetTestCloud(ctrl)
|
||||||
vmSet := testCloud.vmSet
|
vmSet := testCloud.VMSet
|
||||||
expectedVMs := setTestVirtualMachines(testCloud, map[string]string{"vm1": "PowerState/Running"}, false)
|
expectedVMs := setTestVirtualMachines(testCloud, map[string]string{"vm1": "PowerState/Running"}, false)
|
||||||
mockVMsClient := testCloud.VirtualMachinesClient.(*mockvmclient.MockInterface)
|
mockVMsClient := testCloud.VirtualMachinesClient.(*mockvmclient.MockInterface)
|
||||||
for _, vm := range expectedVMs {
|
for _, vm := range expectedVMs {
|
||||||
@ -148,7 +148,7 @@ func TestStandardDetachDisk(t *testing.T) {
|
|||||||
|
|
||||||
for i, test := range testCases {
|
for i, test := range testCases {
|
||||||
testCloud := GetTestCloud(ctrl)
|
testCloud := GetTestCloud(ctrl)
|
||||||
vmSet := testCloud.vmSet
|
vmSet := testCloud.VMSet
|
||||||
expectedVMs := setTestVirtualMachines(testCloud, map[string]string{"vm1": "PowerState/Running"}, false)
|
expectedVMs := setTestVirtualMachines(testCloud, map[string]string{"vm1": "PowerState/Running"}, false)
|
||||||
mockVMsClient := testCloud.VirtualMachinesClient.(*mockvmclient.MockInterface)
|
mockVMsClient := testCloud.VirtualMachinesClient.(*mockvmclient.MockInterface)
|
||||||
for _, vm := range expectedVMs {
|
for _, vm := range expectedVMs {
|
||||||
@ -224,7 +224,7 @@ func TestGetDataDisks(t *testing.T) {
|
|||||||
}
|
}
|
||||||
for i, test := range testCases {
|
for i, test := range testCases {
|
||||||
testCloud := GetTestCloud(ctrl)
|
testCloud := GetTestCloud(ctrl)
|
||||||
vmSet := testCloud.vmSet
|
vmSet := testCloud.VMSet
|
||||||
expectedVMs := setTestVirtualMachines(testCloud, map[string]string{"vm1": "PowerState/Running"}, false)
|
expectedVMs := setTestVirtualMachines(testCloud, map[string]string{"vm1": "PowerState/Running"}, false)
|
||||||
mockVMsClient := testCloud.VirtualMachinesClient.(*mockvmclient.MockInterface)
|
mockVMsClient := testCloud.VirtualMachinesClient.(*mockvmclient.MockInterface)
|
||||||
for _, vm := range expectedVMs {
|
for _, vm := range expectedVMs {
|
||||||
|
@ -83,7 +83,7 @@ func GetTestCloud(ctrl *gomock.Controller) (az *Cloud) {
|
|||||||
az.VirtualMachineScaleSetsClient = mockvmssclient.NewMockInterface(ctrl)
|
az.VirtualMachineScaleSetsClient = mockvmssclient.NewMockInterface(ctrl)
|
||||||
az.VirtualMachineScaleSetVMsClient = mockvmssvmclient.NewMockInterface(ctrl)
|
az.VirtualMachineScaleSetVMsClient = mockvmssvmclient.NewMockInterface(ctrl)
|
||||||
az.VirtualMachinesClient = mockvmclient.NewMockInterface(ctrl)
|
az.VirtualMachinesClient = mockvmclient.NewMockInterface(ctrl)
|
||||||
az.vmSet = newAvailabilitySet(az)
|
az.VMSet = newAvailabilitySet(az)
|
||||||
az.vmCache, _ = az.newVMCache()
|
az.vmCache, _ = az.newVMCache()
|
||||||
az.lbCache, _ = az.newLBCache()
|
az.lbCache, _ = az.newLBCache()
|
||||||
az.nsgCache, _ = az.newNSGCache()
|
az.nsgCache, _ = az.newNSGCache()
|
||||||
|
@ -95,7 +95,7 @@ func (az *Cloud) NodeAddresses(ctx context.Context, name types.NodeName) ([]v1.N
|
|||||||
|
|
||||||
// Not local instance, get addresses from Azure ARM API.
|
// Not local instance, get addresses from Azure ARM API.
|
||||||
if !isLocalInstance {
|
if !isLocalInstance {
|
||||||
if az.vmSet != nil {
|
if az.VMSet != nil {
|
||||||
return az.addressGetter(name)
|
return az.addressGetter(name)
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -168,7 +168,7 @@ func (az *Cloud) NodeAddressesByProviderID(ctx context.Context, providerID strin
|
|||||||
return nil, nil
|
return nil, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
name, err := az.vmSet.GetNodeNameByProviderID(providerID)
|
name, err := az.VMSet.GetNodeNameByProviderID(providerID)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
@ -189,7 +189,7 @@ func (az *Cloud) InstanceExistsByProviderID(ctx context.Context, providerID stri
|
|||||||
return true, nil
|
return true, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
name, err := az.vmSet.GetNodeNameByProviderID(providerID)
|
name, err := az.VMSet.GetNodeNameByProviderID(providerID)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
if err == cloudprovider.InstanceNotFound {
|
if err == cloudprovider.InstanceNotFound {
|
||||||
return false, nil
|
return false, nil
|
||||||
@ -214,7 +214,7 @@ func (az *Cloud) InstanceShutdownByProviderID(ctx context.Context, providerID st
|
|||||||
return false, nil
|
return false, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
nodeName, err := az.vmSet.GetNodeNameByProviderID(providerID)
|
nodeName, err := az.VMSet.GetNodeNameByProviderID(providerID)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
// Returns false, so the controller manager will continue to check InstanceExistsByProviderID().
|
// Returns false, so the controller manager will continue to check InstanceExistsByProviderID().
|
||||||
if err == cloudprovider.InstanceNotFound {
|
if err == cloudprovider.InstanceNotFound {
|
||||||
@ -224,7 +224,7 @@ func (az *Cloud) InstanceShutdownByProviderID(ctx context.Context, providerID st
|
|||||||
return false, err
|
return false, err
|
||||||
}
|
}
|
||||||
|
|
||||||
powerStatus, err := az.vmSet.GetPowerStatusByNodeName(string(nodeName))
|
powerStatus, err := az.VMSet.GetPowerStatusByNodeName(string(nodeName))
|
||||||
if err != nil {
|
if err != nil {
|
||||||
// Returns false, so the controller manager will continue to check InstanceExistsByProviderID().
|
// Returns false, so the controller manager will continue to check InstanceExistsByProviderID().
|
||||||
if err == cloudprovider.InstanceNotFound {
|
if err == cloudprovider.InstanceNotFound {
|
||||||
@ -292,8 +292,8 @@ func (az *Cloud) InstanceID(ctx context.Context, name types.NodeName) (string, e
|
|||||||
|
|
||||||
// Not local instance, get instanceID from Azure ARM API.
|
// Not local instance, get instanceID from Azure ARM API.
|
||||||
if !isLocalInstance {
|
if !isLocalInstance {
|
||||||
if az.vmSet != nil {
|
if az.VMSet != nil {
|
||||||
return az.vmSet.GetInstanceIDByNodeName(nodeName)
|
return az.VMSet.GetInstanceIDByNodeName(nodeName)
|
||||||
}
|
}
|
||||||
|
|
||||||
// vmSet == nil indicates credentials are not provided.
|
// vmSet == nil indicates credentials are not provided.
|
||||||
@ -302,7 +302,7 @@ func (az *Cloud) InstanceID(ctx context.Context, name types.NodeName) (string, e
|
|||||||
return az.getLocalInstanceProviderID(metadata, nodeName)
|
return az.getLocalInstanceProviderID(metadata, nodeName)
|
||||||
}
|
}
|
||||||
|
|
||||||
return az.vmSet.GetInstanceIDByNodeName(nodeName)
|
return az.VMSet.GetInstanceIDByNodeName(nodeName)
|
||||||
}
|
}
|
||||||
|
|
||||||
func (az *Cloud) getLocalInstanceProviderID(metadata *InstanceMetadata, nodeName string) (string, error) {
|
func (az *Cloud) getLocalInstanceProviderID(metadata *InstanceMetadata, nodeName string) (string, error) {
|
||||||
@ -342,7 +342,7 @@ func (az *Cloud) InstanceTypeByProviderID(ctx context.Context, providerID string
|
|||||||
return "", nil
|
return "", nil
|
||||||
}
|
}
|
||||||
|
|
||||||
name, err := az.vmSet.GetNodeNameByProviderID(providerID)
|
name, err := az.VMSet.GetNodeNameByProviderID(providerID)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return "", err
|
return "", err
|
||||||
}
|
}
|
||||||
@ -380,8 +380,8 @@ func (az *Cloud) InstanceType(ctx context.Context, name types.NodeName) (string,
|
|||||||
return "", err
|
return "", err
|
||||||
}
|
}
|
||||||
if !isLocalInstance {
|
if !isLocalInstance {
|
||||||
if az.vmSet != nil {
|
if az.VMSet != nil {
|
||||||
return az.vmSet.GetInstanceTypeByNodeName(string(name))
|
return az.VMSet.GetInstanceTypeByNodeName(string(name))
|
||||||
}
|
}
|
||||||
|
|
||||||
// vmSet == nil indicates credentials are not provided.
|
// vmSet == nil indicates credentials are not provided.
|
||||||
@ -393,7 +393,7 @@ func (az *Cloud) InstanceType(ctx context.Context, name types.NodeName) (string,
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
return az.vmSet.GetInstanceTypeByNodeName(string(name))
|
return az.VMSet.GetInstanceTypeByNodeName(string(name))
|
||||||
}
|
}
|
||||||
|
|
||||||
// AddSSHKeyToAllInstances adds an SSH public key as a legal identity for all instances
|
// AddSSHKeyToAllInstances adds an SSH public key as a legal identity for all instances
|
||||||
|
@ -174,7 +174,7 @@ func TestInstanceID(t *testing.T) {
|
|||||||
expectedErrMsg: fmt.Errorf("failure of getting instance metadata"),
|
expectedErrMsg: fmt.Errorf("failure of getting instance metadata"),
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
name: "NodeAddresses should report error if cloud.vmSet is nil",
|
name: "NodeAddresses should report error if cloud.VMSet is nil",
|
||||||
nodeName: "vm1",
|
nodeName: "vm1",
|
||||||
vmType: vmTypeStandard,
|
vmType: vmTypeStandard,
|
||||||
useInstanceMetadata: true,
|
useInstanceMetadata: true,
|
||||||
@ -194,9 +194,9 @@ func TestInstanceID(t *testing.T) {
|
|||||||
|
|
||||||
for _, test := range testcases {
|
for _, test := range testcases {
|
||||||
if test.nilVMSet {
|
if test.nilVMSet {
|
||||||
cloud.vmSet = nil
|
cloud.VMSet = nil
|
||||||
} else {
|
} else {
|
||||||
cloud.vmSet = newAvailabilitySet(cloud)
|
cloud.VMSet = newAvailabilitySet(cloud)
|
||||||
}
|
}
|
||||||
cloud.Config.VMType = test.vmType
|
cloud.Config.VMType = test.vmType
|
||||||
cloud.Config.UseInstanceMetadata = test.useInstanceMetadata
|
cloud.Config.UseInstanceMetadata = test.useInstanceMetadata
|
||||||
@ -445,7 +445,7 @@ func TestNodeAddresses(t *testing.T) {
|
|||||||
expectedErrMsg: fmt.Errorf("getError"),
|
expectedErrMsg: fmt.Errorf("getError"),
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
name: "NodeAddresses should report error if cloud.vmSet is nil",
|
name: "NodeAddresses should report error if cloud.VMSet is nil",
|
||||||
nodeName: "vm1",
|
nodeName: "vm1",
|
||||||
vmType: vmTypeStandard,
|
vmType: vmTypeStandard,
|
||||||
useInstanceMetadata: true,
|
useInstanceMetadata: true,
|
||||||
@ -518,9 +518,9 @@ func TestNodeAddresses(t *testing.T) {
|
|||||||
|
|
||||||
for _, test := range testcases {
|
for _, test := range testcases {
|
||||||
if test.nilVMSet {
|
if test.nilVMSet {
|
||||||
cloud.vmSet = nil
|
cloud.VMSet = nil
|
||||||
} else {
|
} else {
|
||||||
cloud.vmSet = newAvailabilitySet(cloud)
|
cloud.VMSet = newAvailabilitySet(cloud)
|
||||||
}
|
}
|
||||||
cloud.Config.VMType = test.vmType
|
cloud.Config.VMType = test.vmType
|
||||||
cloud.Config.UseInstanceMetadata = test.useInstanceMetadata
|
cloud.Config.UseInstanceMetadata = test.useInstanceMetadata
|
||||||
@ -571,175 +571,6 @@ func TestNodeAddresses(t *testing.T) {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestIsCurrentInstance(t *testing.T) {
|
|
||||||
cloud := &Cloud{
|
|
||||||
Config: Config{
|
|
||||||
VMType: vmTypeStandard,
|
|
||||||
},
|
|
||||||
}
|
|
||||||
testcases := []struct {
|
|
||||||
nodeName string
|
|
||||||
metadataVMName string
|
|
||||||
expected bool
|
|
||||||
expectedErrMsg error
|
|
||||||
}{
|
|
||||||
{
|
|
||||||
nodeName: "node1",
|
|
||||||
metadataVMName: "node1",
|
|
||||||
expected: true,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
nodeName: "node1",
|
|
||||||
metadataVMName: "node2",
|
|
||||||
expected: false,
|
|
||||||
},
|
|
||||||
}
|
|
||||||
|
|
||||||
for _, test := range testcases {
|
|
||||||
real, err := cloud.isCurrentInstance(types.NodeName(test.nodeName), test.metadataVMName)
|
|
||||||
assert.Equal(t, test.expectedErrMsg, err)
|
|
||||||
assert.Equal(t, test.expected, real)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
func TestInstanceTypeByProviderID(t *testing.T) {
|
|
||||||
ctrl := gomock.NewController(t)
|
|
||||||
defer ctrl.Finish()
|
|
||||||
cloud := GetTestCloud(ctrl)
|
|
||||||
cloud.Config.UseInstanceMetadata = true
|
|
||||||
|
|
||||||
testcases := []struct {
|
|
||||||
name string
|
|
||||||
vmList []string
|
|
||||||
vmSize string
|
|
||||||
nodeName string
|
|
||||||
vmType string
|
|
||||||
providerID string
|
|
||||||
metadataName string
|
|
||||||
metadataTemplate string
|
|
||||||
useCustomImsCache bool
|
|
||||||
expectedVMsize string
|
|
||||||
expectedErrMsg error
|
|
||||||
}{
|
|
||||||
{
|
|
||||||
name: "InstanceTypeByProviderID should get InstanceType from Azure API if metadata.Compute.VMSize is nil",
|
|
||||||
vmList: []string{"vm1"},
|
|
||||||
nodeName: "vm1",
|
|
||||||
metadataName: "vm1",
|
|
||||||
vmType: vmTypeStandard,
|
|
||||||
providerID: "azure:///subscriptions/subscription/resourceGroups/rg/providers/Microsoft.Compute/virtualMachines/vm1",
|
|
||||||
expectedVMsize: "Standard_A0",
|
|
||||||
},
|
|
||||||
{
|
|
||||||
name: "InstanceTypeByProviderID should get InstanceType from metedata if node's name are equal to metadataName and metadata.Compute.VMSize is not nil",
|
|
||||||
vmList: []string{"vm1"},
|
|
||||||
vmSize: "Standard_A0",
|
|
||||||
nodeName: "vm1",
|
|
||||||
metadataName: "vm1",
|
|
||||||
vmType: vmTypeStandard,
|
|
||||||
providerID: "azure:///subscriptions/subscription/resourceGroups/rg/providers/Microsoft.Compute/virtualMachines/vm1",
|
|
||||||
expectedVMsize: "Standard_A0",
|
|
||||||
},
|
|
||||||
{
|
|
||||||
name: "InstanceTypeByProviderID should get InstanceType from Azure API if node is not local instance",
|
|
||||||
vmList: []string{"vm2"},
|
|
||||||
nodeName: "vm2",
|
|
||||||
metadataName: "vm1",
|
|
||||||
vmType: vmTypeStandard,
|
|
||||||
providerID: "azure:///subscriptions/subscription/resourceGroups/rg/providers/Microsoft.Compute/virtualMachines/vm2",
|
|
||||||
expectedVMsize: "Standard_A0",
|
|
||||||
},
|
|
||||||
{
|
|
||||||
name: "InstanceTypeByProviderID should return nil if node is unmanaged",
|
|
||||||
providerID: "/subscriptions/subscription/resourceGroups/rg/providers/Microsoft.Compute/virtualMachine/vm1",
|
|
||||||
},
|
|
||||||
{
|
|
||||||
name: "InstanceTypeByProviderID should report error if node doesn't exist",
|
|
||||||
vmList: []string{"vm1"},
|
|
||||||
nodeName: "vm3",
|
|
||||||
providerID: "azure:///subscriptions/subscription/resourceGroups/rg/providers/Microsoft.Compute/virtualMachines/vm3",
|
|
||||||
expectedErrMsg: fmt.Errorf("instance not found"),
|
|
||||||
},
|
|
||||||
{
|
|
||||||
name: "InstanceTypeByProviderID should report error if providerID is invalid",
|
|
||||||
providerID: "azure:///subscriptions/subscription/resourceGroups/rg/providers/Microsoft.Compute/virtualMachine/vm3",
|
|
||||||
expectedErrMsg: fmt.Errorf("error splitting providerID"),
|
|
||||||
},
|
|
||||||
{
|
|
||||||
name: "InstanceTypeByProviderID should report error if providerID is null",
|
|
||||||
expectedErrMsg: fmt.Errorf("providerID is empty, the node is not initialized yet"),
|
|
||||||
},
|
|
||||||
{
|
|
||||||
name: "InstanceTypeByProviderID should report error if metadata.Compute is nil",
|
|
||||||
nodeName: "vm1",
|
|
||||||
metadataName: "vm1",
|
|
||||||
providerID: "azure:///subscriptions/subscription/resourceGroups/rg/providers/Microsoft.Compute/virtualMachines/vm1",
|
|
||||||
metadataTemplate: `{"network":{"interface":[]}}`,
|
|
||||||
expectedErrMsg: fmt.Errorf("failure of getting instance metadata"),
|
|
||||||
},
|
|
||||||
{
|
|
||||||
name: "NodeAddresses should report error when invoke GetMetadata",
|
|
||||||
nodeName: "vm1",
|
|
||||||
metadataName: "vm1",
|
|
||||||
providerID: "azure:///subscriptions/subscription/resourceGroups/rg/providers/Microsoft.Compute/virtualMachines/vm1",
|
|
||||||
vmType: vmTypeStandard,
|
|
||||||
useCustomImsCache: true,
|
|
||||||
expectedErrMsg: fmt.Errorf("getError"),
|
|
||||||
},
|
|
||||||
}
|
|
||||||
|
|
||||||
for _, test := range testcases {
|
|
||||||
cloud.Config.VMType = test.vmType
|
|
||||||
listener, err := net.Listen("tcp", "127.0.0.1:0")
|
|
||||||
if err != nil {
|
|
||||||
t.Errorf("Test [%s] unexpected error: %v", test.name, err)
|
|
||||||
}
|
|
||||||
|
|
||||||
mux := http.NewServeMux()
|
|
||||||
mux.Handle("/", http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {
|
|
||||||
if test.metadataTemplate != "" {
|
|
||||||
fmt.Fprintf(w, test.metadataTemplate)
|
|
||||||
} else {
|
|
||||||
fmt.Fprintf(w, "{\"compute\":{\"name\":\"%s\",\"vmsize\":\"%s\",\"subscriptionId\":\"subscription\",\"resourceGroupName\":\"rg\"}}", test.metadataName, test.vmSize)
|
|
||||||
}
|
|
||||||
}))
|
|
||||||
go func() {
|
|
||||||
http.Serve(listener, mux)
|
|
||||||
}()
|
|
||||||
defer listener.Close()
|
|
||||||
|
|
||||||
cloud.metadata, err = NewInstanceMetadataService("http://" + listener.Addr().String() + "/")
|
|
||||||
if err != nil {
|
|
||||||
t.Errorf("Test [%s] unexpected error: %v", test.name, err)
|
|
||||||
}
|
|
||||||
|
|
||||||
if test.useCustomImsCache {
|
|
||||||
cloud.metadata.imsCache, err = azcache.NewTimedcache(metadataCacheTTL, func(key string) (interface{}, error) {
|
|
||||||
return nil, fmt.Errorf("getError")
|
|
||||||
})
|
|
||||||
if err != nil {
|
|
||||||
t.Errorf("Test [%s] unexpected error: %v", test.name, err)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
vmListWithPowerState := make(map[string]string)
|
|
||||||
for _, vm := range test.vmList {
|
|
||||||
vmListWithPowerState[vm] = ""
|
|
||||||
}
|
|
||||||
expectedVMs := setTestVirtualMachines(cloud, vmListWithPowerState, false)
|
|
||||||
mockVMsClient := cloud.VirtualMachinesClient.(*mockvmclient.MockInterface)
|
|
||||||
for _, vm := range expectedVMs {
|
|
||||||
mockVMsClient.EXPECT().Get(gomock.Any(), cloud.ResourceGroup, *vm.Name, gomock.Any()).Return(vm, nil).AnyTimes()
|
|
||||||
}
|
|
||||||
mockVMsClient.EXPECT().Get(gomock.Any(), cloud.ResourceGroup, "vm3", gomock.Any()).Return(compute.VirtualMachine{}, &retry.Error{HTTPStatusCode: http.StatusNotFound, RawError: cloudprovider.InstanceNotFound}).AnyTimes()
|
|
||||||
mockVMsClient.EXPECT().Update(gomock.Any(), cloud.ResourceGroup, gomock.Any(), gomock.Any(), gomock.Any()).Return(nil).AnyTimes()
|
|
||||||
|
|
||||||
instanceType, err := cloud.InstanceTypeByProviderID(context.Background(), test.providerID)
|
|
||||||
assert.Equal(t, test.expectedErrMsg, err, test.name)
|
|
||||||
assert.Equal(t, test.expectedVMsize, instanceType, test.name)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
func TestInstanceExistsByProviderID(t *testing.T) {
|
func TestInstanceExistsByProviderID(t *testing.T) {
|
||||||
ctrl := gomock.NewController(t)
|
ctrl := gomock.NewController(t)
|
||||||
defer ctrl.Finish()
|
defer ctrl.Finish()
|
||||||
|
@ -258,7 +258,7 @@ func (az *Cloud) getLoadBalancerResourceGroup() string {
|
|||||||
func (az *Cloud) getServiceLoadBalancer(service *v1.Service, clusterName string, nodes []*v1.Node, wantLb bool) (lb *network.LoadBalancer, status *v1.LoadBalancerStatus, exists bool, err error) {
|
func (az *Cloud) getServiceLoadBalancer(service *v1.Service, clusterName string, nodes []*v1.Node, wantLb bool) (lb *network.LoadBalancer, status *v1.LoadBalancerStatus, exists bool, err error) {
|
||||||
isInternal := requiresInternalLoadBalancer(service)
|
isInternal := requiresInternalLoadBalancer(service)
|
||||||
var defaultLB *network.LoadBalancer
|
var defaultLB *network.LoadBalancer
|
||||||
primaryVMSetName := az.vmSet.GetPrimaryVMSetName()
|
primaryVMSetName := az.VMSet.GetPrimaryVMSetName()
|
||||||
defaultLBName := az.getAzureLoadBalancerName(clusterName, primaryVMSetName, isInternal)
|
defaultLBName := az.getAzureLoadBalancerName(clusterName, primaryVMSetName, isInternal)
|
||||||
|
|
||||||
existingLBs, err := az.ListLB(service)
|
existingLBs, err := az.ListLB(service)
|
||||||
@ -329,7 +329,7 @@ func (az *Cloud) selectLoadBalancer(clusterName string, service *v1.Service, exi
|
|||||||
isInternal := requiresInternalLoadBalancer(service)
|
isInternal := requiresInternalLoadBalancer(service)
|
||||||
serviceName := getServiceName(service)
|
serviceName := getServiceName(service)
|
||||||
klog.V(2).Infof("selectLoadBalancer for service (%s): isInternal(%v) - start", serviceName, isInternal)
|
klog.V(2).Infof("selectLoadBalancer for service (%s): isInternal(%v) - start", serviceName, isInternal)
|
||||||
vmSetNames, err := az.vmSet.GetVMSetNames(service, nodes)
|
vmSetNames, err := az.VMSet.GetVMSetNames(service, nodes)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
klog.Errorf("az.selectLoadBalancer: cluster(%s) service(%s) isInternal(%t) - az.GetVMSetNames failed, err=(%v)", clusterName, serviceName, isInternal, err)
|
klog.Errorf("az.selectLoadBalancer: cluster(%s) service(%s) isInternal(%t) - az.GetVMSetNames failed, err=(%v)", clusterName, serviceName, isInternal, err)
|
||||||
return nil, false, err
|
return nil, false, err
|
||||||
@ -935,7 +935,7 @@ func (az *Cloud) reconcileLoadBalancer(clusterName string, service *v1.Service,
|
|||||||
// Remove backend pools from vmSets. This is required for virtual machine scale sets before removing the LB.
|
// Remove backend pools from vmSets. This is required for virtual machine scale sets before removing the LB.
|
||||||
vmSetName := az.mapLoadBalancerNameToVMSet(lbName, clusterName)
|
vmSetName := az.mapLoadBalancerNameToVMSet(lbName, clusterName)
|
||||||
klog.V(10).Infof("EnsureBackendPoolDeleted(%s,%s) for service %s: start", lbBackendPoolID, vmSetName, serviceName)
|
klog.V(10).Infof("EnsureBackendPoolDeleted(%s,%s) for service %s: start", lbBackendPoolID, vmSetName, serviceName)
|
||||||
err := az.vmSet.EnsureBackendPoolDeleted(service, lbBackendPoolID, vmSetName, lb.BackendAddressPools)
|
err := az.VMSet.EnsureBackendPoolDeleted(service, lbBackendPoolID, vmSetName, lb.BackendAddressPools)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
klog.Errorf("EnsureBackendPoolDeleted(%s) for service %s failed: %v", lbBackendPoolID, serviceName, err)
|
klog.Errorf("EnsureBackendPoolDeleted(%s) for service %s failed: %v", lbBackendPoolID, serviceName, err)
|
||||||
return nil, err
|
return nil, err
|
||||||
@ -979,7 +979,7 @@ func (az *Cloud) reconcileLoadBalancer(clusterName string, service *v1.Service,
|
|||||||
vmSetName := az.mapLoadBalancerNameToVMSet(lbName, clusterName)
|
vmSetName := az.mapLoadBalancerNameToVMSet(lbName, clusterName)
|
||||||
// Etag would be changed when updating backend pools, so invalidate lbCache after it.
|
// Etag would be changed when updating backend pools, so invalidate lbCache after it.
|
||||||
defer az.lbCache.Delete(lbName)
|
defer az.lbCache.Delete(lbName)
|
||||||
err := az.vmSet.EnsureHostsInPool(service, nodes, lbBackendPoolID, vmSetName, isInternal)
|
err := az.VMSet.EnsureHostsInPool(service, nodes, lbBackendPoolID, vmSetName, isInternal)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
|
@ -297,11 +297,15 @@ func (c *ManagedDiskController) ResizeDisk(diskURI string, oldSize resource.Quan
|
|||||||
return newSizeQuant, nil
|
return newSizeQuant, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
result.DiskProperties.DiskSizeGB = &requestGiB
|
diskParameter := compute.DiskUpdate{
|
||||||
|
DiskUpdateProperties: &compute.DiskUpdateProperties{
|
||||||
|
DiskSizeGB: &requestGiB,
|
||||||
|
},
|
||||||
|
}
|
||||||
|
|
||||||
ctx, cancel = getContextWithCancel()
|
ctx, cancel = getContextWithCancel()
|
||||||
defer cancel()
|
defer cancel()
|
||||||
if rerr := c.common.cloud.DisksClient.CreateOrUpdate(ctx, resourceGroup, diskName, result); rerr != nil {
|
if rerr := c.common.cloud.DisksClient.Update(ctx, resourceGroup, diskName, diskParameter); rerr != nil {
|
||||||
return oldSize, rerr.Error()
|
return oldSize, rerr.Error()
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -369,9 +369,9 @@ func TestResizeDisk(t *testing.T) {
|
|||||||
mockDisksClient.EXPECT().Get(gomock.Any(), testCloud.ResourceGroup, test.diskName).Return(test.existedDisk, nil).AnyTimes()
|
mockDisksClient.EXPECT().Get(gomock.Any(), testCloud.ResourceGroup, test.diskName).Return(test.existedDisk, nil).AnyTimes()
|
||||||
}
|
}
|
||||||
if test.diskName == fakeCreateDiskFailed {
|
if test.diskName == fakeCreateDiskFailed {
|
||||||
mockDisksClient.EXPECT().CreateOrUpdate(gomock.Any(), testCloud.ResourceGroup, test.diskName, gomock.Any()).Return(&retry.Error{RawError: fmt.Errorf("Create Disk failed")}).AnyTimes()
|
mockDisksClient.EXPECT().Update(gomock.Any(), testCloud.ResourceGroup, test.diskName, gomock.Any()).Return(&retry.Error{RawError: fmt.Errorf("Create Disk failed")}).AnyTimes()
|
||||||
} else {
|
} else {
|
||||||
mockDisksClient.EXPECT().CreateOrUpdate(gomock.Any(), testCloud.ResourceGroup, test.diskName, gomock.Any()).Return(nil).AnyTimes()
|
mockDisksClient.EXPECT().Update(gomock.Any(), testCloud.ResourceGroup, test.diskName, gomock.Any()).Return(nil).AnyTimes()
|
||||||
}
|
}
|
||||||
|
|
||||||
result, err := managedDiskController.ResizeDisk(diskURI, test.oldSize, test.newSize)
|
result, err := managedDiskController.ResizeDisk(diskURI, test.oldSize, test.newSize)
|
||||||
|
@ -119,7 +119,7 @@ func TestCreateRoute(t *testing.T) {
|
|||||||
|
|
||||||
cloud := &Cloud{
|
cloud := &Cloud{
|
||||||
RouteTablesClient: routeTableClient,
|
RouteTablesClient: routeTableClient,
|
||||||
vmSet: mockVMSet,
|
VMSet: mockVMSet,
|
||||||
Config: Config{
|
Config: Config{
|
||||||
RouteTableResourceGroup: "foo",
|
RouteTableResourceGroup: "foo",
|
||||||
RouteTableName: "bar",
|
RouteTableName: "bar",
|
||||||
@ -526,7 +526,7 @@ func TestListRoutes(t *testing.T) {
|
|||||||
|
|
||||||
cloud := &Cloud{
|
cloud := &Cloud{
|
||||||
RouteTablesClient: routeTableClient,
|
RouteTablesClient: routeTableClient,
|
||||||
vmSet: mockVMSet,
|
VMSet: mockVMSet,
|
||||||
Config: Config{
|
Config: Config{
|
||||||
RouteTableResourceGroup: "foo",
|
RouteTableResourceGroup: "foo",
|
||||||
RouteTableName: "bar",
|
RouteTableName: "bar",
|
||||||
|
@ -135,7 +135,7 @@ func (az *Cloud) getNetworkResourceSubscriptionID() string {
|
|||||||
func (az *Cloud) mapLoadBalancerNameToVMSet(lbName string, clusterName string) (vmSetName string) {
|
func (az *Cloud) mapLoadBalancerNameToVMSet(lbName string, clusterName string) (vmSetName string) {
|
||||||
vmSetName = strings.TrimSuffix(lbName, InternalLoadBalancerNameSuffix)
|
vmSetName = strings.TrimSuffix(lbName, InternalLoadBalancerNameSuffix)
|
||||||
if strings.EqualFold(clusterName, vmSetName) {
|
if strings.EqualFold(clusterName, vmSetName) {
|
||||||
vmSetName = az.vmSet.GetPrimaryVMSetName()
|
vmSetName = az.VMSet.GetPrimaryVMSetName()
|
||||||
}
|
}
|
||||||
|
|
||||||
return vmSetName
|
return vmSetName
|
||||||
@ -150,7 +150,7 @@ func (az *Cloud) getAzureLoadBalancerName(clusterName string, vmSetName string,
|
|||||||
clusterName = az.LoadBalancerName
|
clusterName = az.LoadBalancerName
|
||||||
}
|
}
|
||||||
lbNamePrefix := vmSetName
|
lbNamePrefix := vmSetName
|
||||||
if strings.EqualFold(vmSetName, az.vmSet.GetPrimaryVMSetName()) || az.useStandardLoadBalancer() {
|
if strings.EqualFold(vmSetName, az.VMSet.GetPrimaryVMSetName()) || az.useStandardLoadBalancer() {
|
||||||
lbNamePrefix = clusterName
|
lbNamePrefix = clusterName
|
||||||
}
|
}
|
||||||
if isInternal {
|
if isInternal {
|
||||||
@ -732,7 +732,7 @@ func (as *availabilitySet) EnsureHostInPool(service *v1.Service, nodeName types.
|
|||||||
return "", "", "", nil, nil
|
return "", "", "", nil, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
klog.Errorf("error: az.EnsureHostInPool(%s), az.vmSet.GetPrimaryInterface.Get(%s, %s), err=%v", nodeName, vmName, vmSetName, err)
|
klog.Errorf("error: az.EnsureHostInPool(%s), az.VMSet.GetPrimaryInterface.Get(%s, %s), err=%v", nodeName, vmName, vmSetName, err)
|
||||||
return "", "", "", nil, err
|
return "", "", "", nil, err
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -908,7 +908,7 @@ func TestGetStandardInstanceIDByNodeName(t *testing.T) {
|
|||||||
ID: to.StringPtr(invalidResouceID),
|
ID: to.StringPtr(invalidResouceID),
|
||||||
}, nil).AnyTimes()
|
}, nil).AnyTimes()
|
||||||
|
|
||||||
instanceID, err := cloud.vmSet.GetInstanceIDByNodeName(test.nodeName)
|
instanceID, err := cloud.VMSet.GetInstanceIDByNodeName(test.nodeName)
|
||||||
assert.Equal(t, test.expectedErrMsg, err, test.name)
|
assert.Equal(t, test.expectedErrMsg, err, test.name)
|
||||||
assert.Equal(t, test.expectedID, instanceID, test.name)
|
assert.Equal(t, test.expectedID, instanceID, test.name)
|
||||||
}
|
}
|
||||||
@ -979,7 +979,7 @@ func TestGetStandardVMPowerStatusByNodeName(t *testing.T) {
|
|||||||
mockVMClient := cloud.VirtualMachinesClient.(*mockvmclient.MockInterface)
|
mockVMClient := cloud.VirtualMachinesClient.(*mockvmclient.MockInterface)
|
||||||
mockVMClient.EXPECT().Get(gomock.Any(), cloud.ResourceGroup, test.nodeName, gomock.Any()).Return(test.vm, test.getErr).AnyTimes()
|
mockVMClient.EXPECT().Get(gomock.Any(), cloud.ResourceGroup, test.nodeName, gomock.Any()).Return(test.vm, test.getErr).AnyTimes()
|
||||||
|
|
||||||
powerState, err := cloud.vmSet.GetPowerStatusByNodeName(test.nodeName)
|
powerState, err := cloud.VMSet.GetPowerStatusByNodeName(test.nodeName)
|
||||||
assert.Equal(t, test.expectedErrMsg, err, test.name)
|
assert.Equal(t, test.expectedErrMsg, err, test.name)
|
||||||
assert.Equal(t, test.expectedStatus, powerState, test.name)
|
assert.Equal(t, test.expectedStatus, powerState, test.name)
|
||||||
}
|
}
|
||||||
@ -1064,7 +1064,7 @@ func TestGetStandardVMZoneByNodeName(t *testing.T) {
|
|||||||
mockVMClient := cloud.VirtualMachinesClient.(*mockvmclient.MockInterface)
|
mockVMClient := cloud.VirtualMachinesClient.(*mockvmclient.MockInterface)
|
||||||
mockVMClient.EXPECT().Get(gomock.Any(), cloud.ResourceGroup, test.nodeName, gomock.Any()).Return(test.vm, test.getErr).AnyTimes()
|
mockVMClient.EXPECT().Get(gomock.Any(), cloud.ResourceGroup, test.nodeName, gomock.Any()).Return(test.vm, test.getErr).AnyTimes()
|
||||||
|
|
||||||
zone, err := cloud.vmSet.GetZoneByNodeName(test.nodeName)
|
zone, err := cloud.VMSet.GetZoneByNodeName(test.nodeName)
|
||||||
assert.Equal(t, test.expectedErrMsg, err, test.name)
|
assert.Equal(t, test.expectedErrMsg, err, test.name)
|
||||||
assert.Equal(t, test.expectedZone, zone, test.name)
|
assert.Equal(t, test.expectedZone, zone, test.name)
|
||||||
}
|
}
|
||||||
@ -1166,7 +1166,7 @@ func TestGetStandardVMSetNames(t *testing.T) {
|
|||||||
mockVMClient := cloud.VirtualMachinesClient.(*mockvmclient.MockInterface)
|
mockVMClient := cloud.VirtualMachinesClient.(*mockvmclient.MockInterface)
|
||||||
mockVMClient.EXPECT().List(gomock.Any(), cloud.ResourceGroup).Return(test.vm, nil).AnyTimes()
|
mockVMClient.EXPECT().List(gomock.Any(), cloud.ResourceGroup).Return(test.vm, nil).AnyTimes()
|
||||||
|
|
||||||
vmSetNames, err := cloud.vmSet.GetVMSetNames(test.service, test.nodes)
|
vmSetNames, err := cloud.VMSet.GetVMSetNames(test.service, test.nodes)
|
||||||
assert.Equal(t, test.expectedErrMsg, err, test.name)
|
assert.Equal(t, test.expectedErrMsg, err, test.name)
|
||||||
assert.Equal(t, test.expectedVMSetNames, vmSetNames, test.name)
|
assert.Equal(t, test.expectedVMSetNames, vmSetNames, test.name)
|
||||||
}
|
}
|
||||||
@ -1339,7 +1339,7 @@ func TestStandardEnsureHostInPool(t *testing.T) {
|
|||||||
mockInterfaceClient.EXPECT().Get(gomock.Any(), cloud.ResourceGroup, test.nicName, gomock.Any()).Return(testNIC, nil).AnyTimes()
|
mockInterfaceClient.EXPECT().Get(gomock.Any(), cloud.ResourceGroup, test.nicName, gomock.Any()).Return(testNIC, nil).AnyTimes()
|
||||||
mockInterfaceClient.EXPECT().CreateOrUpdate(gomock.Any(), cloud.ResourceGroup, gomock.Any(), gomock.Any()).Return(nil).AnyTimes()
|
mockInterfaceClient.EXPECT().CreateOrUpdate(gomock.Any(), cloud.ResourceGroup, gomock.Any(), gomock.Any()).Return(nil).AnyTimes()
|
||||||
|
|
||||||
_, _, _, vm, err := cloud.vmSet.EnsureHostInPool(test.service, test.nodeName, test.backendPoolID, test.vmSetName, false)
|
_, _, _, vm, err := cloud.VMSet.EnsureHostInPool(test.service, test.nodeName, test.backendPoolID, test.vmSetName, false)
|
||||||
assert.Equal(t, test.expectedErrMsg, err, test.name)
|
assert.Equal(t, test.expectedErrMsg, err, test.name)
|
||||||
assert.Nil(t, vm, test.name)
|
assert.Nil(t, vm, test.name)
|
||||||
}
|
}
|
||||||
@ -1501,7 +1501,7 @@ func TestStandardEnsureHostsInPool(t *testing.T) {
|
|||||||
mockInterfaceClient.EXPECT().Get(gomock.Any(), cloud.ResourceGroup, test.nicName, gomock.Any()).Return(testNIC, nil).AnyTimes()
|
mockInterfaceClient.EXPECT().Get(gomock.Any(), cloud.ResourceGroup, test.nicName, gomock.Any()).Return(testNIC, nil).AnyTimes()
|
||||||
mockInterfaceClient.EXPECT().CreateOrUpdate(gomock.Any(), cloud.ResourceGroup, gomock.Any(), gomock.Any()).Return(nil).AnyTimes()
|
mockInterfaceClient.EXPECT().CreateOrUpdate(gomock.Any(), cloud.ResourceGroup, gomock.Any(), gomock.Any()).Return(nil).AnyTimes()
|
||||||
|
|
||||||
err := cloud.vmSet.EnsureHostsInPool(test.service, test.nodes, test.backendPoolID, test.vmSetName, false)
|
err := cloud.VMSet.EnsureHostsInPool(test.service, test.nodes, test.backendPoolID, test.vmSetName, false)
|
||||||
if test.expectedErr {
|
if test.expectedErr {
|
||||||
assert.Equal(t, test.expectedErrMsg, err.Error(), test.name)
|
assert.Equal(t, test.expectedErrMsg, err.Error(), test.name)
|
||||||
} else {
|
} else {
|
||||||
|
@ -2180,7 +2180,7 @@ func TestGetNodeNameByProviderID(t *testing.T) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
for _, test := range providers {
|
for _, test := range providers {
|
||||||
name, err := az.vmSet.GetNodeNameByProviderID(test.providerID)
|
name, err := az.VMSet.GetNodeNameByProviderID(test.providerID)
|
||||||
if (err != nil) != test.fail {
|
if (err != nil) != test.fail {
|
||||||
t.Errorf("Expected to fail=%t, with pattern %v", test.fail, test)
|
t.Errorf("Expected to fail=%t, with pattern %v", test.fail, test)
|
||||||
}
|
}
|
||||||
|
@ -87,7 +87,7 @@ func (az *Cloud) GetZone(ctx context.Context) (cloudprovider.Zone, error) {
|
|||||||
if err != nil {
|
if err != nil {
|
||||||
return cloudprovider.Zone{}, fmt.Errorf("failure getting hostname from kernel")
|
return cloudprovider.Zone{}, fmt.Errorf("failure getting hostname from kernel")
|
||||||
}
|
}
|
||||||
return az.vmSet.GetZoneByNodeName(strings.ToLower(hostname))
|
return az.VMSet.GetZoneByNodeName(strings.ToLower(hostname))
|
||||||
}
|
}
|
||||||
|
|
||||||
// GetZoneByProviderID implements Zones.GetZoneByProviderID
|
// GetZoneByProviderID implements Zones.GetZoneByProviderID
|
||||||
@ -104,7 +104,7 @@ func (az *Cloud) GetZoneByProviderID(ctx context.Context, providerID string) (cl
|
|||||||
return cloudprovider.Zone{}, nil
|
return cloudprovider.Zone{}, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
nodeName, err := az.vmSet.GetNodeNameByProviderID(providerID)
|
nodeName, err := az.VMSet.GetNodeNameByProviderID(providerID)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return cloudprovider.Zone{}, err
|
return cloudprovider.Zone{}, err
|
||||||
}
|
}
|
||||||
@ -126,5 +126,5 @@ func (az *Cloud) GetZoneByNodeName(ctx context.Context, nodeName types.NodeName)
|
|||||||
return cloudprovider.Zone{}, nil
|
return cloudprovider.Zone{}, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
return az.vmSet.GetZoneByNodeName(string(nodeName))
|
return az.VMSet.GetZoneByNodeName(string(nodeName))
|
||||||
}
|
}
|
||||||
|
@ -19,6 +19,7 @@ go_library(
|
|||||||
"//vendor/github.com/Azure/azure-sdk-for-go/services/compute/mgmt/2019-12-01/compute:go_default_library",
|
"//vendor/github.com/Azure/azure-sdk-for-go/services/compute/mgmt/2019-12-01/compute:go_default_library",
|
||||||
"//vendor/github.com/Azure/go-autorest/autorest:go_default_library",
|
"//vendor/github.com/Azure/go-autorest/autorest:go_default_library",
|
||||||
"//vendor/github.com/Azure/go-autorest/autorest/azure:go_default_library",
|
"//vendor/github.com/Azure/go-autorest/autorest/azure:go_default_library",
|
||||||
|
"//vendor/github.com/Azure/go-autorest/autorest/to:go_default_library",
|
||||||
"//vendor/k8s.io/klog/v2:go_default_library",
|
"//vendor/k8s.io/klog/v2:go_default_library",
|
||||||
],
|
],
|
||||||
)
|
)
|
||||||
|
@ -20,12 +20,14 @@ package diskclient
|
|||||||
|
|
||||||
import (
|
import (
|
||||||
"context"
|
"context"
|
||||||
|
"fmt"
|
||||||
"net/http"
|
"net/http"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
"github.com/Azure/azure-sdk-for-go/services/compute/mgmt/2019-12-01/compute"
|
"github.com/Azure/azure-sdk-for-go/services/compute/mgmt/2019-12-01/compute"
|
||||||
"github.com/Azure/go-autorest/autorest"
|
"github.com/Azure/go-autorest/autorest"
|
||||||
"github.com/Azure/go-autorest/autorest/azure"
|
"github.com/Azure/go-autorest/autorest/azure"
|
||||||
|
"github.com/Azure/go-autorest/autorest/to"
|
||||||
|
|
||||||
"k8s.io/client-go/util/flowcontrol"
|
"k8s.io/client-go/util/flowcontrol"
|
||||||
"k8s.io/klog/v2"
|
"k8s.io/klog/v2"
|
||||||
@ -204,6 +206,74 @@ func (c *Client) createOrUpdateResponder(resp *http.Response) (*compute.Disk, *r
|
|||||||
return result, retry.GetError(resp, err)
|
return result, retry.GetError(resp, err)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// Update creates or updates a Disk.
|
||||||
|
func (c *Client) Update(ctx context.Context, resourceGroupName string, diskName string, diskParameter compute.DiskUpdate) *retry.Error {
|
||||||
|
mc := metrics.NewMetricContext("disks", "update", resourceGroupName, c.subscriptionID, "")
|
||||||
|
|
||||||
|
// Report errors if the client is rate limited.
|
||||||
|
if !c.rateLimiterWriter.TryAccept() {
|
||||||
|
mc.RateLimitedCount()
|
||||||
|
return retry.GetRateLimitError(true, "DiskUpdate")
|
||||||
|
}
|
||||||
|
|
||||||
|
// Report errors if the client is throttled.
|
||||||
|
if c.RetryAfterWriter.After(time.Now()) {
|
||||||
|
mc.ThrottledCount()
|
||||||
|
rerr := retry.GetThrottlingError("DiskUpdate", "client throttled", c.RetryAfterWriter)
|
||||||
|
return rerr
|
||||||
|
}
|
||||||
|
|
||||||
|
rerr := c.updateDisk(ctx, resourceGroupName, diskName, diskParameter)
|
||||||
|
mc.Observe(rerr.Error())
|
||||||
|
if rerr != nil {
|
||||||
|
if rerr.IsThrottled() {
|
||||||
|
// Update RetryAfterReader so that no more requests would be sent until RetryAfter expires.
|
||||||
|
c.RetryAfterWriter = rerr.RetryAfter
|
||||||
|
}
|
||||||
|
|
||||||
|
return rerr
|
||||||
|
}
|
||||||
|
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// updateDisk updates a Disk.
|
||||||
|
func (c *Client) updateDisk(ctx context.Context, resourceGroupName string, diskName string, diskParameter compute.DiskUpdate) *retry.Error {
|
||||||
|
resourceID := armclient.GetResourceID(
|
||||||
|
c.subscriptionID,
|
||||||
|
resourceGroupName,
|
||||||
|
"Microsoft.Compute/disks",
|
||||||
|
diskName,
|
||||||
|
)
|
||||||
|
|
||||||
|
response, rerr := c.armClient.PatchResource(ctx, resourceID, diskParameter)
|
||||||
|
defer c.armClient.CloseResponse(ctx, response)
|
||||||
|
if rerr != nil {
|
||||||
|
klog.V(5).Infof("Received error in %s: resourceID: %s, error: %s", "disk.put.request", resourceID, rerr.Error())
|
||||||
|
return rerr
|
||||||
|
}
|
||||||
|
|
||||||
|
if response != nil && response.StatusCode != http.StatusNoContent {
|
||||||
|
_, rerr = c.updateResponder(response)
|
||||||
|
if rerr != nil {
|
||||||
|
klog.V(5).Infof("Received error in %s: resourceID: %s, error: %s", "disk.put.respond", resourceID, rerr.Error())
|
||||||
|
return rerr
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *Client) updateResponder(resp *http.Response) (*compute.Disk, *retry.Error) {
|
||||||
|
result := &compute.Disk{}
|
||||||
|
err := autorest.Respond(
|
||||||
|
resp,
|
||||||
|
azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusCreated),
|
||||||
|
autorest.ByUnmarshallingJSON(&result))
|
||||||
|
result.Response = autorest.Response{Response: resp}
|
||||||
|
return result, retry.GetError(resp, err)
|
||||||
|
}
|
||||||
|
|
||||||
// Delete deletes a Disk by name.
|
// Delete deletes a Disk by name.
|
||||||
func (c *Client) Delete(ctx context.Context, resourceGroupName string, diskName string) *retry.Error {
|
func (c *Client) Delete(ctx context.Context, resourceGroupName string, diskName string) *retry.Error {
|
||||||
mc := metrics.NewMetricContext("disks", "delete", resourceGroupName, c.subscriptionID, "")
|
mc := metrics.NewMetricContext("disks", "delete", resourceGroupName, c.subscriptionID, "")
|
||||||
@ -246,3 +316,126 @@ func (c *Client) deleteDisk(ctx context.Context, resourceGroupName string, diskN
|
|||||||
|
|
||||||
return c.armClient.DeleteResource(ctx, resourceID, "")
|
return c.armClient.DeleteResource(ctx, resourceID, "")
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// ListByResourceGroup lists all the disks under a resource group.
|
||||||
|
func (c *Client) ListByResourceGroup(ctx context.Context, resourceGroupName string) ([]compute.Disk, *retry.Error) {
|
||||||
|
resourceID := fmt.Sprintf("/subscriptions/%s/resourceGroups/%s/providers/Microsoft.Compute/disks",
|
||||||
|
autorest.Encode("path", c.subscriptionID),
|
||||||
|
autorest.Encode("path", resourceGroupName))
|
||||||
|
|
||||||
|
result := make([]compute.Disk, 0)
|
||||||
|
page := &DiskListPage{}
|
||||||
|
page.fn = c.listNextResults
|
||||||
|
|
||||||
|
resp, rerr := c.armClient.GetResource(ctx, resourceID, "")
|
||||||
|
defer c.armClient.CloseResponse(ctx, resp)
|
||||||
|
if rerr != nil {
|
||||||
|
klog.V(5).Infof("Received error in %s: resourceID: %s, error: %s", "disk.list.request", resourceID, rerr.Error())
|
||||||
|
return result, rerr
|
||||||
|
}
|
||||||
|
|
||||||
|
var err error
|
||||||
|
page.dl, err = c.listResponder(resp)
|
||||||
|
if err != nil {
|
||||||
|
klog.V(5).Infof("Received error in %s: resourceID: %s, error: %s", "disk.list.respond", resourceID, err)
|
||||||
|
return result, retry.GetError(resp, err)
|
||||||
|
}
|
||||||
|
|
||||||
|
for page.NotDone() {
|
||||||
|
result = append(result, *page.Response().Value...)
|
||||||
|
if err = page.NextWithContext(ctx); err != nil {
|
||||||
|
klog.V(5).Infof("Received error in %s: resourceID: %s, error: %s", "disk.list.next", resourceID, err)
|
||||||
|
return result, retry.GetError(page.Response().Response.Response, err)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return result, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// listNextResults retrieves the next set of results, if any.
|
||||||
|
func (c *Client) listNextResults(ctx context.Context, lastResults compute.DiskList) (result compute.DiskList, err error) {
|
||||||
|
req, err := c.diskListPreparer(ctx, lastResults)
|
||||||
|
if err != nil {
|
||||||
|
return result, autorest.NewErrorWithError(err, "diskclient", "listNextResults", nil, "Failure preparing next results request")
|
||||||
|
}
|
||||||
|
if req == nil {
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
resp, rerr := c.armClient.Send(ctx, req)
|
||||||
|
defer c.armClient.CloseResponse(ctx, resp)
|
||||||
|
if rerr != nil {
|
||||||
|
result.Response = autorest.Response{Response: resp}
|
||||||
|
return result, autorest.NewErrorWithError(rerr.Error(), "diskclient", "listNextResults", resp, "Failure sending next results request")
|
||||||
|
}
|
||||||
|
|
||||||
|
result, err = c.listResponder(resp)
|
||||||
|
if err != nil {
|
||||||
|
err = autorest.NewErrorWithError(err, "diskclient", "listNextResults", resp, "Failure responding to next results request")
|
||||||
|
}
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
// listResponder handles the response to the List request. The method always
|
||||||
|
// closes the http.Response Body.
|
||||||
|
func (c *Client) listResponder(resp *http.Response) (result compute.DiskList, err error) {
|
||||||
|
err = autorest.Respond(
|
||||||
|
resp,
|
||||||
|
azure.WithErrorUnlessStatusCode(http.StatusOK),
|
||||||
|
autorest.ByUnmarshallingJSON(&result),
|
||||||
|
autorest.ByClosing())
|
||||||
|
result.Response = autorest.Response{Response: resp}
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *Client) diskListPreparer(ctx context.Context, lr compute.DiskList) (*http.Request, error) {
|
||||||
|
if lr.NextLink == nil || len(to.String(lr.NextLink)) < 1 {
|
||||||
|
return nil, nil
|
||||||
|
}
|
||||||
|
return autorest.Prepare((&http.Request{}).WithContext(ctx),
|
||||||
|
autorest.AsJSON(),
|
||||||
|
autorest.AsGet(),
|
||||||
|
autorest.WithBaseURL(to.String(lr.NextLink)))
|
||||||
|
}
|
||||||
|
|
||||||
|
// DiskListPage contains a page of Disk values.
|
||||||
|
type DiskListPage struct {
|
||||||
|
fn func(context.Context, compute.DiskList) (compute.DiskList, error)
|
||||||
|
dl compute.DiskList
|
||||||
|
}
|
||||||
|
|
||||||
|
// NextWithContext advances to the next page of values. If there was an error making
|
||||||
|
// the request the page does not advance and the error is returned.
|
||||||
|
func (page *DiskListPage) NextWithContext(ctx context.Context) (err error) {
|
||||||
|
next, err := page.fn(ctx, page.dl)
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
page.dl = next
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// Next advances to the next page of values. If there was an error making
|
||||||
|
// the request the page does not advance and the error is returned.
|
||||||
|
// Deprecated: Use NextWithContext() instead.
|
||||||
|
func (page *DiskListPage) Next() error {
|
||||||
|
return page.NextWithContext(context.Background())
|
||||||
|
}
|
||||||
|
|
||||||
|
// NotDone returns true if the page enumeration should be started or is not yet complete.
|
||||||
|
func (page DiskListPage) NotDone() bool {
|
||||||
|
return !page.dl.IsEmpty()
|
||||||
|
}
|
||||||
|
|
||||||
|
// Response returns the raw server response from the last page request.
|
||||||
|
func (page DiskListPage) Response() compute.DiskList {
|
||||||
|
return page.dl
|
||||||
|
}
|
||||||
|
|
||||||
|
// Values returns the slice of values for the current page or nil if there are no values.
|
||||||
|
func (page DiskListPage) Values() []compute.Disk {
|
||||||
|
if page.dl.IsEmpty() {
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
return *page.dl.Value
|
||||||
|
}
|
||||||
|
@ -161,6 +161,49 @@ func TestCreateOrUpdate(t *testing.T) {
|
|||||||
assert.Equal(t, throttleErr, rerr)
|
assert.Equal(t, throttleErr, rerr)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func TestUpdate(t *testing.T) {
|
||||||
|
ctrl := gomock.NewController(t)
|
||||||
|
defer ctrl.Finish()
|
||||||
|
|
||||||
|
resourceID := "/subscriptions/subscriptionID/resourceGroups/rg/providers/Microsoft.Compute/disks/disk1"
|
||||||
|
diskUpdate := getTestDiskUpdate()
|
||||||
|
armClient := mockarmclient.NewMockInterface(ctrl)
|
||||||
|
response := &http.Response{
|
||||||
|
StatusCode: http.StatusOK,
|
||||||
|
Body: ioutil.NopCloser(bytes.NewReader([]byte(""))),
|
||||||
|
}
|
||||||
|
armClient.EXPECT().PatchResource(gomock.Any(), resourceID, diskUpdate).Return(response, nil).Times(1)
|
||||||
|
armClient.EXPECT().CloseResponse(gomock.Any(), gomock.Any()).Times(1)
|
||||||
|
|
||||||
|
diskClient := getTestDiskClient(armClient)
|
||||||
|
rerr := diskClient.Update(context.TODO(), "rg", "disk1", diskUpdate)
|
||||||
|
assert.Nil(t, rerr)
|
||||||
|
|
||||||
|
response = &http.Response{
|
||||||
|
StatusCode: http.StatusTooManyRequests,
|
||||||
|
Body: ioutil.NopCloser(bytes.NewReader([]byte("{}"))),
|
||||||
|
}
|
||||||
|
throttleErr := &retry.Error{
|
||||||
|
HTTPStatusCode: http.StatusTooManyRequests,
|
||||||
|
RawError: fmt.Errorf("error"),
|
||||||
|
Retriable: true,
|
||||||
|
RetryAfter: time.Unix(100, 0),
|
||||||
|
}
|
||||||
|
|
||||||
|
armClient.EXPECT().PatchResource(gomock.Any(), resourceID, diskUpdate).Return(response, throttleErr).Times(1)
|
||||||
|
armClient.EXPECT().CloseResponse(gomock.Any(), gomock.Any()).Times(1)
|
||||||
|
rerr = diskClient.Update(context.TODO(), "rg", "disk1", diskUpdate)
|
||||||
|
assert.Equal(t, throttleErr, rerr)
|
||||||
|
}
|
||||||
|
|
||||||
|
func getTestDiskUpdate() compute.DiskUpdate {
|
||||||
|
return compute.DiskUpdate{
|
||||||
|
DiskUpdateProperties: &compute.DiskUpdateProperties{
|
||||||
|
DiskSizeGB: to.Int32Ptr(100),
|
||||||
|
},
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
func TestDelete(t *testing.T) {
|
func TestDelete(t *testing.T) {
|
||||||
ctrl := gomock.NewController(t)
|
ctrl := gomock.NewController(t)
|
||||||
defer ctrl.Finish()
|
defer ctrl.Finish()
|
||||||
|
@ -40,6 +40,12 @@ type Interface interface {
|
|||||||
// CreateOrUpdate creates or updates a Disk.
|
// CreateOrUpdate creates or updates a Disk.
|
||||||
CreateOrUpdate(ctx context.Context, resourceGroupName string, diskName string, diskParameter compute.Disk) *retry.Error
|
CreateOrUpdate(ctx context.Context, resourceGroupName string, diskName string, diskParameter compute.Disk) *retry.Error
|
||||||
|
|
||||||
|
// Update updates a Disk.
|
||||||
|
Update(ctx context.Context, resourceGroupName string, diskName string, diskParameter compute.DiskUpdate) *retry.Error
|
||||||
|
|
||||||
// Delete deletes a Disk by name.
|
// Delete deletes a Disk by name.
|
||||||
Delete(ctx context.Context, resourceGroupName string, diskName string) *retry.Error
|
Delete(ctx context.Context, resourceGroupName string, diskName string) *retry.Error
|
||||||
|
|
||||||
|
// ListByResourceGroup lists all the disks under a resource group.
|
||||||
|
ListByResourceGroup(ctx context.Context, resourceGroupName string) ([]compute.Disk, *retry.Error)
|
||||||
}
|
}
|
||||||
|
@ -79,6 +79,20 @@ func (mr *MockInterfaceMockRecorder) CreateOrUpdate(ctx, resourceGroupName, disk
|
|||||||
return mr.mock.ctrl.RecordCallWithMethodType(mr.mock, "CreateOrUpdate", reflect.TypeOf((*MockInterface)(nil).CreateOrUpdate), ctx, resourceGroupName, diskName, diskParameter)
|
return mr.mock.ctrl.RecordCallWithMethodType(mr.mock, "CreateOrUpdate", reflect.TypeOf((*MockInterface)(nil).CreateOrUpdate), ctx, resourceGroupName, diskName, diskParameter)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// Update mocks base method
|
||||||
|
func (m *MockInterface) Update(ctx context.Context, resourceGroupName, diskName string, diskParameter compute.DiskUpdate) *retry.Error {
|
||||||
|
m.ctrl.T.Helper()
|
||||||
|
ret := m.ctrl.Call(m, "Update", ctx, resourceGroupName, diskName, diskParameter)
|
||||||
|
ret0, _ := ret[0].(*retry.Error)
|
||||||
|
return ret0
|
||||||
|
}
|
||||||
|
|
||||||
|
// Update indicates an expected call of Update
|
||||||
|
func (mr *MockInterfaceMockRecorder) Update(ctx, resourceGroupName, diskName, diskParameter interface{}) *gomock.Call {
|
||||||
|
mr.mock.ctrl.T.Helper()
|
||||||
|
return mr.mock.ctrl.RecordCallWithMethodType(mr.mock, "Update", reflect.TypeOf((*MockInterface)(nil).Update), ctx, resourceGroupName, diskName, diskParameter)
|
||||||
|
}
|
||||||
|
|
||||||
// Delete mocks base method
|
// Delete mocks base method
|
||||||
func (m *MockInterface) Delete(ctx context.Context, resourceGroupName, diskName string) *retry.Error {
|
func (m *MockInterface) Delete(ctx context.Context, resourceGroupName, diskName string) *retry.Error {
|
||||||
m.ctrl.T.Helper()
|
m.ctrl.T.Helper()
|
||||||
@ -92,3 +106,18 @@ func (mr *MockInterfaceMockRecorder) Delete(ctx, resourceGroupName, diskName int
|
|||||||
mr.mock.ctrl.T.Helper()
|
mr.mock.ctrl.T.Helper()
|
||||||
return mr.mock.ctrl.RecordCallWithMethodType(mr.mock, "Delete", reflect.TypeOf((*MockInterface)(nil).Delete), ctx, resourceGroupName, diskName)
|
return mr.mock.ctrl.RecordCallWithMethodType(mr.mock, "Delete", reflect.TypeOf((*MockInterface)(nil).Delete), ctx, resourceGroupName, diskName)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// ListByResourceGroup mocks base method
|
||||||
|
func (m *MockInterface) ListByResourceGroup(ctx context.Context, resourceGroupName string) ([]compute.Disk, *retry.Error) {
|
||||||
|
m.ctrl.T.Helper()
|
||||||
|
ret := m.ctrl.Call(m, "ListByResourceGroup", ctx, resourceGroupName)
|
||||||
|
ret0, _ := ret[0].([]compute.Disk)
|
||||||
|
ret1, _ := ret[1].(*retry.Error)
|
||||||
|
return ret0, ret1
|
||||||
|
}
|
||||||
|
|
||||||
|
// ListByResourceGroup indicates an expected call of ListByResourceGroup
|
||||||
|
func (mr *MockInterfaceMockRecorder) ListByResourceGroup(ctx, resourceGroupName interface{}) *gomock.Call {
|
||||||
|
mr.mock.ctrl.T.Helper()
|
||||||
|
return mr.mock.ctrl.RecordCallWithMethodType(mr.mock, "ListByResourceGroup", reflect.TypeOf((*MockInterface)(nil).Delete), ctx, resourceGroupName)
|
||||||
|
}
|
||||||
|
Loading…
Reference in New Issue
Block a user