mirror of https://github.com/k3s-io/k3s
211 lines
7.7 KiB
Go
211 lines
7.7 KiB
Go
// +build !providerless
|
|
|
|
/*
|
|
Copyright 2018 The Kubernetes Authors.
|
|
|
|
Licensed under the Apache License, Version 2.0 (the "License");
|
|
you may not use this file except in compliance with the License.
|
|
You may obtain a copy of the License at
|
|
|
|
http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
Unless required by applicable law or agreed to in writing, software
|
|
distributed under the License is distributed on an "AS IS" BASIS,
|
|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
See the License for the specific language governing permissions and
|
|
limitations under the License.
|
|
*/
|
|
|
|
package azure
|
|
|
|
import (
|
|
"net/http"
|
|
"strings"
|
|
|
|
"github.com/Azure/azure-sdk-for-go/services/compute/mgmt/2019-07-01/compute"
|
|
"github.com/Azure/go-autorest/autorest/to"
|
|
|
|
"k8s.io/apimachinery/pkg/types"
|
|
"k8s.io/klog"
|
|
azcache "k8s.io/legacy-cloud-providers/azure/cache"
|
|
)
|
|
|
|
// AttachDisk attaches a vhd to vm
|
|
// the vhd must exist, can be identified by diskName, diskURI, and lun.
|
|
func (ss *scaleSet) AttachDisk(isManagedDisk bool, diskName, diskURI string, nodeName types.NodeName, lun int32, cachingMode compute.CachingTypes, diskEncryptionSetID string, writeAcceleratorEnabled bool) error {
|
|
vmName := mapNodeNameToVMName(nodeName)
|
|
ssName, instanceID, vm, err := ss.getVmssVM(vmName, azcache.CacheReadTypeDefault)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
nodeResourceGroup, err := ss.GetNodeResourceGroup(vmName)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
disks := []compute.DataDisk{}
|
|
if vm.StorageProfile != nil && vm.StorageProfile.DataDisks != nil {
|
|
disks = make([]compute.DataDisk, len(*vm.StorageProfile.DataDisks))
|
|
copy(disks, *vm.StorageProfile.DataDisks)
|
|
}
|
|
if isManagedDisk {
|
|
managedDisk := &compute.ManagedDiskParameters{ID: &diskURI}
|
|
if diskEncryptionSetID == "" {
|
|
if vm.StorageProfile.OsDisk != nil &&
|
|
vm.StorageProfile.OsDisk.ManagedDisk != nil &&
|
|
vm.StorageProfile.OsDisk.ManagedDisk.DiskEncryptionSet != nil &&
|
|
vm.StorageProfile.OsDisk.ManagedDisk.DiskEncryptionSet.ID != nil {
|
|
// set diskEncryptionSet as value of os disk by default
|
|
diskEncryptionSetID = *vm.StorageProfile.OsDisk.ManagedDisk.DiskEncryptionSet.ID
|
|
}
|
|
}
|
|
if diskEncryptionSetID != "" {
|
|
managedDisk.DiskEncryptionSet = &compute.DiskEncryptionSetParameters{ID: &diskEncryptionSetID}
|
|
}
|
|
disks = append(disks,
|
|
compute.DataDisk{
|
|
Name: &diskName,
|
|
Lun: &lun,
|
|
Caching: compute.CachingTypes(cachingMode),
|
|
CreateOption: "attach",
|
|
ManagedDisk: managedDisk,
|
|
WriteAcceleratorEnabled: to.BoolPtr(writeAcceleratorEnabled),
|
|
})
|
|
} else {
|
|
disks = append(disks,
|
|
compute.DataDisk{
|
|
Name: &diskName,
|
|
Vhd: &compute.VirtualHardDisk{
|
|
URI: &diskURI,
|
|
},
|
|
Lun: &lun,
|
|
Caching: compute.CachingTypes(cachingMode),
|
|
CreateOption: "attach",
|
|
})
|
|
}
|
|
newVM := compute.VirtualMachineScaleSetVM{
|
|
Sku: vm.Sku,
|
|
Location: vm.Location,
|
|
VirtualMachineScaleSetVMProperties: &compute.VirtualMachineScaleSetVMProperties{
|
|
HardwareProfile: vm.HardwareProfile,
|
|
StorageProfile: &compute.StorageProfile{
|
|
OsDisk: vm.StorageProfile.OsDisk,
|
|
DataDisks: &disks,
|
|
},
|
|
},
|
|
}
|
|
|
|
ctx, cancel := getContextWithCancel()
|
|
defer cancel()
|
|
|
|
// Invalidate the cache right after updating
|
|
defer ss.deleteCacheForNode(vmName)
|
|
|
|
klog.V(2).Infof("azureDisk - update(%s): vm(%s) - attach disk(%s, %s) with DiskEncryptionSetID(%s)", nodeResourceGroup, nodeName, diskName, diskURI, diskEncryptionSetID)
|
|
rerr := ss.VirtualMachineScaleSetVMsClient.Update(ctx, nodeResourceGroup, ssName, instanceID, newVM, "attach_disk")
|
|
if rerr != nil {
|
|
klog.Errorf("azureDisk - attach disk(%s, %s) on rg(%s) vm(%s) failed, err: %v", diskName, diskURI, nodeResourceGroup, nodeName, rerr)
|
|
if rerr.HTTPStatusCode == http.StatusNotFound {
|
|
klog.Errorf("azureDisk - begin to filterNonExistingDisks(%s, %s) on rg(%s) vm(%s)", diskName, diskURI, nodeResourceGroup, nodeName)
|
|
disks := ss.filterNonExistingDisks(ctx, *newVM.VirtualMachineScaleSetVMProperties.StorageProfile.DataDisks)
|
|
newVM.VirtualMachineScaleSetVMProperties.StorageProfile.DataDisks = &disks
|
|
rerr = ss.VirtualMachineScaleSetVMsClient.Update(ctx, nodeResourceGroup, ssName, instanceID, newVM, "attach_disk")
|
|
}
|
|
}
|
|
|
|
klog.V(2).Infof("azureDisk - update(%s): vm(%s) - attach disk(%s, %s) returned with %v", nodeResourceGroup, nodeName, diskName, diskURI, rerr)
|
|
if rerr != nil {
|
|
return rerr.Error()
|
|
}
|
|
return nil
|
|
}
|
|
|
|
// DetachDisk detaches a disk from host
|
|
// the vhd can be identified by diskName or diskURI
|
|
func (ss *scaleSet) DetachDisk(diskName, diskURI string, nodeName types.NodeName) error {
|
|
vmName := mapNodeNameToVMName(nodeName)
|
|
ssName, instanceID, vm, err := ss.getVmssVM(vmName, azcache.CacheReadTypeDefault)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
nodeResourceGroup, err := ss.GetNodeResourceGroup(vmName)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
disks := []compute.DataDisk{}
|
|
if vm.StorageProfile != nil && vm.StorageProfile.DataDisks != nil {
|
|
disks = make([]compute.DataDisk, len(*vm.StorageProfile.DataDisks))
|
|
copy(disks, *vm.StorageProfile.DataDisks)
|
|
}
|
|
bFoundDisk := false
|
|
for i, disk := range disks {
|
|
if disk.Lun != nil && (disk.Name != nil && diskName != "" && strings.EqualFold(*disk.Name, diskName)) ||
|
|
(disk.Vhd != nil && disk.Vhd.URI != nil && diskURI != "" && strings.EqualFold(*disk.Vhd.URI, diskURI)) ||
|
|
(disk.ManagedDisk != nil && diskURI != "" && strings.EqualFold(*disk.ManagedDisk.ID, diskURI)) {
|
|
// found the disk
|
|
klog.V(2).Infof("azureDisk - detach disk: name %q uri %q", diskName, diskURI)
|
|
disks[i].ToBeDetached = to.BoolPtr(true)
|
|
bFoundDisk = true
|
|
break
|
|
}
|
|
}
|
|
|
|
if !bFoundDisk {
|
|
// only log here, next action is to update VM status with original meta data
|
|
klog.Errorf("detach azure disk: disk %s not found, diskURI: %s", diskName, diskURI)
|
|
}
|
|
|
|
newVM := compute.VirtualMachineScaleSetVM{
|
|
Sku: vm.Sku,
|
|
Location: vm.Location,
|
|
VirtualMachineScaleSetVMProperties: &compute.VirtualMachineScaleSetVMProperties{
|
|
HardwareProfile: vm.HardwareProfile,
|
|
StorageProfile: &compute.StorageProfile{
|
|
OsDisk: vm.StorageProfile.OsDisk,
|
|
DataDisks: &disks,
|
|
},
|
|
},
|
|
}
|
|
|
|
ctx, cancel := getContextWithCancel()
|
|
defer cancel()
|
|
|
|
// Invalidate the cache right after updating
|
|
defer ss.deleteCacheForNode(vmName)
|
|
|
|
klog.V(2).Infof("azureDisk - update(%s): vm(%s) - detach disk(%s, %s)", nodeResourceGroup, nodeName, diskName, diskURI)
|
|
rerr := ss.VirtualMachineScaleSetVMsClient.Update(ctx, nodeResourceGroup, ssName, instanceID, newVM, "detach_disk")
|
|
if rerr != nil {
|
|
klog.Errorf("azureDisk - detach disk(%s, %s) on rg(%s) vm(%s) failed, err: %v", diskName, diskURI, nodeResourceGroup, nodeName, rerr)
|
|
if rerr.HTTPStatusCode == http.StatusNotFound {
|
|
klog.Errorf("azureDisk - begin to filterNonExistingDisks(%s, %s) on rg(%s) vm(%s)", diskName, diskURI, nodeResourceGroup, nodeName)
|
|
disks := ss.filterNonExistingDisks(ctx, *newVM.VirtualMachineScaleSetVMProperties.StorageProfile.DataDisks)
|
|
newVM.VirtualMachineScaleSetVMProperties.StorageProfile.DataDisks = &disks
|
|
rerr = ss.VirtualMachineScaleSetVMsClient.Update(ctx, nodeResourceGroup, ssName, instanceID, newVM, "detach_disk")
|
|
}
|
|
}
|
|
|
|
klog.V(2).Infof("azureDisk - update(%s): vm(%s) - detach disk(%s, %s) returned with %v", nodeResourceGroup, nodeName, diskName, diskURI, rerr)
|
|
if rerr != nil {
|
|
return rerr.Error()
|
|
}
|
|
return nil
|
|
}
|
|
|
|
// GetDataDisks gets a list of data disks attached to the node.
|
|
func (ss *scaleSet) GetDataDisks(nodeName types.NodeName, crt azcache.AzureCacheReadType) ([]compute.DataDisk, error) {
|
|
_, _, vm, err := ss.getVmssVM(string(nodeName), crt)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
if vm.StorageProfile == nil || vm.StorageProfile.DataDisks == nil {
|
|
return nil, nil
|
|
}
|
|
|
|
return *vm.StorageProfile.DataDisks, nil
|
|
}
|