See issue #33128 We can't rely on the device name provided by Cinder, and thus must perform detection based on the drive serial number (aka It's cinder ID) on the kubelet itself. This patch re-works the cinder volume attacher to ignore the supplied deviceName, and instead defer to the pre-existing GetDevicePath method to discover the device path based on it's serial number and /dev/disk/by-id mapping. This new behavior is controller by a config option, as falling back to the cinder value when we can't discover a device would risk devices not showing up, falling back to cinder's guess, and detecting the wrong disk as attached.
308 lines
9.2 KiB
Go
308 lines
9.2 KiB
Go
/*
|
|
Copyright 2016 The Kubernetes Authors.
|
|
|
|
Licensed under the Apache License, Version 2.0 (the "License");
|
|
you may not use this file except in compliance with the License.
|
|
You may obtain a copy of the License at
|
|
|
|
http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
Unless required by applicable law or agreed to in writing, software
|
|
distributed under the License is distributed on an "AS IS" BASIS,
|
|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
See the License for the specific language governing permissions and
|
|
limitations under the License.
|
|
*/
|
|
|
|
package cinder
|
|
|
|
import (
|
|
"fmt"
|
|
"os"
|
|
"path"
|
|
"strings"
|
|
"time"
|
|
|
|
"github.com/golang/glog"
|
|
"k8s.io/kubernetes/pkg/types"
|
|
"k8s.io/kubernetes/pkg/util/exec"
|
|
"k8s.io/kubernetes/pkg/util/mount"
|
|
"k8s.io/kubernetes/pkg/volume"
|
|
volumeutil "k8s.io/kubernetes/pkg/volume/util"
|
|
)
|
|
|
|
type cinderDiskAttacher struct {
|
|
host volume.VolumeHost
|
|
cinderProvider CinderProvider
|
|
}
|
|
|
|
var _ volume.Attacher = &cinderDiskAttacher{}
|
|
|
|
var _ volume.AttachableVolumePlugin = &cinderPlugin{}
|
|
|
|
const (
|
|
checkSleepDuration = time.Second
|
|
)
|
|
|
|
func (plugin *cinderPlugin) NewAttacher() (volume.Attacher, error) {
|
|
cinder, err := getCloudProvider(plugin.host.GetCloudProvider())
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
return &cinderDiskAttacher{
|
|
host: plugin.host,
|
|
cinderProvider: cinder,
|
|
}, nil
|
|
}
|
|
|
|
func (plugin *cinderPlugin) GetDeviceMountRefs(deviceMountPath string) ([]string, error) {
|
|
mounter := plugin.host.GetMounter()
|
|
return mount.GetMountRefs(mounter, deviceMountPath)
|
|
}
|
|
|
|
func (attacher *cinderDiskAttacher) Attach(spec *volume.Spec, nodeName types.NodeName) (string, error) {
|
|
volumeSource, _, err := getVolumeSource(spec)
|
|
if err != nil {
|
|
return "", err
|
|
}
|
|
|
|
volumeID := volumeSource.VolumeID
|
|
|
|
instances, res := attacher.cinderProvider.Instances()
|
|
if !res {
|
|
return "", fmt.Errorf("failed to list openstack instances")
|
|
}
|
|
instanceid, err := instances.InstanceID(nodeName)
|
|
if err != nil {
|
|
return "", err
|
|
}
|
|
if ind := strings.LastIndex(instanceid, "/"); ind >= 0 {
|
|
instanceid = instanceid[(ind + 1):]
|
|
}
|
|
attached, err := attacher.cinderProvider.DiskIsAttached(volumeID, instanceid)
|
|
if err != nil {
|
|
// Log error and continue with attach
|
|
glog.Warningf(
|
|
"Error checking if volume (%q) is already attached to current node (%q). Will continue and try attach anyway. err=%v",
|
|
volumeID, instanceid, err)
|
|
}
|
|
|
|
if err == nil && attached {
|
|
// Volume is already attached to node.
|
|
glog.Infof("Attach operation is successful. volume %q is already attached to node %q.", volumeID, instanceid)
|
|
} else {
|
|
_, err = attacher.cinderProvider.AttachDisk(instanceid, volumeID)
|
|
if err == nil {
|
|
glog.Infof("Attach operation successful: volume %q attached to node %q.", volumeID, instanceid)
|
|
} else {
|
|
glog.Infof("Attach volume %q to instance %q failed with %v", volumeID, instanceid, err)
|
|
return "", err
|
|
}
|
|
}
|
|
|
|
devicePath, err := attacher.cinderProvider.GetAttachmentDiskPath(instanceid, volumeID)
|
|
if err != nil {
|
|
glog.Infof("Attach volume %q to instance %q failed with %v", volumeID, instanceid, err)
|
|
return "", err
|
|
}
|
|
|
|
return devicePath, err
|
|
}
|
|
|
|
func (attacher *cinderDiskAttacher) VolumesAreAttached(specs []*volume.Spec, nodeName types.NodeName) (map[*volume.Spec]bool, error) {
|
|
volumesAttachedCheck := make(map[*volume.Spec]bool)
|
|
volumeSpecMap := make(map[string]*volume.Spec)
|
|
volumeIDList := []string{}
|
|
for _, spec := range specs {
|
|
volumeSource, _, err := getVolumeSource(spec)
|
|
if err != nil {
|
|
glog.Errorf("Error getting volume (%q) source : %v", spec.Name(), err)
|
|
continue
|
|
}
|
|
|
|
volumeIDList = append(volumeIDList, volumeSource.VolumeID)
|
|
volumesAttachedCheck[spec] = true
|
|
volumeSpecMap[volumeSource.VolumeID] = spec
|
|
}
|
|
attachedResult, err := attacher.cinderProvider.DisksAreAttached(volumeIDList, string(nodeName))
|
|
if err != nil {
|
|
// Log error and continue with attach
|
|
glog.Errorf(
|
|
"Error checking if Volumes (%v) are already attached to current node (%q). Will continue and try attach anyway. err=%v",
|
|
volumeIDList, nodeName, err)
|
|
return volumesAttachedCheck, err
|
|
}
|
|
|
|
for volumeID, attached := range attachedResult {
|
|
if !attached {
|
|
spec := volumeSpecMap[volumeID]
|
|
volumesAttachedCheck[spec] = false
|
|
glog.V(2).Infof("VolumesAreAttached: check volume %q (specName: %q) is no longer attached", volumeID, spec.Name())
|
|
}
|
|
}
|
|
return volumesAttachedCheck, nil
|
|
}
|
|
|
|
func (attacher *cinderDiskAttacher) WaitForAttach(spec *volume.Spec, devicePath string, timeout time.Duration) (string, error) {
|
|
// NOTE: devicePath is is path as reported by Cinder, which may be incorrect and should not be used. See Issue #33128
|
|
volumeSource, _, err := getVolumeSource(spec)
|
|
if err != nil {
|
|
return "", err
|
|
}
|
|
|
|
volumeID := volumeSource.VolumeID
|
|
|
|
if devicePath == "" {
|
|
return "", fmt.Errorf("WaitForAttach failed for Cinder disk %q: devicePath is empty.", volumeID)
|
|
}
|
|
|
|
ticker := time.NewTicker(checkSleepDuration)
|
|
defer ticker.Stop()
|
|
timer := time.NewTimer(timeout)
|
|
defer timer.Stop()
|
|
|
|
for {
|
|
probeAttachedVolume()
|
|
select {
|
|
case <-ticker.C:
|
|
glog.V(5).Infof("Checking Cinder disk %q is attached.", volumeID)
|
|
probeAttachedVolume()
|
|
if !attacher.cinderProvider.ShouldTrustDevicePath() {
|
|
// Using the Cinder volume ID, find the real device path (See Issue #33128)
|
|
devicePath = attacher.cinderProvider.GetDevicePath(volumeID)
|
|
}
|
|
exists, err := volumeutil.PathExists(devicePath)
|
|
if exists && err == nil {
|
|
glog.Infof("Successfully found attached Cinder disk %q at %v.", volumeID, devicePath)
|
|
return devicePath, nil
|
|
} else {
|
|
// Log an error, and continue checking periodically
|
|
glog.Errorf("Error: could not find attached Cinder disk %q: %v", volumeID, err)
|
|
}
|
|
case <-timer.C:
|
|
return "", fmt.Errorf("Could not find attached Cinder disk %q. Timeout waiting for mount paths to be created.", volumeID)
|
|
}
|
|
}
|
|
}
|
|
|
|
func (attacher *cinderDiskAttacher) GetDeviceMountPath(
|
|
spec *volume.Spec) (string, error) {
|
|
volumeSource, _, err := getVolumeSource(spec)
|
|
if err != nil {
|
|
return "", err
|
|
}
|
|
|
|
return makeGlobalPDName(attacher.host, volumeSource.VolumeID), nil
|
|
}
|
|
|
|
// FIXME: this method can be further pruned.
|
|
func (attacher *cinderDiskAttacher) MountDevice(spec *volume.Spec, devicePath string, deviceMountPath string) error {
|
|
mounter := attacher.host.GetMounter()
|
|
notMnt, err := mounter.IsLikelyNotMountPoint(deviceMountPath)
|
|
if err != nil {
|
|
if os.IsNotExist(err) {
|
|
if err := os.MkdirAll(deviceMountPath, 0750); err != nil {
|
|
return err
|
|
}
|
|
notMnt = true
|
|
} else {
|
|
return err
|
|
}
|
|
}
|
|
|
|
volumeSource, readOnly, err := getVolumeSource(spec)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
options := []string{}
|
|
if readOnly {
|
|
options = append(options, "ro")
|
|
}
|
|
if notMnt {
|
|
diskMounter := &mount.SafeFormatAndMount{Interface: mounter, Runner: exec.New()}
|
|
err = diskMounter.FormatAndMount(devicePath, deviceMountPath, volumeSource.FSType, options)
|
|
if err != nil {
|
|
os.Remove(deviceMountPath)
|
|
return err
|
|
}
|
|
}
|
|
return nil
|
|
}
|
|
|
|
type cinderDiskDetacher struct {
|
|
mounter mount.Interface
|
|
cinderProvider CinderProvider
|
|
}
|
|
|
|
var _ volume.Detacher = &cinderDiskDetacher{}
|
|
|
|
func (plugin *cinderPlugin) NewDetacher() (volume.Detacher, error) {
|
|
cinder, err := getCloudProvider(plugin.host.GetCloudProvider())
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
return &cinderDiskDetacher{
|
|
mounter: plugin.host.GetMounter(),
|
|
cinderProvider: cinder,
|
|
}, nil
|
|
}
|
|
|
|
func (detacher *cinderDiskDetacher) Detach(deviceMountPath string, nodeName types.NodeName) error {
|
|
volumeID := path.Base(deviceMountPath)
|
|
instances, res := detacher.cinderProvider.Instances()
|
|
if !res {
|
|
return fmt.Errorf("failed to list openstack instances")
|
|
}
|
|
instanceid, err := instances.InstanceID(nodeName)
|
|
if ind := strings.LastIndex(instanceid, "/"); ind >= 0 {
|
|
instanceid = instanceid[(ind + 1):]
|
|
}
|
|
|
|
attached, err := detacher.cinderProvider.DiskIsAttached(volumeID, instanceid)
|
|
if err != nil {
|
|
// Log error and continue with detach
|
|
glog.Errorf(
|
|
"Error checking if volume (%q) is already attached to current node (%q). Will continue and try detach anyway. err=%v",
|
|
volumeID, nodeName, err)
|
|
}
|
|
|
|
if err == nil && !attached {
|
|
// Volume is already detached from node.
|
|
glog.Infof("detach operation was successful. volume %q is already detached from node %q.", volumeID, nodeName)
|
|
return nil
|
|
}
|
|
|
|
if err = detacher.cinderProvider.DetachDisk(instanceid, volumeID); err != nil {
|
|
glog.Errorf("Error detaching volume %q: %v", volumeID, err)
|
|
return err
|
|
}
|
|
glog.Infof("detatached volume %q from instance %q", volumeID, instanceid)
|
|
return nil
|
|
}
|
|
|
|
func (detacher *cinderDiskDetacher) WaitForDetach(devicePath string, timeout time.Duration) error {
|
|
ticker := time.NewTicker(checkSleepDuration)
|
|
defer ticker.Stop()
|
|
timer := time.NewTimer(timeout)
|
|
defer timer.Stop()
|
|
|
|
for {
|
|
select {
|
|
case <-ticker.C:
|
|
glog.V(5).Infof("Checking device %q is detached.", devicePath)
|
|
if pathExists, err := volumeutil.PathExists(devicePath); err != nil {
|
|
return fmt.Errorf("Error checking if device path exists: %v", err)
|
|
} else if !pathExists {
|
|
return nil
|
|
}
|
|
case <-timer.C:
|
|
return fmt.Errorf("Timeout reached; PD Device %v is still attached", devicePath)
|
|
}
|
|
}
|
|
}
|
|
|
|
func (detacher *cinderDiskDetacher) UnmountDevice(deviceMountPath string) error {
|
|
return volumeutil.UnmountPath(deviceMountPath, detacher.mounter)
|
|
}
|