
This changes the text registration so that tags for which the framework has a dedicated API (features, feature gates, slow, serial, etc.) those APIs are used. Arbitrary, custom tags are still left in place for now.
249 lines
8.4 KiB
Go
249 lines
8.4 KiB
Go
/*
|
|
Copyright 2017 The Kubernetes Authors.
|
|
|
|
Licensed under the Apache License, Version 2.0 (the "License");
|
|
you may not use this file except in compliance with the License.
|
|
You may obtain a copy of the License at
|
|
|
|
http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
Unless required by applicable law or agreed to in writing, software
|
|
distributed under the License is distributed on an "AS IS" BASIS,
|
|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
See the License for the specific language governing permissions and
|
|
limitations under the License.
|
|
*/
|
|
|
|
package storage
|
|
|
|
import (
|
|
"context"
|
|
"fmt"
|
|
"path"
|
|
|
|
"time"
|
|
|
|
v1 "k8s.io/api/core/v1"
|
|
metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
|
|
"k8s.io/apimachinery/pkg/util/wait"
|
|
clientset "k8s.io/client-go/kubernetes"
|
|
"k8s.io/kubernetes/test/e2e/feature"
|
|
"k8s.io/kubernetes/test/e2e/framework"
|
|
e2enode "k8s.io/kubernetes/test/e2e/framework/node"
|
|
e2epod "k8s.io/kubernetes/test/e2e/framework/pod"
|
|
e2eskipper "k8s.io/kubernetes/test/e2e/framework/skipper"
|
|
"k8s.io/kubernetes/test/e2e/storage/utils"
|
|
imageutils "k8s.io/kubernetes/test/utils/image"
|
|
admissionapi "k8s.io/pod-security-admission/api"
|
|
|
|
"github.com/onsi/ginkgo/v2"
|
|
)
|
|
|
|
var (
|
|
// BusyBoxImage is the image URI of BusyBox.
|
|
BusyBoxImage = imageutils.GetE2EImage(imageutils.BusyBox)
|
|
durationForStuckMount = 110 * time.Second
|
|
)
|
|
|
|
var _ = utils.SIGDescribe(feature.Flexvolumes, "Detaching volumes", func() {
|
|
f := framework.NewDefaultFramework("flexvolume")
|
|
f.NamespacePodSecurityLevel = admissionapi.LevelPrivileged
|
|
|
|
// note that namespace deletion is handled by delete-namespace flag
|
|
|
|
var cs clientset.Interface
|
|
var ns *v1.Namespace
|
|
var node *v1.Node
|
|
var suffix string
|
|
|
|
ginkgo.BeforeEach(func(ctx context.Context) {
|
|
e2eskipper.SkipUnlessProviderIs("gce", "local")
|
|
e2eskipper.SkipUnlessMasterOSDistroIs("debian", "ubuntu", "gci", "custom")
|
|
e2eskipper.SkipUnlessNodeOSDistroIs("debian", "ubuntu", "gci", "custom")
|
|
e2eskipper.SkipUnlessSSHKeyPresent()
|
|
|
|
cs = f.ClientSet
|
|
ns = f.Namespace
|
|
var err error
|
|
node, err = e2enode.GetRandomReadySchedulableNode(ctx, f.ClientSet)
|
|
framework.ExpectNoError(err)
|
|
suffix = ns.Name
|
|
})
|
|
|
|
f.It("should not work when mount is in progress", f.WithSlow(), func(ctx context.Context) {
|
|
e2eskipper.SkipUnlessSSHKeyPresent()
|
|
|
|
driver := "attachable-with-long-mount"
|
|
driverInstallAs := driver + "-" + suffix
|
|
|
|
ginkgo.By(fmt.Sprintf("installing flexvolume %s on node %s as %s", path.Join(driverDir, driver), node.Name, driverInstallAs))
|
|
installFlex(ctx, cs, node, "k8s", driverInstallAs, path.Join(driverDir, driver))
|
|
ginkgo.By(fmt.Sprintf("installing flexvolume %s on master as %s", path.Join(driverDir, driver), driverInstallAs))
|
|
installFlex(ctx, cs, nil, "k8s", driverInstallAs, path.Join(driverDir, driver))
|
|
volumeSource := v1.VolumeSource{
|
|
FlexVolume: &v1.FlexVolumeSource{
|
|
Driver: "k8s/" + driverInstallAs,
|
|
},
|
|
}
|
|
|
|
clientPod := getFlexVolumePod(volumeSource, node.Name)
|
|
ginkgo.By("Creating pod that uses slow format volume")
|
|
pod, err := cs.CoreV1().Pods(ns.Name).Create(ctx, clientPod, metav1.CreateOptions{})
|
|
framework.ExpectNoError(err)
|
|
|
|
uniqueVolumeName := getUniqueVolumeName(pod, driverInstallAs)
|
|
|
|
ginkgo.By("waiting for volumes to be attached to node")
|
|
err = waitForVolumesAttached(ctx, cs, node.Name, uniqueVolumeName)
|
|
framework.ExpectNoError(err, "while waiting for volume to attach to %s node", node.Name)
|
|
|
|
ginkgo.By("waiting for volume-in-use on the node after pod creation")
|
|
err = waitForVolumesInUse(ctx, cs, node.Name, uniqueVolumeName)
|
|
framework.ExpectNoError(err, "while waiting for volume in use")
|
|
|
|
ginkgo.By("waiting for kubelet to start mounting the volume")
|
|
time.Sleep(20 * time.Second)
|
|
|
|
ginkgo.By("Deleting the flexvolume pod")
|
|
err = e2epod.DeletePodWithWait(ctx, cs, pod)
|
|
framework.ExpectNoError(err, "in deleting the pod")
|
|
|
|
// Wait a bit for node to sync the volume status
|
|
time.Sleep(30 * time.Second)
|
|
|
|
ginkgo.By("waiting for volume-in-use on the node after pod deletion")
|
|
err = waitForVolumesInUse(ctx, cs, node.Name, uniqueVolumeName)
|
|
framework.ExpectNoError(err, "while waiting for volume in use")
|
|
|
|
// Wait for 110s because mount device operation has a sleep of 120 seconds
|
|
// we previously already waited for 30s.
|
|
time.Sleep(durationForStuckMount)
|
|
|
|
ginkgo.By("waiting for volume to disappear from node in-use")
|
|
err = waitForVolumesNotInUse(ctx, cs, node.Name, uniqueVolumeName)
|
|
framework.ExpectNoError(err, "while waiting for volume to be removed from in-use")
|
|
|
|
ginkgo.By(fmt.Sprintf("uninstalling flexvolume %s from node %s", driverInstallAs, node.Name))
|
|
uninstallFlex(ctx, cs, node, "k8s", driverInstallAs)
|
|
ginkgo.By(fmt.Sprintf("uninstalling flexvolume %s from master", driverInstallAs))
|
|
uninstallFlex(ctx, cs, nil, "k8s", driverInstallAs)
|
|
})
|
|
})
|
|
|
|
func getUniqueVolumeName(pod *v1.Pod, driverName string) string {
|
|
return fmt.Sprintf("k8s/%s/%s", driverName, pod.Spec.Volumes[0].Name)
|
|
}
|
|
|
|
func waitForVolumesNotInUse(ctx context.Context, client clientset.Interface, nodeName, volumeName string) error {
|
|
waitErr := wait.PollUntilContextTimeout(ctx, 10*time.Second, 60*time.Second, true, func(ctx context.Context) (bool, error) {
|
|
node, err := client.CoreV1().Nodes().Get(ctx, nodeName, metav1.GetOptions{})
|
|
if err != nil {
|
|
return false, fmt.Errorf("error fetching node %s with %v", nodeName, err)
|
|
}
|
|
volumeInUse := node.Status.VolumesInUse
|
|
for _, volume := range volumeInUse {
|
|
if string(volume) == volumeName {
|
|
return false, nil
|
|
}
|
|
}
|
|
return true, nil
|
|
})
|
|
if waitErr != nil {
|
|
return fmt.Errorf("error waiting for volumes to not be in use: %v", waitErr)
|
|
}
|
|
return nil
|
|
}
|
|
|
|
func waitForVolumesAttached(ctx context.Context, client clientset.Interface, nodeName, volumeName string) error {
|
|
waitErr := wait.PollUntilContextTimeout(ctx, 2*time.Second, 2*time.Minute, true, func(ctx context.Context) (bool, error) {
|
|
node, err := client.CoreV1().Nodes().Get(ctx, nodeName, metav1.GetOptions{})
|
|
if err != nil {
|
|
return false, fmt.Errorf("error fetching node %s with %v", nodeName, err)
|
|
}
|
|
volumeAttached := node.Status.VolumesAttached
|
|
for _, volume := range volumeAttached {
|
|
if string(volume.Name) == volumeName {
|
|
return true, nil
|
|
}
|
|
}
|
|
return false, nil
|
|
})
|
|
if waitErr != nil {
|
|
return fmt.Errorf("error waiting for volume %v to attach to node %v: %v", volumeName, nodeName, waitErr)
|
|
}
|
|
return nil
|
|
}
|
|
|
|
func waitForVolumesInUse(ctx context.Context, client clientset.Interface, nodeName, volumeName string) error {
|
|
waitErr := wait.PollUntilContextTimeout(ctx, 10*time.Second, 60*time.Second, true, func(ctx context.Context) (bool, error) {
|
|
node, err := client.CoreV1().Nodes().Get(ctx, nodeName, metav1.GetOptions{})
|
|
if err != nil {
|
|
return false, fmt.Errorf("error fetching node %s with %v", nodeName, err)
|
|
}
|
|
volumeInUse := node.Status.VolumesInUse
|
|
for _, volume := range volumeInUse {
|
|
if string(volume) == volumeName {
|
|
return true, nil
|
|
}
|
|
}
|
|
return false, nil
|
|
})
|
|
if waitErr != nil {
|
|
return fmt.Errorf("error waiting for volume %v to be in use on node %v: %v", volumeName, nodeName, waitErr)
|
|
}
|
|
return nil
|
|
}
|
|
|
|
func getFlexVolumePod(volumeSource v1.VolumeSource, nodeName string) *v1.Pod {
|
|
var gracePeriod int64
|
|
clientPod := &v1.Pod{
|
|
TypeMeta: metav1.TypeMeta{
|
|
Kind: "Pod",
|
|
APIVersion: "v1",
|
|
},
|
|
ObjectMeta: metav1.ObjectMeta{
|
|
Name: "flexvolume-detach-test" + "-client",
|
|
Labels: map[string]string{
|
|
"role": "flexvolume-detach-test" + "-client",
|
|
},
|
|
},
|
|
Spec: v1.PodSpec{
|
|
Containers: []v1.Container{
|
|
{
|
|
Name: "flexvolume-detach-test" + "-client",
|
|
Image: BusyBoxImage,
|
|
WorkingDir: "/opt",
|
|
// An imperative and easily debuggable container which reads vol contents for
|
|
// us to scan in the tests or by eye.
|
|
// We expect that /opt is empty in the minimal containers which we use in this test.
|
|
Command: []string{
|
|
"/bin/sh",
|
|
"-c",
|
|
"while true ; do cat /opt/foo/index.html ; sleep 2 ; ls -altrh /opt/ ; sleep 2 ; done ",
|
|
},
|
|
VolumeMounts: []v1.VolumeMount{
|
|
{
|
|
Name: "test-long-detach-flex",
|
|
MountPath: "/opt/foo",
|
|
},
|
|
},
|
|
},
|
|
},
|
|
TerminationGracePeriodSeconds: &gracePeriod,
|
|
SecurityContext: &v1.PodSecurityContext{
|
|
SELinuxOptions: &v1.SELinuxOptions{
|
|
Level: "s0:c0,c1",
|
|
},
|
|
},
|
|
Volumes: []v1.Volume{
|
|
{
|
|
Name: "test-long-detach-flex",
|
|
VolumeSource: volumeSource,
|
|
},
|
|
},
|
|
NodeName: nodeName,
|
|
},
|
|
}
|
|
return clientPod
|
|
}
|