123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483 |
- /*
- Copyright 2018 The Kubernetes Authors.
- Licensed under the Apache License, Version 2.0 (the "License");
- you may not use this file except in compliance with the License.
- You may obtain a copy of the License at
- http://www.apache.org/licenses/LICENSE-2.0
- Unless required by applicable law or agreed to in writing, software
- distributed under the License is distributed on an "AS IS" BASIS,
- WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- See the License for the specific language governing permissions and
- limitations under the License.
- */
- package testsuites
- import (
- "context"
- "fmt"
- "path/filepath"
- "strings"
- "github.com/onsi/ginkgo"
- "github.com/onsi/gomega"
- v1 "k8s.io/api/core/v1"
- storagev1 "k8s.io/api/storage/v1"
- metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
- "k8s.io/apimachinery/pkg/fields"
- "k8s.io/apimachinery/pkg/util/errors"
- clientset "k8s.io/client-go/kubernetes"
- volevents "k8s.io/kubernetes/pkg/controller/volume/events"
- "k8s.io/kubernetes/pkg/kubelet/events"
- "k8s.io/kubernetes/test/e2e/framework"
- e2eevents "k8s.io/kubernetes/test/e2e/framework/events"
- e2epod "k8s.io/kubernetes/test/e2e/framework/pod"
- e2epv "k8s.io/kubernetes/test/e2e/framework/pv"
- e2eskipper "k8s.io/kubernetes/test/e2e/framework/skipper"
- "k8s.io/kubernetes/test/e2e/framework/volume"
- "k8s.io/kubernetes/test/e2e/storage/testpatterns"
- "k8s.io/kubernetes/test/e2e/storage/utils"
- )
- const (
- noProvisioner = "kubernetes.io/no-provisioner"
- pvNamePrefix = "pv"
- )
- type volumeModeTestSuite struct {
- tsInfo TestSuiteInfo
- }
- var _ TestSuite = &volumeModeTestSuite{}
- // InitVolumeModeTestSuite returns volumeModeTestSuite that implements TestSuite interface
- func InitVolumeModeTestSuite() TestSuite {
- return &volumeModeTestSuite{
- tsInfo: TestSuiteInfo{
- Name: "volumeMode",
- TestPatterns: []testpatterns.TestPattern{
- testpatterns.FsVolModePreprovisionedPV,
- testpatterns.FsVolModeDynamicPV,
- testpatterns.BlockVolModePreprovisionedPV,
- testpatterns.BlockVolModeDynamicPV,
- },
- SupportedSizeRange: volume.SizeRange{
- Min: "1Mi",
- },
- },
- }
- }
- func (t *volumeModeTestSuite) GetTestSuiteInfo() TestSuiteInfo {
- return t.tsInfo
- }
- func (t *volumeModeTestSuite) SkipRedundantSuite(driver TestDriver, pattern testpatterns.TestPattern) {
- }
- func (t *volumeModeTestSuite) DefineTests(driver TestDriver, pattern testpatterns.TestPattern) {
- type local struct {
- config *PerTestConfig
- driverCleanup func()
- cs clientset.Interface
- ns *v1.Namespace
- // VolumeResource contains pv, pvc, sc, etc., owns cleaning that up
- VolumeResource
- intreeOps opCounts
- migratedOps opCounts
- }
- var (
- dInfo = driver.GetDriverInfo()
- l local
- )
- // No preconditions to test. Normally they would be in a BeforeEach here.
- // This intentionally comes after checking the preconditions because it
- // registers its own BeforeEach which creates the namespace. Beware that it
- // also registers an AfterEach which renders f unusable. Any code using
- // f must run inside an It or Context callback.
- f := framework.NewDefaultFramework("volumemode")
- init := func() {
- l = local{}
- l.ns = f.Namespace
- l.cs = f.ClientSet
- // Now do the more expensive test initialization.
- l.config, l.driverCleanup = driver.PrepareTest(f)
- l.intreeOps, l.migratedOps = getMigrationVolumeOpCounts(f.ClientSet, dInfo.InTreePluginName)
- }
- // manualInit initializes l.VolumeResource without creating the PV & PVC objects.
- manualInit := func() {
- init()
- fsType := pattern.FsType
- volBindMode := storagev1.VolumeBindingImmediate
- var (
- scName string
- pvSource *v1.PersistentVolumeSource
- volumeNodeAffinity *v1.VolumeNodeAffinity
- )
- l.VolumeResource = VolumeResource{
- Config: l.config,
- Pattern: pattern,
- }
- // Create volume for pre-provisioned volume tests
- l.Volume = CreateVolume(driver, l.config, pattern.VolType)
- switch pattern.VolType {
- case testpatterns.PreprovisionedPV:
- if pattern.VolMode == v1.PersistentVolumeBlock {
- scName = fmt.Sprintf("%s-%s-sc-for-block", l.ns.Name, dInfo.Name)
- } else if pattern.VolMode == v1.PersistentVolumeFilesystem {
- scName = fmt.Sprintf("%s-%s-sc-for-file", l.ns.Name, dInfo.Name)
- }
- if pDriver, ok := driver.(PreprovisionedPVTestDriver); ok {
- pvSource, volumeNodeAffinity = pDriver.GetPersistentVolumeSource(false, fsType, l.Volume)
- if pvSource == nil {
- e2eskipper.Skipf("Driver %q does not define PersistentVolumeSource - skipping", dInfo.Name)
- }
- storageClass, pvConfig, pvcConfig := generateConfigsForPreprovisionedPVTest(scName, volBindMode, pattern.VolMode, *pvSource, volumeNodeAffinity)
- l.Sc = storageClass
- l.Pv = e2epv.MakePersistentVolume(pvConfig)
- l.Pvc = e2epv.MakePersistentVolumeClaim(pvcConfig, l.ns.Name)
- }
- case testpatterns.DynamicPV:
- if dDriver, ok := driver.(DynamicPVTestDriver); ok {
- l.Sc = dDriver.GetDynamicProvisionStorageClass(l.config, fsType)
- if l.Sc == nil {
- e2eskipper.Skipf("Driver %q does not define Dynamic Provision StorageClass - skipping", dInfo.Name)
- }
- l.Sc.VolumeBindingMode = &volBindMode
- testVolumeSizeRange := t.GetTestSuiteInfo().SupportedSizeRange
- driverVolumeSizeRange := dInfo.SupportedSizeRange
- claimSize, err := getSizeRangesIntersection(testVolumeSizeRange, driverVolumeSizeRange)
- framework.ExpectNoError(err, "determine intersection of test size range %+v and driver size range %+v", testVolumeSizeRange, driverVolumeSizeRange)
- l.Pvc = e2epv.MakePersistentVolumeClaim(e2epv.PersistentVolumeClaimConfig{
- ClaimSize: claimSize,
- StorageClassName: &(l.Sc.Name),
- VolumeMode: &pattern.VolMode,
- }, l.ns.Name)
- }
- default:
- framework.Failf("Volume mode test doesn't support: %s", pattern.VolType)
- }
- }
- cleanup := func() {
- var errs []error
- errs = append(errs, l.CleanupResource())
- errs = append(errs, tryFunc(l.driverCleanup))
- l.driverCleanup = nil
- framework.ExpectNoError(errors.NewAggregate(errs), "while cleaning up resource")
- validateMigrationVolumeOpCounts(f.ClientSet, dInfo.InTreePluginName, l.intreeOps, l.migratedOps)
- }
- // We register different tests depending on the drive
- isBlockSupported := dInfo.Capabilities[CapBlock]
- switch pattern.VolType {
- case testpatterns.PreprovisionedPV:
- if pattern.VolMode == v1.PersistentVolumeBlock && !isBlockSupported {
- ginkgo.It("should fail to create pod by failing to mount volume [Slow]", func() {
- manualInit()
- defer cleanup()
- var err error
- ginkgo.By("Creating sc")
- l.Sc, err = l.cs.StorageV1().StorageClasses().Create(context.TODO(), l.Sc, metav1.CreateOptions{})
- framework.ExpectNoError(err, "Failed to create sc")
- ginkgo.By("Creating pv and pvc")
- l.Pv, err = l.cs.CoreV1().PersistentVolumes().Create(context.TODO(), l.Pv, metav1.CreateOptions{})
- framework.ExpectNoError(err, "Failed to create pv")
- // Prebind pv
- l.Pvc.Spec.VolumeName = l.Pv.Name
- l.Pvc, err = l.cs.CoreV1().PersistentVolumeClaims(l.ns.Name).Create(context.TODO(), l.Pvc, metav1.CreateOptions{})
- framework.ExpectNoError(err, "Failed to create pvc")
- framework.ExpectNoError(e2epv.WaitOnPVandPVC(l.cs, l.ns.Name, l.Pv, l.Pvc), "Failed to bind pv and pvc")
- ginkgo.By("Creating pod")
- pod := e2epod.MakeSecPod(l.ns.Name, []*v1.PersistentVolumeClaim{l.Pvc}, nil, false, "", false, false, e2epv.SELinuxLabel, nil)
- // Setting node
- e2epod.SetNodeSelection(pod, l.config.ClientNodeSelection)
- pod, err = l.cs.CoreV1().Pods(l.ns.Name).Create(context.TODO(), pod, metav1.CreateOptions{})
- framework.ExpectNoError(err, "Failed to create pod")
- defer func() {
- framework.ExpectNoError(e2epod.DeletePodWithWait(l.cs, pod), "Failed to delete pod")
- }()
- eventSelector := fields.Set{
- "involvedObject.kind": "Pod",
- "involvedObject.name": pod.Name,
- "involvedObject.namespace": l.ns.Name,
- "reason": events.FailedMountVolume,
- }.AsSelector().String()
- msg := "Unable to attach or mount volumes"
- err = e2eevents.WaitTimeoutForEvent(l.cs, l.ns.Name, eventSelector, msg, framework.PodStartTimeout)
- // Events are unreliable, don't depend on the event. It's used only to speed up the test.
- if err != nil {
- framework.Logf("Warning: did not get event about FailedMountVolume")
- }
- // Check the pod is still not running
- p, err := l.cs.CoreV1().Pods(l.ns.Name).Get(context.TODO(), pod.Name, metav1.GetOptions{})
- framework.ExpectNoError(err, "could not re-read the pod after event (or timeout)")
- framework.ExpectEqual(p.Status.Phase, v1.PodPending, "Pod phase isn't pending")
- })
- }
- case testpatterns.DynamicPV:
- if pattern.VolMode == v1.PersistentVolumeBlock && !isBlockSupported {
- ginkgo.It("should fail in binding dynamic provisioned PV to PVC [Slow]", func() {
- manualInit()
- defer cleanup()
- var err error
- ginkgo.By("Creating sc")
- l.Sc, err = l.cs.StorageV1().StorageClasses().Create(context.TODO(), l.Sc, metav1.CreateOptions{})
- framework.ExpectNoError(err, "Failed to create sc")
- ginkgo.By("Creating pv and pvc")
- l.Pvc, err = l.cs.CoreV1().PersistentVolumeClaims(l.ns.Name).Create(context.TODO(), l.Pvc, metav1.CreateOptions{})
- framework.ExpectNoError(err, "Failed to create pvc")
- eventSelector := fields.Set{
- "involvedObject.kind": "PersistentVolumeClaim",
- "involvedObject.name": l.Pvc.Name,
- "involvedObject.namespace": l.ns.Name,
- "reason": volevents.ProvisioningFailed,
- }.AsSelector().String()
- msg := "does not support block volume provisioning"
- err = e2eevents.WaitTimeoutForEvent(l.cs, l.ns.Name, eventSelector, msg, framework.ClaimProvisionTimeout)
- // Events are unreliable, don't depend on the event. It's used only to speed up the test.
- if err != nil {
- framework.Logf("Warning: did not get event about provisioing failed")
- }
- // Check the pvc is still pending
- pvc, err := l.cs.CoreV1().PersistentVolumeClaims(l.ns.Name).Get(context.TODO(), l.Pvc.Name, metav1.GetOptions{})
- framework.ExpectNoError(err, "Failed to re-read the pvc after event (or timeout)")
- framework.ExpectEqual(pvc.Status.Phase, v1.ClaimPending, "PVC phase isn't pending")
- })
- }
- default:
- framework.Failf("Volume mode test doesn't support volType: %v", pattern.VolType)
- }
- ginkgo.It("should fail to use a volume in a pod with mismatched mode [Slow]", func() {
- skipTestIfBlockNotSupported(driver)
- init()
- testVolumeSizeRange := t.GetTestSuiteInfo().SupportedSizeRange
- l.VolumeResource = *CreateVolumeResource(driver, l.config, pattern, testVolumeSizeRange)
- defer cleanup()
- ginkgo.By("Creating pod")
- var err error
- pod := e2epod.MakeSecPod(l.ns.Name, []*v1.PersistentVolumeClaim{l.Pvc}, nil, false, "", false, false, e2epv.SELinuxLabel, nil)
- // Change volumeMounts to volumeDevices and the other way around
- pod = swapVolumeMode(pod)
- // Run the pod
- pod, err = l.cs.CoreV1().Pods(l.ns.Name).Create(context.TODO(), pod, metav1.CreateOptions{})
- framework.ExpectNoError(err, "Failed to create pod")
- defer func() {
- framework.ExpectNoError(e2epod.DeletePodWithWait(l.cs, pod), "Failed to delete pod")
- }()
- ginkgo.By("Waiting for the pod to fail")
- // Wait for an event that the pod is invalid.
- eventSelector := fields.Set{
- "involvedObject.kind": "Pod",
- "involvedObject.name": pod.Name,
- "involvedObject.namespace": l.ns.Name,
- "reason": events.FailedMountVolume,
- }.AsSelector().String()
- var msg string
- if pattern.VolMode == v1.PersistentVolumeBlock {
- msg = "has volumeMode Block, but is specified in volumeMounts"
- } else {
- msg = "has volumeMode Filesystem, but is specified in volumeDevices"
- }
- err = e2eevents.WaitTimeoutForEvent(l.cs, l.ns.Name, eventSelector, msg, framework.PodStartTimeout)
- // Events are unreliable, don't depend on them. They're used only to speed up the test.
- if err != nil {
- framework.Logf("Warning: did not get event about mismatched volume use")
- }
- // Check the pod is still not running
- p, err := l.cs.CoreV1().Pods(l.ns.Name).Get(context.TODO(), pod.Name, metav1.GetOptions{})
- framework.ExpectNoError(err, "could not re-read the pod after event (or timeout)")
- framework.ExpectEqual(p.Status.Phase, v1.PodPending, "Pod phase isn't pending")
- })
- ginkgo.It("should not mount / map unused volumes in a pod", func() {
- if pattern.VolMode == v1.PersistentVolumeBlock {
- skipTestIfBlockNotSupported(driver)
- }
- init()
- testVolumeSizeRange := t.GetTestSuiteInfo().SupportedSizeRange
- l.VolumeResource = *CreateVolumeResource(driver, l.config, pattern, testVolumeSizeRange)
- defer cleanup()
- ginkgo.By("Creating pod")
- var err error
- pod := e2epod.MakeSecPod(l.ns.Name, []*v1.PersistentVolumeClaim{l.Pvc}, nil, false, "", false, false, e2epv.SELinuxLabel, nil)
- for i := range pod.Spec.Containers {
- pod.Spec.Containers[i].VolumeDevices = nil
- pod.Spec.Containers[i].VolumeMounts = nil
- }
- // Run the pod
- pod, err = l.cs.CoreV1().Pods(l.ns.Name).Create(context.TODO(), pod, metav1.CreateOptions{})
- framework.ExpectNoError(err)
- defer func() {
- framework.ExpectNoError(e2epod.DeletePodWithWait(l.cs, pod))
- }()
- err = e2epod.WaitForPodNameRunningInNamespace(l.cs, pod.Name, pod.Namespace)
- framework.ExpectNoError(err)
- // Reload the pod to get its node
- pod, err = l.cs.CoreV1().Pods(l.ns.Name).Get(context.TODO(), pod.Name, metav1.GetOptions{})
- framework.ExpectNoError(err)
- framework.ExpectNotEqual(pod.Spec.NodeName, "", "pod should be scheduled to a node")
- node, err := l.cs.CoreV1().Nodes().Get(context.TODO(), pod.Spec.NodeName, metav1.GetOptions{})
- framework.ExpectNoError(err)
- ginkgo.By("Listing mounted volumes in the pod")
- hostExec := utils.NewHostExec(f)
- defer hostExec.Cleanup()
- volumePaths, devicePaths, err := listPodVolumePluginDirectory(hostExec, pod, node)
- framework.ExpectNoError(err)
- driverInfo := driver.GetDriverInfo()
- volumePlugin := driverInfo.InTreePluginName
- if len(volumePlugin) == 0 {
- // TODO: check if it's a CSI volume first
- volumePlugin = "kubernetes.io/csi"
- }
- ginkgo.By(fmt.Sprintf("Checking that volume plugin %s is not used in pod directory", volumePlugin))
- safeVolumePlugin := strings.ReplaceAll(volumePlugin, "/", "~")
- for _, path := range volumePaths {
- gomega.Expect(path).NotTo(gomega.ContainSubstring(safeVolumePlugin), fmt.Sprintf("no %s volume should be mounted into pod directory", volumePlugin))
- }
- for _, path := range devicePaths {
- gomega.Expect(path).NotTo(gomega.ContainSubstring(safeVolumePlugin), fmt.Sprintf("no %s volume should be symlinked into pod directory", volumePlugin))
- }
- })
- }
- func generateConfigsForPreprovisionedPVTest(scName string, volBindMode storagev1.VolumeBindingMode,
- volMode v1.PersistentVolumeMode, pvSource v1.PersistentVolumeSource, volumeNodeAffinity *v1.VolumeNodeAffinity) (*storagev1.StorageClass,
- e2epv.PersistentVolumeConfig, e2epv.PersistentVolumeClaimConfig) {
- // StorageClass
- scConfig := &storagev1.StorageClass{
- ObjectMeta: metav1.ObjectMeta{
- Name: scName,
- },
- Provisioner: noProvisioner,
- VolumeBindingMode: &volBindMode,
- }
- // PV
- pvConfig := e2epv.PersistentVolumeConfig{
- PVSource: pvSource,
- NodeAffinity: volumeNodeAffinity,
- NamePrefix: pvNamePrefix,
- StorageClassName: scName,
- VolumeMode: &volMode,
- }
- // PVC
- pvcConfig := e2epv.PersistentVolumeClaimConfig{
- AccessModes: []v1.PersistentVolumeAccessMode{v1.ReadWriteOnce},
- StorageClassName: &scName,
- VolumeMode: &volMode,
- }
- return scConfig, pvConfig, pvcConfig
- }
- // swapVolumeMode changes volumeMounts to volumeDevices and the other way around
- func swapVolumeMode(podTemplate *v1.Pod) *v1.Pod {
- pod := podTemplate.DeepCopy()
- for c := range pod.Spec.Containers {
- container := &pod.Spec.Containers[c]
- container.VolumeDevices = []v1.VolumeDevice{}
- container.VolumeMounts = []v1.VolumeMount{}
- // Change VolumeMounts to VolumeDevices
- for _, volumeMount := range podTemplate.Spec.Containers[c].VolumeMounts {
- container.VolumeDevices = append(container.VolumeDevices, v1.VolumeDevice{
- Name: volumeMount.Name,
- DevicePath: volumeMount.MountPath,
- })
- }
- // Change VolumeDevices to VolumeMounts
- for _, volumeDevice := range podTemplate.Spec.Containers[c].VolumeDevices {
- container.VolumeMounts = append(container.VolumeMounts, v1.VolumeMount{
- Name: volumeDevice.Name,
- MountPath: volumeDevice.DevicePath,
- })
- }
- }
- return pod
- }
- // listPodVolumePluginDirectory returns all volumes in /var/lib/kubelet/pods/<pod UID>/volumes/* and
- // /var/lib/kubelet/pods/<pod UID>/volumeDevices/*
- // Sample output:
- // /var/lib/kubelet/pods/a4717a30-000a-4081-a7a8-f51adf280036/volumes/kubernetes.io~secret/default-token-rphdt
- // /var/lib/kubelet/pods/4475b7a3-4a55-4716-9119-fd0053d9d4a6/volumeDevices/kubernetes.io~aws-ebs/pvc-5f9f80f5-c90b-4586-9966-83f91711e1c0
- func listPodVolumePluginDirectory(h utils.HostExec, pod *v1.Pod, node *v1.Node) (mounts []string, devices []string, err error) {
- mountPath := filepath.Join("/var/lib/kubelet/pods/", string(pod.UID), "volumes")
- devicePath := filepath.Join("/var/lib/kubelet/pods/", string(pod.UID), "volumeDevices")
- mounts, err = listPodDirectory(h, mountPath, node)
- if err != nil {
- return nil, nil, err
- }
- devices, err = listPodDirectory(h, devicePath, node)
- if err != nil {
- return nil, nil, err
- }
- return mounts, devices, nil
- }
- func listPodDirectory(h utils.HostExec, path string, node *v1.Node) ([]string, error) {
- // Return no error if the directory does not exist (e.g. there are no block volumes used)
- _, err := h.IssueCommandWithResult("test ! -d "+path, node)
- if err == nil {
- // The directory does not exist
- return nil, nil
- }
- // The directory either exists or a real error happened (e.g. "access denied").
- // Ignore the error, "find" will hit the error again and we report it there.
- // Inside /var/lib/kubelet/pods/<pod>/volumes, look for <volume_plugin>/<volume-name>, hence depth 2
- cmd := fmt.Sprintf("find %s -mindepth 2 -maxdepth 2", path)
- out, err := h.IssueCommandWithResult(cmd, node)
- if err != nil {
- return nil, fmt.Errorf("error checking directory %s on node %s: %s", path, node.Name, err)
- }
- return strings.Split(out, "\n"), nil
- }
|