123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702703704705706707708709710711712713714715716717718719720721722723724725726727728729730731732733734735736737738739740741742743744745746747748749750751752753754755756757758 |
- /*
- Copyright 2017 The Kubernetes Authors.
- Licensed under the Apache License, Version 2.0 (the "License");
- you may not use this file except in compliance with the License.
- You may obtain a copy of the License at
- http://www.apache.org/licenses/LICENSE-2.0
- Unless required by applicable law or agreed to in writing, software
- distributed under the License is distributed on an "AS IS" BASIS,
- WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- See the License for the specific language governing permissions and
- limitations under the License.
- */
- package storage
- import (
- "context"
- "fmt"
- "time"
- "github.com/onsi/ginkgo"
- "github.com/onsi/gomega"
- v1 "k8s.io/api/core/v1"
- storagev1 "k8s.io/api/storage/v1"
- metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
- "k8s.io/apimachinery/pkg/util/wait"
- clientset "k8s.io/client-go/kubernetes"
- "k8s.io/component-base/metrics/testutil"
- kubeletmetrics "k8s.io/kubernetes/pkg/kubelet/metrics"
- "k8s.io/kubernetes/test/e2e/framework"
- "k8s.io/kubernetes/test/e2e/framework/metrics"
- e2epod "k8s.io/kubernetes/test/e2e/framework/pod"
- e2epv "k8s.io/kubernetes/test/e2e/framework/pv"
- e2eskipper "k8s.io/kubernetes/test/e2e/framework/skipper"
- "k8s.io/kubernetes/test/e2e/storage/testsuites"
- "k8s.io/kubernetes/test/e2e/storage/utils"
- )
- // This test needs to run in serial because other tests could interfere
- // with metrics being tested here.
- var _ = utils.SIGDescribe("[Serial] Volume metrics", func() {
- var (
- c clientset.Interface
- ns string
- pvc *v1.PersistentVolumeClaim
- metricsGrabber *metrics.Grabber
- invalidSc *storagev1.StorageClass
- defaultScName string
- )
- f := framework.NewDefaultFramework("pv")
- ginkgo.BeforeEach(func() {
- c = f.ClientSet
- ns = f.Namespace.Name
- var err error
- e2eskipper.SkipUnlessProviderIs("gce", "gke", "aws")
- defaultScName, err = e2epv.GetDefaultStorageClassName(c)
- if err != nil {
- framework.Failf(err.Error())
- }
- test := testsuites.StorageClassTest{
- Name: "default",
- ClaimSize: "2Gi",
- }
- pvc = e2epv.MakePersistentVolumeClaim(e2epv.PersistentVolumeClaimConfig{
- ClaimSize: test.ClaimSize,
- VolumeMode: &test.VolumeMode,
- }, ns)
- metricsGrabber, err = metrics.NewMetricsGrabber(c, nil, true, false, true, false, false)
- if err != nil {
- framework.Failf("Error creating metrics grabber : %v", err)
- }
- })
- ginkgo.AfterEach(func() {
- newPvc, err := c.CoreV1().PersistentVolumeClaims(pvc.Namespace).Get(context.TODO(), pvc.Name, metav1.GetOptions{})
- if err != nil {
- framework.Logf("Failed to get pvc %s/%s: %v", pvc.Namespace, pvc.Name, err)
- } else {
- e2epv.DeletePersistentVolumeClaim(c, newPvc.Name, newPvc.Namespace)
- if newPvc.Spec.VolumeName != "" {
- err = framework.WaitForPersistentVolumeDeleted(c, newPvc.Spec.VolumeName, 5*time.Second, 5*time.Minute)
- framework.ExpectNoError(err, "Persistent Volume %v not deleted by dynamic provisioner", newPvc.Spec.VolumeName)
- }
- }
- if invalidSc != nil {
- err := c.StorageV1().StorageClasses().Delete(context.TODO(), invalidSc.Name, nil)
- framework.ExpectNoError(err, "Error deleting storageclass %v: %v", invalidSc.Name, err)
- invalidSc = nil
- }
- })
- ginkgo.It("should create prometheus metrics for volume provisioning and attach/detach", func() {
- var err error
- if !metricsGrabber.HasRegisteredMaster() {
- e2eskipper.Skipf("Environment does not support getting controller-manager metrics - skipping")
- }
- ginkgo.By("Getting plugin name")
- defaultClass, err := c.StorageV1().StorageClasses().Get(context.TODO(), defaultScName, metav1.GetOptions{})
- framework.ExpectNoError(err, "Error getting default storageclass: %v", err)
- pluginName := defaultClass.Provisioner
- controllerMetrics, err := metricsGrabber.GrabFromControllerManager()
- framework.ExpectNoError(err, "Error getting c-m metrics : %v", err)
- storageOpMetrics := getControllerStorageMetrics(controllerMetrics, pluginName)
- pvc, err = c.CoreV1().PersistentVolumeClaims(pvc.Namespace).Create(context.TODO(), pvc, metav1.CreateOptions{})
- framework.ExpectNoError(err)
- framework.ExpectNotEqual(pvc, nil)
- claims := []*v1.PersistentVolumeClaim{pvc}
- pod := e2epod.MakePod(ns, nil, claims, false, "")
- pod, err = c.CoreV1().Pods(ns).Create(context.TODO(), pod, metav1.CreateOptions{})
- framework.ExpectNoError(err)
- err = e2epod.WaitForPodRunningInNamespace(c, pod)
- framework.ExpectNoError(err, "Error starting pod %s", pod.Name)
- framework.Logf("Deleting pod %q/%q", pod.Namespace, pod.Name)
- framework.ExpectNoError(e2epod.DeletePodWithWait(c, pod))
- updatedStorageMetrics := waitForDetachAndGrabMetrics(storageOpMetrics, metricsGrabber, pluginName)
- framework.ExpectNotEqual(len(updatedStorageMetrics.latencyMetrics), 0, "Error fetching c-m updated storage metrics")
- framework.ExpectNotEqual(len(updatedStorageMetrics.statusMetrics), 0, "Error fetching c-m updated storage metrics")
- volumeOperations := []string{"volume_provision", "volume_detach", "volume_attach"}
- for _, volumeOp := range volumeOperations {
- verifyMetricCount(storageOpMetrics, updatedStorageMetrics, volumeOp, false)
- }
- })
- ginkgo.It("should create prometheus metrics for volume provisioning errors [Slow]", func() {
- var err error
- if !metricsGrabber.HasRegisteredMaster() {
- e2eskipper.Skipf("Environment does not support getting controller-manager metrics - skipping")
- }
- ginkgo.By("Geting default storageclass")
- defaultClass, err := c.StorageV1().StorageClasses().Get(context.TODO(), defaultScName, metav1.GetOptions{})
- framework.ExpectNoError(err, "Error getting default storageclass: %v", err)
- pluginName := defaultClass.Provisioner
- controllerMetrics, err := metricsGrabber.GrabFromControllerManager()
- framework.ExpectNoError(err, "Error getting c-m metrics : %v", err)
- storageOpMetrics := getControllerStorageMetrics(controllerMetrics, pluginName)
- invalidSc = &storagev1.StorageClass{
- ObjectMeta: metav1.ObjectMeta{
- Name: fmt.Sprintf("fail-metrics-invalid-sc-%s", pvc.Namespace),
- },
- Provisioner: defaultClass.Provisioner,
- Parameters: map[string]string{
- "invalidparam": "invalidvalue",
- },
- }
- _, err = c.StorageV1().StorageClasses().Create(context.TODO(), invalidSc, metav1.CreateOptions{})
- framework.ExpectNoError(err, "Error creating new storageclass: %v", err)
- pvc.Spec.StorageClassName = &invalidSc.Name
- pvc, err = c.CoreV1().PersistentVolumeClaims(pvc.Namespace).Create(context.TODO(), pvc, metav1.CreateOptions{})
- framework.ExpectNoError(err, "failed to create PVC %s/%s", pvc.Namespace, pvc.Name)
- framework.ExpectNotEqual(pvc, nil)
- claims := []*v1.PersistentVolumeClaim{pvc}
- ginkgo.By("Creating a pod and expecting it to fail")
- pod := e2epod.MakePod(ns, nil, claims, false, "")
- pod, err = c.CoreV1().Pods(ns).Create(context.TODO(), pod, metav1.CreateOptions{})
- framework.ExpectNoError(err, "failed to create Pod %s/%s", pod.Namespace, pod.Name)
- err = e2epod.WaitTimeoutForPodRunningInNamespace(c, pod.Name, pod.Namespace, framework.PodStartShortTimeout)
- framework.ExpectError(err)
- framework.Logf("Deleting pod %q/%q", pod.Namespace, pod.Name)
- framework.ExpectNoError(e2epod.DeletePodWithWait(c, pod))
- ginkgo.By("Checking failure metrics")
- updatedControllerMetrics, err := metricsGrabber.GrabFromControllerManager()
- framework.ExpectNoError(err, "failed to get controller manager metrics")
- updatedStorageMetrics := getControllerStorageMetrics(updatedControllerMetrics, pluginName)
- framework.ExpectNotEqual(len(updatedStorageMetrics.statusMetrics), 0, "Error fetching c-m updated storage metrics")
- verifyMetricCount(storageOpMetrics, updatedStorageMetrics, "volume_provision", true)
- })
- ginkgo.It("should create volume metrics with the correct PVC ref", func() {
- var err error
- pvc, err = c.CoreV1().PersistentVolumeClaims(pvc.Namespace).Create(context.TODO(), pvc, metav1.CreateOptions{})
- framework.ExpectNoError(err)
- framework.ExpectNotEqual(pvc, nil)
- claims := []*v1.PersistentVolumeClaim{pvc}
- pod := e2epod.MakePod(ns, nil, claims, false, "")
- pod, err = c.CoreV1().Pods(ns).Create(context.TODO(), pod, metav1.CreateOptions{})
- framework.ExpectNoError(err)
- err = e2epod.WaitForPodRunningInNamespace(c, pod)
- framework.ExpectNoError(err, "Error starting pod ", pod.Name)
- pod, err = c.CoreV1().Pods(ns).Get(context.TODO(), pod.Name, metav1.GetOptions{})
- framework.ExpectNoError(err)
- // Verify volume stat metrics were collected for the referenced PVC
- volumeStatKeys := []string{
- kubeletmetrics.VolumeStatsUsedBytesKey,
- kubeletmetrics.VolumeStatsCapacityBytesKey,
- kubeletmetrics.VolumeStatsAvailableBytesKey,
- kubeletmetrics.VolumeStatsUsedBytesKey,
- kubeletmetrics.VolumeStatsInodesFreeKey,
- kubeletmetrics.VolumeStatsInodesUsedKey,
- }
- key := volumeStatKeys[0]
- kubeletKeyName := fmt.Sprintf("%s_%s", kubeletmetrics.KubeletSubsystem, key)
- // Poll kubelet metrics waiting for the volume to be picked up
- // by the volume stats collector
- var kubeMetrics metrics.KubeletMetrics
- waitErr := wait.Poll(30*time.Second, 5*time.Minute, func() (bool, error) {
- framework.Logf("Grabbing Kubelet metrics")
- // Grab kubelet metrics from the node the pod was scheduled on
- var err error
- kubeMetrics, err = metricsGrabber.GrabFromKubelet(pod.Spec.NodeName)
- if err != nil {
- framework.Logf("Error fetching kubelet metrics")
- return false, err
- }
- if !findVolumeStatMetric(kubeletKeyName, pvc.Namespace, pvc.Name, kubeMetrics) {
- return false, nil
- }
- return true, nil
- })
- framework.ExpectNoError(waitErr, "Unable to find metric %s for PVC %s/%s", kubeletKeyName, pvc.Namespace, pvc.Name)
- for _, key := range volumeStatKeys {
- kubeletKeyName := fmt.Sprintf("%s_%s", kubeletmetrics.KubeletSubsystem, key)
- found := findVolumeStatMetric(kubeletKeyName, pvc.Namespace, pvc.Name, kubeMetrics)
- framework.ExpectEqual(found, true, "PVC %s, Namespace %s not found for %s", pvc.Name, pvc.Namespace, kubeletKeyName)
- }
- framework.Logf("Deleting pod %q/%q", pod.Namespace, pod.Name)
- framework.ExpectNoError(e2epod.DeletePodWithWait(c, pod))
- })
- ginkgo.It("should create metrics for total time taken in volume operations in P/V Controller", func() {
- var err error
- pvc, err = c.CoreV1().PersistentVolumeClaims(pvc.Namespace).Create(context.TODO(), pvc, metav1.CreateOptions{})
- framework.ExpectNoError(err)
- framework.ExpectNotEqual(pvc, nil)
- claims := []*v1.PersistentVolumeClaim{pvc}
- pod := e2epod.MakePod(ns, nil, claims, false, "")
- pod, err = c.CoreV1().Pods(ns).Create(context.TODO(), pod, metav1.CreateOptions{})
- framework.ExpectNoError(err)
- err = e2epod.WaitForPodRunningInNamespace(c, pod)
- framework.ExpectNoError(err, "Error starting pod ", pod.Name)
- pod, err = c.CoreV1().Pods(ns).Get(context.TODO(), pod.Name, metav1.GetOptions{})
- framework.ExpectNoError(err)
- controllerMetrics, err := metricsGrabber.GrabFromControllerManager()
- if err != nil {
- e2eskipper.Skipf("Could not get controller-manager metrics - skipping")
- }
- metricKey := "volume_operation_total_seconds_count"
- dimensions := []string{"operation_name", "plugin_name"}
- err = testutil.ValidateMetrics(testutil.Metrics(controllerMetrics), metricKey, dimensions...)
- framework.ExpectNoError(err, "Invalid metric in P/V Controller metrics: %q", metricKey)
- framework.Logf("Deleting pod %q/%q", pod.Namespace, pod.Name)
- framework.ExpectNoError(e2epod.DeletePodWithWait(c, pod))
- })
- ginkgo.It("should create volume metrics in Volume Manager", func() {
- var err error
- pvc, err = c.CoreV1().PersistentVolumeClaims(pvc.Namespace).Create(context.TODO(), pvc, metav1.CreateOptions{})
- framework.ExpectNoError(err)
- framework.ExpectNotEqual(pvc, nil)
- claims := []*v1.PersistentVolumeClaim{pvc}
- pod := e2epod.MakePod(ns, nil, claims, false, "")
- pod, err = c.CoreV1().Pods(ns).Create(context.TODO(), pod, metav1.CreateOptions{})
- framework.ExpectNoError(err)
- err = e2epod.WaitForPodRunningInNamespace(c, pod)
- framework.ExpectNoError(err, "Error starting pod ", pod.Name)
- pod, err = c.CoreV1().Pods(ns).Get(context.TODO(), pod.Name, metav1.GetOptions{})
- framework.ExpectNoError(err)
- kubeMetrics, err := metricsGrabber.GrabFromKubelet(pod.Spec.NodeName)
- framework.ExpectNoError(err)
- // Metrics should have dimensions plugin_name and state available
- totalVolumesKey := "volume_manager_total_volumes"
- dimensions := []string{"state", "plugin_name"}
- err = testutil.ValidateMetrics(testutil.Metrics(kubeMetrics), totalVolumesKey, dimensions...)
- framework.ExpectNoError(err, "Invalid metric in Volume Manager metrics: %q", totalVolumesKey)
- framework.Logf("Deleting pod %q/%q", pod.Namespace, pod.Name)
- framework.ExpectNoError(e2epod.DeletePodWithWait(c, pod))
- })
- ginkgo.It("should create metrics for total number of volumes in A/D Controller", func() {
- var err error
- pvc, err = c.CoreV1().PersistentVolumeClaims(pvc.Namespace).Create(context.TODO(), pvc, metav1.CreateOptions{})
- framework.ExpectNoError(err)
- framework.ExpectNotEqual(pvc, nil)
- claims := []*v1.PersistentVolumeClaim{pvc}
- pod := e2epod.MakePod(ns, nil, claims, false, "")
- // Get metrics
- controllerMetrics, err := metricsGrabber.GrabFromControllerManager()
- if err != nil {
- e2eskipper.Skipf("Could not get controller-manager metrics - skipping")
- }
- // Create pod
- pod, err = c.CoreV1().Pods(ns).Create(context.TODO(), pod, metav1.CreateOptions{})
- framework.ExpectNoError(err)
- err = e2epod.WaitForPodRunningInNamespace(c, pod)
- framework.ExpectNoError(err, "Error starting pod ", pod.Name)
- pod, err = c.CoreV1().Pods(ns).Get(context.TODO(), pod.Name, metav1.GetOptions{})
- framework.ExpectNoError(err)
- // Get updated metrics
- updatedControllerMetrics, err := metricsGrabber.GrabFromControllerManager()
- if err != nil {
- e2eskipper.Skipf("Could not get controller-manager metrics - skipping")
- }
- // Forced detach metric should be present
- forceDetachKey := "attachdetach_controller_forced_detaches"
- _, ok := updatedControllerMetrics[forceDetachKey]
- framework.ExpectEqual(ok, true, "Key %q not found in A/D Controller metrics", forceDetachKey)
- // Wait and validate
- totalVolumesKey := "attachdetach_controller_total_volumes"
- states := []string{"actual_state_of_world", "desired_state_of_world"}
- dimensions := []string{"state", "plugin_name"}
- waitForADControllerStatesMetrics(metricsGrabber, totalVolumesKey, dimensions, states)
- // Total number of volumes in both ActualStateofWorld and DesiredStateOfWorld
- // states should be higher or equal than it used to be
- oldStates := getStatesMetrics(totalVolumesKey, testutil.Metrics(controllerMetrics))
- updatedStates := getStatesMetrics(totalVolumesKey, testutil.Metrics(updatedControllerMetrics))
- for _, stateName := range states {
- if _, ok := oldStates[stateName]; !ok {
- continue
- }
- for pluginName, numVolumes := range updatedStates[stateName] {
- oldNumVolumes := oldStates[stateName][pluginName]
- gomega.Expect(numVolumes).To(gomega.BeNumerically(">=", oldNumVolumes),
- "Wrong number of volumes in state %q, plugin %q: wanted >=%d, got %d",
- stateName, pluginName, oldNumVolumes, numVolumes)
- }
- }
- framework.Logf("Deleting pod %q/%q", pod.Namespace, pod.Name)
- framework.ExpectNoError(e2epod.DeletePodWithWait(c, pod))
- })
- // Test for pv controller metrics, concretely: bound/unbound pv/pvc count.
- ginkgo.Describe("PVController", func() {
- const (
- classKey = "storage_class"
- namespaceKey = "namespace"
- boundPVKey = "pv_collector_bound_pv_count"
- unboundPVKey = "pv_collector_unbound_pv_count"
- boundPVCKey = "pv_collector_bound_pvc_count"
- unboundPVCKey = "pv_collector_unbound_pvc_count"
- )
- var (
- pv *v1.PersistentVolume
- pvc *v1.PersistentVolumeClaim
- className = "bound-unbound-count-test-sc"
- pvConfig = e2epv.PersistentVolumeConfig{
- PVSource: v1.PersistentVolumeSource{
- HostPath: &v1.HostPathVolumeSource{Path: "/data"},
- },
- NamePrefix: "pv-test-",
- StorageClassName: className,
- }
- pvcConfig = e2epv.PersistentVolumeClaimConfig{StorageClassName: &className}
- metrics = []struct {
- name string
- dimension string
- }{
- {boundPVKey, classKey},
- {unboundPVKey, classKey},
- {boundPVCKey, namespaceKey},
- {unboundPVCKey, namespaceKey},
- }
- // Original metric values before we create any PV/PVCs. The length should be 4,
- // and the elements should be bound pv count, unbound pv count, bound pvc count,
- // unbound pvc count in turn.
- // We use these values to calculate relative increment of each test.
- originMetricValues []map[string]int64
- )
- // validator used to validate each metric's values, the length of metricValues
- // should be 4, and the elements should be bound pv count, unbound pv count, bound
- // pvc count, unbound pvc count in turn.
- validator := func(metricValues []map[string]int64) {
- framework.ExpectEqual(len(metricValues), 4, "Wrong metric size: %d", len(metricValues))
- controllerMetrics, err := metricsGrabber.GrabFromControllerManager()
- framework.ExpectNoError(err, "Error getting c-m metricValues: %v", err)
- for i, metric := range metrics {
- expectValues := metricValues[i]
- if expectValues == nil {
- expectValues = make(map[string]int64)
- }
- // We using relative increment value instead of absolute value to reduce unexpected flakes.
- // Concretely, we expect the difference of the updated values and original values for each
- // test suit are equal to expectValues.
- actualValues := calculateRelativeValues(originMetricValues[i],
- testutil.GetMetricValuesForLabel(testutil.Metrics(controllerMetrics), metric.name, metric.dimension))
- framework.ExpectEqual(actualValues, expectValues, "Wrong pv controller metric %s(%s): wanted %v, got %v",
- metric.name, metric.dimension, expectValues, actualValues)
- }
- }
- ginkgo.BeforeEach(func() {
- if !metricsGrabber.HasRegisteredMaster() {
- e2eskipper.Skipf("Environment does not support getting controller-manager metrics - skipping")
- }
- pv = e2epv.MakePersistentVolume(pvConfig)
- pvc = e2epv.MakePersistentVolumeClaim(pvcConfig, ns)
- // Initializes all original metric values.
- controllerMetrics, err := metricsGrabber.GrabFromControllerManager()
- framework.ExpectNoError(err, "Error getting c-m metricValues: %v", err)
- for _, metric := range metrics {
- originMetricValues = append(originMetricValues,
- testutil.GetMetricValuesForLabel(testutil.Metrics(controllerMetrics), metric.name, metric.dimension))
- }
- })
- ginkgo.AfterEach(func() {
- if err := e2epv.DeletePersistentVolume(c, pv.Name); err != nil {
- framework.Failf("Error deleting pv: %v", err)
- }
- if err := e2epv.DeletePersistentVolumeClaim(c, pvc.Name, pvc.Namespace); err != nil {
- framework.Failf("Error deleting pvc: %v", err)
- }
- // Clear original metric values.
- originMetricValues = nil
- })
- ginkgo.It("should create none metrics for pvc controller before creating any PV or PVC", func() {
- validator([]map[string]int64{nil, nil, nil, nil})
- })
- ginkgo.It("should create unbound pv count metrics for pvc controller after creating pv only",
- func() {
- var err error
- pv, err = e2epv.CreatePV(c, pv)
- framework.ExpectNoError(err, "Error creating pv: %v", err)
- waitForPVControllerSync(metricsGrabber, unboundPVKey, classKey)
- validator([]map[string]int64{nil, {className: 1}, nil, nil})
- })
- ginkgo.It("should create unbound pvc count metrics for pvc controller after creating pvc only",
- func() {
- var err error
- pvc, err = e2epv.CreatePVC(c, ns, pvc)
- framework.ExpectNoError(err, "Error creating pvc: %v", err)
- waitForPVControllerSync(metricsGrabber, unboundPVCKey, namespaceKey)
- validator([]map[string]int64{nil, nil, nil, {ns: 1}})
- })
- ginkgo.It("should create bound pv/pvc count metrics for pvc controller after creating both pv and pvc",
- func() {
- var err error
- pv, pvc, err = e2epv.CreatePVPVC(c, pvConfig, pvcConfig, ns, true)
- framework.ExpectNoError(err, "Error creating pv pvc: %v", err)
- waitForPVControllerSync(metricsGrabber, boundPVKey, classKey)
- waitForPVControllerSync(metricsGrabber, boundPVCKey, namespaceKey)
- validator([]map[string]int64{{className: 1}, nil, {ns: 1}, nil})
- })
- })
- })
- type storageControllerMetrics struct {
- latencyMetrics map[string]int64
- statusMetrics map[string]statusMetricCounts
- }
- type statusMetricCounts struct {
- successCount int64
- failCount int64
- otherCount int64
- }
- func newStorageControllerMetrics() *storageControllerMetrics {
- return &storageControllerMetrics{
- latencyMetrics: make(map[string]int64),
- statusMetrics: make(map[string]statusMetricCounts),
- }
- }
- func waitForDetachAndGrabMetrics(oldMetrics *storageControllerMetrics, metricsGrabber *metrics.Grabber, pluginName string) *storageControllerMetrics {
- backoff := wait.Backoff{
- Duration: 10 * time.Second,
- Factor: 1.2,
- Steps: 21,
- }
- updatedStorageMetrics := newStorageControllerMetrics()
- oldDetachCount, ok := oldMetrics.latencyMetrics["volume_detach"]
- if !ok {
- oldDetachCount = 0
- }
- verifyMetricFunc := func() (bool, error) {
- updatedMetrics, err := metricsGrabber.GrabFromControllerManager()
- if err != nil {
- framework.Logf("Error fetching controller-manager metrics")
- return false, err
- }
- updatedStorageMetrics = getControllerStorageMetrics(updatedMetrics, pluginName)
- newDetachCount, ok := updatedStorageMetrics.latencyMetrics["volume_detach"]
- // if detach metrics are not yet there, we need to retry
- if !ok {
- return false, nil
- }
- // if old Detach count is more or equal to new detach count, that means detach
- // event has not been observed yet.
- if oldDetachCount >= newDetachCount {
- return false, nil
- }
- return true, nil
- }
- waitErr := wait.ExponentialBackoff(backoff, verifyMetricFunc)
- framework.ExpectNoError(waitErr, "Unable to get updated metrics for plugin %s", pluginName)
- return updatedStorageMetrics
- }
- func verifyMetricCount(oldMetrics, newMetrics *storageControllerMetrics, metricName string, expectFailure bool) {
- oldLatencyCount, ok := oldMetrics.latencyMetrics[metricName]
- // if metric does not exist in oldMap, it probably hasn't been emitted yet.
- if !ok {
- oldLatencyCount = 0
- }
- oldStatusCount := int64(0)
- if oldStatusCounts, ok := oldMetrics.statusMetrics[metricName]; ok {
- if expectFailure {
- oldStatusCount = oldStatusCounts.failCount
- } else {
- oldStatusCount = oldStatusCounts.successCount
- }
- }
- newLatencyCount, ok := newMetrics.latencyMetrics[metricName]
- if !expectFailure {
- framework.ExpectEqual(ok, true, "Error getting updated latency metrics for %s", metricName)
- }
- newStatusCounts, ok := newMetrics.statusMetrics[metricName]
- framework.ExpectEqual(ok, true, "Error getting updated status metrics for %s", metricName)
- newStatusCount := int64(0)
- if expectFailure {
- newStatusCount = newStatusCounts.failCount
- } else {
- newStatusCount = newStatusCounts.successCount
- }
- // It appears that in a busy cluster some spurious detaches are unavoidable
- // even if the test is run serially. We really just verify if new count
- // is greater than old count
- if !expectFailure {
- gomega.Expect(newLatencyCount).To(gomega.BeNumerically(">", oldLatencyCount), "New latency count %d should be more than old count %d for action %s", newLatencyCount, oldLatencyCount, metricName)
- }
- gomega.Expect(newStatusCount).To(gomega.BeNumerically(">", oldStatusCount), "New status count %d should be more than old count %d for action %s", newStatusCount, oldStatusCount, metricName)
- }
- func getControllerStorageMetrics(ms metrics.ControllerManagerMetrics, pluginName string) *storageControllerMetrics {
- result := newStorageControllerMetrics()
- for method, samples := range ms {
- switch method {
- case "storage_operation_duration_seconds_count":
- for _, sample := range samples {
- count := int64(sample.Value)
- operation := string(sample.Metric["operation_name"])
- metricPluginName := string(sample.Metric["volume_plugin"])
- if len(pluginName) > 0 && pluginName != metricPluginName {
- continue
- }
- result.latencyMetrics[operation] = count
- }
- case "storage_operation_status_count":
- for _, sample := range samples {
- count := int64(sample.Value)
- operation := string(sample.Metric["operation_name"])
- status := string(sample.Metric["status"])
- statusCounts := result.statusMetrics[operation]
- metricPluginName := string(sample.Metric["volume_plugin"])
- if len(pluginName) > 0 && pluginName != metricPluginName {
- continue
- }
- switch status {
- case "success":
- statusCounts.successCount = count
- case "fail-unknown":
- statusCounts.failCount = count
- default:
- statusCounts.otherCount = count
- }
- result.statusMetrics[operation] = statusCounts
- }
- }
- }
- return result
- }
- // Finds the sample in the specified metric from `KubeletMetrics` tagged with
- // the specified namespace and pvc name
- func findVolumeStatMetric(metricKeyName string, namespace string, pvcName string, kubeletMetrics metrics.KubeletMetrics) bool {
- found := false
- errCount := 0
- framework.Logf("Looking for sample in metric `%s` tagged with namespace `%s`, PVC `%s`", metricKeyName, namespace, pvcName)
- if samples, ok := kubeletMetrics[metricKeyName]; ok {
- for _, sample := range samples {
- framework.Logf("Found sample %s", sample.String())
- samplePVC, ok := sample.Metric["persistentvolumeclaim"]
- if !ok {
- framework.Logf("Error getting pvc for metric %s, sample %s", metricKeyName, sample.String())
- errCount++
- }
- sampleNS, ok := sample.Metric["namespace"]
- if !ok {
- framework.Logf("Error getting namespace for metric %s, sample %s", metricKeyName, sample.String())
- errCount++
- }
- if string(samplePVC) == pvcName && string(sampleNS) == namespace {
- found = true
- break
- }
- }
- }
- framework.ExpectEqual(errCount, 0, "Found invalid samples")
- return found
- }
- // Wait for the count of a pv controller's metric specified by metricName and dimension bigger than zero.
- func waitForPVControllerSync(metricsGrabber *metrics.Grabber, metricName, dimension string) {
- backoff := wait.Backoff{
- Duration: 10 * time.Second,
- Factor: 1.2,
- Steps: 21,
- }
- verifyMetricFunc := func() (bool, error) {
- updatedMetrics, err := metricsGrabber.GrabFromControllerManager()
- if err != nil {
- framework.Logf("Error fetching controller-manager metrics")
- return false, err
- }
- return len(testutil.GetMetricValuesForLabel(testutil.Metrics(updatedMetrics), metricName, dimension)) > 0, nil
- }
- waitErr := wait.ExponentialBackoff(backoff, verifyMetricFunc)
- framework.ExpectNoError(waitErr, "Unable to get pv controller metrics")
- }
- func calculateRelativeValues(originValues, updatedValues map[string]int64) map[string]int64 {
- relativeValues := make(map[string]int64)
- for key, value := range updatedValues {
- relativeValue := value - originValues[key]
- if relativeValue != 0 {
- relativeValues[key] = relativeValue
- }
- }
- for key, value := range originValues {
- if _, exist := updatedValues[key]; !exist && value > 0 {
- relativeValues[key] = -value
- }
- }
- return relativeValues
- }
- func getStatesMetrics(metricKey string, givenMetrics testutil.Metrics) map[string]map[string]int64 {
- states := make(map[string]map[string]int64)
- for _, sample := range givenMetrics[metricKey] {
- framework.Logf("Found sample %q", sample.String())
- state := string(sample.Metric["state"])
- pluginName := string(sample.Metric["plugin_name"])
- states[state] = map[string]int64{pluginName: int64(sample.Value)}
- }
- return states
- }
- func waitForADControllerStatesMetrics(metricsGrabber *metrics.Grabber, metricName string, dimensions []string, stateNames []string) {
- backoff := wait.Backoff{
- Duration: 10 * time.Second,
- Factor: 1.2,
- Steps: 21,
- }
- verifyMetricFunc := func() (bool, error) {
- updatedMetrics, err := metricsGrabber.GrabFromControllerManager()
- if err != nil {
- e2eskipper.Skipf("Could not get controller-manager metrics - skipping")
- return false, err
- }
- err = testutil.ValidateMetrics(testutil.Metrics(updatedMetrics), metricName, dimensions...)
- if err != nil {
- return false, fmt.Errorf("could not get valid metrics: %v ", err)
- }
- states := getStatesMetrics(metricName, testutil.Metrics(updatedMetrics))
- for _, name := range stateNames {
- if _, ok := states[name]; !ok {
- return false, fmt.Errorf("could not get state %q from A/D Controller metrics", name)
- }
- }
- return true, nil
- }
- waitErr := wait.ExponentialBackoff(backoff, verifyMetricFunc)
- framework.ExpectNoError(waitErr, "Unable to get A/D controller metrics")
- }
|