123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278 |
- /*
- Copyright 2018 The Kubernetes Authors.
- Licensed under the Apache License, Version 2.0 (the "License");
- you may not use this file except in compliance with the License.
- You may obtain a copy of the License at
- http://www.apache.org/licenses/LICENSE-2.0
- Unless required by applicable law or agreed to in writing, software
- distributed under the License is distributed on an "AS IS" BASIS,
- WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- See the License for the specific language governing permissions and
- limitations under the License.
- */
- package windows
- import (
- "context"
- "fmt"
- "sort"
- "sync"
- "time"
- v1 "k8s.io/api/core/v1"
- metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
- "k8s.io/apimachinery/pkg/labels"
- "k8s.io/apimachinery/pkg/runtime"
- "k8s.io/apimachinery/pkg/util/uuid"
- "k8s.io/apimachinery/pkg/watch"
- "k8s.io/client-go/tools/cache"
- "k8s.io/kubernetes/test/e2e/framework"
- e2emetrics "k8s.io/kubernetes/test/e2e/framework/metrics"
- e2epod "k8s.io/kubernetes/test/e2e/framework/pod"
- imageutils "k8s.io/kubernetes/test/utils/image"
- "github.com/onsi/ginkgo"
- "github.com/onsi/gomega"
- )
- var _ = SIGDescribe("[Feature:Windows] Density [Serial] [Slow]", func() {
- f := framework.NewDefaultFramework("density-test-windows")
- ginkgo.Context("create a batch of pods", func() {
- // TODO(coufon): the values are generous, set more precise limits with benchmark data
- // and add more tests
- dTests := []densityTest{
- {
- podsNr: 10,
- interval: 0 * time.Millisecond,
- // percentile limit of single pod startup latency
- podStartupLimits: e2emetrics.LatencyMetric{
- Perc50: 30 * time.Second,
- Perc90: 54 * time.Second,
- Perc99: 59 * time.Second,
- },
- // upbound of startup latency of a batch of pods
- podBatchStartupLimit: 10 * time.Minute,
- },
- }
- for _, testArg := range dTests {
- itArg := testArg
- desc := fmt.Sprintf("latency/resource should be within limit when create %d pods with %v interval", itArg.podsNr, itArg.interval)
- ginkgo.It(desc, func() {
- itArg.createMethod = "batch"
- runDensityBatchTest(f, itArg)
- })
- }
- })
- })
- type densityTest struct {
- // number of pods
- podsNr int
- // interval between creating pod (rate control)
- interval time.Duration
- // create pods in 'batch' or 'sequence'
- createMethod string
- // API QPS limit
- APIQPSLimit int
- // performance limits
- podStartupLimits e2emetrics.LatencyMetric
- podBatchStartupLimit time.Duration
- }
- // runDensityBatchTest runs the density batch pod creation test
- func runDensityBatchTest(f *framework.Framework, testArg densityTest) (time.Duration, []e2emetrics.PodLatencyData) {
- const (
- podType = "density_test_pod"
- )
- var (
- mutex = &sync.Mutex{}
- watchTimes = make(map[string]metav1.Time)
- stopCh = make(chan struct{})
- )
- // create test pod data structure
- pods := newDensityTestPods(testArg.podsNr, false, imageutils.GetPauseImageName(), podType)
- // the controller watches the change of pod status
- controller := newInformerWatchPod(f, mutex, watchTimes, podType)
- go controller.Run(stopCh)
- defer close(stopCh)
- ginkgo.By("Creating a batch of pods")
- // It returns a map['pod name']'creation time' containing the creation timestamps
- createTimes := createBatchPodWithRateControl(f, pods, testArg.interval)
- ginkgo.By("Waiting for all Pods to be observed by the watch...")
- gomega.Eventually(func() bool {
- return len(watchTimes) == testArg.podsNr
- }, 10*time.Minute, 10*time.Second).Should(gomega.BeTrue())
- if len(watchTimes) < testArg.podsNr {
- framework.Failf("Timeout reached waiting for all Pods to be observed by the watch.")
- }
- // Analyze results
- var (
- firstCreate metav1.Time
- lastRunning metav1.Time
- init = true
- e2eLags = make([]e2emetrics.PodLatencyData, 0)
- )
- for name, create := range createTimes {
- watch, ok := watchTimes[name]
- framework.ExpectEqual(ok, true)
- e2eLags = append(e2eLags,
- e2emetrics.PodLatencyData{Name: name, Latency: watch.Time.Sub(create.Time)})
- if !init {
- if firstCreate.Time.After(create.Time) {
- firstCreate = create
- }
- if lastRunning.Time.Before(watch.Time) {
- lastRunning = watch
- }
- } else {
- init = false
- firstCreate, lastRunning = create, watch
- }
- }
- sort.Sort(e2emetrics.LatencySlice(e2eLags))
- batchLag := lastRunning.Time.Sub(firstCreate.Time)
- deletePodsSync(f, pods)
- return batchLag, e2eLags
- }
- // createBatchPodWithRateControl creates a batch of pods concurrently, uses one goroutine for each creation.
- // between creations there is an interval for throughput control
- func createBatchPodWithRateControl(f *framework.Framework, pods []*v1.Pod, interval time.Duration) map[string]metav1.Time {
- createTimes := make(map[string]metav1.Time)
- for _, pod := range pods {
- createTimes[pod.ObjectMeta.Name] = metav1.Now()
- go f.PodClient().Create(pod)
- time.Sleep(interval)
- }
- return createTimes
- }
- // newInformerWatchPod creates an informer to check whether all pods are running.
- func newInformerWatchPod(f *framework.Framework, mutex *sync.Mutex, watchTimes map[string]metav1.Time, podType string) cache.Controller {
- ns := f.Namespace.Name
- checkPodRunning := func(p *v1.Pod) {
- mutex.Lock()
- defer mutex.Unlock()
- defer ginkgo.GinkgoRecover()
- if p.Status.Phase == v1.PodRunning {
- if _, found := watchTimes[p.Name]; !found {
- watchTimes[p.Name] = metav1.Now()
- }
- }
- }
- _, controller := cache.NewInformer(
- &cache.ListWatch{
- ListFunc: func(options metav1.ListOptions) (runtime.Object, error) {
- options.LabelSelector = labels.SelectorFromSet(labels.Set{"type": podType}).String()
- obj, err := f.ClientSet.CoreV1().Pods(ns).List(context.TODO(), options)
- return runtime.Object(obj), err
- },
- WatchFunc: func(options metav1.ListOptions) (watch.Interface, error) {
- options.LabelSelector = labels.SelectorFromSet(labels.Set{"type": podType}).String()
- return f.ClientSet.CoreV1().Pods(ns).Watch(context.TODO(), options)
- },
- },
- &v1.Pod{},
- 0,
- cache.ResourceEventHandlerFuncs{
- AddFunc: func(obj interface{}) {
- p, ok := obj.(*v1.Pod)
- framework.ExpectEqual(ok, true)
- go checkPodRunning(p)
- },
- UpdateFunc: func(oldObj, newObj interface{}) {
- p, ok := newObj.(*v1.Pod)
- framework.ExpectEqual(ok, true)
- go checkPodRunning(p)
- },
- },
- )
- return controller
- }
- // newDensityTestPods creates a list of pods (specification) for test.
- func newDensityTestPods(numPods int, volume bool, imageName, podType string) []*v1.Pod {
- var pods []*v1.Pod
- for i := 0; i < numPods; i++ {
- podName := "test-" + string(uuid.NewUUID())
- pod := v1.Pod{
- ObjectMeta: metav1.ObjectMeta{
- Name: podName,
- Labels: map[string]string{
- "type": podType,
- "name": podName,
- },
- },
- Spec: v1.PodSpec{
- // Restart policy is always (default).
- Containers: []v1.Container{
- {
- Image: imageName,
- Name: podName,
- },
- },
- NodeSelector: map[string]string{
- "kubernetes.io/os": "windows",
- },
- },
- }
- if volume {
- pod.Spec.Containers[0].VolumeMounts = []v1.VolumeMount{
- {MountPath: "/test-volume-mnt", Name: podName + "-volume"},
- }
- pod.Spec.Volumes = []v1.Volume{
- {Name: podName + "-volume", VolumeSource: v1.VolumeSource{EmptyDir: &v1.EmptyDirVolumeSource{}}},
- }
- }
- pods = append(pods, &pod)
- }
- return pods
- }
- // deletePodsSync deletes a list of pods and block until pods disappear.
- func deletePodsSync(f *framework.Framework, pods []*v1.Pod) {
- var wg sync.WaitGroup
- for _, pod := range pods {
- wg.Add(1)
- go func(pod *v1.Pod) {
- defer ginkgo.GinkgoRecover()
- defer wg.Done()
- err := f.PodClient().Delete(context.TODO(), pod.ObjectMeta.Name, metav1.NewDeleteOptions(30))
- framework.ExpectNoError(err)
- err = e2epod.WaitForPodToDisappear(f.ClientSet, f.Namespace.Name, pod.ObjectMeta.Name, labels.Everything(),
- 30*time.Second, 10*time.Minute)
- framework.ExpectNoError(err)
- }(pod)
- }
- wg.Wait()
- }
|