123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702703704705706707708709710711712713714715716717718719720721722723724725726727728729730731732733734735736737738739740741742743744745746747748749750751752753754755756757758759760761762763764765766767768769770771772773774775776777778779780781782783784785786787788789790791792793794795796797798799800801802803804805806807808809810811812813814815816817818819820821822823824825826827828829830831832833834835836837838839840841842843844845846847848849850851852853854855856857858859860861862863864865866867868869870871872873874875876877878879880881882883884885886887888889890891892893894895896897898899900901902903904905906907908909910911912913914915916917918919920921922923924925926927928929930931932933934935936937938939940941942943944945946947948949950951952953954955956957958959960961962963964965966967968969970971972973974975976977978979980981982983984985986987988989990991992993994995996997998999100010011002100310041005100610071008100910101011101210131014101510161017101810191020102110221023102410251026102710281029103010311032103310341035103610371038103910401041104210431044104510461047104810491050105110521053105410551056105710581059106010611062106310641065106610671068106910701071107210731074107510761077107810791080108110821083108410851086108710881089109010911092109310941095109610971098109911001101110211031104110511061107110811091110111111121113111411151116111711181119112011211122112311241125112611271128112911301131113211331134113511361137113811391140114111421143114411451146114711481149115011511152115311541155115611571158115911601161116211631164116511661167116811691170117111721173117411751176117711781179118011811182118311841185118611871188118911901191119211931194119511961197119811991200120112021203120412051206120712081209121012111212121312141215121612171218121912201221122212231224122512261227122812291230123112321233123412351236123712381239124012411242124312441245124612471248124912501251125212531254125512561257125812591260126112621263126412651266126712681269127012711272127312741275127612771278127912801281128212831284128512861287128812891290129112921293129412951296129712981299130013011302130313041305130613071308130913101311131213131314131513161317131813191320132113221323132413251326132713281329133013311332133313341335133613371338133913401341134213431344134513461347134813491350135113521353135413551356135713581359136013611362136313641365136613671368136913701371137213731374137513761377137813791380138113821383138413851386138713881389139013911392139313941395139613971398139914001401140214031404140514061407140814091410141114121413141414151416141714181419142014211422142314241425142614271428142914301431143214331434143514361437143814391440144114421443144414451446144714481449145014511452145314541455145614571458145914601461146214631464146514661467146814691470147114721473147414751476147714781479148014811482148314841485148614871488148914901491149214931494149514961497149814991500150115021503150415051506150715081509151015111512151315141515151615171518151915201521152215231524152515261527152815291530153115321533153415351536153715381539154015411542154315441545154615471548154915501551155215531554155515561557155815591560156115621563156415651566156715681569157015711572157315741575157615771578157915801581158215831584158515861587158815891590159115921593159415951596159715981599160016011602160316041605160616071608160916101611 |
- /*
- Copyright 2016 The Kubernetes Authors.
- Licensed under the Apache License, Version 2.0 (the "License");
- you may not use this file except in compliance with the License.
- You may obtain a copy of the License at
- http://www.apache.org/licenses/LICENSE-2.0
- Unless required by applicable law or agreed to in writing, software
- distributed under the License is distributed on an "AS IS" BASIS,
- WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- See the License for the specific language governing permissions and
- limitations under the License.
- */
- package replicaset
- import (
- "errors"
- "fmt"
- "math/rand"
- "net/http/httptest"
- "net/url"
- "reflect"
- "strings"
- "sync"
- "testing"
- "time"
- apps "k8s.io/api/apps/v1"
- "k8s.io/api/core/v1"
- apiequality "k8s.io/apimachinery/pkg/api/equality"
- metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
- "k8s.io/apimachinery/pkg/runtime"
- "k8s.io/apimachinery/pkg/runtime/schema"
- "k8s.io/apimachinery/pkg/util/sets"
- "k8s.io/apimachinery/pkg/util/uuid"
- "k8s.io/apimachinery/pkg/util/wait"
- "k8s.io/apimachinery/pkg/watch"
- "k8s.io/client-go/informers"
- clientset "k8s.io/client-go/kubernetes"
- "k8s.io/client-go/kubernetes/fake"
- restclient "k8s.io/client-go/rest"
- core "k8s.io/client-go/testing"
- "k8s.io/client-go/tools/cache"
- utiltesting "k8s.io/client-go/util/testing"
- "k8s.io/client-go/util/workqueue"
- "k8s.io/kubernetes/pkg/controller"
- . "k8s.io/kubernetes/pkg/controller/testutil"
- "k8s.io/kubernetes/pkg/securitycontext"
- )
- func testNewReplicaSetControllerFromClient(client clientset.Interface, stopCh chan struct{}, burstReplicas int) (*ReplicaSetController, informers.SharedInformerFactory) {
- informers := informers.NewSharedInformerFactory(client, controller.NoResyncPeriodFunc())
- ret := NewReplicaSetController(
- informers.Apps().V1().ReplicaSets(),
- informers.Core().V1().Pods(),
- client,
- burstReplicas,
- )
- ret.podListerSynced = alwaysReady
- ret.rsListerSynced = alwaysReady
- return ret, informers
- }
- func skipListerFunc(verb string, url url.URL) bool {
- if verb != "GET" {
- return false
- }
- if strings.HasSuffix(url.Path, "/pods") || strings.Contains(url.Path, "/replicasets") {
- return true
- }
- return false
- }
- var alwaysReady = func() bool { return true }
- func newReplicaSet(replicas int, selectorMap map[string]string) *apps.ReplicaSet {
- rs := &apps.ReplicaSet{
- TypeMeta: metav1.TypeMeta{APIVersion: "v1", Kind: "ReplicaSet"},
- ObjectMeta: metav1.ObjectMeta{
- UID: uuid.NewUUID(),
- Name: "foobar",
- Namespace: metav1.NamespaceDefault,
- ResourceVersion: "18",
- },
- Spec: apps.ReplicaSetSpec{
- Replicas: func() *int32 { i := int32(replicas); return &i }(),
- Selector: &metav1.LabelSelector{MatchLabels: selectorMap},
- Template: v1.PodTemplateSpec{
- ObjectMeta: metav1.ObjectMeta{
- Labels: map[string]string{
- "name": "foo",
- "type": "production",
- },
- },
- Spec: v1.PodSpec{
- Containers: []v1.Container{
- {
- Image: "foo/bar",
- TerminationMessagePath: v1.TerminationMessagePathDefault,
- ImagePullPolicy: v1.PullIfNotPresent,
- SecurityContext: securitycontext.ValidSecurityContextWithContainerDefaults(),
- },
- },
- RestartPolicy: v1.RestartPolicyAlways,
- DNSPolicy: v1.DNSDefault,
- NodeSelector: map[string]string{
- "baz": "blah",
- },
- },
- },
- },
- }
- return rs
- }
- // create a pod with the given phase for the given rs (same selectors and namespace)
- func newPod(name string, rs *apps.ReplicaSet, status v1.PodPhase, lastTransitionTime *metav1.Time, properlyOwned bool) *v1.Pod {
- var conditions []v1.PodCondition
- if status == v1.PodRunning {
- condition := v1.PodCondition{Type: v1.PodReady, Status: v1.ConditionTrue}
- if lastTransitionTime != nil {
- condition.LastTransitionTime = *lastTransitionTime
- }
- conditions = append(conditions, condition)
- }
- var controllerReference metav1.OwnerReference
- if properlyOwned {
- var trueVar = true
- controllerReference = metav1.OwnerReference{UID: rs.UID, APIVersion: "v1beta1", Kind: "ReplicaSet", Name: rs.Name, Controller: &trueVar}
- }
- return &v1.Pod{
- ObjectMeta: metav1.ObjectMeta{
- Name: name,
- Namespace: rs.Namespace,
- Labels: rs.Spec.Selector.MatchLabels,
- OwnerReferences: []metav1.OwnerReference{controllerReference},
- },
- Status: v1.PodStatus{Phase: status, Conditions: conditions},
- }
- }
- // create count pods with the given phase for the given ReplicaSet (same selectors and namespace), and add them to the store.
- func newPodList(store cache.Store, count int, status v1.PodPhase, labelMap map[string]string, rs *apps.ReplicaSet, name string) *v1.PodList {
- pods := []v1.Pod{}
- var trueVar = true
- controllerReference := metav1.OwnerReference{UID: rs.UID, APIVersion: "v1beta1", Kind: "ReplicaSet", Name: rs.Name, Controller: &trueVar}
- for i := 0; i < count; i++ {
- pod := newPod(fmt.Sprintf("%s%d", name, i), rs, status, nil, false)
- pod.ObjectMeta.Labels = labelMap
- pod.OwnerReferences = []metav1.OwnerReference{controllerReference}
- if store != nil {
- store.Add(pod)
- }
- pods = append(pods, *pod)
- }
- return &v1.PodList{
- Items: pods,
- }
- }
- // processSync initiates a sync via processNextWorkItem() to test behavior that
- // depends on both functions (such as re-queueing on sync error).
- func processSync(rsc *ReplicaSetController, key string) error {
- // Save old syncHandler and replace with one that captures the error.
- oldSyncHandler := rsc.syncHandler
- defer func() {
- rsc.syncHandler = oldSyncHandler
- }()
- var syncErr error
- rsc.syncHandler = func(key string) error {
- syncErr = oldSyncHandler(key)
- return syncErr
- }
- rsc.queue.Add(key)
- rsc.processNextWorkItem()
- return syncErr
- }
- func validateSyncReplicaSet(t *testing.T, fakePodControl *controller.FakePodControl, expectedCreates, expectedDeletes, expectedPatches int) {
- if e, a := expectedCreates, len(fakePodControl.Templates); e != a {
- t.Errorf("Unexpected number of creates. Expected %d, saw %d\n", e, a)
- }
- if e, a := expectedDeletes, len(fakePodControl.DeletePodName); e != a {
- t.Errorf("Unexpected number of deletes. Expected %d, saw %d\n", e, a)
- }
- if e, a := expectedPatches, len(fakePodControl.Patches); e != a {
- t.Errorf("Unexpected number of patches. Expected %d, saw %d\n", e, a)
- }
- }
- func TestSyncReplicaSetDoesNothing(t *testing.T) {
- client := clientset.NewForConfigOrDie(&restclient.Config{Host: "", ContentConfig: restclient.ContentConfig{GroupVersion: &schema.GroupVersion{Group: "", Version: "v1"}}})
- fakePodControl := controller.FakePodControl{}
- stopCh := make(chan struct{})
- defer close(stopCh)
- manager, informers := testNewReplicaSetControllerFromClient(client, stopCh, BurstReplicas)
- // 2 running pods, a controller with 2 replicas, sync is a no-op
- labelMap := map[string]string{"foo": "bar"}
- rsSpec := newReplicaSet(2, labelMap)
- informers.Apps().V1().ReplicaSets().Informer().GetIndexer().Add(rsSpec)
- newPodList(informers.Core().V1().Pods().Informer().GetIndexer(), 2, v1.PodRunning, labelMap, rsSpec, "pod")
- manager.podControl = &fakePodControl
- manager.syncReplicaSet(GetKey(rsSpec, t))
- validateSyncReplicaSet(t, &fakePodControl, 0, 0, 0)
- }
- func TestDeleteFinalStateUnknown(t *testing.T) {
- client := clientset.NewForConfigOrDie(&restclient.Config{Host: "", ContentConfig: restclient.ContentConfig{GroupVersion: &schema.GroupVersion{Group: "", Version: "v1"}}})
- fakePodControl := controller.FakePodControl{}
- stopCh := make(chan struct{})
- defer close(stopCh)
- manager, informers := testNewReplicaSetControllerFromClient(client, stopCh, BurstReplicas)
- manager.podControl = &fakePodControl
- received := make(chan string)
- manager.syncHandler = func(key string) error {
- received <- key
- return nil
- }
- // The DeletedFinalStateUnknown object should cause the ReplicaSet manager to insert
- // the controller matching the selectors of the deleted pod into the work queue.
- labelMap := map[string]string{"foo": "bar"}
- rsSpec := newReplicaSet(1, labelMap)
- informers.Apps().V1().ReplicaSets().Informer().GetIndexer().Add(rsSpec)
- pods := newPodList(nil, 1, v1.PodRunning, labelMap, rsSpec, "pod")
- manager.deletePod(cache.DeletedFinalStateUnknown{Key: "foo", Obj: &pods.Items[0]})
- go manager.worker()
- expected := GetKey(rsSpec, t)
- select {
- case key := <-received:
- if key != expected {
- t.Errorf("Unexpected sync all for ReplicaSet %v, expected %v", key, expected)
- }
- case <-time.After(wait.ForeverTestTimeout):
- t.Errorf("Processing DeleteFinalStateUnknown took longer than expected")
- }
- }
- // Tell the rs to create 100 replicas, but simulate a limit (like a quota limit)
- // of 10, and verify that the rs doesn't make 100 create calls per sync pass
- func TestSyncReplicaSetCreateFailures(t *testing.T) {
- fakePodControl := controller.FakePodControl{}
- fakePodControl.CreateLimit = 10
- labelMap := map[string]string{"foo": "bar"}
- rs := newReplicaSet(fakePodControl.CreateLimit*10, labelMap)
- client := fake.NewSimpleClientset(rs)
- stopCh := make(chan struct{})
- defer close(stopCh)
- manager, informers := testNewReplicaSetControllerFromClient(client, stopCh, BurstReplicas)
- informers.Apps().V1().ReplicaSets().Informer().GetIndexer().Add(rs)
- manager.podControl = &fakePodControl
- manager.syncReplicaSet(GetKey(rs, t))
- validateSyncReplicaSet(t, &fakePodControl, fakePodControl.CreateLimit, 0, 0)
- expectedLimit := 0
- for pass := uint8(0); expectedLimit <= fakePodControl.CreateLimit; pass++ {
- expectedLimit += controller.SlowStartInitialBatchSize << pass
- }
- if fakePodControl.CreateCallCount > expectedLimit {
- t.Errorf("Unexpected number of create calls. Expected <= %d, saw %d\n", fakePodControl.CreateLimit*2, fakePodControl.CreateCallCount)
- }
- }
- func TestSyncReplicaSetDormancy(t *testing.T) {
- // Setup a test server so we can lie about the current state of pods
- fakeHandler := utiltesting.FakeHandler{
- StatusCode: 200,
- ResponseBody: "{}",
- SkipRequestFn: skipListerFunc,
- T: t,
- }
- testServer := httptest.NewServer(&fakeHandler)
- defer testServer.Close()
- client := clientset.NewForConfigOrDie(&restclient.Config{Host: testServer.URL, ContentConfig: restclient.ContentConfig{GroupVersion: &schema.GroupVersion{Group: "", Version: "v1"}}})
- fakePodControl := controller.FakePodControl{}
- stopCh := make(chan struct{})
- defer close(stopCh)
- manager, informers := testNewReplicaSetControllerFromClient(client, stopCh, BurstReplicas)
- manager.podControl = &fakePodControl
- labelMap := map[string]string{"foo": "bar"}
- rsSpec := newReplicaSet(2, labelMap)
- informers.Apps().V1().ReplicaSets().Informer().GetIndexer().Add(rsSpec)
- newPodList(informers.Core().V1().Pods().Informer().GetIndexer(), 1, v1.PodRunning, labelMap, rsSpec, "pod")
- // Creates a replica and sets expectations
- rsSpec.Status.Replicas = 1
- rsSpec.Status.ReadyReplicas = 1
- rsSpec.Status.AvailableReplicas = 1
- manager.syncReplicaSet(GetKey(rsSpec, t))
- validateSyncReplicaSet(t, &fakePodControl, 1, 0, 0)
- // Expectations prevents replicas but not an update on status
- rsSpec.Status.Replicas = 0
- rsSpec.Status.ReadyReplicas = 0
- rsSpec.Status.AvailableReplicas = 0
- fakePodControl.Clear()
- manager.syncReplicaSet(GetKey(rsSpec, t))
- validateSyncReplicaSet(t, &fakePodControl, 0, 0, 0)
- // Get the key for the controller
- rsKey, err := controller.KeyFunc(rsSpec)
- if err != nil {
- t.Errorf("Couldn't get key for object %#v: %v", rsSpec, err)
- }
- // Lowering expectations should lead to a sync that creates a replica, however the
- // fakePodControl error will prevent this, leaving expectations at 0, 0
- manager.expectations.CreationObserved(rsKey)
- rsSpec.Status.Replicas = 1
- rsSpec.Status.ReadyReplicas = 1
- rsSpec.Status.AvailableReplicas = 1
- fakePodControl.Clear()
- fakePodControl.Err = fmt.Errorf("Fake Error")
- manager.syncReplicaSet(GetKey(rsSpec, t))
- validateSyncReplicaSet(t, &fakePodControl, 1, 0, 0)
- // This replica should not need a Lowering of expectations, since the previous create failed
- fakePodControl.Clear()
- fakePodControl.Err = nil
- manager.syncReplicaSet(GetKey(rsSpec, t))
- validateSyncReplicaSet(t, &fakePodControl, 1, 0, 0)
- // 2 PUT for the ReplicaSet status during dormancy window.
- // Note that the pod creates go through pod control so they're not recorded.
- fakeHandler.ValidateRequestCount(t, 2)
- }
- func TestPodControllerLookup(t *testing.T) {
- stopCh := make(chan struct{})
- defer close(stopCh)
- manager, informers := testNewReplicaSetControllerFromClient(clientset.NewForConfigOrDie(&restclient.Config{Host: "", ContentConfig: restclient.ContentConfig{GroupVersion: &schema.GroupVersion{Group: "", Version: "v1"}}}), stopCh, BurstReplicas)
- testCases := []struct {
- inRSs []*apps.ReplicaSet
- pod *v1.Pod
- outRSName string
- }{
- // pods without labels don't match any ReplicaSets
- {
- inRSs: []*apps.ReplicaSet{
- {ObjectMeta: metav1.ObjectMeta{Name: "basic"}}},
- pod: &v1.Pod{ObjectMeta: metav1.ObjectMeta{Name: "foo1", Namespace: metav1.NamespaceAll}},
- outRSName: "",
- },
- // Matching labels, not namespace
- {
- inRSs: []*apps.ReplicaSet{
- {
- ObjectMeta: metav1.ObjectMeta{Name: "foo"},
- Spec: apps.ReplicaSetSpec{
- Selector: &metav1.LabelSelector{MatchLabels: map[string]string{"foo": "bar"}},
- },
- },
- },
- pod: &v1.Pod{
- ObjectMeta: metav1.ObjectMeta{
- Name: "foo2", Namespace: "ns", Labels: map[string]string{"foo": "bar"}}},
- outRSName: "",
- },
- // Matching ns and labels returns the key to the ReplicaSet, not the ReplicaSet name
- {
- inRSs: []*apps.ReplicaSet{
- {
- ObjectMeta: metav1.ObjectMeta{Name: "bar", Namespace: "ns"},
- Spec: apps.ReplicaSetSpec{
- Selector: &metav1.LabelSelector{MatchLabels: map[string]string{"foo": "bar"}},
- },
- },
- },
- pod: &v1.Pod{
- ObjectMeta: metav1.ObjectMeta{
- Name: "foo3", Namespace: "ns", Labels: map[string]string{"foo": "bar"}}},
- outRSName: "bar",
- },
- }
- for _, c := range testCases {
- for _, r := range c.inRSs {
- informers.Apps().V1().ReplicaSets().Informer().GetIndexer().Add(r)
- }
- if rss := manager.getPodReplicaSets(c.pod); rss != nil {
- if len(rss) != 1 {
- t.Errorf("len(rss) = %v, want %v", len(rss), 1)
- continue
- }
- rs := rss[0]
- if c.outRSName != rs.Name {
- t.Errorf("Got replica set %+v expected %+v", rs.Name, c.outRSName)
- }
- } else if c.outRSName != "" {
- t.Errorf("Expected a replica set %v pod %v, found none", c.outRSName, c.pod.Name)
- }
- }
- }
- func TestWatchControllers(t *testing.T) {
- fakeWatch := watch.NewFake()
- client := fake.NewSimpleClientset()
- client.PrependWatchReactor("replicasets", core.DefaultWatchReactor(fakeWatch, nil))
- stopCh := make(chan struct{})
- defer close(stopCh)
- informers := informers.NewSharedInformerFactory(client, controller.NoResyncPeriodFunc())
- manager := NewReplicaSetController(
- informers.Apps().V1().ReplicaSets(),
- informers.Core().V1().Pods(),
- client,
- BurstReplicas,
- )
- informers.Start(stopCh)
- var testRSSpec apps.ReplicaSet
- received := make(chan string)
- // The update sent through the fakeWatcher should make its way into the workqueue,
- // and eventually into the syncHandler. The handler validates the received controller
- // and closes the received channel to indicate that the test can finish.
- manager.syncHandler = func(key string) error {
- obj, exists, err := informers.Apps().V1().ReplicaSets().Informer().GetIndexer().GetByKey(key)
- if !exists || err != nil {
- t.Errorf("Expected to find replica set under key %v", key)
- }
- rsSpec := *obj.(*apps.ReplicaSet)
- if !apiequality.Semantic.DeepDerivative(rsSpec, testRSSpec) {
- t.Errorf("Expected %#v, but got %#v", testRSSpec, rsSpec)
- }
- close(received)
- return nil
- }
- // Start only the ReplicaSet watcher and the workqueue, send a watch event,
- // and make sure it hits the sync method.
- go wait.Until(manager.worker, 10*time.Millisecond, stopCh)
- testRSSpec.Name = "foo"
- fakeWatch.Add(&testRSSpec)
- select {
- case <-received:
- case <-time.After(wait.ForeverTestTimeout):
- t.Errorf("unexpected timeout from result channel")
- }
- }
- func TestWatchPods(t *testing.T) {
- client := fake.NewSimpleClientset()
- fakeWatch := watch.NewFake()
- client.PrependWatchReactor("pods", core.DefaultWatchReactor(fakeWatch, nil))
- stopCh := make(chan struct{})
- defer close(stopCh)
- manager, informers := testNewReplicaSetControllerFromClient(client, stopCh, BurstReplicas)
- // Put one ReplicaSet into the shared informer
- labelMap := map[string]string{"foo": "bar"}
- testRSSpec := newReplicaSet(1, labelMap)
- informers.Apps().V1().ReplicaSets().Informer().GetIndexer().Add(testRSSpec)
- received := make(chan string)
- // The pod update sent through the fakeWatcher should figure out the managing ReplicaSet and
- // send it into the syncHandler.
- manager.syncHandler = func(key string) error {
- namespace, name, err := cache.SplitMetaNamespaceKey(key)
- if err != nil {
- t.Errorf("Error splitting key: %v", err)
- }
- rsSpec, err := manager.rsLister.ReplicaSets(namespace).Get(name)
- if err != nil {
- t.Errorf("Expected to find replica set under key %v: %v", key, err)
- }
- if !apiequality.Semantic.DeepDerivative(rsSpec, testRSSpec) {
- t.Errorf("\nExpected %#v,\nbut got %#v", testRSSpec, rsSpec)
- }
- close(received)
- return nil
- }
- // Start only the pod watcher and the workqueue, send a watch event,
- // and make sure it hits the sync method for the right ReplicaSet.
- go informers.Core().V1().Pods().Informer().Run(stopCh)
- go manager.Run(1, stopCh)
- pods := newPodList(nil, 1, v1.PodRunning, labelMap, testRSSpec, "pod")
- testPod := pods.Items[0]
- testPod.Status.Phase = v1.PodFailed
- fakeWatch.Add(&testPod)
- select {
- case <-received:
- case <-time.After(wait.ForeverTestTimeout):
- t.Errorf("unexpected timeout from result channel")
- }
- }
- func TestUpdatePods(t *testing.T) {
- stopCh := make(chan struct{})
- defer close(stopCh)
- manager, informers := testNewReplicaSetControllerFromClient(fake.NewSimpleClientset(), stopCh, BurstReplicas)
- received := make(chan string)
- manager.syncHandler = func(key string) error {
- namespace, name, err := cache.SplitMetaNamespaceKey(key)
- if err != nil {
- t.Errorf("Error splitting key: %v", err)
- }
- rsSpec, err := manager.rsLister.ReplicaSets(namespace).Get(name)
- if err != nil {
- t.Errorf("Expected to find replica set under key %v: %v", key, err)
- }
- received <- rsSpec.Name
- return nil
- }
- go wait.Until(manager.worker, 10*time.Millisecond, stopCh)
- // Put 2 ReplicaSets and one pod into the informers
- labelMap1 := map[string]string{"foo": "bar"}
- testRSSpec1 := newReplicaSet(1, labelMap1)
- informers.Apps().V1().ReplicaSets().Informer().GetIndexer().Add(testRSSpec1)
- testRSSpec2 := *testRSSpec1
- labelMap2 := map[string]string{"bar": "foo"}
- testRSSpec2.Spec.Selector = &metav1.LabelSelector{MatchLabels: labelMap2}
- testRSSpec2.Name = "barfoo"
- informers.Apps().V1().ReplicaSets().Informer().GetIndexer().Add(&testRSSpec2)
- isController := true
- controllerRef1 := metav1.OwnerReference{UID: testRSSpec1.UID, APIVersion: "v1", Kind: "ReplicaSet", Name: testRSSpec1.Name, Controller: &isController}
- controllerRef2 := metav1.OwnerReference{UID: testRSSpec2.UID, APIVersion: "v1", Kind: "ReplicaSet", Name: testRSSpec2.Name, Controller: &isController}
- // case 1: Pod with a ControllerRef
- pod1 := newPodList(informers.Core().V1().Pods().Informer().GetIndexer(), 1, v1.PodRunning, labelMap1, testRSSpec1, "pod").Items[0]
- pod1.OwnerReferences = []metav1.OwnerReference{controllerRef1}
- pod1.ResourceVersion = "1"
- pod2 := pod1
- pod2.Labels = labelMap2
- pod2.ResourceVersion = "2"
- manager.updatePod(&pod1, &pod2)
- expected := sets.NewString(testRSSpec1.Name)
- for _, name := range expected.List() {
- t.Logf("Expecting update for %+v", name)
- select {
- case got := <-received:
- if !expected.Has(got) {
- t.Errorf("Expected keys %#v got %v", expected, got)
- }
- case <-time.After(wait.ForeverTestTimeout):
- t.Errorf("Expected update notifications for replica sets")
- }
- }
- // case 2: Remove ControllerRef (orphan). Expect to sync label-matching RS.
- pod1 = newPodList(informers.Core().V1().Pods().Informer().GetIndexer(), 1, v1.PodRunning, labelMap1, testRSSpec1, "pod").Items[0]
- pod1.ResourceVersion = "1"
- pod1.Labels = labelMap2
- pod1.OwnerReferences = []metav1.OwnerReference{controllerRef2}
- pod2 = pod1
- pod2.OwnerReferences = nil
- pod2.ResourceVersion = "2"
- manager.updatePod(&pod1, &pod2)
- expected = sets.NewString(testRSSpec2.Name)
- for _, name := range expected.List() {
- t.Logf("Expecting update for %+v", name)
- select {
- case got := <-received:
- if !expected.Has(got) {
- t.Errorf("Expected keys %#v got %v", expected, got)
- }
- case <-time.After(wait.ForeverTestTimeout):
- t.Errorf("Expected update notifications for replica sets")
- }
- }
- // case 2: Remove ControllerRef (orphan). Expect to sync both former owner and
- // any label-matching RS.
- pod1 = newPodList(informers.Core().V1().Pods().Informer().GetIndexer(), 1, v1.PodRunning, labelMap1, testRSSpec1, "pod").Items[0]
- pod1.ResourceVersion = "1"
- pod1.Labels = labelMap2
- pod1.OwnerReferences = []metav1.OwnerReference{controllerRef1}
- pod2 = pod1
- pod2.OwnerReferences = nil
- pod2.ResourceVersion = "2"
- manager.updatePod(&pod1, &pod2)
- expected = sets.NewString(testRSSpec1.Name, testRSSpec2.Name)
- for _, name := range expected.List() {
- t.Logf("Expecting update for %+v", name)
- select {
- case got := <-received:
- if !expected.Has(got) {
- t.Errorf("Expected keys %#v got %v", expected, got)
- }
- case <-time.After(wait.ForeverTestTimeout):
- t.Errorf("Expected update notifications for replica sets")
- }
- }
- // case 4: Keep ControllerRef, change labels. Expect to sync owning RS.
- pod1 = newPodList(informers.Core().V1().Pods().Informer().GetIndexer(), 1, v1.PodRunning, labelMap1, testRSSpec1, "pod").Items[0]
- pod1.ResourceVersion = "1"
- pod1.Labels = labelMap1
- pod1.OwnerReferences = []metav1.OwnerReference{controllerRef2}
- pod2 = pod1
- pod2.Labels = labelMap2
- pod2.ResourceVersion = "2"
- manager.updatePod(&pod1, &pod2)
- expected = sets.NewString(testRSSpec2.Name)
- for _, name := range expected.List() {
- t.Logf("Expecting update for %+v", name)
- select {
- case got := <-received:
- if !expected.Has(got) {
- t.Errorf("Expected keys %#v got %v", expected, got)
- }
- case <-time.After(wait.ForeverTestTimeout):
- t.Errorf("Expected update notifications for replica sets")
- }
- }
- }
- func TestControllerUpdateRequeue(t *testing.T) {
- // This server should force a requeue of the controller because it fails to update status.Replicas.
- labelMap := map[string]string{"foo": "bar"}
- rs := newReplicaSet(1, labelMap)
- client := fake.NewSimpleClientset(rs)
- client.PrependReactor("update", "replicasets",
- func(action core.Action) (bool, runtime.Object, error) {
- if action.GetSubresource() != "status" {
- return false, nil, nil
- }
- return true, nil, errors.New("failed to update status")
- })
- stopCh := make(chan struct{})
- defer close(stopCh)
- manager, informers := testNewReplicaSetControllerFromClient(client, stopCh, BurstReplicas)
- informers.Apps().V1().ReplicaSets().Informer().GetIndexer().Add(rs)
- rs.Status = apps.ReplicaSetStatus{Replicas: 2}
- newPodList(informers.Core().V1().Pods().Informer().GetIndexer(), 1, v1.PodRunning, labelMap, rs, "pod")
- fakePodControl := controller.FakePodControl{}
- manager.podControl = &fakePodControl
- // Enqueue once. Then process it. Disable rate-limiting for this.
- manager.queue = workqueue.NewRateLimitingQueue(workqueue.NewMaxOfRateLimiter())
- manager.enqueueReplicaSet(rs)
- manager.processNextWorkItem()
- // It should have been requeued.
- if got, want := manager.queue.Len(), 1; got != want {
- t.Errorf("queue.Len() = %v, want %v", got, want)
- }
- }
- func TestControllerUpdateStatusWithFailure(t *testing.T) {
- rs := newReplicaSet(1, map[string]string{"foo": "bar"})
- fakeClient := &fake.Clientset{}
- fakeClient.AddReactor("get", "replicasets", func(action core.Action) (bool, runtime.Object, error) { return true, rs, nil })
- fakeClient.AddReactor("*", "*", func(action core.Action) (bool, runtime.Object, error) {
- return true, &apps.ReplicaSet{}, fmt.Errorf("Fake error")
- })
- fakeRSClient := fakeClient.AppsV1().ReplicaSets("default")
- numReplicas := int32(10)
- newStatus := apps.ReplicaSetStatus{Replicas: numReplicas}
- updateReplicaSetStatus(fakeRSClient, rs, newStatus)
- updates, gets := 0, 0
- for _, a := range fakeClient.Actions() {
- if a.GetResource().Resource != "replicasets" {
- t.Errorf("Unexpected action %+v", a)
- continue
- }
- switch action := a.(type) {
- case core.GetAction:
- gets++
- // Make sure the get is for the right ReplicaSet even though the update failed.
- if action.GetName() != rs.Name {
- t.Errorf("Expected get for ReplicaSet %v, got %+v instead", rs.Name, action.GetName())
- }
- case core.UpdateAction:
- updates++
- // Confirm that the update has the right status.Replicas even though the Get
- // returned a ReplicaSet with replicas=1.
- if c, ok := action.GetObject().(*apps.ReplicaSet); !ok {
- t.Errorf("Expected a ReplicaSet as the argument to update, got %T", c)
- } else if c.Status.Replicas != numReplicas {
- t.Errorf("Expected update for ReplicaSet to contain replicas %v, got %v instead",
- numReplicas, c.Status.Replicas)
- }
- default:
- t.Errorf("Unexpected action %+v", a)
- break
- }
- }
- if gets != 1 || updates != 2 {
- t.Errorf("Expected 1 get and 2 updates, got %d gets %d updates", gets, updates)
- }
- }
- // TODO: This test is too hairy for a unittest. It should be moved to an E2E suite.
- func doTestControllerBurstReplicas(t *testing.T, burstReplicas, numReplicas int) {
- labelMap := map[string]string{"foo": "bar"}
- rsSpec := newReplicaSet(numReplicas, labelMap)
- client := fake.NewSimpleClientset(rsSpec)
- fakePodControl := controller.FakePodControl{}
- stopCh := make(chan struct{})
- defer close(stopCh)
- manager, informers := testNewReplicaSetControllerFromClient(client, stopCh, burstReplicas)
- manager.podControl = &fakePodControl
- informers.Apps().V1().ReplicaSets().Informer().GetIndexer().Add(rsSpec)
- expectedPods := int32(0)
- pods := newPodList(nil, numReplicas, v1.PodPending, labelMap, rsSpec, "pod")
- rsKey, err := controller.KeyFunc(rsSpec)
- if err != nil {
- t.Errorf("Couldn't get key for object %#v: %v", rsSpec, err)
- }
- // Size up the controller, then size it down, and confirm the expected create/delete pattern
- for _, replicas := range []int32{int32(numReplicas), 0} {
- *(rsSpec.Spec.Replicas) = replicas
- informers.Apps().V1().ReplicaSets().Informer().GetIndexer().Add(rsSpec)
- for i := 0; i < numReplicas; i += burstReplicas {
- manager.syncReplicaSet(GetKey(rsSpec, t))
- // The store accrues active pods. It's also used by the ReplicaSet to determine how many
- // replicas to create.
- activePods := int32(len(informers.Core().V1().Pods().Informer().GetIndexer().List()))
- if replicas != 0 {
- // This is the number of pods currently "in flight". They were created by the
- // ReplicaSet controller above, which then puts the ReplicaSet to sleep till
- // all of them have been observed.
- expectedPods = replicas - activePods
- if expectedPods > int32(burstReplicas) {
- expectedPods = int32(burstReplicas)
- }
- // This validates the ReplicaSet manager sync actually created pods
- validateSyncReplicaSet(t, &fakePodControl, int(expectedPods), 0, 0)
- // This simulates the watch events for all but 1 of the expected pods.
- // None of these should wake the controller because it has expectations==BurstReplicas.
- for i := int32(0); i < expectedPods-1; i++ {
- informers.Core().V1().Pods().Informer().GetIndexer().Add(&pods.Items[i])
- manager.addPod(&pods.Items[i])
- }
- podExp, exists, err := manager.expectations.GetExpectations(rsKey)
- if !exists || err != nil {
- t.Fatalf("Did not find expectations for rs.")
- }
- if add, _ := podExp.GetExpectations(); add != 1 {
- t.Fatalf("Expectations are wrong %v", podExp)
- }
- } else {
- expectedPods = (replicas - activePods) * -1
- if expectedPods > int32(burstReplicas) {
- expectedPods = int32(burstReplicas)
- }
- validateSyncReplicaSet(t, &fakePodControl, 0, int(expectedPods), 0)
- // To accurately simulate a watch we must delete the exact pods
- // the rs is waiting for.
- expectedDels := manager.expectations.GetUIDs(GetKey(rsSpec, t))
- podsToDelete := []*v1.Pod{}
- isController := true
- for _, key := range expectedDels.List() {
- nsName := strings.Split(key, "/")
- podsToDelete = append(podsToDelete, &v1.Pod{
- ObjectMeta: metav1.ObjectMeta{
- Name: nsName[1],
- Namespace: nsName[0],
- Labels: rsSpec.Spec.Selector.MatchLabels,
- OwnerReferences: []metav1.OwnerReference{
- {UID: rsSpec.UID, APIVersion: "v1", Kind: "ReplicaSet", Name: rsSpec.Name, Controller: &isController},
- },
- },
- })
- }
- // Don't delete all pods because we confirm that the last pod
- // has exactly one expectation at the end, to verify that we
- // don't double delete.
- for i := range podsToDelete[1:] {
- informers.Core().V1().Pods().Informer().GetIndexer().Delete(podsToDelete[i])
- manager.deletePod(podsToDelete[i])
- }
- podExp, exists, err := manager.expectations.GetExpectations(rsKey)
- if !exists || err != nil {
- t.Fatalf("Did not find expectations for ReplicaSet.")
- }
- if _, del := podExp.GetExpectations(); del != 1 {
- t.Fatalf("Expectations are wrong %v", podExp)
- }
- }
- // Check that the ReplicaSet didn't take any action for all the above pods
- fakePodControl.Clear()
- manager.syncReplicaSet(GetKey(rsSpec, t))
- validateSyncReplicaSet(t, &fakePodControl, 0, 0, 0)
- // Create/Delete the last pod
- // The last add pod will decrease the expectation of the ReplicaSet to 0,
- // which will cause it to create/delete the remaining replicas up to burstReplicas.
- if replicas != 0 {
- informers.Core().V1().Pods().Informer().GetIndexer().Add(&pods.Items[expectedPods-1])
- manager.addPod(&pods.Items[expectedPods-1])
- } else {
- expectedDel := manager.expectations.GetUIDs(GetKey(rsSpec, t))
- if expectedDel.Len() != 1 {
- t.Fatalf("Waiting on unexpected number of deletes.")
- }
- nsName := strings.Split(expectedDel.List()[0], "/")
- isController := true
- lastPod := &v1.Pod{
- ObjectMeta: metav1.ObjectMeta{
- Name: nsName[1],
- Namespace: nsName[0],
- Labels: rsSpec.Spec.Selector.MatchLabels,
- OwnerReferences: []metav1.OwnerReference{
- {UID: rsSpec.UID, APIVersion: "v1", Kind: "ReplicaSet", Name: rsSpec.Name, Controller: &isController},
- },
- },
- }
- informers.Core().V1().Pods().Informer().GetIndexer().Delete(lastPod)
- manager.deletePod(lastPod)
- }
- pods.Items = pods.Items[expectedPods:]
- }
- // Confirm that we've created the right number of replicas
- activePods := int32(len(informers.Core().V1().Pods().Informer().GetIndexer().List()))
- if activePods != *(rsSpec.Spec.Replicas) {
- t.Fatalf("Unexpected number of active pods, expected %d, got %d", *(rsSpec.Spec.Replicas), activePods)
- }
- // Replenish the pod list, since we cut it down sizing up
- pods = newPodList(nil, int(replicas), v1.PodRunning, labelMap, rsSpec, "pod")
- }
- }
- func TestControllerBurstReplicas(t *testing.T) {
- doTestControllerBurstReplicas(t, 5, 30)
- doTestControllerBurstReplicas(t, 5, 12)
- doTestControllerBurstReplicas(t, 3, 2)
- }
- type FakeRSExpectations struct {
- *controller.ControllerExpectations
- satisfied bool
- expSatisfied func()
- }
- func (fe FakeRSExpectations) SatisfiedExpectations(controllerKey string) bool {
- fe.expSatisfied()
- return fe.satisfied
- }
- // TestRSSyncExpectations tests that a pod cannot sneak in between counting active pods
- // and checking expectations.
- func TestRSSyncExpectations(t *testing.T) {
- client := clientset.NewForConfigOrDie(&restclient.Config{Host: "", ContentConfig: restclient.ContentConfig{GroupVersion: &schema.GroupVersion{Group: "", Version: "v1"}}})
- fakePodControl := controller.FakePodControl{}
- stopCh := make(chan struct{})
- defer close(stopCh)
- manager, informers := testNewReplicaSetControllerFromClient(client, stopCh, 2)
- manager.podControl = &fakePodControl
- labelMap := map[string]string{"foo": "bar"}
- rsSpec := newReplicaSet(2, labelMap)
- informers.Apps().V1().ReplicaSets().Informer().GetIndexer().Add(rsSpec)
- pods := newPodList(nil, 2, v1.PodPending, labelMap, rsSpec, "pod")
- informers.Core().V1().Pods().Informer().GetIndexer().Add(&pods.Items[0])
- postExpectationsPod := pods.Items[1]
- manager.expectations = controller.NewUIDTrackingControllerExpectations(FakeRSExpectations{
- controller.NewControllerExpectations(), true, func() {
- // If we check active pods before checking expectataions, the
- // ReplicaSet will create a new replica because it doesn't see
- // this pod, but has fulfilled its expectations.
- informers.Core().V1().Pods().Informer().GetIndexer().Add(&postExpectationsPod)
- },
- })
- manager.syncReplicaSet(GetKey(rsSpec, t))
- validateSyncReplicaSet(t, &fakePodControl, 0, 0, 0)
- }
- func TestDeleteControllerAndExpectations(t *testing.T) {
- rs := newReplicaSet(1, map[string]string{"foo": "bar"})
- client := fake.NewSimpleClientset(rs)
- stopCh := make(chan struct{})
- defer close(stopCh)
- manager, informers := testNewReplicaSetControllerFromClient(client, stopCh, 10)
- informers.Apps().V1().ReplicaSets().Informer().GetIndexer().Add(rs)
- fakePodControl := controller.FakePodControl{}
- manager.podControl = &fakePodControl
- // This should set expectations for the ReplicaSet
- manager.syncReplicaSet(GetKey(rs, t))
- validateSyncReplicaSet(t, &fakePodControl, 1, 0, 0)
- fakePodControl.Clear()
- // Get the ReplicaSet key
- rsKey, err := controller.KeyFunc(rs)
- if err != nil {
- t.Errorf("Couldn't get key for object %#v: %v", rs, err)
- }
- // This is to simulate a concurrent addPod, that has a handle on the expectations
- // as the controller deletes it.
- podExp, exists, err := manager.expectations.GetExpectations(rsKey)
- if !exists || err != nil {
- t.Errorf("No expectations found for ReplicaSet")
- }
- informers.Apps().V1().ReplicaSets().Informer().GetIndexer().Delete(rs)
- manager.syncReplicaSet(GetKey(rs, t))
- if _, exists, err = manager.expectations.GetExpectations(rsKey); exists {
- t.Errorf("Found expectaions, expected none since the ReplicaSet has been deleted.")
- }
- // This should have no effect, since we've deleted the ReplicaSet.
- podExp.Add(-1, 0)
- informers.Core().V1().Pods().Informer().GetIndexer().Replace(make([]interface{}, 0), "0")
- manager.syncReplicaSet(GetKey(rs, t))
- validateSyncReplicaSet(t, &fakePodControl, 0, 0, 0)
- }
- // shuffle returns a new shuffled list of container controllers.
- func shuffle(controllers []*apps.ReplicaSet) []*apps.ReplicaSet {
- numControllers := len(controllers)
- randIndexes := rand.Perm(numControllers)
- shuffled := make([]*apps.ReplicaSet, numControllers)
- for i := 0; i < numControllers; i++ {
- shuffled[i] = controllers[randIndexes[i]]
- }
- return shuffled
- }
- func TestOverlappingRSs(t *testing.T) {
- client := clientset.NewForConfigOrDie(&restclient.Config{Host: "", ContentConfig: restclient.ContentConfig{GroupVersion: &schema.GroupVersion{Group: "", Version: "v1"}}})
- labelMap := map[string]string{"foo": "bar"}
- stopCh := make(chan struct{})
- defer close(stopCh)
- manager, informers := testNewReplicaSetControllerFromClient(client, stopCh, 10)
- // Create 10 ReplicaSets, shuffled them randomly and insert them into the
- // ReplicaSet controller's store.
- // All use the same CreationTimestamp since ControllerRef should be able
- // to handle that.
- timestamp := metav1.Date(2014, time.December, 0, 0, 0, 0, 0, time.Local)
- var controllers []*apps.ReplicaSet
- for j := 1; j < 10; j++ {
- rsSpec := newReplicaSet(1, labelMap)
- rsSpec.CreationTimestamp = timestamp
- rsSpec.Name = fmt.Sprintf("rs%d", j)
- controllers = append(controllers, rsSpec)
- }
- shuffledControllers := shuffle(controllers)
- for j := range shuffledControllers {
- informers.Apps().V1().ReplicaSets().Informer().GetIndexer().Add(shuffledControllers[j])
- }
- // Add a pod with a ControllerRef and make sure only the corresponding
- // ReplicaSet is synced. Pick a RS in the middle since the old code used to
- // sort by name if all timestamps were equal.
- rs := controllers[3]
- pods := newPodList(nil, 1, v1.PodPending, labelMap, rs, "pod")
- pod := &pods.Items[0]
- isController := true
- pod.OwnerReferences = []metav1.OwnerReference{
- {UID: rs.UID, APIVersion: "v1", Kind: "ReplicaSet", Name: rs.Name, Controller: &isController},
- }
- rsKey := GetKey(rs, t)
- manager.addPod(pod)
- queueRS, _ := manager.queue.Get()
- if queueRS != rsKey {
- t.Fatalf("Expected to find key %v in queue, found %v", rsKey, queueRS)
- }
- }
- func TestDeletionTimestamp(t *testing.T) {
- c := clientset.NewForConfigOrDie(&restclient.Config{Host: "", ContentConfig: restclient.ContentConfig{GroupVersion: &schema.GroupVersion{Group: "", Version: "v1"}}})
- labelMap := map[string]string{"foo": "bar"}
- stopCh := make(chan struct{})
- defer close(stopCh)
- manager, informers := testNewReplicaSetControllerFromClient(c, stopCh, 10)
- rs := newReplicaSet(1, labelMap)
- informers.Apps().V1().ReplicaSets().Informer().GetIndexer().Add(rs)
- rsKey, err := controller.KeyFunc(rs)
- if err != nil {
- t.Errorf("Couldn't get key for object %#v: %v", rs, err)
- }
- pod := newPodList(nil, 1, v1.PodPending, labelMap, rs, "pod").Items[0]
- pod.DeletionTimestamp = &metav1.Time{Time: time.Now()}
- pod.ResourceVersion = "1"
- manager.expectations.ExpectDeletions(rsKey, []string{controller.PodKey(&pod)})
- // A pod added with a deletion timestamp should decrement deletions, not creations.
- manager.addPod(&pod)
- queueRS, _ := manager.queue.Get()
- if queueRS != rsKey {
- t.Fatalf("Expected to find key %v in queue, found %v", rsKey, queueRS)
- }
- manager.queue.Done(rsKey)
- podExp, exists, err := manager.expectations.GetExpectations(rsKey)
- if !exists || err != nil || !podExp.Fulfilled() {
- t.Fatalf("Wrong expectations %#v", podExp)
- }
- // An update from no deletion timestamp to having one should be treated
- // as a deletion.
- oldPod := newPodList(nil, 1, v1.PodPending, labelMap, rs, "pod").Items[0]
- oldPod.ResourceVersion = "2"
- manager.expectations.ExpectDeletions(rsKey, []string{controller.PodKey(&pod)})
- manager.updatePod(&oldPod, &pod)
- queueRS, _ = manager.queue.Get()
- if queueRS != rsKey {
- t.Fatalf("Expected to find key %v in queue, found %v", rsKey, queueRS)
- }
- manager.queue.Done(rsKey)
- podExp, exists, err = manager.expectations.GetExpectations(rsKey)
- if !exists || err != nil || !podExp.Fulfilled() {
- t.Fatalf("Wrong expectations %#v", podExp)
- }
- // An update to the pod (including an update to the deletion timestamp)
- // should not be counted as a second delete.
- isController := true
- secondPod := &v1.Pod{
- ObjectMeta: metav1.ObjectMeta{
- Namespace: pod.Namespace,
- Name: "secondPod",
- Labels: pod.Labels,
- OwnerReferences: []metav1.OwnerReference{
- {UID: rs.UID, APIVersion: "v1", Kind: "ReplicaSet", Name: rs.Name, Controller: &isController},
- },
- },
- }
- manager.expectations.ExpectDeletions(rsKey, []string{controller.PodKey(secondPod)})
- oldPod.DeletionTimestamp = &metav1.Time{Time: time.Now()}
- oldPod.ResourceVersion = "2"
- manager.updatePod(&oldPod, &pod)
- podExp, exists, err = manager.expectations.GetExpectations(rsKey)
- if !exists || err != nil || podExp.Fulfilled() {
- t.Fatalf("Wrong expectations %#v", podExp)
- }
- // A pod with a non-nil deletion timestamp should also be ignored by the
- // delete handler, because it's already been counted in the update.
- manager.deletePod(&pod)
- podExp, exists, err = manager.expectations.GetExpectations(rsKey)
- if !exists || err != nil || podExp.Fulfilled() {
- t.Fatalf("Wrong expectations %#v", podExp)
- }
- // Deleting the second pod should clear expectations.
- manager.deletePod(secondPod)
- queueRS, _ = manager.queue.Get()
- if queueRS != rsKey {
- t.Fatalf("Expected to find key %v in queue, found %v", rsKey, queueRS)
- }
- manager.queue.Done(rsKey)
- podExp, exists, err = manager.expectations.GetExpectations(rsKey)
- if !exists || err != nil || !podExp.Fulfilled() {
- t.Fatalf("Wrong expectations %#v", podExp)
- }
- }
- // setupManagerWithGCEnabled creates a RS manager with a fakePodControl
- func setupManagerWithGCEnabled(stopCh chan struct{}, objs ...runtime.Object) (manager *ReplicaSetController, fakePodControl *controller.FakePodControl, informers informers.SharedInformerFactory) {
- c := fake.NewSimpleClientset(objs...)
- fakePodControl = &controller.FakePodControl{}
- manager, informers = testNewReplicaSetControllerFromClient(c, stopCh, BurstReplicas)
- manager.podControl = fakePodControl
- return manager, fakePodControl, informers
- }
- func TestDoNotPatchPodWithOtherControlRef(t *testing.T) {
- labelMap := map[string]string{"foo": "bar"}
- rs := newReplicaSet(2, labelMap)
- stopCh := make(chan struct{})
- defer close(stopCh)
- manager, fakePodControl, informers := setupManagerWithGCEnabled(stopCh, rs)
- informers.Apps().V1().ReplicaSets().Informer().GetIndexer().Add(rs)
- var trueVar = true
- otherControllerReference := metav1.OwnerReference{UID: uuid.NewUUID(), APIVersion: "v1beta1", Kind: "ReplicaSet", Name: "AnotherRS", Controller: &trueVar}
- // add to podLister a matching Pod controlled by another controller. Expect no patch.
- pod := newPod("pod", rs, v1.PodRunning, nil, true)
- pod.OwnerReferences = []metav1.OwnerReference{otherControllerReference}
- informers.Core().V1().Pods().Informer().GetIndexer().Add(pod)
- err := manager.syncReplicaSet(GetKey(rs, t))
- if err != nil {
- t.Fatal(err)
- }
- // because the matching pod already has a controller, so 2 pods should be created.
- validateSyncReplicaSet(t, fakePodControl, 2, 0, 0)
- }
- func TestPatchPodFails(t *testing.T) {
- labelMap := map[string]string{"foo": "bar"}
- rs := newReplicaSet(2, labelMap)
- stopCh := make(chan struct{})
- defer close(stopCh)
- manager, fakePodControl, informers := setupManagerWithGCEnabled(stopCh, rs)
- informers.Apps().V1().ReplicaSets().Informer().GetIndexer().Add(rs)
- // add to podLister two matching pods. Expect two patches to take control
- // them.
- informers.Core().V1().Pods().Informer().GetIndexer().Add(newPod("pod1", rs, v1.PodRunning, nil, false))
- informers.Core().V1().Pods().Informer().GetIndexer().Add(newPod("pod2", rs, v1.PodRunning, nil, false))
- // let both patches fail. The rs controller will assume it fails to take
- // control of the pods and requeue to try again.
- fakePodControl.Err = fmt.Errorf("Fake Error")
- rsKey := GetKey(rs, t)
- err := processSync(manager, rsKey)
- if err == nil || !strings.Contains(err.Error(), "Fake Error") {
- t.Errorf("expected Fake Error, got %+v", err)
- }
- // 2 patches to take control of pod1 and pod2 (both fail).
- validateSyncReplicaSet(t, fakePodControl, 0, 0, 2)
- // RS should requeue itself.
- queueRS, _ := manager.queue.Get()
- if queueRS != rsKey {
- t.Fatalf("Expected to find key %v in queue, found %v", rsKey, queueRS)
- }
- }
- // RS controller shouldn't adopt or create more pods if the rc is about to be
- // deleted.
- func TestDoNotAdoptOrCreateIfBeingDeleted(t *testing.T) {
- labelMap := map[string]string{"foo": "bar"}
- rs := newReplicaSet(2, labelMap)
- now := metav1.Now()
- rs.DeletionTimestamp = &now
- stopCh := make(chan struct{})
- defer close(stopCh)
- manager, fakePodControl, informers := setupManagerWithGCEnabled(stopCh, rs)
- informers.Apps().V1().ReplicaSets().Informer().GetIndexer().Add(rs)
- pod1 := newPod("pod1", rs, v1.PodRunning, nil, false)
- informers.Core().V1().Pods().Informer().GetIndexer().Add(pod1)
- // no patch, no create
- err := manager.syncReplicaSet(GetKey(rs, t))
- if err != nil {
- t.Fatal(err)
- }
- validateSyncReplicaSet(t, fakePodControl, 0, 0, 0)
- }
- func TestDoNotAdoptOrCreateIfBeingDeletedRace(t *testing.T) {
- labelMap := map[string]string{"foo": "bar"}
- // Bare client says it IS deleted.
- rs := newReplicaSet(2, labelMap)
- now := metav1.Now()
- rs.DeletionTimestamp = &now
- stopCh := make(chan struct{})
- defer close(stopCh)
- manager, fakePodControl, informers := setupManagerWithGCEnabled(stopCh, rs)
- // Lister (cache) says it's NOT deleted.
- rs2 := *rs
- rs2.DeletionTimestamp = nil
- informers.Apps().V1().ReplicaSets().Informer().GetIndexer().Add(&rs2)
- // Recheck occurs if a matching orphan is present.
- pod1 := newPod("pod1", rs, v1.PodRunning, nil, false)
- informers.Core().V1().Pods().Informer().GetIndexer().Add(pod1)
- // sync should abort.
- err := manager.syncReplicaSet(GetKey(rs, t))
- if err == nil {
- t.Error("syncReplicaSet() err = nil, expected non-nil")
- }
- // no patch, no create.
- validateSyncReplicaSet(t, fakePodControl, 0, 0, 0)
- }
- var (
- imagePullBackOff apps.ReplicaSetConditionType = "ImagePullBackOff"
- condImagePullBackOff = func() apps.ReplicaSetCondition {
- return apps.ReplicaSetCondition{
- Type: imagePullBackOff,
- Status: v1.ConditionTrue,
- Reason: "NonExistentImage",
- }
- }
- condReplicaFailure = func() apps.ReplicaSetCondition {
- return apps.ReplicaSetCondition{
- Type: apps.ReplicaSetReplicaFailure,
- Status: v1.ConditionTrue,
- Reason: "OtherFailure",
- }
- }
- condReplicaFailure2 = func() apps.ReplicaSetCondition {
- return apps.ReplicaSetCondition{
- Type: apps.ReplicaSetReplicaFailure,
- Status: v1.ConditionTrue,
- Reason: "AnotherFailure",
- }
- }
- status = func() *apps.ReplicaSetStatus {
- return &apps.ReplicaSetStatus{
- Conditions: []apps.ReplicaSetCondition{condReplicaFailure()},
- }
- }
- )
- func TestGetCondition(t *testing.T) {
- exampleStatus := status()
- tests := []struct {
- name string
- status apps.ReplicaSetStatus
- condType apps.ReplicaSetConditionType
- expected bool
- }{
- {
- name: "condition exists",
- status: *exampleStatus,
- condType: apps.ReplicaSetReplicaFailure,
- expected: true,
- },
- {
- name: "condition does not exist",
- status: *exampleStatus,
- condType: imagePullBackOff,
- expected: false,
- },
- }
- for _, test := range tests {
- cond := GetCondition(test.status, test.condType)
- exists := cond != nil
- if exists != test.expected {
- t.Errorf("%s: expected condition to exist: %t, got: %t", test.name, test.expected, exists)
- }
- }
- }
- func TestSetCondition(t *testing.T) {
- tests := []struct {
- name string
- status *apps.ReplicaSetStatus
- cond apps.ReplicaSetCondition
- expectedStatus *apps.ReplicaSetStatus
- }{
- {
- name: "set for the first time",
- status: &apps.ReplicaSetStatus{},
- cond: condReplicaFailure(),
- expectedStatus: &apps.ReplicaSetStatus{Conditions: []apps.ReplicaSetCondition{condReplicaFailure()}},
- },
- {
- name: "simple set",
- status: &apps.ReplicaSetStatus{Conditions: []apps.ReplicaSetCondition{condImagePullBackOff()}},
- cond: condReplicaFailure(),
- expectedStatus: &apps.ReplicaSetStatus{Conditions: []apps.ReplicaSetCondition{condImagePullBackOff(), condReplicaFailure()}},
- },
- {
- name: "overwrite",
- status: &apps.ReplicaSetStatus{Conditions: []apps.ReplicaSetCondition{condReplicaFailure()}},
- cond: condReplicaFailure2(),
- expectedStatus: &apps.ReplicaSetStatus{Conditions: []apps.ReplicaSetCondition{condReplicaFailure2()}},
- },
- }
- for _, test := range tests {
- SetCondition(test.status, test.cond)
- if !reflect.DeepEqual(test.status, test.expectedStatus) {
- t.Errorf("%s: expected status: %v, got: %v", test.name, test.expectedStatus, test.status)
- }
- }
- }
- func TestRemoveCondition(t *testing.T) {
- tests := []struct {
- name string
- status *apps.ReplicaSetStatus
- condType apps.ReplicaSetConditionType
- expectedStatus *apps.ReplicaSetStatus
- }{
- {
- name: "remove from empty status",
- status: &apps.ReplicaSetStatus{},
- condType: apps.ReplicaSetReplicaFailure,
- expectedStatus: &apps.ReplicaSetStatus{},
- },
- {
- name: "simple remove",
- status: &apps.ReplicaSetStatus{Conditions: []apps.ReplicaSetCondition{condReplicaFailure()}},
- condType: apps.ReplicaSetReplicaFailure,
- expectedStatus: &apps.ReplicaSetStatus{},
- },
- {
- name: "doesn't remove anything",
- status: status(),
- condType: imagePullBackOff,
- expectedStatus: status(),
- },
- }
- for _, test := range tests {
- RemoveCondition(test.status, test.condType)
- if !reflect.DeepEqual(test.status, test.expectedStatus) {
- t.Errorf("%s: expected status: %v, got: %v", test.name, test.expectedStatus, test.status)
- }
- }
- }
- func TestSlowStartBatch(t *testing.T) {
- fakeErr := fmt.Errorf("fake error")
- callCnt := 0
- callLimit := 0
- var lock sync.Mutex
- fn := func() error {
- lock.Lock()
- defer lock.Unlock()
- callCnt++
- if callCnt > callLimit {
- return fakeErr
- }
- return nil
- }
- tests := []struct {
- name string
- count int
- callLimit int
- fn func() error
- expectedSuccesses int
- expectedErr error
- expectedCallCnt int
- }{
- {
- name: "callLimit = 0 (all fail)",
- count: 10,
- callLimit: 0,
- fn: fn,
- expectedSuccesses: 0,
- expectedErr: fakeErr,
- expectedCallCnt: 1, // 1(first batch): function will be called at least once
- },
- {
- name: "callLimit = count (all succeed)",
- count: 10,
- callLimit: 10,
- fn: fn,
- expectedSuccesses: 10,
- expectedErr: nil,
- expectedCallCnt: 10, // 1(first batch) + 2(2nd batch) + 4(3rd batch) + 3(4th batch) = 10
- },
- {
- name: "callLimit < count (some succeed)",
- count: 10,
- callLimit: 5,
- fn: fn,
- expectedSuccesses: 5,
- expectedErr: fakeErr,
- expectedCallCnt: 7, // 1(first batch) + 2(2nd batch) + 4(3rd batch) = 7
- },
- }
- for _, test := range tests {
- callCnt = 0
- callLimit = test.callLimit
- successes, err := slowStartBatch(test.count, 1, test.fn)
- if successes != test.expectedSuccesses {
- t.Errorf("%s: unexpected processed batch size, expected %d, got %d", test.name, test.expectedSuccesses, successes)
- }
- if err != test.expectedErr {
- t.Errorf("%s: unexpected processed batch size, expected %v, got %v", test.name, test.expectedErr, err)
- }
- // verify that slowStartBatch stops trying more calls after a batch fails
- if callCnt != test.expectedCallCnt {
- t.Errorf("%s: slowStartBatch() still tries calls after a batch fails, expected %d calls, got %d", test.name, test.expectedCallCnt, callCnt)
- }
- }
- }
- func TestGetPodsToDelete(t *testing.T) {
- labelMap := map[string]string{"name": "foo"}
- rs := newReplicaSet(1, labelMap)
- // an unscheduled, pending pod
- unscheduledPendingPod := newPod("unscheduled-pending-pod", rs, v1.PodPending, nil, true)
- // a scheduled, pending pod
- scheduledPendingPod := newPod("scheduled-pending-pod", rs, v1.PodPending, nil, true)
- scheduledPendingPod.Spec.NodeName = "fake-node"
- // a scheduled, running, not-ready pod
- scheduledRunningNotReadyPod := newPod("scheduled-running-not-ready-pod", rs, v1.PodRunning, nil, true)
- scheduledRunningNotReadyPod.Spec.NodeName = "fake-node"
- scheduledRunningNotReadyPod.Status.Conditions = []v1.PodCondition{
- {
- Type: v1.PodReady,
- Status: v1.ConditionFalse,
- },
- }
- // a scheduled, running, ready pod
- scheduledRunningReadyPod := newPod("scheduled-running-ready-pod", rs, v1.PodRunning, nil, true)
- scheduledRunningReadyPod.Spec.NodeName = "fake-node"
- scheduledRunningReadyPod.Status.Conditions = []v1.PodCondition{
- {
- Type: v1.PodReady,
- Status: v1.ConditionTrue,
- },
- }
- tests := []struct {
- name string
- pods []*v1.Pod
- diff int
- expectedPodsToDelete []*v1.Pod
- }{
- // Order used when selecting pods for deletion:
- // an unscheduled, pending pod
- // a scheduled, pending pod
- // a scheduled, running, not-ready pod
- // a scheduled, running, ready pod
- // Note that a pending pod cannot be ready
- {
- "len(pods) = 0 (i.e., diff = 0 too)",
- []*v1.Pod{},
- 0,
- []*v1.Pod{},
- },
- {
- "diff = len(pods)",
- []*v1.Pod{
- scheduledRunningNotReadyPod,
- scheduledRunningReadyPod,
- },
- 2,
- []*v1.Pod{scheduledRunningNotReadyPod, scheduledRunningReadyPod},
- },
- {
- "diff < len(pods)",
- []*v1.Pod{
- scheduledRunningReadyPod,
- scheduledRunningNotReadyPod,
- },
- 1,
- []*v1.Pod{scheduledRunningNotReadyPod},
- },
- {
- "various pod phases and conditions, diff = len(pods)",
- []*v1.Pod{
- scheduledRunningReadyPod,
- scheduledRunningNotReadyPod,
- scheduledPendingPod,
- unscheduledPendingPod,
- },
- 4,
- []*v1.Pod{
- scheduledRunningReadyPod,
- scheduledRunningNotReadyPod,
- scheduledPendingPod,
- unscheduledPendingPod,
- },
- },
- {
- "scheduled vs unscheduled, diff < len(pods)",
- []*v1.Pod{
- scheduledPendingPod,
- unscheduledPendingPod,
- },
- 1,
- []*v1.Pod{
- unscheduledPendingPod,
- },
- },
- {
- "ready vs not-ready, diff < len(pods)",
- []*v1.Pod{
- scheduledRunningReadyPod,
- scheduledRunningNotReadyPod,
- scheduledRunningNotReadyPod,
- },
- 2,
- []*v1.Pod{
- scheduledRunningNotReadyPod,
- scheduledRunningNotReadyPod,
- },
- },
- {
- "pending vs running, diff < len(pods)",
- []*v1.Pod{
- scheduledPendingPod,
- scheduledRunningNotReadyPod,
- },
- 1,
- []*v1.Pod{
- scheduledPendingPod,
- },
- },
- {
- "various pod phases and conditions, diff < len(pods)",
- []*v1.Pod{
- scheduledRunningReadyPod,
- scheduledRunningNotReadyPod,
- scheduledPendingPod,
- unscheduledPendingPod,
- },
- 3,
- []*v1.Pod{
- unscheduledPendingPod,
- scheduledPendingPod,
- scheduledRunningNotReadyPod,
- },
- },
- }
- for _, test := range tests {
- podsToDelete := getPodsToDelete(test.pods, test.diff)
- if len(podsToDelete) != len(test.expectedPodsToDelete) {
- t.Errorf("%s: unexpected pods to delete, expected %v, got %v", test.name, test.expectedPodsToDelete, podsToDelete)
- }
- if !reflect.DeepEqual(podsToDelete, test.expectedPodsToDelete) {
- t.Errorf("%s: unexpected pods to delete, expected %v, got %v", test.name, test.expectedPodsToDelete, podsToDelete)
- }
- }
- }
- func TestGetPodKeys(t *testing.T) {
- labelMap := map[string]string{"name": "foo"}
- rs := newReplicaSet(1, labelMap)
- pod1 := newPod("pod1", rs, v1.PodRunning, nil, true)
- pod2 := newPod("pod2", rs, v1.PodRunning, nil, true)
- tests := []struct {
- name string
- pods []*v1.Pod
- expectedPodKeys []string
- }{
- {
- "len(pods) = 0 (i.e., pods = nil)",
- []*v1.Pod{},
- []string{},
- },
- {
- "len(pods) > 0",
- []*v1.Pod{
- pod1,
- pod2,
- },
- []string{"default/pod1", "default/pod2"},
- },
- }
- for _, test := range tests {
- podKeys := getPodKeys(test.pods)
- if len(podKeys) != len(test.expectedPodKeys) {
- t.Errorf("%s: unexpected keys for pods to delete, expected %v, got %v", test.name, test.expectedPodKeys, podKeys)
- }
- for i := 0; i < len(podKeys); i++ {
- if podKeys[i] != test.expectedPodKeys[i] {
- t.Errorf("%s: unexpected keys for pods to delete, expected %v, got %v", test.name, test.expectedPodKeys, podKeys)
- }
- }
- }
- }
|