123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214 |
- /*
- Copyright 2018 The Kubernetes Authors.
- Licensed under the Apache License, Version 2.0 (the "License");
- you may not use this file except in compliance with the License.
- You may obtain a copy of the License at
- http://www.apache.org/licenses/LICENSE-2.0
- Unless required by applicable law or agreed to in writing, software
- distributed under the License is distributed on an "AS IS" BASIS,
- WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- See the License for the specific language governing permissions and
- limitations under the License.
- */
- package common
- import (
- "context"
- "fmt"
- "time"
- coordinationv1 "k8s.io/api/coordination/v1"
- v1 "k8s.io/api/core/v1"
- apiequality "k8s.io/apimachinery/pkg/api/equality"
- metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
- "k8s.io/apimachinery/pkg/util/diff"
- "k8s.io/apimachinery/pkg/util/wait"
- clientset "k8s.io/client-go/kubernetes"
- "k8s.io/kubernetes/test/e2e/framework"
- e2enode "k8s.io/kubernetes/test/e2e/framework/node"
- testutils "k8s.io/kubernetes/test/utils"
- "github.com/onsi/ginkgo"
- "github.com/onsi/gomega"
- )
- var _ = framework.KubeDescribe("NodeLease", func() {
- var nodeName string
- f := framework.NewDefaultFramework("node-lease-test")
- ginkgo.BeforeEach(func() {
- node, err := e2enode.GetRandomReadySchedulableNode(f.ClientSet)
- framework.ExpectNoError(err)
- nodeName = node.Name
- })
- ginkgo.Context("when the NodeLease feature is enabled", func() {
- ginkgo.It("the kubelet should create and update a lease in the kube-node-lease namespace", func() {
- leaseClient := f.ClientSet.CoordinationV1().Leases(v1.NamespaceNodeLease)
- var (
- err error
- lease *coordinationv1.Lease
- )
- ginkgo.By("check that lease for this Kubelet exists in the kube-node-lease namespace")
- gomega.Eventually(func() error {
- lease, err = leaseClient.Get(context.TODO(), nodeName, metav1.GetOptions{})
- if err != nil {
- return err
- }
- return nil
- }, 5*time.Minute, 5*time.Second).Should(gomega.BeNil())
- // check basic expectations for the lease
- gomega.Expect(expectLease(lease, nodeName)).To(gomega.BeNil())
- ginkgo.By("check that node lease is updated at least once within the lease duration")
- gomega.Eventually(func() error {
- newLease, err := leaseClient.Get(context.TODO(), nodeName, metav1.GetOptions{})
- if err != nil {
- return err
- }
- // check basic expectations for the latest lease
- if err := expectLease(newLease, nodeName); err != nil {
- return err
- }
- // check that RenewTime has been updated on the latest lease
- newTime := (*newLease.Spec.RenewTime).Time
- oldTime := (*lease.Spec.RenewTime).Time
- if !newTime.After(oldTime) {
- return fmt.Errorf("new lease has time %v, which is not after old lease time %v", newTime, oldTime)
- }
- return nil
- }, time.Duration(*lease.Spec.LeaseDurationSeconds)*time.Second,
- time.Duration(*lease.Spec.LeaseDurationSeconds/4)*time.Second)
- })
- ginkgo.It("should have OwnerReferences set", func() {
- leaseClient := f.ClientSet.CoordinationV1().Leases(v1.NamespaceNodeLease)
- var (
- err error
- leaseList *coordinationv1.LeaseList
- )
- gomega.Eventually(func() error {
- leaseList, err = leaseClient.List(context.TODO(), metav1.ListOptions{})
- if err != nil {
- return err
- }
- return nil
- }, 5*time.Minute, 5*time.Second).Should(gomega.BeNil())
- // All the leases should have OwnerReferences set to their corresponding
- // Node object.
- for i := range leaseList.Items {
- lease := &leaseList.Items[i]
- ownerRefs := lease.ObjectMeta.OwnerReferences
- framework.ExpectEqual(len(ownerRefs), 1)
- framework.ExpectEqual(ownerRefs[0].Kind, v1.SchemeGroupVersion.WithKind("Node").Kind)
- framework.ExpectEqual(ownerRefs[0].APIVersion, v1.SchemeGroupVersion.WithKind("Node").Version)
- }
- })
- ginkgo.It("the kubelet should report node status infrequently", func() {
- ginkgo.By("wait until node is ready")
- e2enode.WaitForNodeToBeReady(f.ClientSet, nodeName, 5*time.Minute)
- ginkgo.By("wait until there is node lease")
- var err error
- var lease *coordinationv1.Lease
- gomega.Eventually(func() error {
- lease, err = f.ClientSet.CoordinationV1().Leases(v1.NamespaceNodeLease).Get(context.TODO(), nodeName, metav1.GetOptions{})
- if err != nil {
- return err
- }
- return nil
- }, 5*time.Minute, 5*time.Second).Should(gomega.BeNil())
- // check basic expectations for the lease
- gomega.Expect(expectLease(lease, nodeName)).To(gomega.BeNil())
- leaseDuration := time.Duration(*lease.Spec.LeaseDurationSeconds) * time.Second
- ginkgo.By("verify NodeStatus report period is longer than lease duration")
- // NodeStatus is reported from node to master when there is some change or
- // enough time has passed. So for here, keep checking the time diff
- // between 2 NodeStatus report, until it is longer than lease duration
- // (the same as nodeMonitorGracePeriod), or it doesn't change for at least leaseDuration
- lastHeartbeatTime, lastStatus := getHeartbeatTimeAndStatus(f.ClientSet, nodeName)
- lastObserved := time.Now()
- err = wait.Poll(time.Second, 5*time.Minute, func() (bool, error) {
- currentHeartbeatTime, currentStatus := getHeartbeatTimeAndStatus(f.ClientSet, nodeName)
- currentObserved := time.Now()
- if currentHeartbeatTime == lastHeartbeatTime {
- if currentObserved.Sub(lastObserved) > 2*leaseDuration {
- // heartbeat hasn't changed while watching for at least 2*leaseDuration, success!
- framework.Logf("node status heartbeat is unchanged for %s, was waiting for at least %s, success!", currentObserved.Sub(lastObserved), 2*leaseDuration)
- return true, nil
- }
- framework.Logf("node status heartbeat is unchanged for %s, waiting for %s", currentObserved.Sub(lastObserved), 2*leaseDuration)
- return false, nil
- }
- if currentHeartbeatTime.Sub(lastHeartbeatTime) >= leaseDuration {
- // heartbeat time changed, but the diff was greater than leaseDuration, success!
- framework.Logf("node status heartbeat changed in %s, was waiting for at least %s, success!", currentHeartbeatTime.Sub(lastHeartbeatTime), leaseDuration)
- return true, nil
- }
- if !apiequality.Semantic.DeepEqual(lastStatus, currentStatus) {
- // heartbeat time changed, but there were relevant changes in the status, keep waiting
- framework.Logf("node status heartbeat changed in %s (with other status changes), waiting for %s", currentHeartbeatTime.Sub(lastHeartbeatTime), leaseDuration)
- framework.Logf("%s", diff.ObjectReflectDiff(lastStatus, currentStatus))
- lastHeartbeatTime = currentHeartbeatTime
- lastObserved = currentObserved
- lastStatus = currentStatus
- return false, nil
- }
- // heartbeat time changed, with no other status changes, in less time than we expected, so fail.
- return false, fmt.Errorf("node status heartbeat changed in %s (with no other status changes), was waiting for %s", currentHeartbeatTime.Sub(lastHeartbeatTime), leaseDuration)
- })
- // a timeout is acceptable, since it means we waited 5 minutes and didn't see any unwarranted node status updates
- if err != nil && err != wait.ErrWaitTimeout {
- framework.ExpectNoError(err, "error waiting for infrequent nodestatus update")
- }
- ginkgo.By("verify node is still in ready status even though node status report is infrequent")
- // This check on node status is only meaningful when this e2e test is
- // running as cluster e2e test, because node e2e test does not create and
- // run controller manager, i.e., no node lifecycle controller.
- node, err := f.ClientSet.CoreV1().Nodes().Get(context.TODO(), nodeName, metav1.GetOptions{})
- gomega.Expect(err).To(gomega.BeNil())
- _, readyCondition := testutils.GetNodeCondition(&node.Status, v1.NodeReady)
- framework.ExpectEqual(readyCondition.Status, v1.ConditionTrue)
- })
- })
- })
- func getHeartbeatTimeAndStatus(clientSet clientset.Interface, nodeName string) (time.Time, v1.NodeStatus) {
- node, err := clientSet.CoreV1().Nodes().Get(context.TODO(), nodeName, metav1.GetOptions{})
- gomega.Expect(err).To(gomega.BeNil())
- _, readyCondition := testutils.GetNodeCondition(&node.Status, v1.NodeReady)
- framework.ExpectEqual(readyCondition.Status, v1.ConditionTrue)
- heartbeatTime := readyCondition.LastHeartbeatTime.Time
- readyCondition.LastHeartbeatTime = metav1.Time{}
- return heartbeatTime, node.Status
- }
- func expectLease(lease *coordinationv1.Lease, nodeName string) error {
- // expect values for HolderIdentity, LeaseDurationSeconds, and RenewTime
- if lease.Spec.HolderIdentity == nil {
- return fmt.Errorf("Spec.HolderIdentity should not be nil")
- }
- if lease.Spec.LeaseDurationSeconds == nil {
- return fmt.Errorf("Spec.LeaseDurationSeconds should not be nil")
- }
- if lease.Spec.RenewTime == nil {
- return fmt.Errorf("Spec.RenewTime should not be nil")
- }
- // ensure that the HolderIdentity matches the node name
- if *lease.Spec.HolderIdentity != nodeName {
- return fmt.Errorf("Spec.HolderIdentity (%v) should match the node name (%v)", *lease.Spec.HolderIdentity, nodeName)
- }
- return nil
- }
|