core.sh 79 KB

12345678910111213141516171819202122232425262728293031323334353637383940414243444546474849505152535455565758596061626364656667686970717273747576777879808182838485868788899091929394959697989910010110210310410510610710810911011111211311411511611711811912012112212312412512612712812913013113213313413513613713813914014114214314414514614714814915015115215315415515615715815916016116216316416516616716816917017117217317417517617717817918018118218318418518618718818919019119219319419519619719819920020120220320420520620720820921021121221321421521621721821922022122222322422522622722822923023123223323423523623723823924024124224324424524624724824925025125225325425525625725825926026126226326426526626726826927027127227327427527627727827928028128228328428528628728828929029129229329429529629729829930030130230330430530630730830931031131231331431531631731831932032132232332432532632732832933033133233333433533633733833934034134234334434534634734834935035135235335435535635735835936036136236336436536636736836937037137237337437537637737837938038138238338438538638738838939039139239339439539639739839940040140240340440540640740840941041141241341441541641741841942042142242342442542642742842943043143243343443543643743843944044144244344444544644744844945045145245345445545645745845946046146246346446546646746846947047147247347447547647747847948048148248348448548648748848949049149249349449549649749849950050150250350450550650750850951051151251351451551651751851952052152252352452552652752852953053153253353453553653753853954054154254354454554654754854955055155255355455555655755855956056156256356456556656756856957057157257357457557657757857958058158258358458558658758858959059159259359459559659759859960060160260360460560660760860961061161261361461561661761861962062162262362462562662762862963063163263363463563663763863964064164264364464564664764864965065165265365465565665765865966066166266366466566666766866967067167267367467567667767867968068168268368468568668768868969069169269369469569669769869970070170270370470570670770870971071171271371471571671771871972072172272372472572672772872973073173273373473573673773873974074174274374474574674774874975075175275375475575675775875976076176276376476576676776876977077177277377477577677777877978078178278378478578678778878979079179279379479579679779879980080180280380480580680780880981081181281381481581681781881982082182282382482582682782882983083183283383483583683783883984084184284384484584684784884985085185285385485585685785885986086186286386486586686786886987087187287387487587687787887988088188288388488588688788888989089189289389489589689789889990090190290390490590690790890991091191291391491591691791891992092192292392492592692792892993093193293393493593693793893994094194294394494594694794894995095195295395495595695795895996096196296396496596696796896997097197297397497597697797897998098198298398498598698798898999099199299399499599699799899910001001100210031004100510061007100810091010101110121013101410151016101710181019102010211022102310241025102610271028102910301031103210331034103510361037103810391040104110421043104410451046104710481049105010511052105310541055105610571058105910601061106210631064106510661067106810691070107110721073107410751076107710781079108010811082108310841085108610871088108910901091109210931094109510961097109810991100110111021103110411051106110711081109111011111112111311141115111611171118111911201121112211231124112511261127112811291130113111321133113411351136113711381139114011411142114311441145114611471148114911501151115211531154115511561157115811591160116111621163116411651166116711681169117011711172117311741175117611771178117911801181118211831184118511861187118811891190119111921193119411951196119711981199120012011202120312041205120612071208120912101211121212131214121512161217121812191220122112221223122412251226122712281229123012311232123312341235123612371238123912401241124212431244124512461247124812491250125112521253125412551256125712581259126012611262126312641265126612671268126912701271127212731274127512761277127812791280128112821283128412851286128712881289129012911292129312941295129612971298129913001301130213031304130513061307130813091310131113121313131413151316131713181319132013211322132313241325132613271328132913301331133213331334133513361337133813391340134113421343134413451346134713481349135013511352135313541355135613571358135913601361136213631364136513661367136813691370137113721373137413751376137713781379138013811382138313841385138613871388138913901391139213931394139513961397139813991400140114021403140414051406140714081409141014111412141314141415141614171418141914201421142214231424142514261427142814291430143114321433143414351436143714381439144014411442
  1. #!/usr/bin/env bash
  2. # Copyright 2018 The Kubernetes Authors.
  3. #
  4. # Licensed under the Apache License, Version 2.0 (the "License");
  5. # you may not use this file except in compliance with the License.
  6. # You may obtain a copy of the License at
  7. #
  8. # http://www.apache.org/licenses/LICENSE-2.0
  9. #
  10. # Unless required by applicable law or agreed to in writing, software
  11. # distributed under the License is distributed on an "AS IS" BASIS,
  12. # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
  13. # See the License for the specific language governing permissions and
  14. # limitations under the License.
  15. set -o errexit
  16. set -o nounset
  17. set -o pipefail
  18. run_configmap_tests() {
  19. set -o nounset
  20. set -o errexit
  21. create_and_use_new_namespace
  22. kube::log::status "Testing configmaps"
  23. kubectl create -f test/fixtures/doc-yaml/user-guide/configmap/configmap.yaml
  24. kube::test::get_object_assert 'configmap/test-configmap' "{{${id_field:?}}}" 'test-configmap'
  25. kubectl delete configmap test-configmap "${kube_flags[@]:?}"
  26. ### Create a new namespace
  27. # Pre-condition: the test-configmaps namespace does not exist
  28. kube::test::get_object_assert 'namespaces' "{{range.items}}{{ if eq $id_field \\\"test-configmaps\\\" }}found{{end}}{{end}}:" ':'
  29. # Command
  30. kubectl create namespace test-configmaps
  31. # Post-condition: namespace 'test-configmaps' is created.
  32. kube::test::get_object_assert 'namespaces/test-configmaps' "{{$id_field}}" 'test-configmaps'
  33. ### Create a generic configmap in a specific namespace
  34. # Pre-condition: configmap test-configmap and test-binary-configmap does not exist
  35. kube::test::get_object_assert 'configmaps' "{{range.items}}{{ if eq $id_field \\\"test-configmap\\\" }}found{{end}}{{end}}:" ':'
  36. kube::test::get_object_assert 'configmaps' "{{range.items}}{{ if eq $id_field \\\"test-binary-configmap\\\" }}found{{end}}{{end}}:" ':'
  37. # Command
  38. kubectl create configmap test-configmap --from-literal=key1=value1 --namespace=test-configmaps
  39. kubectl create configmap test-binary-configmap --from-file <( head -c 256 /dev/urandom ) --namespace=test-configmaps
  40. # Post-condition: configmap exists and has expected values
  41. kube::test::get_object_assert 'configmap/test-configmap --namespace=test-configmaps' "{{$id_field}}" 'test-configmap'
  42. kube::test::get_object_assert 'configmap/test-binary-configmap --namespace=test-configmaps' "{{$id_field}}" 'test-binary-configmap'
  43. grep -q "key1: value1" <<< "$(kubectl get configmap/test-configmap --namespace=test-configmaps -o yaml "${kube_flags[@]}")"
  44. grep -q "binaryData" <<< "$(kubectl get configmap/test-binary-configmap --namespace=test-configmaps -o yaml "${kube_flags[@]}")"
  45. # Clean-up
  46. kubectl delete configmap test-configmap --namespace=test-configmaps
  47. kubectl delete configmap test-binary-configmap --namespace=test-configmaps
  48. kubectl delete namespace test-configmaps
  49. set +o nounset
  50. set +o errexit
  51. }
  52. # Runs all pod related tests.
  53. run_pod_tests() {
  54. set -o nounset
  55. set -o errexit
  56. kube::log::status "Testing kubectl(v1:pods)"
  57. ### Create POD valid-pod from JSON
  58. # Pre-condition: no POD exists
  59. create_and_use_new_namespace
  60. kube::test::get_object_assert pods "{{range.items}}{{$id_field}}:{{end}}" ''
  61. # Command
  62. kubectl create "${kube_flags[@]}" -f test/fixtures/doc-yaml/admin/limitrange/valid-pod.yaml
  63. # Post-condition: valid-pod POD is created
  64. kubectl get "${kube_flags[@]}" pods -o json
  65. kube::test::get_object_assert pods "{{range.items}}{{$id_field}}:{{end}}" 'valid-pod:'
  66. kube::test::get_object_assert 'pod valid-pod' "{{$id_field}}" 'valid-pod'
  67. kube::test::get_object_assert 'pod/valid-pod' "{{$id_field}}" 'valid-pod'
  68. kube::test::get_object_assert 'pods/valid-pod' "{{$id_field}}" 'valid-pod'
  69. # Repeat above test using jsonpath template
  70. kube::test::get_object_jsonpath_assert pods "{.items[*]$id_field}" 'valid-pod'
  71. kube::test::get_object_jsonpath_assert 'pod valid-pod' "{$id_field}" 'valid-pod'
  72. kube::test::get_object_jsonpath_assert 'pod/valid-pod' "{$id_field}" 'valid-pod'
  73. kube::test::get_object_jsonpath_assert 'pods/valid-pod' "{$id_field}" 'valid-pod'
  74. # Describe command should print detailed information
  75. kube::test::describe_object_assert pods 'valid-pod' "Name:" "Image:" "Node:" "Labels:" "Status:"
  76. # Describe command should print events information by default
  77. kube::test::describe_object_events_assert pods 'valid-pod'
  78. # Describe command should not print events information when show-events=false
  79. kube::test::describe_object_events_assert pods 'valid-pod' false
  80. # Describe command should print events information when show-events=true
  81. kube::test::describe_object_events_assert pods 'valid-pod' true
  82. # Describe command (resource only) should print detailed information
  83. kube::test::describe_resource_assert pods "Name:" "Image:" "Node:" "Labels:" "Status:"
  84. # Describe command should print events information by default
  85. kube::test::describe_resource_events_assert pods
  86. # Describe command should not print events information when show-events=false
  87. kube::test::describe_resource_events_assert pods false
  88. # Describe command should print events information when show-events=true
  89. kube::test::describe_resource_events_assert pods true
  90. ### Validate Export ###
  91. kube::test::get_object_assert 'pods/valid-pod' "{{.metadata.namespace}} {{.metadata.name}}" '<no value> valid-pod' "--export=true"
  92. ### Dump current valid-pod POD
  93. output_pod=$(kubectl get pod valid-pod -o yaml "${kube_flags[@]}")
  94. ### Delete POD valid-pod by id
  95. # Pre-condition: valid-pod POD exists
  96. kube::test::get_object_assert pods "{{range.items}}{{$id_field}}:{{end}}" 'valid-pod:'
  97. # Command
  98. kubectl delete pod valid-pod "${kube_flags[@]}" --grace-period=0 --force
  99. # Post-condition: valid-pod POD doesn't exist
  100. kube::test::get_object_assert pods "{{range.items}}{{$id_field}}:{{end}}" ''
  101. ### Delete POD valid-pod by id with --now
  102. # Pre-condition: valid-pod POD exists
  103. kubectl create "${kube_flags[@]}" -f test/fixtures/doc-yaml/admin/limitrange/valid-pod.yaml
  104. kube::test::get_object_assert pods "{{range.items}}{{$id_field}}:{{end}}" 'valid-pod:'
  105. # Command
  106. kubectl delete pod valid-pod "${kube_flags[@]}" --now
  107. # Post-condition: valid-pod POD doesn't exist
  108. kube::test::get_object_assert pods "{{range.items}}{{$id_field}}:{{end}}" ''
  109. ### Delete POD valid-pod by id with --grace-period=0
  110. # Pre-condition: valid-pod POD exists
  111. kubectl create "${kube_flags[@]}" -f test/fixtures/doc-yaml/admin/limitrange/valid-pod.yaml
  112. kube::test::get_object_assert pods "{{range.items}}{{$id_field}}:{{end}}" 'valid-pod:'
  113. # Command succeeds without --force by waiting
  114. kubectl delete pod valid-pod "${kube_flags[@]}" --grace-period=0
  115. # Post-condition: valid-pod POD doesn't exist
  116. kube::test::get_object_assert pods "{{range.items}}{{$id_field}}:{{end}}" ''
  117. ### Create POD valid-pod from dumped YAML
  118. # Pre-condition: no POD exists
  119. create_and_use_new_namespace
  120. kube::test::get_object_assert pods "{{range.items}}{{$id_field}}:{{end}}" ''
  121. # Command
  122. echo "${output_pod}" | ${SED} '/namespace:/d' | kubectl create -f - "${kube_flags[@]}"
  123. # Post-condition: valid-pod POD is created
  124. kube::test::get_object_assert pods "{{range.items}}{{$id_field}}:{{end}}" 'valid-pod:'
  125. ### Delete POD valid-pod from JSON
  126. # Pre-condition: valid-pod POD exists
  127. kube::test::get_object_assert pods "{{range.items}}{{$id_field}}:{{end}}" 'valid-pod:'
  128. # Command
  129. kubectl delete -f test/fixtures/doc-yaml/admin/limitrange/valid-pod.yaml "${kube_flags[@]}" --grace-period=0 --force
  130. # Post-condition: valid-pod POD doesn't exist
  131. kube::test::get_object_assert pods "{{range.items}}{{$id_field}}:{{end}}" ''
  132. ### Create POD valid-pod from JSON
  133. # Pre-condition: no POD exists
  134. create_and_use_new_namespace
  135. kube::test::get_object_assert pods "{{range.items}}{{$id_field}}:{{end}}" ''
  136. # Command
  137. kubectl create -f test/fixtures/doc-yaml/admin/limitrange/valid-pod.yaml "${kube_flags[@]}"
  138. # Post-condition: valid-pod POD is created
  139. kube::test::get_object_assert pods "{{range.items}}{{$id_field}}:{{end}}" 'valid-pod:'
  140. ### Delete POD valid-pod with label
  141. # Pre-condition: valid-pod POD exists
  142. kube::test::get_object_assert "pods -l'name in (valid-pod)'" "{{range.items}}{{$id_field}}:{{end}}" 'valid-pod:'
  143. # Command
  144. kubectl delete pods -l'name in (valid-pod)' "${kube_flags[@]}" --grace-period=0 --force
  145. # Post-condition: valid-pod POD doesn't exist
  146. kube::test::get_object_assert "pods -l'name in (valid-pod)'" "{{range.items}}{{$id_field}}:{{end}}" ''
  147. ### Create POD valid-pod from YAML
  148. # Pre-condition: no POD exists
  149. create_and_use_new_namespace
  150. kube::test::get_object_assert pods "{{range.items}}{{$id_field}}:{{end}}" ''
  151. # Command
  152. kubectl create -f test/fixtures/doc-yaml/admin/limitrange/valid-pod.yaml "${kube_flags[@]}"
  153. # Post-condition: valid-pod POD is created
  154. kube::test::get_object_assert pods "{{range.items}}{{$id_field}}:{{end}}" 'valid-pod:'
  155. # Command
  156. output_message=$(kubectl get pods --field-selector metadata.name=valid-pod "${kube_flags[@]}")
  157. kube::test::if_has_string "${output_message}" "valid-pod"
  158. # Command
  159. phase=$(kubectl get "${kube_flags[@]}" pod valid-pod -o go-template='{{ .status.phase }}')
  160. output_message=$(kubectl get pods --field-selector status.phase="${phase}" "${kube_flags[@]}")
  161. kube::test::if_has_string "${output_message}" "valid-pod"
  162. ### Delete PODs with no parameter mustn't kill everything
  163. # Pre-condition: valid-pod POD exists
  164. kube::test::get_object_assert pods "{{range.items}}{{$id_field}}:{{end}}" 'valid-pod:'
  165. # Command
  166. ! kubectl delete pods "${kube_flags[@]}" || exit 1
  167. # Post-condition: valid-pod POD exists
  168. kube::test::get_object_assert pods "{{range.items}}{{$id_field}}:{{end}}" 'valid-pod:'
  169. ### Delete PODs with --all and a label selector is not permitted
  170. # Pre-condition: valid-pod POD exists
  171. kube::test::get_object_assert pods "{{range.items}}{{$id_field}}:{{end}}" 'valid-pod:'
  172. # Command
  173. ! kubectl delete --all pods -l'name in (valid-pod)' "${kube_flags[@]}" || exit 1
  174. # Post-condition: valid-pod POD exists
  175. kube::test::get_object_assert pods "{{range.items}}{{$id_field}}:{{end}}" 'valid-pod:'
  176. ### Delete all PODs
  177. # Pre-condition: valid-pod POD exists
  178. kube::test::get_object_assert pods "{{range.items}}{{$id_field}}:{{end}}" 'valid-pod:'
  179. # Command
  180. kubectl delete --all pods "${kube_flags[@]}" --grace-period=0 --force # --all remove all the pods
  181. # Post-condition: no POD exists
  182. kube::test::get_object_assert "pods -l'name in (valid-pod)'" "{{range.items}}{{$id_field}}:{{end}}" ''
  183. # Detailed tests for describe pod output
  184. ### Create a new namespace
  185. # Pre-condition: the test-secrets namespace does not exist
  186. kube::test::get_object_assert 'namespaces' "{{range.items}}{{ if eq $id_field \\\"test-kubectl-describe-pod\\\" }}found{{end}}{{end}}:" ':'
  187. # Command
  188. kubectl create namespace test-kubectl-describe-pod
  189. # Post-condition: namespace 'test-secrets' is created.
  190. kube::test::get_object_assert 'namespaces/test-kubectl-describe-pod' "{{$id_field}}" 'test-kubectl-describe-pod'
  191. ### Create a generic secret
  192. # Pre-condition: no SECRET exists
  193. kube::test::get_object_assert 'secrets --namespace=test-kubectl-describe-pod' "{{range.items}}{{$id_field}}:{{end}}" ''
  194. # Command
  195. kubectl create secret generic test-secret --from-literal=key-1=value1 --type=test-type --namespace=test-kubectl-describe-pod
  196. # Post-condition: secret exists and has expected values
  197. kube::test::get_object_assert 'secret/test-secret --namespace=test-kubectl-describe-pod' "{{$id_field}}" 'test-secret'
  198. kube::test::get_object_assert 'secret/test-secret --namespace=test-kubectl-describe-pod' "{{${secret_type:?}}}" 'test-type'
  199. ### Create a generic configmap
  200. # Pre-condition: CONFIGMAP test-configmap does not exist
  201. #kube::test::get_object_assert 'configmap/test-configmap --namespace=test-kubectl-describe-pod' "{{$id_field}}" ''
  202. kube::test::get_object_assert 'configmaps --namespace=test-kubectl-describe-pod' "{{range.items}}{{ if eq $id_field \\\"test-configmap\\\" }}found{{end}}{{end}}:" ':'
  203. #kube::test::get_object_assert 'configmaps --namespace=test-kubectl-describe-pod' "{{range.items}}{{$id_field}}:{{end}}" ''
  204. # Command
  205. kubectl create configmap test-configmap --from-literal=key-2=value2 --namespace=test-kubectl-describe-pod
  206. # Post-condition: configmap exists and has expected values
  207. kube::test::get_object_assert 'configmap/test-configmap --namespace=test-kubectl-describe-pod' "{{$id_field}}" 'test-configmap'
  208. ### Create a pod disruption budget with minAvailable
  209. # Command
  210. kubectl create pdb test-pdb-1 --selector=app=rails --min-available=2 --namespace=test-kubectl-describe-pod
  211. # Post-condition: pdb exists and has expected values
  212. kube::test::get_object_assert 'pdb/test-pdb-1 --namespace=test-kubectl-describe-pod' "{{${pdb_min_available:?}}}" '2'
  213. # Command
  214. kubectl create pdb test-pdb-2 --selector=app=rails --min-available=50% --namespace=test-kubectl-describe-pod
  215. # Post-condition: pdb exists and has expected values
  216. kube::test::get_object_assert 'pdb/test-pdb-2 --namespace=test-kubectl-describe-pod' "{{$pdb_min_available}}" '50%'
  217. ### Create a pod disruption budget with maxUnavailable
  218. # Command
  219. kubectl create pdb test-pdb-3 --selector=app=rails --max-unavailable=2 --namespace=test-kubectl-describe-pod
  220. # Post-condition: pdb exists and has expected values
  221. kube::test::get_object_assert 'pdb/test-pdb-3 --namespace=test-kubectl-describe-pod' "{{${pdb_max_unavailable:?}}}" '2'
  222. # Command
  223. kubectl create pdb test-pdb-4 --selector=app=rails --max-unavailable=50% --namespace=test-kubectl-describe-pod
  224. # Post-condition: pdb exists and has expected values
  225. kube::test::get_object_assert 'pdb/test-pdb-4 --namespace=test-kubectl-describe-pod' "{{$pdb_max_unavailable}}" '50%'
  226. ### Fail creating a pod disruption budget if both maxUnavailable and minAvailable specified
  227. ! kubectl create pdb test-pdb --selector=app=rails --min-available=2 --max-unavailable=3 --namespace=test-kubectl-describe-pod || exit 1
  228. # Create a pod that consumes secret, configmap, and downward API keys as envs
  229. kube::test::get_object_assert 'pods --namespace=test-kubectl-describe-pod' "{{range.items}}{{$id_field}}:{{end}}" ''
  230. kubectl create -f hack/testdata/pod-with-api-env.yaml --namespace=test-kubectl-describe-pod
  231. kube::test::describe_object_assert 'pods --namespace=test-kubectl-describe-pod' 'env-test-pod' "TEST_CMD_1" "<set to the key 'key-1' in secret 'test-secret'>" "TEST_CMD_2" "<set to the key 'key-2' of config map 'test-configmap'>" "TEST_CMD_3" "env-test-pod (v1:metadata.name)"
  232. # Describe command (resource only) should print detailed information about environment variables
  233. kube::test::describe_resource_assert 'pods --namespace=test-kubectl-describe-pod' "TEST_CMD_1" "<set to the key 'key-1' in secret 'test-secret'>" "TEST_CMD_2" "<set to the key 'key-2' of config map 'test-configmap'>" "TEST_CMD_3" "env-test-pod (v1:metadata.name)"
  234. # Clean-up
  235. kubectl delete pod env-test-pod --namespace=test-kubectl-describe-pod
  236. kubectl delete secret test-secret --namespace=test-kubectl-describe-pod
  237. kubectl delete configmap test-configmap --namespace=test-kubectl-describe-pod
  238. kubectl delete pdb/test-pdb-1 pdb/test-pdb-2 pdb/test-pdb-3 pdb/test-pdb-4 --namespace=test-kubectl-describe-pod
  239. kubectl delete namespace test-kubectl-describe-pod
  240. ### Create two PODs
  241. # Pre-condition: no POD exists
  242. create_and_use_new_namespace
  243. kube::test::get_object_assert pods "{{range.items}}{{$id_field}}:{{end}}" ''
  244. # Command
  245. kubectl create -f test/fixtures/doc-yaml/admin/limitrange/valid-pod.yaml "${kube_flags[@]}"
  246. kubectl create -f test/e2e/testing-manifests/kubectl/agnhost-master-pod.yaml "${kube_flags[@]}"
  247. # Post-condition: valid-pod and agnhost-master PODs are created
  248. kube::test::get_object_assert pods "{{range.items}}{{$id_field}}:{{end}}" 'agnhost-master:valid-pod:'
  249. ### Delete multiple PODs at once
  250. # Pre-condition: valid-pod and agnhost-master PODs exist
  251. kube::test::get_object_assert pods "{{range.items}}{{$id_field}}:{{end}}" 'agnhost-master:valid-pod:'
  252. # Command
  253. kubectl delete pods valid-pod agnhost-master "${kube_flags[@]}" --grace-period=0 --force # delete multiple pods at once
  254. # Post-condition: no POD exists
  255. kube::test::get_object_assert pods "{{range.items}}{{$id_field}}:{{end}}" ''
  256. ### Create valid-pod POD
  257. # Pre-condition: no POD exists
  258. create_and_use_new_namespace
  259. kube::test::get_object_assert pods "{{range.items}}{{$id_field}}:{{end}}" ''
  260. # Command
  261. kubectl create -f test/fixtures/doc-yaml/admin/limitrange/valid-pod.yaml "${kube_flags[@]}"
  262. # Post-condition: valid-pod POD is created
  263. kube::test::get_object_assert pods "{{range.items}}{{$id_field}}:{{end}}" 'valid-pod:'
  264. ### Label the valid-pod POD
  265. # Pre-condition: valid-pod is not labelled
  266. kube::test::get_object_assert 'pod valid-pod' "{{range${labels_field:?}}}{{.}}:{{end}}" 'valid-pod:'
  267. # Command
  268. kubectl label pods valid-pod new-name=new-valid-pod "${kube_flags[@]}"
  269. # Post-condition: valid-pod is labelled
  270. kube::test::get_object_assert 'pod valid-pod' "{{range$labels_field}}{{.}}:{{end}}" 'valid-pod:new-valid-pod:'
  271. ### Label the valid-pod POD with empty label value
  272. # Pre-condition: valid-pod does not have label "emptylabel"
  273. kube::test::get_object_assert 'pod valid-pod' "{{range$labels_field}}{{.}}:{{end}}" 'valid-pod:new-valid-pod:'
  274. # Command
  275. kubectl label pods valid-pod emptylabel="" "${kube_flags[@]}"
  276. # Post-condition: valid pod contains "emptylabel" with no value
  277. kube::test::get_object_assert 'pod valid-pod' "{{${labels_field}.emptylabel}}" ''
  278. ### Annotate the valid-pod POD with empty annotation value
  279. # Pre-condition: valid-pod does not have annotation "emptyannotation"
  280. kube::test::get_object_assert 'pod valid-pod' "{{${annotations_field:?}.emptyannotation}}" '<no value>'
  281. # Command
  282. kubectl annotate pods valid-pod emptyannotation="" "${kube_flags[@]}"
  283. # Post-condition: valid pod contains "emptyannotation" with no value
  284. kube::test::get_object_assert 'pod valid-pod' "{{${annotations_field}.emptyannotation}}" ''
  285. ### Record label change
  286. # Pre-condition: valid-pod does not have record annotation
  287. kube::test::get_object_assert 'pod valid-pod' "{{range.items}}{{$annotations_field}}:{{end}}" ''
  288. # Command
  289. kubectl label pods valid-pod record-change=true --record=true "${kube_flags[@]}"
  290. # Post-condition: valid-pod has record annotation
  291. kube::test::get_object_assert 'pod valid-pod' "{{range$annotations_field}}{{.}}:{{end}}" ".*--record=true.*"
  292. ### Do not record label change
  293. # Command
  294. kubectl label pods valid-pod no-record-change=true --record=false "${kube_flags[@]}"
  295. # Post-condition: valid-pod's record annotation still contains command with --record=true
  296. kube::test::get_object_assert 'pod valid-pod' "{{range$annotations_field}}{{.}}:{{end}}" ".*--record=true.*"
  297. ### Record label change with specified flag and previous change already recorded
  298. ### we are no longer tricked by data from another user into revealing more information about our client
  299. # Command
  300. kubectl label pods valid-pod new-record-change=true --record=true "${kube_flags[@]}"
  301. # Post-condition: valid-pod's record annotation contains new change
  302. kube::test::get_object_assert 'pod valid-pod' "{{range$annotations_field}}{{.}}:{{end}}" ".*new-record-change=true.*"
  303. ### Delete POD by label
  304. # Pre-condition: valid-pod POD exists
  305. kube::test::get_object_assert pods "{{range.items}}{{$id_field}}:{{end}}" 'valid-pod:'
  306. # Command
  307. kubectl delete pods -lnew-name=new-valid-pod --grace-period=0 --force "${kube_flags[@]}"
  308. # Post-condition: valid-pod POD doesn't exist
  309. kube::test::get_object_assert pods "{{range.items}}{{$id_field}}:{{end}}" ''
  310. ### Create pod-with-precision POD
  311. # Pre-condition: no POD is running
  312. kube::test::get_object_assert pods "{{range.items}}{{$id_field}}:{{end}}" ''
  313. # Command
  314. kubectl create -f hack/testdata/pod-with-precision.json "${kube_flags[@]}"
  315. # Post-condition: valid-pod POD is running
  316. kube::test::get_object_assert pods "{{range.items}}{{$id_field}}:{{end}}" 'pod-with-precision:'
  317. ## Patch preserves precision
  318. # Command
  319. kubectl patch "${kube_flags[@]}" pod pod-with-precision -p='{"metadata":{"annotations":{"patchkey": "patchvalue"}}}'
  320. # Post-condition: pod-with-precision POD has patched annotation
  321. kube::test::get_object_assert 'pod pod-with-precision' "{{${annotations_field}.patchkey}}" 'patchvalue'
  322. # Command
  323. kubectl label pods pod-with-precision labelkey=labelvalue "${kube_flags[@]}"
  324. # Post-condition: pod-with-precision POD has label
  325. kube::test::get_object_assert 'pod pod-with-precision' "{{${labels_field}.labelkey}}" 'labelvalue'
  326. # Command
  327. kubectl annotate pods pod-with-precision annotatekey=annotatevalue "${kube_flags[@]}"
  328. # Post-condition: pod-with-precision POD has annotation
  329. kube::test::get_object_assert 'pod pod-with-precision' "{{${annotations_field}.annotatekey}}" 'annotatevalue'
  330. # Cleanup
  331. kubectl delete pod pod-with-precision "${kube_flags[@]}"
  332. ### Annotate POD YAML file locally without effecting the live pod.
  333. kubectl create -f hack/testdata/pod.yaml "${kube_flags[@]}"
  334. # Command
  335. kubectl annotate -f hack/testdata/pod.yaml annotatekey=annotatevalue "${kube_flags[@]}"
  336. # Pre-condition: annotationkey is annotationvalue
  337. kube::test::get_object_assert 'pod test-pod' "{{${annotations_field}.annotatekey}}" 'annotatevalue'
  338. # Command
  339. output_message=$(kubectl annotate --local -f hack/testdata/pod.yaml annotatekey=localvalue -o yaml "${kube_flags[@]}")
  340. echo "$output_message"
  341. # Post-condition: annotationkey is still annotationvalue in the live pod, but command output is the new value
  342. kube::test::get_object_assert 'pod test-pod' "{{${annotations_field}.annotatekey}}" 'annotatevalue'
  343. kube::test::if_has_string "${output_message}" "localvalue"
  344. # Cleanup
  345. kubectl delete -f hack/testdata/pod.yaml "${kube_flags[@]}"
  346. ### Create valid-pod POD
  347. # Pre-condition: no services and no rcs exist
  348. kube::test::get_object_assert service "{{range.items}}{{$id_field}}:{{end}}" ''
  349. kube::test::get_object_assert rc "{{range.items}}{{$id_field}}:{{end}}" ''
  350. ## kubectl create --edit can update the label filed of multiple resources. tmp-editor.sh is a fake editor
  351. TEMP=$(mktemp /tmp/tmp-editor-XXXXXXXX.sh)
  352. echo -e "#!/usr/bin/env bash\n${SED} -i \"s/mock/modified/g\" \$1" > "${TEMP}"
  353. chmod +x "${TEMP}"
  354. # Command
  355. EDITOR=${TEMP} kubectl create --edit -f hack/testdata/multi-resource-json.json "${kube_flags[@]}"
  356. # Post-condition: service named modified and rc named modified are created
  357. kube::test::get_object_assert service "{{range.items}}{{$id_field}}:{{end}}" 'modified:'
  358. kube::test::get_object_assert rc "{{range.items}}{{$id_field}}:{{end}}" 'modified:'
  359. # Clean up
  360. kubectl delete service/modified "${kube_flags[@]}"
  361. kubectl delete rc/modified "${kube_flags[@]}"
  362. # Pre-condition: no services and no rcs exist
  363. kube::test::get_object_assert service "{{range.items}}{{$id_field}}:{{end}}" ''
  364. kube::test::get_object_assert rc "{{range.items}}{{$id_field}}:{{end}}" ''
  365. # Command
  366. EDITOR=${TEMP} kubectl create --edit -f hack/testdata/multi-resource-list.json "${kube_flags[@]}"
  367. # Post-condition: service named modified and rc named modified are created
  368. kube::test::get_object_assert service "{{range.items}}{{$id_field}}:{{end}}" 'modified:'
  369. kube::test::get_object_assert rc "{{range.items}}{{$id_field}}:{{end}}" 'modified:'
  370. # Clean up
  371. rm "${TEMP}"
  372. kubectl delete service/modified "${kube_flags[@]}"
  373. kubectl delete rc/modified "${kube_flags[@]}"
  374. ## kubectl create --edit won't create anything if user makes no changes
  375. grep -q 'Edit cancelled' <<< "$(EDITOR="cat" kubectl create --edit -f test/fixtures/doc-yaml/admin/limitrange/valid-pod.yaml -o json 2>&1)"
  376. ## Create valid-pod POD
  377. # Pre-condition: no POD exists
  378. kube::test::wait_object_assert pods "{{range.items}}{{$id_field}}:{{end}}" ''
  379. # Command
  380. kubectl create -f test/fixtures/doc-yaml/admin/limitrange/valid-pod.yaml "${kube_flags[@]}"
  381. # Post-condition: valid-pod POD is created
  382. kube::test::get_object_assert pods "{{range.items}}{{$id_field}}:{{end}}" 'valid-pod:'
  383. ## Patch can modify a local object
  384. kubectl patch --local -f test/fixtures/pkg/kubectl/cmd/patch/validPod.yaml --patch='{"spec": {"restartPolicy":"Never"}}' -o yaml | grep -q "Never"
  385. ## Patch fails with type restore error and exit code 1
  386. output_message=$(! kubectl patch "${kube_flags[@]}" pod valid-pod -p='{"metadata":{"labels":"invalid"}}' 2>&1)
  387. kube::test::if_has_string "${output_message}" 'cannot restore map from string'
  388. ## Patch exits with error message "patched (no change)" and exit code 0 when no-op occurs
  389. output_message=$(kubectl patch "${kube_flags[@]}" pod valid-pod -p='{"metadata":{"labels":{"name":"valid-pod"}}}' 2>&1)
  390. kube::test::if_has_string "${output_message}" 'patched (no change)'
  391. ## Patch pod can change image
  392. # Command
  393. kubectl patch "${kube_flags[@]}" pod valid-pod --record -p='{"spec":{"containers":[{"name": "kubernetes-serve-hostname", "image": "nginx"}]}}'
  394. # Post-condition: valid-pod POD has image nginx
  395. kube::test::get_object_assert pods "{{range.items}}{{${image_field:?}}}:{{end}}" 'nginx:'
  396. # Post-condition: valid-pod has the record annotation
  397. kube::test::get_object_assert pods "{{range.items}}{{$annotations_field}}:{{end}}" "${change_cause_annotation:?}"
  398. # prove that patch can use different types
  399. kubectl patch "${kube_flags[@]}" pod valid-pod --type="json" -p='[{"op": "replace", "path": "/spec/containers/0/image", "value":"nginx2"}]'
  400. # Post-condition: valid-pod POD has image nginx
  401. kube::test::get_object_assert pods "{{range.items}}{{$image_field}}:{{end}}" 'nginx2:'
  402. # prove that patch can use different types
  403. kubectl patch "${kube_flags[@]}" pod valid-pod --type="json" -p='[{"op": "replace", "path": "/spec/containers/0/image", "value":"nginx"}]'
  404. # Post-condition: valid-pod POD has image nginx
  405. kube::test::get_object_assert pods "{{range.items}}{{$image_field}}:{{end}}" 'nginx:'
  406. # prove that yaml input works too
  407. YAML_PATCH=$'spec:\n containers:\n - name: kubernetes-serve-hostname\n image: changed-with-yaml\n'
  408. kubectl patch "${kube_flags[@]}" pod valid-pod -p="${YAML_PATCH}"
  409. # Post-condition: valid-pod POD has image nginx
  410. kube::test::get_object_assert pods "{{range.items}}{{$image_field}}:{{end}}" 'changed-with-yaml:'
  411. ## Patch pod from JSON can change image
  412. # Command
  413. kubectl patch "${kube_flags[@]}" -f test/fixtures/doc-yaml/admin/limitrange/valid-pod.yaml -p='{"spec":{"containers":[{"name": "kubernetes-serve-hostname", "image": "k8s.gcr.io/pause:3.1"}]}}'
  414. # Post-condition: valid-pod POD has expected image
  415. kube::test::get_object_assert pods "{{range.items}}{{$image_field}}:{{end}}" 'k8s.gcr.io/pause:3.1:'
  416. ## If resourceVersion is specified in the patch, it will be treated as a precondition, i.e., if the resourceVersion is different from that is stored in the server, the Patch should be rejected
  417. ERROR_FILE="${KUBE_TEMP}/conflict-error"
  418. ## If the resourceVersion is the same as the one stored in the server, the patch will be applied.
  419. # Command
  420. # Needs to retry because other party may change the resource.
  421. for count in {0..3}; do
  422. resourceVersion=$(kubectl get "${kube_flags[@]}" pod valid-pod -o go-template='{{ .metadata.resourceVersion }}')
  423. kubectl patch "${kube_flags[@]}" pod valid-pod -p='{"spec":{"containers":[{"name": "kubernetes-serve-hostname", "image": "nginx"}]},"metadata":{"resourceVersion":"'"${resourceVersion}"'"}}' 2> "${ERROR_FILE}" || true
  424. if grep -q "the object has been modified" "${ERROR_FILE}"; then
  425. kube::log::status "retry $1, error: $(cat "${ERROR_FILE}")"
  426. rm "${ERROR_FILE}"
  427. sleep $((2**count))
  428. else
  429. rm "${ERROR_FILE}"
  430. kube::test::get_object_assert pods "{{range.items}}{{$image_field}}:{{end}}" 'nginx:'
  431. break
  432. fi
  433. done
  434. ## If the resourceVersion is the different from the one stored in the server, the patch will be rejected.
  435. resourceVersion=$(kubectl get "${kube_flags[@]}" pod valid-pod -o go-template='{{ .metadata.resourceVersion }}')
  436. ((resourceVersion+=100))
  437. # Command
  438. kubectl patch "${kube_flags[@]}" pod valid-pod -p='{"spec":{"containers":[{"name": "kubernetes-serve-hostname", "image": "nginx"}]},"metadata":{"resourceVersion":"'$resourceVersion'"}}' 2> "${ERROR_FILE}" || true
  439. # Post-condition: should get an error reporting the conflict
  440. if grep -q "please apply your changes to the latest version and try again" "${ERROR_FILE}"; then
  441. kube::log::status "\"kubectl patch with resourceVersion $resourceVersion\" returns error as expected: $(cat "${ERROR_FILE}")"
  442. else
  443. kube::log::status "\"kubectl patch with resourceVersion $resourceVersion\" returns unexpected error or non-error: $(cat "${ERROR_FILE}")"
  444. exit 1
  445. fi
  446. rm "${ERROR_FILE}"
  447. ## --force replace pod can change other field, e.g., spec.container.name
  448. # Command
  449. kubectl get "${kube_flags[@]}" pod valid-pod -o json | ${SED} 's/"kubernetes-serve-hostname"/"replaced-k8s-serve-hostname"/g' > /tmp/tmp-valid-pod.json
  450. kubectl replace "${kube_flags[@]}" --force -f /tmp/tmp-valid-pod.json
  451. # Post-condition: spec.container.name = "replaced-k8s-serve-hostname"
  452. kube::test::get_object_assert 'pod valid-pod' "{{(index .spec.containers 0).name}}" 'replaced-k8s-serve-hostname'
  453. ## check replace --grace-period requires --force
  454. output_message=$(! kubectl replace "${kube_flags[@]}" --grace-period=1 -f /tmp/tmp-valid-pod.json 2>&1)
  455. kube::test::if_has_string "${output_message}" '\-\-grace-period must have \-\-force specified'
  456. ## check replace --timeout requires --force
  457. output_message=$(! kubectl replace "${kube_flags[@]}" --timeout=1s -f /tmp/tmp-valid-pod.json 2>&1)
  458. kube::test::if_has_string "${output_message}" '\-\-timeout must have \-\-force specified'
  459. #cleaning
  460. rm /tmp/tmp-valid-pod.json
  461. ## replace of a cluster scoped resource can succeed
  462. # Pre-condition: a node exists
  463. kubectl create -f - "${kube_flags[@]}" << __EOF__
  464. {
  465. "kind": "Node",
  466. "apiVersion": "v1",
  467. "metadata": {
  468. "name": "node-v1-test"
  469. }
  470. }
  471. __EOF__
  472. kubectl replace -f - "${kube_flags[@]}" << __EOF__
  473. {
  474. "kind": "Node",
  475. "apiVersion": "v1",
  476. "metadata": {
  477. "name": "node-v1-test",
  478. "annotations": {"a":"b"},
  479. "resourceVersion": "0"
  480. }
  481. }
  482. __EOF__
  483. # Post-condition: the node command succeeds
  484. kube::test::get_object_assert "node node-v1-test" "{{.metadata.annotations.a}}" 'b'
  485. kubectl delete node node-v1-test "${kube_flags[@]}"
  486. ## kubectl edit can update the image field of a POD. tmp-editor.sh is a fake editor
  487. echo -e "#!/usr/bin/env bash\n${SED} -i \"s/nginx/k8s.gcr.io\/serve_hostname/g\" \$1" > /tmp/tmp-editor.sh
  488. chmod +x /tmp/tmp-editor.sh
  489. # Pre-condition: valid-pod POD has image nginx
  490. kube::test::get_object_assert pods "{{range.items}}{{$image_field}}:{{end}}" 'nginx:'
  491. grep -q 'Patch:' <<< "$(EDITOR=/tmp/tmp-editor.sh kubectl edit "${kube_flags[@]}" pods/valid-pod --output-patch=true)"
  492. # Post-condition: valid-pod POD has image k8s.gcr.io/serve_hostname
  493. kube::test::get_object_assert pods "{{range.items}}{{$image_field}}:{{end}}" 'k8s.gcr.io/serve_hostname:'
  494. # cleaning
  495. rm /tmp/tmp-editor.sh
  496. ## kubectl edit should work on Windows
  497. grep -q 'Edit cancelled' <<< "$(EDITOR="cat" kubectl edit pod/valid-pod 2>&1)"
  498. grep -q 'name: valid-pod' <<< "$(EDITOR="cat" kubectl edit pod/valid-pod)"
  499. grep -q CRLF <<< "$(EDITOR="cat" kubectl edit --windows-line-endings pod/valid-pod | file - )"
  500. ! grep -q CRLF <<< "$(EDITOR="cat" kubectl edit --windows-line-endings=false pod/valid-pod | file - )" || exit 1
  501. grep -q 'kind: List' <<< "$(EDITOR="cat" kubectl edit ns)"
  502. ### Label POD YAML file locally without effecting the live pod.
  503. # Pre-condition: name is valid-pod
  504. kube::test::get_object_assert 'pod valid-pod' "{{${labels_field}.name}}" 'valid-pod'
  505. # Command
  506. output_message=$(kubectl label --local --overwrite -f hack/testdata/pod.yaml name=localonlyvalue -o yaml "${kube_flags[@]}")
  507. echo "$output_message"
  508. # Post-condition: name is still valid-pod in the live pod, but command output is the new value
  509. kube::test::get_object_assert 'pod valid-pod' "{{${labels_field}.name}}" 'valid-pod'
  510. kube::test::if_has_string "${output_message}" "localonlyvalue"
  511. ### Overwriting an existing label is not permitted
  512. # Pre-condition: name is valid-pod
  513. kube::test::get_object_assert 'pod valid-pod' "{{${labels_field}.name}}" 'valid-pod'
  514. # Command
  515. ! kubectl label pods valid-pod name=valid-pod-super-sayan "${kube_flags[@]}" || exit 1
  516. # Post-condition: name is still valid-pod
  517. kube::test::get_object_assert 'pod valid-pod' "{{${labels_field}.name}}" 'valid-pod'
  518. ### --overwrite must be used to overwrite existing label, can be applied to all resources
  519. # Pre-condition: name is valid-pod
  520. kube::test::get_object_assert 'pod valid-pod' "{{${labels_field}.name}}" 'valid-pod'
  521. # Command
  522. kubectl label --overwrite pods --all name=valid-pod-super-sayan "${kube_flags[@]}"
  523. # Post-condition: name is valid-pod-super-sayan
  524. kube::test::get_object_assert 'pod valid-pod' "{{${labels_field}.name}}" 'valid-pod-super-sayan'
  525. ### Delete POD by label
  526. # Pre-condition: valid-pod POD exists
  527. kube::test::get_object_assert pods "{{range.items}}{{$id_field}}:{{end}}" 'valid-pod:'
  528. # Command
  529. kubectl delete pods -l'name in (valid-pod-super-sayan)' --grace-period=0 --force "${kube_flags[@]}"
  530. # Post-condition: valid-pod POD doesn't exist
  531. kube::test::get_object_assert pods "{{range.items}}{{$id_field}}:{{end}}" ''
  532. ### Create two PODs from 1 yaml file
  533. # Pre-condition: no POD exists
  534. create_and_use_new_namespace
  535. kube::test::get_object_assert pods "{{range.items}}{{$id_field}}:{{end}}" ''
  536. # Command
  537. kubectl create -f test/fixtures/doc-yaml/user-guide/multi-pod.yaml "${kube_flags[@]}"
  538. # Post-condition: redis-master and valid-pod PODs exist
  539. kube::test::get_object_assert pods "{{range.items}}{{$id_field}}:{{end}}" 'redis-master:valid-pod:'
  540. ### Delete two PODs from 1 yaml file
  541. # Pre-condition: redis-master and valid-pod PODs exist
  542. kube::test::get_object_assert pods "{{range.items}}{{$id_field}}:{{end}}" 'redis-master:valid-pod:'
  543. # Command
  544. kubectl delete -f test/fixtures/doc-yaml/user-guide/multi-pod.yaml "${kube_flags[@]}"
  545. # Post-condition: no PODs exist
  546. kube::test::get_object_assert pods "{{range.items}}{{$id_field}}:{{end}}" ''
  547. ## kubectl apply should update configuration annotations only if apply is already called
  548. ## 1. kubectl create doesn't set the annotation
  549. # Pre-Condition: no POD exists
  550. create_and_use_new_namespace
  551. kube::test::get_object_assert pods "{{range.items}}{{$id_field}}:{{end}}" ''
  552. # Command: create a pod "test-pod"
  553. kubectl create -f hack/testdata/pod.yaml "${kube_flags[@]}"
  554. # Post-Condition: pod "test-pod" is created
  555. kube::test::get_object_assert 'pods test-pod' "{{${labels_field}.name}}" 'test-pod-label'
  556. # Post-Condition: pod "test-pod" doesn't have configuration annotation
  557. ! grep -q kubectl.kubernetes.io/last-applied-configuration <<< "$(kubectl get pods test-pod -o yaml "${kube_flags[@]}" )" || exit 1
  558. ## 2. kubectl replace doesn't set the annotation
  559. kubectl get pods test-pod -o yaml "${kube_flags[@]}" | ${SED} 's/test-pod-label/test-pod-replaced/g' > "${KUBE_TEMP}"/test-pod-replace.yaml
  560. # Command: replace the pod "test-pod"
  561. kubectl replace -f "${KUBE_TEMP}"/test-pod-replace.yaml "${kube_flags[@]}"
  562. # Post-Condition: pod "test-pod" is replaced
  563. kube::test::get_object_assert 'pods test-pod' "{{${labels_field}.name}}" 'test-pod-replaced'
  564. # Post-Condition: pod "test-pod" doesn't have configuration annotation
  565. ! grep -q kubectl.kubernetes.io/last-applied-configuration <<< "$(kubectl get pods test-pod -o yaml "${kube_flags[@]}")" || exit 1
  566. ## 3. kubectl apply does set the annotation
  567. # Command: apply the pod "test-pod"
  568. kubectl apply -f hack/testdata/pod-apply.yaml "${kube_flags[@]}"
  569. # Post-Condition: pod "test-pod" is applied
  570. kube::test::get_object_assert 'pods test-pod' "{{${labels_field}.name}}" 'test-pod-applied'
  571. # Post-Condition: pod "test-pod" has configuration annotation
  572. grep -q kubectl.kubernetes.io/last-applied-configuration <<< "$(kubectl get pods test-pod -o yaml "${kube_flags[@]}")"
  573. kubectl get pods test-pod -o yaml "${kube_flags[@]}" | grep kubectl.kubernetes.io/last-applied-configuration > "${KUBE_TEMP}"/annotation-configuration
  574. ## 4. kubectl replace updates an existing annotation
  575. kubectl get pods test-pod -o yaml "${kube_flags[@]}" | ${SED} 's/test-pod-applied/test-pod-replaced/g' > "${KUBE_TEMP}"/test-pod-replace.yaml
  576. # Command: replace the pod "test-pod"
  577. kubectl replace -f "${KUBE_TEMP}"/test-pod-replace.yaml "${kube_flags[@]}"
  578. # Post-Condition: pod "test-pod" is replaced
  579. kube::test::get_object_assert 'pods test-pod' "{{${labels_field}.name}}" 'test-pod-replaced'
  580. # Post-Condition: pod "test-pod" has configuration annotation, and it's updated (different from the annotation when it's applied)
  581. grep -q kubectl.kubernetes.io/last-applied-configuration <<< "$(kubectl get pods test-pod -o yaml "${kube_flags[@]}" )"
  582. kubectl get pods test-pod -o yaml "${kube_flags[@]}" | grep kubectl.kubernetes.io/last-applied-configuration > "${KUBE_TEMP}"/annotation-configuration-replaced
  583. ! [[ $(diff -q "${KUBE_TEMP}"/annotation-configuration "${KUBE_TEMP}"/annotation-configuration-replaced > /dev/null) ]] || exit 1
  584. # Clean up
  585. rm "${KUBE_TEMP}"/test-pod-replace.yaml "${KUBE_TEMP}"/annotation-configuration "${KUBE_TEMP}"/annotation-configuration-replaced
  586. kubectl delete pods test-pod "${kube_flags[@]}"
  587. set +o nounset
  588. set +o errexit
  589. }
  590. # runs specific kubectl create tests
  591. run_create_secret_tests() {
  592. set -o nounset
  593. set -o errexit
  594. ### Create generic secret with explicit namespace
  595. # Pre-condition: secret 'mysecret' does not exist
  596. output_message=$(! kubectl get secrets mysecret 2>&1 "${kube_flags[@]}")
  597. kube::test::if_has_string "${output_message}" 'secrets "mysecret" not found'
  598. # Command
  599. output_message=$(kubectl create "${kube_flags[@]}" secret generic mysecret --dry-run --from-literal=foo=bar -o jsonpath='{.metadata.namespace}' --namespace=user-specified)
  600. # Post-condition: mysecret still not created since --dry-run was used
  601. # Output from 'create' command should contain the specified --namespace value
  602. failure_message=$(! kubectl get secrets mysecret 2>&1 "${kube_flags[@]}")
  603. kube::test::if_has_string "${failure_message}" 'secrets "mysecret" not found'
  604. kube::test::if_has_string "${output_message}" 'user-specified'
  605. # Command
  606. output_message=$(kubectl create "${kube_flags[@]}" secret generic mysecret --dry-run --from-literal=foo=bar -o jsonpath='{.metadata.namespace}')
  607. # Post-condition: jsonpath for .metadata.namespace should be empty for object since --namespace was not explicitly specified
  608. kube::test::if_empty_string "${output_message}"
  609. # check to make sure that replace correctly PUTs to a URL
  610. kubectl create configmap tester-update-cm -o json --dry-run | kubectl create "${kube_flags[@]}" --raw /api/v1/namespaces/default/configmaps -f -
  611. output_message=$(kubectl create configmap tester-update-cm --from-literal=key1=config1 -o json --dry-run | kubectl replace "${kube_flags[@]}" --raw /api/v1/namespaces/default/configmaps/tester-update-cm -f -)
  612. # the message should show the body returned which will include a UID not present in the input
  613. kube::test::if_has_string "${output_message}" 'uid'
  614. # if the PUT was well-formed, the server will now have a key and value we can retrieve on GET
  615. output_message=$(kubectl get "${kube_flags[@]}" --raw /api/v1/namespaces/default/configmaps/tester-update-cm 2>&1 "${kube_flags[@]}")
  616. kube::test::if_has_string "${output_message}" 'config1'
  617. # if DELETE raw works correctly, this will delete the configmap
  618. kubectl delete "${kube_flags[@]}" --raw /api/v1/namespaces/default/configmaps/tester-update-cm
  619. output_message=$(! kubectl get "${kube_flags[@]}" configmap tester-update-cm 2>&1 "${kube_flags[@]}")
  620. kube::test::if_has_string "${output_message}" 'configmaps "tester-update-cm" not found'
  621. set +o nounset
  622. set +o errexit
  623. }
  624. run_secrets_test() {
  625. set -o nounset
  626. set -o errexit
  627. create_and_use_new_namespace
  628. kube::log::status "Testing secrets"
  629. # Ensure dry run succeeds and includes kind, apiVersion and data, and doesn't require a server connection
  630. output_message=$(kubectl create secret generic test --from-literal=key1=value1 --dry-run -o yaml --server=example.com --v=6)
  631. kube::test::if_has_string "${output_message}" 'kind: Secret'
  632. kube::test::if_has_string "${output_message}" 'apiVersion: v1'
  633. kube::test::if_has_string "${output_message}" 'key1: dmFsdWUx'
  634. kube::test::if_has_not_string "${output_message}" 'example.com'
  635. ### Create a new namespace
  636. # Pre-condition: the test-secrets namespace does not exist
  637. kube::test::get_object_assert 'namespaces' "{{range.items}}{{ if eq $id_field \\\"test-secrets\\\" }}found{{end}}{{end}}:" ':'
  638. # Command
  639. kubectl create namespace test-secrets
  640. # Post-condition: namespace 'test-secrets' is created.
  641. kube::test::get_object_assert 'namespaces/test-secrets' "{{$id_field}}" 'test-secrets'
  642. ### Create a generic secret in a specific namespace
  643. # Pre-condition: no SECRET exists
  644. kube::test::get_object_assert 'secrets --namespace=test-secrets' "{{range.items}}{{$id_field}}:{{end}}" ''
  645. # Command
  646. kubectl create secret generic test-secret --from-literal=key1=value1 --type=test-type --namespace=test-secrets
  647. # Post-condition: secret exists and has expected values
  648. kube::test::get_object_assert 'secret/test-secret --namespace=test-secrets' "{{$id_field}}" 'test-secret'
  649. kube::test::get_object_assert 'secret/test-secret --namespace=test-secrets' "{{$secret_type}}" 'test-type'
  650. grep -q 'key1: dmFsdWUx' <<< "$(kubectl get secret/test-secret --namespace=test-secrets -o yaml "${kube_flags[@]}")"
  651. # Clean-up
  652. kubectl delete secret test-secret --namespace=test-secrets
  653. ### Create a docker-registry secret in a specific namespace
  654. if [[ "${WAIT_FOR_DELETION:-}" == "true" ]]; then
  655. kube::test::wait_object_assert 'secrets --namespace=test-secrets' "{{range.items}}{{$id_field}}:{{end}}" ''
  656. fi
  657. # Pre-condition: no SECRET exists
  658. kube::test::get_object_assert 'secrets --namespace=test-secrets' "{{range.items}}{{$id_field}}:{{end}}" ''
  659. # Command
  660. kubectl create secret docker-registry test-secret --docker-username=test-user --docker-password=test-password --docker-email='test-user@test.com' --namespace=test-secrets
  661. # Post-condition: secret exists and has expected values
  662. kube::test::get_object_assert 'secret/test-secret --namespace=test-secrets' "{{$id_field}}" 'test-secret'
  663. kube::test::get_object_assert 'secret/test-secret --namespace=test-secrets' "{{$secret_type}}" 'kubernetes.io/dockerconfigjson'
  664. grep -q '.dockerconfigjson: eyJhdXRocyI6eyJodHRwczovL2luZGV4LmRvY2tlci5pby92MS8iOnsidXNlcm5hbWUiOiJ0ZXN0LXVzZXIiLCJwYXNzd29yZCI6InRlc3QtcGFzc3dvcmQiLCJlbWFpbCI6InRlc3QtdXNlckB0ZXN0LmNvbSIsImF1dGgiOiJkR1Z6ZEMxMWMyVnlPblJsYzNRdGNHRnpjM2R2Y21RPSJ9fX0=' <<< "$(kubectl get secret/test-secret --namespace=test-secrets -o yaml "${kube_flags[@]}")"
  665. # Clean-up
  666. kubectl delete secret test-secret --namespace=test-secrets
  667. ### Create a tls secret
  668. if [[ "${WAIT_FOR_DELETION:-}" == "true" ]]; then
  669. kube::test::wait_object_assert 'secrets --namespace=test-secrets' "{{range.items}}{{$id_field}}:{{end}}" ''
  670. fi
  671. # Pre-condition: no SECRET exists
  672. kube::test::get_object_assert 'secrets --namespace=test-secrets' "{{range.items}}{{$id_field}}:{{end}}" ''
  673. # Command
  674. kubectl create secret tls test-secret --namespace=test-secrets --key=hack/testdata/tls.key --cert=hack/testdata/tls.crt
  675. kube::test::get_object_assert 'secret/test-secret --namespace=test-secrets' "{{$id_field}}" 'test-secret'
  676. kube::test::get_object_assert 'secret/test-secret --namespace=test-secrets' "{{$secret_type}}" 'kubernetes.io/tls'
  677. # Clean-up
  678. kubectl delete secret test-secret --namespace=test-secrets
  679. # Command with process substitution
  680. kubectl create secret tls test-secret --namespace=test-secrets --key <(cat hack/testdata/tls.key) --cert <(cat hack/testdata/tls.crt)
  681. kube::test::get_object_assert 'secret/test-secret --namespace=test-secrets' "{{$id_field}}" 'test-secret'
  682. kube::test::get_object_assert 'secret/test-secret --namespace=test-secrets' "{{$secret_type}}" 'kubernetes.io/tls'
  683. # Clean-up
  684. kubectl delete secret test-secret --namespace=test-secrets
  685. # Create a secret using stringData
  686. kubectl create --namespace=test-secrets -f - "${kube_flags[@]}" << __EOF__
  687. {
  688. "kind": "Secret",
  689. "apiVersion": "v1",
  690. "metadata": {
  691. "name": "secret-string-data"
  692. },
  693. "data": {
  694. "k1":"djE=",
  695. "k2":""
  696. },
  697. "stringData": {
  698. "k2":"v2"
  699. }
  700. }
  701. __EOF__
  702. # Post-condition: secret-string-data secret is created with expected data, merged/overridden data from stringData, and a cleared stringData field
  703. kube::test::get_object_assert 'secret/secret-string-data --namespace=test-secrets ' '{{.data}}' '.*k1:djE=.*'
  704. kube::test::get_object_assert 'secret/secret-string-data --namespace=test-secrets ' '{{.data}}' '.*k2:djI=.*'
  705. kube::test::get_object_assert 'secret/secret-string-data --namespace=test-secrets ' '{{.stringData}}' '<no value>'
  706. # Clean up
  707. kubectl delete secret secret-string-data --namespace=test-secrets
  708. ### Create a secret using output flags
  709. if [[ "${WAIT_FOR_DELETION:-}" == "true" ]]; then
  710. kube::test::wait_object_assert 'secrets --namespace=test-secrets' "{{range.items}}{{$id_field}}:{{end}}" ''
  711. fi
  712. # Pre-condition: no secret exists
  713. kube::test::get_object_assert 'secrets --namespace=test-secrets' "{{range.items}}{{$id_field}}:{{end}}" ''
  714. # Command
  715. grep -q 'test-secret:' <<< "$(kubectl create secret generic test-secret --namespace=test-secrets --from-literal=key1=value1 --output=go-template --template="{{.metadata.name}}:")"
  716. ## Clean-up
  717. kubectl delete secret test-secret --namespace=test-secrets
  718. # Clean up
  719. kubectl delete namespace test-secrets
  720. set +o nounset
  721. set +o errexit
  722. }
  723. run_service_accounts_tests() {
  724. set -o nounset
  725. set -o errexit
  726. create_and_use_new_namespace
  727. kube::log::status "Testing service accounts"
  728. ### Create a new namespace
  729. # Pre-condition: the test-service-accounts namespace does not exist
  730. kube::test::get_object_assert 'namespaces' "{{range.items}}{{ if eq $id_field \\\"test-service-accounts\\\" }}found{{end}}{{end}}:" ':'
  731. # Command
  732. kubectl create namespace test-service-accounts
  733. # Post-condition: namespace 'test-service-accounts' is created.
  734. kube::test::get_object_assert 'namespaces/test-service-accounts' "{{$id_field}}" 'test-service-accounts'
  735. ### Create a service account in a specific namespace
  736. # Command
  737. kubectl create serviceaccount test-service-account --namespace=test-service-accounts
  738. # Post-condition: secret exists and has expected values
  739. kube::test::get_object_assert 'serviceaccount/test-service-account --namespace=test-service-accounts' "{{$id_field}}" 'test-service-account'
  740. # Clean-up
  741. kubectl delete serviceaccount test-service-account --namespace=test-service-accounts
  742. # Clean up
  743. kubectl delete namespace test-service-accounts
  744. set +o nounset
  745. set +o errexit
  746. }
  747. run_service_tests() {
  748. set -o nounset
  749. set -o errexit
  750. # switch back to the default namespace
  751. kubectl config set-context "${CONTEXT}" --namespace=""
  752. kube::log::status "Testing kubectl(v1:services)"
  753. ### Create redis-master service from JSON
  754. # Pre-condition: Only the default kubernetes services exist
  755. kube::test::get_object_assert services "{{range.items}}{{$id_field}}:{{end}}" 'kubernetes:'
  756. # Command
  757. kubectl create -f test/e2e/testing-manifests/guestbook/redis-master-service.yaml "${kube_flags[@]}"
  758. # Post-condition: redis-master service exists
  759. kube::test::get_object_assert services "{{range.items}}{{$id_field}}:{{end}}" 'kubernetes:redis-master:'
  760. # Describe command should print detailed information
  761. kube::test::describe_object_assert services 'redis-master' "Name:" "Labels:" "Selector:" "IP:" "Port:" "Endpoints:" "Session Affinity:"
  762. # Describe command should print events information by default
  763. kube::test::describe_object_events_assert services 'redis-master'
  764. # Describe command should not print events information when show-events=false
  765. kube::test::describe_object_events_assert services 'redis-master' false
  766. # Describe command should print events information when show-events=true
  767. kube::test::describe_object_events_assert services 'redis-master' true
  768. # Describe command (resource only) should print detailed information
  769. kube::test::describe_resource_assert services "Name:" "Labels:" "Selector:" "IP:" "Port:" "Endpoints:" "Session Affinity:"
  770. # Describe command should print events information by default
  771. kube::test::describe_resource_events_assert services
  772. # Describe command should not print events information when show-events=false
  773. kube::test::describe_resource_events_assert services false
  774. # Describe command should print events information when show-events=true
  775. kube::test::describe_resource_events_assert services true
  776. ### set selector
  777. # prove role=master
  778. kube::test::get_object_assert 'services redis-master' "{{range${service_selector_field:?}}}{{.}}:{{end}}" "redis:master:backend:"
  779. # Set selector of a local file without talking to the server
  780. kubectl set selector -f test/e2e/testing-manifests/guestbook/redis-master-service.yaml role=padawan --local -o yaml "${kube_flags[@]}"
  781. kubectl set selector -f test/e2e/testing-manifests/guestbook/redis-master-service.yaml role=padawan --dry-run -o yaml "${kube_flags[@]}"
  782. # Set command to change the selector.
  783. kubectl set selector -f test/e2e/testing-manifests/guestbook/redis-master-service.yaml role=padawan
  784. # prove role=padawan
  785. kube::test::get_object_assert 'services redis-master' "{{range$service_selector_field}}{{.}}:{{end}}" "padawan:"
  786. # Set command to reset the selector back to the original one.
  787. kubectl set selector -f test/e2e/testing-manifests/guestbook/redis-master-service.yaml app=redis,role=master,tier=backend
  788. # prove role=master
  789. kube::test::get_object_assert 'services redis-master' "{{range$service_selector_field}}{{.}}:{{end}}" "redis:master:backend:"
  790. # Show dry-run works on running selector
  791. kubectl set selector services redis-master role=padawan --dry-run -o yaml "${kube_flags[@]}"
  792. ! kubectl set selector services redis-master role=padawan --local -o yaml "${kube_flags[@]}" || exit 1
  793. kube::test::get_object_assert 'services redis-master' "{{range$service_selector_field}}{{.}}:{{end}}" "redis:master:backend:"
  794. # --resource-version=<current-resource-version> succeeds
  795. rv=$(kubectl get services redis-master -o jsonpath='{.metadata.resourceVersion}' "${kube_flags[@]}")
  796. kubectl set selector services redis-master rvtest1=true "--resource-version=${rv}" "${kube_flags[@]}"
  797. # --resource-version=<non-current-resource-version> fails
  798. output_message=$(! kubectl set selector services redis-master rvtest1=true --resource-version=1 2>&1 "${kube_flags[@]}")
  799. kube::test::if_has_string "${output_message}" 'Conflict'
  800. ### Dump current redis-master service
  801. output_service=$(kubectl get service redis-master -o json "${kube_flags[@]}")
  802. ### Delete redis-master-service by id
  803. # Pre-condition: redis-master service exists
  804. kube::test::get_object_assert services "{{range.items}}{{$id_field}}:{{end}}" 'kubernetes:redis-master:'
  805. # Command
  806. kubectl delete service redis-master "${kube_flags[@]}"
  807. if [[ "${WAIT_FOR_DELETION:-}" == "true" ]]; then
  808. kube::test::wait_object_assert services "{{range.items}}{{$id_field}}:{{end}}" 'kubernetes:'
  809. fi
  810. # Post-condition: Only the default kubernetes services exist
  811. kube::test::get_object_assert services "{{range.items}}{{$id_field}}:{{end}}" 'kubernetes:'
  812. ### Create redis-master-service from dumped JSON
  813. # Pre-condition: Only the default kubernetes services exist
  814. kube::test::get_object_assert services "{{range.items}}{{$id_field}}:{{end}}" 'kubernetes:'
  815. # Command
  816. echo "${output_service}" | kubectl create -f - "${kube_flags[@]}"
  817. # Post-condition: redis-master service is created
  818. kube::test::get_object_assert services "{{range.items}}{{$id_field}}:{{end}}" 'kubernetes:redis-master:'
  819. ### Create redis-master-v1-test service
  820. # Pre-condition: redis-master-service service exists
  821. kube::test::get_object_assert services "{{range.items}}{{$id_field}}:{{end}}" 'kubernetes:redis-master:'
  822. # Command
  823. kubectl create -f - "${kube_flags[@]}" << __EOF__
  824. {
  825. "kind": "Service",
  826. "apiVersion": "v1",
  827. "metadata": {
  828. "name": "service-v1-test"
  829. },
  830. "spec": {
  831. "ports": [
  832. {
  833. "protocol": "TCP",
  834. "port": 80,
  835. "targetPort": 80
  836. }
  837. ]
  838. }
  839. }
  840. __EOF__
  841. # Post-condition: service-v1-test service is created
  842. kube::test::get_object_assert services "{{range.items}}{{$id_field}}:{{end}}" 'kubernetes:redis-master:service-.*-test:'
  843. ### Identity
  844. kubectl get service "${kube_flags[@]}" service-v1-test -o json | kubectl replace "${kube_flags[@]}" -f -
  845. ### Delete services by id
  846. # Pre-condition: service-v1-test exists
  847. kube::test::get_object_assert services "{{range.items}}{{$id_field}}:{{end}}" 'kubernetes:redis-master:service-.*-test:'
  848. # Command
  849. kubectl delete service redis-master "${kube_flags[@]}"
  850. kubectl delete service "service-v1-test" "${kube_flags[@]}"
  851. if [[ "${WAIT_FOR_DELETION:-}" == "true" ]]; then
  852. kube::test::wait_object_assert services "{{range.items}}{{$id_field}}:{{end}}" 'kubernetes:'
  853. fi
  854. # Post-condition: Only the default kubernetes services exist
  855. kube::test::get_object_assert services "{{range.items}}{{$id_field}}:{{end}}" 'kubernetes:'
  856. ### Create two services
  857. # Pre-condition: Only the default kubernetes services exist
  858. kube::test::get_object_assert services "{{range.items}}{{$id_field}}:{{end}}" 'kubernetes:'
  859. # Command
  860. kubectl create -f test/e2e/testing-manifests/guestbook/redis-master-service.yaml "${kube_flags[@]}"
  861. kubectl create -f test/e2e/testing-manifests/guestbook/redis-slave-service.yaml "${kube_flags[@]}"
  862. # Post-condition: redis-master and redis-slave services are created
  863. kube::test::get_object_assert services "{{range.items}}{{$id_field}}:{{end}}" 'kubernetes:redis-master:redis-slave:'
  864. ### Custom columns can be specified
  865. # Pre-condition: generate output using custom columns
  866. output_message=$(kubectl get services -o=custom-columns=NAME:.metadata.name,RSRC:.metadata.resourceVersion 2>&1 "${kube_flags[@]}")
  867. # Post-condition: should contain name column
  868. kube::test::if_has_string "${output_message}" 'redis-master'
  869. ### Delete multiple services at once
  870. # Pre-condition: redis-master and redis-slave services exist
  871. kube::test::get_object_assert services "{{range.items}}{{$id_field}}:{{end}}" 'kubernetes:redis-master:redis-slave:'
  872. # Command
  873. kubectl delete services redis-master redis-slave "${kube_flags[@]}" # delete multiple services at once
  874. if [[ "${WAIT_FOR_DELETION:-}" == "true" ]]; then
  875. kube::test::wait_object_assert services "{{range.items}}{{$id_field}}:{{end}}" 'kubernetes:'
  876. fi
  877. # Post-condition: Only the default kubernetes services exist
  878. kube::test::get_object_assert services "{{range.items}}{{$id_field}}:{{end}}" 'kubernetes:'
  879. ### Create an ExternalName service
  880. # Pre-condition: Only the default kubernetes service exist
  881. kube::test::get_object_assert services "{{range.items}}{{$id_field}}:{{end}}" 'kubernetes:'
  882. # Command
  883. kubectl create service externalname beep-boop --external-name bar.com
  884. # Post-condition: beep-boop service is created
  885. kube::test::get_object_assert services "{{range.items}}{{$id_field}}:{{end}}" 'beep-boop:kubernetes:'
  886. ### Delete beep-boop service by id
  887. # Pre-condition: beep-boop service exists
  888. kube::test::get_object_assert services "{{range.items}}{{$id_field}}:{{end}}" 'beep-boop:kubernetes:'
  889. # Command
  890. kubectl delete service beep-boop "${kube_flags[@]}"
  891. if [[ "${WAIT_FOR_DELETION:-}" == "true" ]]; then
  892. kube::test::wait_object_assert services "{{range.items}}{{$id_field}}:{{end}}" 'kubernetes:'
  893. fi
  894. # Post-condition: Only the default kubernetes services exist
  895. kube::test::get_object_assert services "{{range.items}}{{$id_field}}:{{end}}" 'kubernetes:'
  896. ### Create pod and service
  897. # Pre-condition: no pod exists
  898. kube::test::wait_object_assert pods "{{range.items}}{{$id_field}}:{{end}}" ''
  899. # Command
  900. kubectl run testmetadata --image=nginx --port=80 --expose --service-overrides='{ "metadata": { "annotations": { "zone-context": "home" } } } '
  901. # Check result
  902. kube::test::get_object_assert pods "{{range.items}}{{$id_field}}:{{end}}" 'testmetadata:'
  903. kube::test::get_object_assert 'service testmetadata' "{{.metadata.annotations}}" "map\[zone-context:home\]"
  904. ### Expose pod as a new service
  905. # Command
  906. kubectl expose pod testmetadata --port=1000 --target-port=80 --type=NodePort --name=exposemetadata --overrides='{ "metadata": { "annotations": { "zone-context": "work" } } } '
  907. # Check result
  908. kube::test::get_object_assert 'service exposemetadata' "{{.metadata.annotations}}" "map\[zone-context:work\]"
  909. # Clean-Up
  910. # Command
  911. kubectl delete service exposemetadata testmetadata "${kube_flags[@]}"
  912. if [[ "${WAIT_FOR_DELETION:-}" == "true" ]]; then
  913. kube::test::wait_object_assert services "{{range.items}}{{$id_field}}:{{end}}" 'kubernetes:'
  914. fi
  915. kubectl delete pod testmetadata "${kube_flags[@]}"
  916. if [[ "${WAIT_FOR_DELETION:-}" == "true" ]]; then
  917. kube::test::wait_object_assert deployment "{{range.items}}{{$id_field}}:{{end}}" ''
  918. fi
  919. set +o nounset
  920. set +o errexit
  921. }
  922. run_rc_tests() {
  923. set -o nounset
  924. set -o errexit
  925. create_and_use_new_namespace
  926. kube::log::status "Testing kubectl(v1:replicationcontrollers)"
  927. ### Create and stop controller, make sure it doesn't leak pods
  928. # Pre-condition: no replication controller exists
  929. kube::test::get_object_assert rc "{{range.items}}{{$id_field}}:{{end}}" ''
  930. # Command
  931. kubectl create -f hack/testdata/frontend-controller.yaml "${kube_flags[@]}"
  932. kubectl delete rc frontend "${kube_flags[@]}"
  933. # Post-condition: no pods from frontend controller
  934. kube::test::wait_object_assert 'pods -l "name=frontend"' "{{range.items}}{{$id_field}}:{{end}}" ''
  935. ### Create replication controller frontend from JSON
  936. # Pre-condition: no replication controller exists
  937. kube::test::get_object_assert rc "{{range.items}}{{$id_field}}:{{end}}" ''
  938. # Command
  939. kubectl create -f hack/testdata/frontend-controller.yaml "${kube_flags[@]}"
  940. # Post-condition: frontend replication controller is created
  941. kube::test::get_object_assert rc "{{range.items}}{{$id_field}}:{{end}}" 'frontend:'
  942. # Describe command should print detailed information
  943. kube::test::describe_object_assert rc 'frontend' "Name:" "Pod Template:" "Labels:" "Selector:" "Replicas:" "Pods Status:" "Volumes:" "GET_HOSTS_FROM:"
  944. # Describe command should print events information by default
  945. kube::test::describe_object_events_assert rc 'frontend'
  946. # Describe command should not print events information when show-events=false
  947. kube::test::describe_object_events_assert rc 'frontend' false
  948. # Describe command should print events information when show-events=true
  949. kube::test::describe_object_events_assert rc 'frontend' true
  950. # Describe command (resource only) should print detailed information
  951. kube::test::describe_resource_assert rc "Name:" "Name:" "Pod Template:" "Labels:" "Selector:" "Replicas:" "Pods Status:" "Volumes:" "GET_HOSTS_FROM:"
  952. # Describe command should print events information by default
  953. kube::test::describe_resource_events_assert rc
  954. # Describe command should not print events information when show-events=false
  955. kube::test::describe_resource_events_assert rc false
  956. # Describe command should print events information when show-events=true
  957. kube::test::describe_resource_events_assert rc true
  958. ### Scale replication controller frontend with current-replicas and replicas
  959. # Pre-condition: 3 replicas
  960. kube::test::get_object_assert 'rc frontend' "{{${rc_replicas_field:?}}}" '3'
  961. # Command
  962. kubectl scale --current-replicas=3 --replicas=2 replicationcontrollers frontend "${kube_flags[@]}"
  963. # Post-condition: 2 replicas
  964. kube::test::get_object_assert 'rc frontend' "{{$rc_replicas_field}}" '2'
  965. ### Scale replication controller frontend with (wrong) current-replicas and replicas
  966. # Pre-condition: 2 replicas
  967. kube::test::get_object_assert 'rc frontend' "{{$rc_replicas_field}}" '2'
  968. # Command
  969. ! kubectl scale --current-replicas=3 --replicas=2 replicationcontrollers frontend "${kube_flags[@]}" || exit 1
  970. # Post-condition: nothing changed
  971. kube::test::get_object_assert 'rc frontend' "{{$rc_replicas_field}}" '2'
  972. ### Scale replication controller frontend with replicas only
  973. # Pre-condition: 2 replicas
  974. kube::test::get_object_assert 'rc frontend' "{{$rc_replicas_field}}" '2'
  975. # Command
  976. kubectl scale --replicas=3 replicationcontrollers frontend "${kube_flags[@]}"
  977. # Post-condition: 3 replicas
  978. kube::test::get_object_assert 'rc frontend' "{{$rc_replicas_field}}" '3'
  979. ### Scale replication controller from JSON with replicas only
  980. # Pre-condition: 3 replicas
  981. kube::test::get_object_assert 'rc frontend' "{{$rc_replicas_field}}" '3'
  982. # Command
  983. kubectl scale --replicas=2 -f hack/testdata/frontend-controller.yaml "${kube_flags[@]}"
  984. # Post-condition: 2 replicas
  985. kube::test::get_object_assert 'rc frontend' "{{$rc_replicas_field}}" '2'
  986. # Clean-up
  987. kubectl delete rc frontend "${kube_flags[@]}"
  988. ### Scale multiple replication controllers
  989. kubectl create -f test/e2e/testing-manifests/guestbook/legacy/redis-master-controller.yaml "${kube_flags[@]}"
  990. kubectl create -f test/e2e/testing-manifests/guestbook/legacy/redis-slave-controller.yaml "${kube_flags[@]}"
  991. # Command
  992. kubectl scale rc/redis-master rc/redis-slave --replicas=4 "${kube_flags[@]}"
  993. # Post-condition: 4 replicas each
  994. kube::test::get_object_assert 'rc redis-master' "{{$rc_replicas_field}}" '4'
  995. kube::test::get_object_assert 'rc redis-slave' "{{$rc_replicas_field}}" '4'
  996. # Clean-up
  997. kubectl delete rc redis-{master,slave} "${kube_flags[@]}"
  998. ### Scale a deployment
  999. kubectl create -f test/fixtures/doc-yaml/user-guide/deployment.yaml "${kube_flags[@]}"
  1000. # Command
  1001. kubectl scale --current-replicas=3 --replicas=1 deployment/nginx-deployment
  1002. # Post-condition: 1 replica for nginx-deployment
  1003. kube::test::get_object_assert 'deployment nginx-deployment' "{{${deployment_replicas:?}}}" '1'
  1004. # Clean-up
  1005. kubectl delete deployment/nginx-deployment "${kube_flags[@]}"
  1006. ### Expose deployments by creating a service
  1007. # Uses deployment selectors for created service
  1008. output_message=$(kubectl expose -f test/fixtures/pkg/kubectl/cmd/expose/appsv1deployment.yaml --port 80 2>&1 "${kube_flags[@]}")
  1009. # Post-condition: service created for deployment.
  1010. kube::test::if_has_string "${output_message}" 'service/expose-test-deployment exposed'
  1011. # Clean-up
  1012. kubectl delete service/expose-test-deployment "${kube_flags[@]}"
  1013. # Contains no selectors, should fail.
  1014. output_message=$(! kubectl expose -f test/fixtures/pkg/kubectl/cmd/expose/appsv1deployment-no-selectors.yaml --port 80 2>&1 "${kube_flags[@]}")
  1015. # Post-condition: service created for deployment.
  1016. kube::test::if_has_string "${output_message}" 'invalid deployment: no selectors'
  1017. ### Expose a deployment as a service
  1018. kubectl create -f test/fixtures/doc-yaml/user-guide/deployment.yaml "${kube_flags[@]}"
  1019. # Pre-condition: 3 replicas
  1020. kube::test::get_object_assert 'deployment nginx-deployment' "{{$deployment_replicas}}" '3'
  1021. # Command
  1022. kubectl expose deployment/nginx-deployment
  1023. # Post-condition: service exists and exposes deployment port (80)
  1024. kube::test::get_object_assert 'service nginx-deployment' "{{${port_field:?}}}" '80'
  1025. # Clean-up
  1026. kubectl delete deployment/nginx-deployment service/nginx-deployment "${kube_flags[@]}"
  1027. ### Expose replication controller as service
  1028. kubectl create -f hack/testdata/frontend-controller.yaml "${kube_flags[@]}"
  1029. # Pre-condition: 3 replicas
  1030. kube::test::get_object_assert 'rc frontend' "{{$rc_replicas_field}}" '3'
  1031. # Command
  1032. kubectl expose rc frontend --port=80 "${kube_flags[@]}"
  1033. # Post-condition: service exists and the port is unnamed
  1034. kube::test::get_object_assert 'service frontend' "{{${port_name:?}}} {{$port_field}}" '<no value> 80'
  1035. # Command
  1036. kubectl expose service frontend --port=443 --name=frontend-2 "${kube_flags[@]}"
  1037. # Post-condition: service exists and the port is unnamed
  1038. kube::test::get_object_assert 'service frontend-2' "{{$port_name}} {{$port_field}}" '<no value> 443'
  1039. # Command
  1040. kubectl create -f test/fixtures/doc-yaml/admin/limitrange/valid-pod.yaml "${kube_flags[@]}"
  1041. kubectl expose pod valid-pod --port=444 --name=frontend-3 "${kube_flags[@]}"
  1042. # Post-condition: service exists and the port is unnamed
  1043. kube::test::get_object_assert 'service frontend-3' "{{$port_name}} {{$port_field}}" '<no value> 444'
  1044. # Create a service using service/v1 generator
  1045. kubectl expose rc frontend --port=80 --name=frontend-4 --generator=service/v1 "${kube_flags[@]}"
  1046. # Post-condition: service exists and the port is named default.
  1047. kube::test::get_object_assert 'service frontend-4' "{{$port_name}} {{$port_field}}" 'default 80'
  1048. # Verify that expose service works without specifying a port.
  1049. kubectl expose service frontend --name=frontend-5 "${kube_flags[@]}"
  1050. # Post-condition: service exists with the same port as the original service.
  1051. kube::test::get_object_assert 'service frontend-5' "{{$port_field}}" '80'
  1052. # Cleanup services
  1053. kubectl delete pod valid-pod "${kube_flags[@]}"
  1054. kubectl delete service frontend{,-2,-3,-4,-5} "${kube_flags[@]}"
  1055. ### Expose negative invalid resource test
  1056. # Pre-condition: don't need
  1057. # Command
  1058. output_message=$(! kubectl expose nodes 127.0.0.1 2>&1 "${kube_flags[@]}")
  1059. # Post-condition: the error message has "cannot expose" string
  1060. kube::test::if_has_string "${output_message}" 'cannot expose'
  1061. ### Try to generate a service with invalid name (exceeding maximum valid size)
  1062. # Pre-condition: use --name flag
  1063. output_message=$(! kubectl expose -f hack/testdata/pod-with-large-name.yaml --name=invalid-large-service-name-that-has-more-than-sixty-three-characters --port=8081 2>&1 "${kube_flags[@]}")
  1064. # Post-condition: should fail due to invalid name
  1065. kube::test::if_has_string "${output_message}" 'metadata.name: Invalid value'
  1066. # Pre-condition: default run without --name flag; should succeed by truncating the inherited name
  1067. output_message=$(kubectl expose -f hack/testdata/pod-with-large-name.yaml --port=8081 2>&1 "${kube_flags[@]}")
  1068. # Post-condition: inherited name from pod has been truncated
  1069. kube::test::if_has_string "${output_message}" 'kubernetes-serve-hostname-testing-sixty-three-characters-in-len exposed'
  1070. # Clean-up
  1071. kubectl delete svc kubernetes-serve-hostname-testing-sixty-three-characters-in-len "${kube_flags[@]}"
  1072. ### Expose multiport object as a new service
  1073. # Pre-condition: don't use --port flag
  1074. output_message=$(kubectl expose -f test/fixtures/doc-yaml/admin/high-availability/etcd.yaml --selector=test=etcd 2>&1 "${kube_flags[@]}")
  1075. # Post-condition: expose succeeded
  1076. kube::test::if_has_string "${output_message}" 'etcd-server exposed'
  1077. # Post-condition: generated service has both ports from the exposed pod
  1078. kube::test::get_object_assert 'service etcd-server' "{{$port_name}} {{$port_field}}" 'port-1 2380'
  1079. kube::test::get_object_assert 'service etcd-server' "{{${second_port_name:?}}} {{${second_port_field:?}}}" 'port-2 2379'
  1080. # Clean-up
  1081. kubectl delete svc etcd-server "${kube_flags[@]}"
  1082. ### Delete replication controller with id
  1083. # Pre-condition: frontend replication controller exists
  1084. kube::test::get_object_assert rc "{{range.items}}{{$id_field}}:{{end}}" 'frontend:'
  1085. # Command
  1086. kubectl delete rc frontend "${kube_flags[@]}"
  1087. # Post-condition: no replication controller exists
  1088. kube::test::get_object_assert rc "{{range.items}}{{$id_field}}:{{end}}" ''
  1089. ### Create two replication controllers
  1090. # Pre-condition: no replication controller exists
  1091. kube::test::get_object_assert rc "{{range.items}}{{$id_field}}:{{end}}" ''
  1092. # Command
  1093. kubectl create -f hack/testdata/frontend-controller.yaml "${kube_flags[@]}"
  1094. kubectl create -f test/e2e/testing-manifests/guestbook/legacy/redis-slave-controller.yaml "${kube_flags[@]}"
  1095. # Post-condition: frontend and redis-slave
  1096. kube::test::get_object_assert rc "{{range.items}}{{$id_field}}:{{end}}" 'frontend:redis-slave:'
  1097. ### Delete multiple controllers at once
  1098. # Pre-condition: frontend and redis-slave
  1099. kube::test::get_object_assert rc "{{range.items}}{{$id_field}}:{{end}}" 'frontend:redis-slave:'
  1100. # Command
  1101. kubectl delete rc frontend redis-slave "${kube_flags[@]}" # delete multiple controllers at once
  1102. # Post-condition: no replication controller exists
  1103. kube::test::get_object_assert rc "{{range.items}}{{$id_field}}:{{end}}" ''
  1104. ### Auto scale replication controller
  1105. # Pre-condition: no replication controller exists
  1106. kube::test::get_object_assert rc "{{range.items}}{{$id_field}}:{{end}}" ''
  1107. # Command
  1108. kubectl create -f hack/testdata/frontend-controller.yaml "${kube_flags[@]}"
  1109. kube::test::get_object_assert rc "{{range.items}}{{$id_field}}:{{end}}" 'frontend:'
  1110. # autoscale 1~2 pods, CPU utilization 70%, rc specified by file
  1111. kubectl autoscale -f hack/testdata/frontend-controller.yaml "${kube_flags[@]}" --max=2 --cpu-percent=70
  1112. kube::test::get_object_assert 'hpa frontend' "{{${hpa_min_field:?}}} {{${hpa_max_field:?}}} {{${hpa_cpu_field:?}}}" '1 2 70'
  1113. kubectl delete hpa frontend "${kube_flags[@]}"
  1114. # autoscale 2~3 pods, no CPU utilization specified, rc specified by name
  1115. kubectl autoscale rc frontend "${kube_flags[@]}" --min=2 --max=3
  1116. kube::test::get_object_assert 'hpa frontend' "{{$hpa_min_field}} {{$hpa_max_field}} {{$hpa_cpu_field}}" '2 3 80'
  1117. kubectl delete hpa frontend "${kube_flags[@]}"
  1118. # autoscale without specifying --max should fail
  1119. ! kubectl autoscale rc frontend "${kube_flags[@]}" || exit 1
  1120. # Clean up
  1121. kubectl delete rc frontend "${kube_flags[@]}"
  1122. ## Set resource limits/request of a deployment
  1123. # Pre-condition: no deployment exists
  1124. kube::test::get_object_assert deployment "{{range.items}}{{$id_field}}:{{end}}" ''
  1125. # Set resources of a local file without talking to the server
  1126. kubectl set resources -f hack/testdata/deployment-multicontainer-resources.yaml -c=perl --limits=cpu=300m --requests=cpu=300m --local -o yaml "${kube_flags[@]}"
  1127. ! kubectl set resources -f hack/testdata/deployment-multicontainer-resources.yaml -c=perl --limits=cpu=300m --requests=cpu=300m --dry-run -o yaml "${kube_flags[@]}" || exit 1
  1128. # Create a deployment
  1129. kubectl create -f hack/testdata/deployment-multicontainer-resources.yaml "${kube_flags[@]}"
  1130. kube::test::get_object_assert deployment "{{range.items}}{{$id_field}}:{{end}}" 'nginx-deployment-resources:'
  1131. kube::test::get_object_assert deployment "{{range.items}}{{${image_field0:?}}}:{{end}}" "${IMAGE_DEPLOYMENT_R1}:"
  1132. kube::test::get_object_assert deployment "{{range.items}}{{${image_field1:?}}}:{{end}}" "${IMAGE_PERL}:"
  1133. # Set the deployment's cpu limits
  1134. kubectl set resources deployment nginx-deployment-resources --limits=cpu=100m "${kube_flags[@]}"
  1135. kube::test::get_object_assert deployment "{{range.items}}{{(index .spec.template.spec.containers 0).resources.limits.cpu}}:{{end}}" "100m:"
  1136. kube::test::get_object_assert deployment "{{range.items}}{{(index .spec.template.spec.containers 1).resources.limits.cpu}}:{{end}}" "100m:"
  1137. # Set a non-existing container should fail
  1138. ! kubectl set resources deployment nginx-deployment-resources -c=redis --limits=cpu=100m || exit 1
  1139. # Set the limit of a specific container in deployment
  1140. kubectl set resources deployment nginx-deployment-resources -c=nginx --limits=cpu=200m "${kube_flags[@]}"
  1141. kube::test::get_object_assert deployment "{{range.items}}{{(index .spec.template.spec.containers 0).resources.limits.cpu}}:{{end}}" "200m:"
  1142. kube::test::get_object_assert deployment "{{range.items}}{{(index .spec.template.spec.containers 1).resources.limits.cpu}}:{{end}}" "100m:"
  1143. # Set limits/requests of a deployment specified by a file
  1144. kubectl set resources -f hack/testdata/deployment-multicontainer-resources.yaml -c=perl --limits=cpu=300m --requests=cpu=300m "${kube_flags[@]}"
  1145. kube::test::get_object_assert deployment "{{range.items}}{{(index .spec.template.spec.containers 0).resources.limits.cpu}}:{{end}}" "200m:"
  1146. kube::test::get_object_assert deployment "{{range.items}}{{(index .spec.template.spec.containers 1).resources.limits.cpu}}:{{end}}" "300m:"
  1147. kube::test::get_object_assert deployment "{{range.items}}{{(index .spec.template.spec.containers 1).resources.requests.cpu}}:{{end}}" "300m:"
  1148. # Show dry-run works on running deployments
  1149. kubectl set resources deployment nginx-deployment-resources -c=perl --limits=cpu=400m --requests=cpu=400m --dry-run -o yaml "${kube_flags[@]}"
  1150. ! kubectl set resources deployment nginx-deployment-resources -c=perl --limits=cpu=400m --requests=cpu=400m --local -o yaml "${kube_flags[@]}" || exit 1
  1151. kube::test::get_object_assert deployment "{{range.items}}{{(index .spec.template.spec.containers 0).resources.limits.cpu}}:{{end}}" "200m:"
  1152. kube::test::get_object_assert deployment "{{range.items}}{{(index .spec.template.spec.containers 1).resources.limits.cpu}}:{{end}}" "300m:"
  1153. kube::test::get_object_assert deployment "{{range.items}}{{(index .spec.template.spec.containers 1).resources.requests.cpu}}:{{end}}" "300m:"
  1154. # Clean up
  1155. kubectl delete deployment nginx-deployment-resources "${kube_flags[@]}"
  1156. set +o nounset
  1157. set +o errexit
  1158. }
  1159. run_namespace_tests() {
  1160. set -o nounset
  1161. set -o errexit
  1162. kube::log::status "Testing kubectl(v1:namespaces)"
  1163. ### Create a new namespace
  1164. # Pre-condition: only the "default" namespace exists
  1165. # The Pre-condition doesn't hold anymore after we create and switch namespaces before creating pods with same name in the test.
  1166. # kube::test::get_object_assert namespaces "{{range.items}}{{$id_field}}:{{end}}" 'default:'
  1167. # Command
  1168. kubectl create namespace my-namespace
  1169. # Post-condition: namespace 'my-namespace' is created.
  1170. kube::test::get_object_assert 'namespaces/my-namespace' "{{$id_field}}" 'my-namespace'
  1171. # Clean up
  1172. kubectl delete namespace my-namespace --wait=false
  1173. # make sure that wait properly waits for finalization
  1174. kubectl wait --for=delete ns/my-namespace
  1175. output_message=$(! kubectl get ns/my-namespace 2>&1 "${kube_flags[@]}")
  1176. kube::test::if_has_string "${output_message}" ' not found'
  1177. kubectl create namespace my-namespace
  1178. kube::test::get_object_assert 'namespaces/my-namespace' "{{$id_field}}" 'my-namespace'
  1179. output_message=$(! kubectl delete namespace -n my-namespace --all 2>&1 "${kube_flags[@]}")
  1180. kube::test::if_has_string "${output_message}" 'warning: deleting cluster-scoped resources'
  1181. kube::test::if_has_string "${output_message}" 'namespace "my-namespace" deleted'
  1182. ######################
  1183. # Pods in Namespaces #
  1184. ######################
  1185. if kube::test::if_supports_resource "${pods:?}" ; then
  1186. ### Create a new namespace
  1187. # Pre-condition: the other namespace does not exist
  1188. kube::test::get_object_assert 'namespaces' "{{range.items}}{{ if eq $id_field \\\"other\\\" }}found{{end}}{{end}}:" ':'
  1189. # Command
  1190. kubectl create namespace other
  1191. # Post-condition: namespace 'other' is created.
  1192. kube::test::get_object_assert 'namespaces/other' "{{$id_field}}" 'other'
  1193. ### Create POD valid-pod in specific namespace
  1194. # Pre-condition: no POD exists
  1195. kube::test::get_object_assert 'pods --namespace=other' "{{range.items}}{{$id_field}}:{{end}}" ''
  1196. # Command
  1197. kubectl create "${kube_flags[@]}" --namespace=other -f test/fixtures/doc-yaml/admin/limitrange/valid-pod.yaml
  1198. # Post-condition: valid-pod POD is created
  1199. kube::test::get_object_assert 'pods --namespace=other' "{{range.items}}{{$id_field}}:{{end}}" 'valid-pod:'
  1200. # Post-condition: verify shorthand `-n other` has the same results as `--namespace=other`
  1201. kube::test::get_object_assert 'pods -n other' "{{range.items}}{{$id_field}}:{{end}}" 'valid-pod:'
  1202. # Post-condition: a resource cannot be retrieved by name across all namespaces
  1203. output_message=$(! kubectl get "${kube_flags[@]}" pod valid-pod --all-namespaces 2>&1)
  1204. kube::test::if_has_string "${output_message}" "a resource cannot be retrieved by name across all namespaces"
  1205. ### Delete POD valid-pod in specific namespace
  1206. # Pre-condition: valid-pod POD exists
  1207. kube::test::get_object_assert 'pods --namespace=other' "{{range.items}}{{$id_field}}:{{end}}" 'valid-pod:'
  1208. # Command
  1209. kubectl delete "${kube_flags[@]}" pod --namespace=other valid-pod --grace-period=0 --force
  1210. # Post-condition: valid-pod POD doesn't exist
  1211. kube::test::get_object_assert 'pods --namespace=other' "{{range.items}}{{$id_field}}:{{end}}" ''
  1212. # Clean up
  1213. kubectl delete namespace other
  1214. fi
  1215. set +o nounset
  1216. set +o errexit
  1217. }
  1218. run_nodes_tests() {
  1219. set -o nounset
  1220. set -o errexit
  1221. kube::log::status "Testing kubectl(v1:nodes)"
  1222. kube::test::get_object_assert nodes "{{range.items}}{{$id_field}}:{{end}}" '127.0.0.1:'
  1223. kube::test::describe_object_assert nodes "127.0.0.1" "Name:" "Labels:" "CreationTimestamp:" "Conditions:" "Addresses:" "Capacity:" "Pods:"
  1224. # Describe command should print events information by default
  1225. kube::test::describe_object_events_assert nodes "127.0.0.1"
  1226. # Describe command should not print events information when show-events=false
  1227. kube::test::describe_object_events_assert nodes "127.0.0.1" false
  1228. # Describe command should print events information when show-events=true
  1229. kube::test::describe_object_events_assert nodes "127.0.0.1" true
  1230. # Describe command (resource only) should print detailed information
  1231. kube::test::describe_resource_assert nodes "Name:" "Labels:" "CreationTimestamp:" "Conditions:" "Addresses:" "Capacity:" "Pods:"
  1232. # Describe command should print events information by default
  1233. kube::test::describe_resource_events_assert nodes
  1234. # Describe command should not print events information when show-events=false
  1235. kube::test::describe_resource_events_assert nodes false
  1236. # Describe command should print events information when show-events=true
  1237. kube::test::describe_resource_events_assert nodes true
  1238. ### kubectl patch update can mark node unschedulable
  1239. # Pre-condition: node is schedulable
  1240. kube::test::get_object_assert "nodes 127.0.0.1" "{{.spec.unschedulable}}" '<no value>'
  1241. kubectl patch "${kube_flags[@]}" nodes "127.0.0.1" -p='{"spec":{"unschedulable":true}}'
  1242. # Post-condition: node is unschedulable
  1243. kube::test::get_object_assert "nodes 127.0.0.1" "{{.spec.unschedulable}}" 'true'
  1244. kubectl patch "${kube_flags[@]}" nodes "127.0.0.1" -p='{"spec":{"unschedulable":null}}'
  1245. # Post-condition: node is schedulable
  1246. kube::test::get_object_assert "nodes 127.0.0.1" "{{.spec.unschedulable}}" '<no value>'
  1247. # check webhook token authentication endpoint, kubectl doesn't actually display the returned object so this isn't super useful
  1248. # but it proves that works
  1249. kubectl create -f test/fixtures/pkg/kubectl/cmd/create/tokenreview-v1beta1.json --validate=false
  1250. kubectl create -f test/fixtures/pkg/kubectl/cmd/create/tokenreview-v1.json --validate=false
  1251. set +o nounset
  1252. set +o errexit
  1253. }
  1254. run_pod_templates_tests() {
  1255. set -o nounset
  1256. set -o errexit
  1257. create_and_use_new_namespace
  1258. kube::log::status "Testing pod templates"
  1259. ### Create PODTEMPLATE
  1260. # Pre-condition: no PODTEMPLATE
  1261. kube::test::get_object_assert podtemplates "{{range.items}}{{.metadata.name}}:{{end}}" ''
  1262. # Command
  1263. kubectl create -f test/fixtures/doc-yaml/user-guide/walkthrough/podtemplate.json "${kube_flags[@]}"
  1264. # Post-condition: nginx PODTEMPLATE is available
  1265. kube::test::get_object_assert podtemplates "{{range.items}}{{.metadata.name}}:{{end}}" 'nginx:'
  1266. ### Printing pod templates works
  1267. kubectl get podtemplates "${kube_flags[@]}"
  1268. grep -q nginx <<< "$(kubectl get podtemplates -o yaml "${kube_flags[@]}")"
  1269. ### Delete nginx pod template by name
  1270. # Pre-condition: nginx pod template is available
  1271. kube::test::get_object_assert podtemplates "{{range.items}}{{.metadata.name}}:{{end}}" 'nginx:'
  1272. # Command
  1273. kubectl delete podtemplate nginx "${kube_flags[@]}"
  1274. # Post-condition: No templates exist
  1275. kube::test::get_object_assert podtemplate "{{range.items}}{{.metadata.name}}:{{end}}" ''
  1276. set +o nounset
  1277. set +o errexit
  1278. }