From b58f60d347542ba92270b69e71b9b3d2d3b70eeb Mon Sep 17 00:00:00 2001 From: Richard Theis Date: Wed, 19 Jun 2019 16:26:48 -0500 Subject: [PATCH] Update conformance results for IKS v1.13 and v1.14 (#626) IBM Cloud Kubernetes Service conformance results updated for version 1.13.7 and 1.14.3. --- v1.13/iks/PRODUCT.yaml | 2 +- v1.13/iks/e2e.log | 17433 ++++++++++++++++++------------------ v1.13/iks/junit_01.xml | 3890 ++++---- v1.13/iks/version.txt | 4 +- v1.14/iks/PRODUCT.yaml | 2 +- v1.14/iks/e2e.log | 18791 ++++++++++++++++++++------------------- v1.14/iks/junit_01.xml | 7172 +++++++-------- v1.14/iks/version.txt | 4 +- 8 files changed, 24029 insertions(+), 23269 deletions(-) diff --git a/v1.13/iks/PRODUCT.yaml b/v1.13/iks/PRODUCT.yaml index 39a393d220..456a70c997 100644 --- a/v1.13/iks/PRODUCT.yaml +++ b/v1.13/iks/PRODUCT.yaml @@ -1,6 +1,6 @@ vendor: IBM name: IBM Cloud Kubernetes Service -version: 1.13.6 +version: 1.13.7 website_url: https://www.ibm.com/cloud/container-service documentation_url: https://cloud.ibm.com/docs/containers?topic=containers-container_index#container_index product_logo_url: https://raw.githubusercontent.com/ibm-containers/logo/master/iks.png diff --git a/v1.13/iks/e2e.log b/v1.13/iks/e2e.log index 926d97fa8f..9ceebd8d35 100644 --- a/v1.13/iks/e2e.log +++ b/v1.13/iks/e2e.log @@ -1,2137 +1,1906 @@ -I0513 18:41:11.252051 16 test_context.go:358] Using a temporary kubeconfig file from in-cluster config : /tmp/kubeconfig-555138423 -I0513 18:41:11.252137 16 e2e.go:224] Starting e2e run "ad7e8b22-75ae-11e9-8f67-2632f168be36" on Ginkgo node 1 +I0618 11:34:31.287091 17 test_context.go:358] Using a temporary kubeconfig file from in-cluster config : /tmp/kubeconfig-953583206 +I0618 11:34:31.287416 17 e2e.go:224] Starting e2e run "097fb1c0-91bd-11e9-bce2-ae54e022189f" on Ginkgo node 1 Running Suite: Kubernetes e2e suite =================================== -Random Seed: 1557772870 - Will randomize all specs +Random Seed: 1560857670 - Will randomize all specs Will run 201 of 1946 specs -May 13 18:41:11.368: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -May 13 18:41:11.370: INFO: Waiting up to 30m0s for all (but 0) nodes to be schedulable -May 13 18:41:11.440: INFO: Waiting up to 10m0s for all pods (need at least 0) in namespace 'kube-system' to be running and ready -May 13 18:41:11.497: INFO: 23 / 23 pods in namespace 'kube-system' are running and ready (0 seconds elapsed) -May 13 18:41:11.497: INFO: expected 11 pod replicas in namespace 'kube-system', 11 are Running and Ready. -May 13 18:41:11.497: INFO: Waiting up to 5m0s for all daemonsets in namespace 'kube-system' to start -May 13 18:41:11.515: INFO: 3 / 3 pods ready in namespace 'kube-system' in daemonset 'calico-node' (0 seconds elapsed) -May 13 18:41:11.515: INFO: 3 / 3 pods ready in namespace 'kube-system' in daemonset 'ibm-keepalived-watcher' (0 seconds elapsed) -May 13 18:41:11.515: INFO: 3 / 3 pods ready in namespace 'kube-system' in daemonset 'ibm-kube-fluentd' (0 seconds elapsed) -May 13 18:41:11.516: INFO: 0 / 0 pods ready in namespace 'kube-system' in daemonset 'ibm-master-proxy' (0 seconds elapsed) -May 13 18:41:11.516: INFO: 0 / 0 pods ready in namespace 'kube-system' in daemonset 'nvidia-driver-installer' (0 seconds elapsed) -May 13 18:41:11.516: INFO: 0 / 0 pods ready in namespace 'kube-system' in daemonset 'nvidia-gpu-device-plugin' (0 seconds elapsed) -May 13 18:41:11.516: INFO: e2e test version: v1.13.0 -May 13 18:41:11.518: INFO: kube-apiserver version: v1.13.6+IKS -[k8s.io] InitContainer [NodeConformance] - should not start app containers and fail the pod if init containers fail on a RestartNever pod [Conformance] +Jun 18 11:34:31.433: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 +Jun 18 11:34:31.436: INFO: Waiting up to 30m0s for all (but 0) nodes to be schedulable +Jun 18 11:34:31.514: INFO: Waiting up to 10m0s for all pods (need at least 0) in namespace 'kube-system' to be running and ready +Jun 18 11:34:31.605: INFO: 23 / 23 pods in namespace 'kube-system' are running and ready (0 seconds elapsed) +Jun 18 11:34:31.605: INFO: expected 11 pod replicas in namespace 'kube-system', 11 are Running and Ready. +Jun 18 11:34:31.605: INFO: Waiting up to 5m0s for all daemonsets in namespace 'kube-system' to start +Jun 18 11:34:31.629: INFO: 3 / 3 pods ready in namespace 'kube-system' in daemonset 'calico-node' (0 seconds elapsed) +Jun 18 11:34:31.629: INFO: 3 / 3 pods ready in namespace 'kube-system' in daemonset 'ibm-keepalived-watcher' (0 seconds elapsed) +Jun 18 11:34:31.629: INFO: 3 / 3 pods ready in namespace 'kube-system' in daemonset 'ibm-kube-fluentd' (0 seconds elapsed) +Jun 18 11:34:31.629: INFO: 0 / 0 pods ready in namespace 'kube-system' in daemonset 'ibm-master-proxy' (0 seconds elapsed) +Jun 18 11:34:31.629: INFO: 0 / 0 pods ready in namespace 'kube-system' in daemonset 'nvidia-driver-installer' (0 seconds elapsed) +Jun 18 11:34:31.629: INFO: 0 / 0 pods ready in namespace 'kube-system' in daemonset 'nvidia-gpu-device-plugin' (0 seconds elapsed) +Jun 18 11:34:31.629: INFO: e2e test version: v1.13.0 +Jun 18 11:34:31.634: INFO: kube-apiserver version: v1.13.7+IKS +SSS +------------------------------ +[sig-scheduling] SchedulerPredicates [Serial] + validates that NodeSelector is respected if matching [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -[BeforeEach] [k8s.io] InitContainer [NodeConformance] +[BeforeEach] [sig-scheduling] SchedulerPredicates [Serial] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client -May 13 18:41:11.518: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -STEP: Building a namespace api object, basename init-container -May 13 18:41:12.529: INFO: Found PodSecurityPolicies; assuming PodSecurityPolicy is enabled. -May 13 18:41:12.550: INFO: Found ClusterRoles; assuming RBAC is enabled. -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-init-container-bs6n8 +Jun 18 11:34:31.635: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 +STEP: Building a namespace api object, basename sched-pred +Jun 18 11:34:32.035: INFO: Found PodSecurityPolicies; assuming PodSecurityPolicy is enabled. +Jun 18 11:34:32.092: INFO: Found ClusterRoles; assuming RBAC is enabled. +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-sched-pred-6gssw STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [k8s.io] InitContainer [NodeConformance] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/init_container.go:43 -[It] should not start app containers and fail the pod if init containers fail on a RestartNever pod [Conformance] +[BeforeEach] [sig-scheduling] SchedulerPredicates [Serial] + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:79 +Jun 18 11:34:32.235: INFO: Waiting up to 1m0s for all (but 0) nodes to be ready +Jun 18 11:34:32.264: INFO: Waiting for terminating namespaces to be deleted... +Jun 18 11:34:32.279: INFO: +Logging pods the kubelet thinks is on node 10.72.74.143 before test +Jun 18 11:34:32.331: INFO: calico-node-fw2l9 from kube-system started at 2019-06-17 21:36:37 +0000 UTC (1 container statuses recorded) +Jun 18 11:34:32.332: INFO: Container calico-node ready: true, restart count 0 +Jun 18 11:34:32.332: INFO: coredns-autoscaler-5c7646547d-dshx6 from kube-system started at 2019-06-17 21:36:47 +0000 UTC (1 container statuses recorded) +Jun 18 11:34:32.332: INFO: Container autoscaler ready: true, restart count 0 +Jun 18 11:34:32.332: INFO: ibm-master-proxy-static-10.72.74.143 from kube-system started at (0 container statuses recorded) +Jun 18 11:34:32.332: INFO: ibm-keepalived-watcher-5z7h2 from kube-system started at 2019-06-17 21:36:37 +0000 UTC (1 container statuses recorded) +Jun 18 11:34:32.332: INFO: Container keepalived-watcher ready: true, restart count 0 +Jun 18 11:34:32.332: INFO: kubernetes-dashboard-6cf8b975c-prz8l from kube-system started at 2019-06-17 21:36:47 +0000 UTC (1 container statuses recorded) +Jun 18 11:34:32.333: INFO: Container kubernetes-dashboard ready: true, restart count 0 +Jun 18 11:34:32.333: INFO: coredns-5545c6ddc4-dxkvs from kube-system started at 2019-06-17 21:36:47 +0000 UTC (1 container statuses recorded) +Jun 18 11:34:32.333: INFO: Container coredns ready: true, restart count 0 +Jun 18 11:34:32.333: INFO: calico-kube-controllers-54d47c87f-kwkh9 from kube-system started at 2019-06-17 21:36:47 +0000 UTC (1 container statuses recorded) +Jun 18 11:34:32.333: INFO: Container calico-kube-controllers ready: true, restart count 0 +Jun 18 11:34:32.333: INFO: sonobuoy-systemd-logs-daemon-set-562f76bc52c447d0-zt85r from heptio-sonobuoy started at 2019-06-18 11:33:59 +0000 UTC (2 container statuses recorded) +Jun 18 11:34:32.333: INFO: Container sonobuoy-worker ready: true, restart count 0 +Jun 18 11:34:32.333: INFO: Container systemd-logs ready: true, restart count 0 +Jun 18 11:34:32.333: INFO: vpn-7f677b8cb5-29tf9 from kube-system started at 2019-06-17 21:36:47 +0000 UTC (1 container statuses recorded) +Jun 18 11:34:32.333: INFO: Container vpn ready: true, restart count 0 +Jun 18 11:34:32.333: INFO: test-k8s-e2e-pvg-master-verification from default started at 2019-06-18 11:33:43 +0000 UTC (1 container statuses recorded) +Jun 18 11:34:32.334: INFO: Container test-k8s-e2e-pvg-master-verification ready: true, restart count 0 +Jun 18 11:34:32.334: INFO: ibm-kube-fluentd-7spm2 from kube-system started at 2019-06-17 21:43:21 +0000 UTC (1 container statuses recorded) +Jun 18 11:34:32.334: INFO: Container fluentd ready: true, restart count 0 +Jun 18 11:34:32.334: INFO: ibm-file-plugin-bf4cc7987-jwdjh from kube-system started at 2019-06-17 21:36:47 +0000 UTC (1 container statuses recorded) +Jun 18 11:34:32.334: INFO: Container ibm-file-plugin-container ready: true, restart count 0 +Jun 18 11:34:32.334: INFO: ibm-storage-watcher-64989c44d-tp68k from kube-system started at 2019-06-17 21:36:47 +0000 UTC (1 container statuses recorded) +Jun 18 11:34:32.334: INFO: Container ibm-storage-watcher-container ready: true, restart count 0 +Jun 18 11:34:32.334: INFO: sonobuoy from heptio-sonobuoy started at 2019-06-18 11:33:50 +0000 UTC (1 container statuses recorded) +Jun 18 11:34:32.334: INFO: Container kube-sonobuoy ready: true, restart count 0 +Jun 18 11:34:32.334: INFO: sonobuoy-e2e-job-4f826760f7504668 from heptio-sonobuoy started at 2019-06-18 11:33:59 +0000 UTC (2 container statuses recorded) +Jun 18 11:34:32.334: INFO: Container e2e ready: true, restart count 0 +Jun 18 11:34:32.334: INFO: Container sonobuoy-worker ready: true, restart count 0 +Jun 18 11:34:32.334: INFO: +Logging pods the kubelet thinks is on node 10.72.74.144 before test +Jun 18 11:34:32.448: INFO: sonobuoy-systemd-logs-daemon-set-562f76bc52c447d0-ct76c from heptio-sonobuoy started at 2019-06-18 11:33:59 +0000 UTC (2 container statuses recorded) +Jun 18 11:34:32.448: INFO: Container sonobuoy-worker ready: true, restart count 0 +Jun 18 11:34:32.448: INFO: Container systemd-logs ready: true, restart count 0 +Jun 18 11:34:32.448: INFO: ibm-master-proxy-static-10.72.74.144 from kube-system started at (0 container statuses recorded) +Jun 18 11:34:32.448: INFO: calico-node-rptvs from kube-system started at 2019-06-17 21:36:43 +0000 UTC (1 container statuses recorded) +Jun 18 11:34:32.448: INFO: Container calico-node ready: true, restart count 0 +Jun 18 11:34:32.448: INFO: coredns-5545c6ddc4-4s87g from kube-system started at 2019-06-17 21:37:04 +0000 UTC (1 container statuses recorded) +Jun 18 11:34:32.448: INFO: Container coredns ready: true, restart count 0 +Jun 18 11:34:32.448: INFO: ibm-keepalived-watcher-drbmt from kube-system started at 2019-06-17 21:36:43 +0000 UTC (1 container statuses recorded) +Jun 18 11:34:32.448: INFO: Container keepalived-watcher ready: true, restart count 0 +Jun 18 11:34:32.448: INFO: ibm-cloud-provider-ip-158-176-120-130-699ff5cfd-z4hhb from ibm-system started at 2019-06-17 21:40:39 +0000 UTC (1 container statuses recorded) +Jun 18 11:34:32.448: INFO: Container ibm-cloud-provider-ip-158-176-120-130 ready: true, restart count 0 +Jun 18 11:34:32.448: INFO: public-cr49a3e8d7011b436d9b4596ba0f279008-alb1-778b7ff477-tpktg from kube-system started at 2019-06-17 21:41:04 +0000 UTC (4 container statuses recorded) +Jun 18 11:34:32.448: INFO: Container ingress-auth-1 ready: true, restart count 0 +Jun 18 11:34:32.448: INFO: Container ingress-auth-2 ready: true, restart count 0 +Jun 18 11:34:32.448: INFO: Container ingress-auth-3 ready: true, restart count 0 +Jun 18 11:34:32.448: INFO: Container nginx-ingress ready: true, restart count 0 +Jun 18 11:34:32.448: INFO: ibm-kube-fluentd-g5hgb from kube-system started at 2019-06-17 21:43:21 +0000 UTC (1 container statuses recorded) +Jun 18 11:34:32.448: INFO: Container fluentd ready: true, restart count 0 +Jun 18 11:34:32.448: INFO: +Logging pods the kubelet thinks is on node 10.72.74.149 before test +Jun 18 11:34:32.638: INFO: ibm-cloud-provider-ip-158-176-120-130-699ff5cfd-td8hg from ibm-system started at 2019-06-17 21:40:39 +0000 UTC (1 container statuses recorded) +Jun 18 11:34:32.638: INFO: Container ibm-cloud-provider-ip-158-176-120-130 ready: true, restart count 0 +Jun 18 11:34:32.638: INFO: public-cr49a3e8d7011b436d9b4596ba0f279008-alb1-778b7ff477-sxttq from kube-system started at 2019-06-17 21:41:04 +0000 UTC (4 container statuses recorded) +Jun 18 11:34:32.638: INFO: Container ingress-auth-1 ready: true, restart count 0 +Jun 18 11:34:32.638: INFO: Container ingress-auth-2 ready: true, restart count 0 +Jun 18 11:34:32.638: INFO: Container ingress-auth-3 ready: true, restart count 0 +Jun 18 11:34:32.638: INFO: Container nginx-ingress ready: true, restart count 0 +Jun 18 11:34:32.638: INFO: ibm-kube-fluentd-c6kth from kube-system started at 2019-06-17 21:43:21 +0000 UTC (1 container statuses recorded) +Jun 18 11:34:32.638: INFO: Container fluentd ready: true, restart count 0 +Jun 18 11:34:32.638: INFO: sonobuoy-systemd-logs-daemon-set-562f76bc52c447d0-btfpp from heptio-sonobuoy started at 2019-06-18 11:33:59 +0000 UTC (2 container statuses recorded) +Jun 18 11:34:32.638: INFO: Container sonobuoy-worker ready: true, restart count 0 +Jun 18 11:34:32.638: INFO: Container systemd-logs ready: true, restart count 0 +Jun 18 11:34:32.638: INFO: ibm-master-proxy-static-10.72.74.149 from kube-system started at (0 container statuses recorded) +Jun 18 11:34:32.638: INFO: metrics-server-6ccf788d5b-6gwxm from kube-system started at 2019-06-17 21:37:11 +0000 UTC (2 container statuses recorded) +Jun 18 11:34:32.638: INFO: Container metrics-server ready: true, restart count 0 +Jun 18 11:34:32.638: INFO: Container metrics-server-nanny ready: true, restart count 0 +Jun 18 11:34:32.638: INFO: ibm-keepalived-watcher-6846v from kube-system started at 2019-06-17 21:36:50 +0000 UTC (1 container statuses recorded) +Jun 18 11:34:32.638: INFO: Container keepalived-watcher ready: true, restart count 0 +Jun 18 11:34:32.638: INFO: calico-node-4pqtj from kube-system started at 2019-06-17 21:36:50 +0000 UTC (1 container statuses recorded) +Jun 18 11:34:32.639: INFO: Container calico-node ready: true, restart count 0 +[It] validates that NodeSelector is respected if matching [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -STEP: creating the pod -May 13 18:41:12.670: INFO: PodSpec: initContainers in spec.initContainers -[AfterEach] [k8s.io] InitContainer [NodeConformance] +STEP: Trying to launch a pod without a label to get a node which can launch it. +STEP: Explicitly delete pod here to free the resource it takes. +STEP: Trying to apply a random label on the found node. +STEP: verifying the node has the label kubernetes.io/e2e-0d356e86-91bd-11e9-bce2-ae54e022189f 42 +STEP: Trying to relaunch the pod, now with labels. +STEP: removing the label kubernetes.io/e2e-0d356e86-91bd-11e9-bce2-ae54e022189f off the node 10.72.74.149 +STEP: verifying the node doesn't have the label kubernetes.io/e2e-0d356e86-91bd-11e9-bce2-ae54e022189f +[AfterEach] [sig-scheduling] SchedulerPredicates [Serial] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 18:41:16.145: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-init-container-bs6n8" for this suite. -May 13 18:41:22.197: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 18:41:22.293: INFO: namespace: e2e-tests-init-container-bs6n8, resource: bindings, ignored listing per whitelist -May 13 18:41:22.445: INFO: namespace e2e-tests-init-container-bs6n8 deletion completed in 6.290767253s +Jun 18 11:34:38.944: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "e2e-tests-sched-pred-6gssw" for this suite. +Jun 18 11:34:59.040: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 11:34:59.600: INFO: namespace: e2e-tests-sched-pred-6gssw, resource: bindings, ignored listing per whitelist +Jun 18 11:34:59.774: INFO: namespace e2e-tests-sched-pred-6gssw deletion completed in 20.78903945s +[AfterEach] [sig-scheduling] SchedulerPredicates [Serial] + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:70 -• [SLOW TEST:10.926 seconds] -[k8s.io] InitContainer [NodeConformance] -/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 - should not start app containers and fail the pod if init containers fail on a RestartNever pod [Conformance] +• [SLOW TEST:28.139 seconds] +[sig-scheduling] SchedulerPredicates [Serial] +/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/framework.go:22 + validates that NodeSelector is respected if matching [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ -SSSSSSSSS +SSSSSSSS ------------------------------ -[sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] - Scaling should happen in predictable order and halt if any stateful pod is unhealthy [Conformance] +[sig-storage] EmptyDir volumes + should support (root,0666,tmpfs) [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -[BeforeEach] [sig-apps] StatefulSet +[BeforeEach] [sig-storage] EmptyDir volumes /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client -May 13 18:41:22.445: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -STEP: Building a namespace api object, basename statefulset -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-statefulset-vhs7b +Jun 18 11:34:59.774: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 +STEP: Building a namespace api object, basename emptydir +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-emptydir-5wsxk STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [sig-apps] StatefulSet - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:59 -[BeforeEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:74 -STEP: Creating service test in namespace e2e-tests-statefulset-vhs7b -[It] Scaling should happen in predictable order and halt if any stateful pod is unhealthy [Conformance] +[It] should support (root,0666,tmpfs) [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -STEP: Initializing watcher for selector baz=blah,foo=bar -STEP: Creating stateful set ss in namespace e2e-tests-statefulset-vhs7b -STEP: Waiting until all stateful set ss replicas will be running in namespace e2e-tests-statefulset-vhs7b -May 13 18:41:22.849: INFO: Found 0 stateful pods, waiting for 1 -May 13 18:41:32.871: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true -STEP: Confirming that stateful set scale up will halt with unhealthy stateful pod -May 13 18:41:32.878: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 exec --namespace=e2e-tests-statefulset-vhs7b ss-0 -- /bin/sh -c mv -v /usr/share/nginx/html/index.html /tmp/ || true' -May 13 18:41:33.197: INFO: stderr: "" -May 13 18:41:33.197: INFO: stdout: "'/usr/share/nginx/html/index.html' -> '/tmp/index.html'\n" -May 13 18:41:33.197: INFO: stdout of mv -v /usr/share/nginx/html/index.html /tmp/ || true on ss-0: '/usr/share/nginx/html/index.html' -> '/tmp/index.html' - -May 13 18:41:33.205: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=true -May 13 18:41:43.226: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=false -May 13 18:41:43.226: INFO: Waiting for statefulset status.replicas updated to 0 -May 13 18:41:43.334: INFO: Verifying statefulset ss doesn't scale past 1 for another 9.999998915s -May 13 18:41:44.343: INFO: Verifying statefulset ss doesn't scale past 1 for another 8.99239426s -May 13 18:41:45.356: INFO: Verifying statefulset ss doesn't scale past 1 for another 7.983535621s -May 13 18:41:46.365: INFO: Verifying statefulset ss doesn't scale past 1 for another 6.970246819s -May 13 18:41:47.374: INFO: Verifying statefulset ss doesn't scale past 1 for another 5.961626761s -May 13 18:41:48.382: INFO: Verifying statefulset ss doesn't scale past 1 for another 4.952328786s -May 13 18:41:49.391: INFO: Verifying statefulset ss doesn't scale past 1 for another 3.944525454s -May 13 18:41:50.399: INFO: Verifying statefulset ss doesn't scale past 1 for another 2.935752056s -May 13 18:41:51.408: INFO: Verifying statefulset ss doesn't scale past 1 for another 1.927516593s -May 13 18:41:52.416: INFO: Verifying statefulset ss doesn't scale past 1 for another 918.673616ms -STEP: Scaling up stateful set ss to 3 replicas and waiting until all of them will be running in namespace e2e-tests-statefulset-vhs7b -May 13 18:41:53.437: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 exec --namespace=e2e-tests-statefulset-vhs7b ss-0 -- /bin/sh -c mv -v /tmp/index.html /usr/share/nginx/html/ || true' -May 13 18:41:53.727: INFO: stderr: "" -May 13 18:41:53.727: INFO: stdout: "'/tmp/index.html' -> '/usr/share/nginx/html/index.html'\n" -May 13 18:41:53.727: INFO: stdout of mv -v /tmp/index.html /usr/share/nginx/html/ || true on ss-0: '/tmp/index.html' -> '/usr/share/nginx/html/index.html' - -May 13 18:41:53.827: INFO: Found 1 stateful pods, waiting for 3 -May 13 18:42:03.850: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true -May 13 18:42:03.850: INFO: Waiting for pod ss-1 to enter Running - Ready=true, currently Running - Ready=true -May 13 18:42:03.850: INFO: Waiting for pod ss-2 to enter Running - Ready=true, currently Running - Ready=true -STEP: Verifying that stateful set ss was scaled up in order -STEP: Scale down will halt with unhealthy stateful pod -May 13 18:42:03.920: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 exec --namespace=e2e-tests-statefulset-vhs7b ss-0 -- /bin/sh -c mv -v /usr/share/nginx/html/index.html /tmp/ || true' -May 13 18:42:04.175: INFO: stderr: "" -May 13 18:42:04.175: INFO: stdout: "'/usr/share/nginx/html/index.html' -> '/tmp/index.html'\n" -May 13 18:42:04.175: INFO: stdout of mv -v /usr/share/nginx/html/index.html /tmp/ || true on ss-0: '/usr/share/nginx/html/index.html' -> '/tmp/index.html' - -May 13 18:42:04.175: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 exec --namespace=e2e-tests-statefulset-vhs7b ss-1 -- /bin/sh -c mv -v /usr/share/nginx/html/index.html /tmp/ || true' -May 13 18:42:04.502: INFO: stderr: "" -May 13 18:42:04.502: INFO: stdout: "'/usr/share/nginx/html/index.html' -> '/tmp/index.html'\n" -May 13 18:42:04.502: INFO: stdout of mv -v /usr/share/nginx/html/index.html /tmp/ || true on ss-1: '/usr/share/nginx/html/index.html' -> '/tmp/index.html' - -May 13 18:42:04.502: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 exec --namespace=e2e-tests-statefulset-vhs7b ss-2 -- /bin/sh -c mv -v /usr/share/nginx/html/index.html /tmp/ || true' -May 13 18:42:04.828: INFO: stderr: "" -May 13 18:42:04.828: INFO: stdout: "'/usr/share/nginx/html/index.html' -> '/tmp/index.html'\n" -May 13 18:42:04.828: INFO: stdout of mv -v /usr/share/nginx/html/index.html /tmp/ || true on ss-2: '/usr/share/nginx/html/index.html' -> '/tmp/index.html' - -May 13 18:42:04.828: INFO: Waiting for statefulset status.replicas updated to 0 -May 13 18:42:04.834: INFO: Waiting for stateful set status.readyReplicas to become 0, currently 1 -May 13 18:42:14.862: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=false -May 13 18:42:14.862: INFO: Waiting for pod ss-1 to enter Running - Ready=false, currently Running - Ready=false -May 13 18:42:14.862: INFO: Waiting for pod ss-2 to enter Running - Ready=false, currently Running - Ready=false -May 13 18:42:14.920: INFO: Verifying statefulset ss doesn't scale past 3 for another 9.999998974s -May 13 18:42:16.020: INFO: Verifying statefulset ss doesn't scale past 3 for another 8.953782895s -May 13 18:42:17.029: INFO: Verifying statefulset ss doesn't scale past 3 for another 7.853669749s -May 13 18:42:18.038: INFO: Verifying statefulset ss doesn't scale past 3 for another 6.844568726s -May 13 18:42:19.048: INFO: Verifying statefulset ss doesn't scale past 3 for another 5.835519558s -May 13 18:42:20.057: INFO: Verifying statefulset ss doesn't scale past 3 for another 4.825283464s -May 13 18:42:21.066: INFO: Verifying statefulset ss doesn't scale past 3 for another 3.816367746s -May 13 18:42:22.075: INFO: Verifying statefulset ss doesn't scale past 3 for another 2.807599301s -May 13 18:42:23.084: INFO: Verifying statefulset ss doesn't scale past 3 for another 1.798430196s -May 13 18:42:24.092: INFO: Verifying statefulset ss doesn't scale past 3 for another 789.787963ms -STEP: Scaling down stateful set ss to 0 replicas and waiting until none of pods will run in namespacee2e-tests-statefulset-vhs7b -May 13 18:42:25.113: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 exec --namespace=e2e-tests-statefulset-vhs7b ss-0 -- /bin/sh -c mv -v /tmp/index.html /usr/share/nginx/html/ || true' -May 13 18:42:25.370: INFO: stderr: "" -May 13 18:42:25.370: INFO: stdout: "'/tmp/index.html' -> '/usr/share/nginx/html/index.html'\n" -May 13 18:42:25.370: INFO: stdout of mv -v /tmp/index.html /usr/share/nginx/html/ || true on ss-0: '/tmp/index.html' -> '/usr/share/nginx/html/index.html' - -May 13 18:42:25.370: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 exec --namespace=e2e-tests-statefulset-vhs7b ss-1 -- /bin/sh -c mv -v /tmp/index.html /usr/share/nginx/html/ || true' -May 13 18:42:25.645: INFO: stderr: "" -May 13 18:42:25.645: INFO: stdout: "'/tmp/index.html' -> '/usr/share/nginx/html/index.html'\n" -May 13 18:42:25.645: INFO: stdout of mv -v /tmp/index.html /usr/share/nginx/html/ || true on ss-1: '/tmp/index.html' -> '/usr/share/nginx/html/index.html' - -May 13 18:42:25.645: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 exec --namespace=e2e-tests-statefulset-vhs7b ss-2 -- /bin/sh -c mv -v /tmp/index.html /usr/share/nginx/html/ || true' -May 13 18:42:25.933: INFO: stderr: "" -May 13 18:42:25.933: INFO: stdout: "'/tmp/index.html' -> '/usr/share/nginx/html/index.html'\n" -May 13 18:42:25.933: INFO: stdout of mv -v /tmp/index.html /usr/share/nginx/html/ || true on ss-2: '/tmp/index.html' -> '/usr/share/nginx/html/index.html' - -May 13 18:42:25.933: INFO: Scaling statefulset ss to 0 -STEP: Verifying that stateful set ss was scaled down in reverse order -[AfterEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:85 -May 13 18:42:46.018: INFO: Deleting all statefulset in ns e2e-tests-statefulset-vhs7b -May 13 18:42:46.024: INFO: Scaling statefulset ss to 0 -May 13 18:42:46.053: INFO: Waiting for statefulset status.replicas updated to 0 -May 13 18:42:46.058: INFO: Deleting statefulset ss -[AfterEach] [sig-apps] StatefulSet +STEP: Creating a pod to test emptydir 0666 on tmpfs +Jun 18 11:35:00.253: INFO: Waiting up to 5m0s for pod "pod-1b2cd48c-91bd-11e9-bce2-ae54e022189f" in namespace "e2e-tests-emptydir-5wsxk" to be "success or failure" +Jun 18 11:35:00.268: INFO: Pod "pod-1b2cd48c-91bd-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 14.299289ms +Jun 18 11:35:02.289: INFO: Pod "pod-1b2cd48c-91bd-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 2.035284058s +Jun 18 11:35:04.303: INFO: Pod "pod-1b2cd48c-91bd-11e9-bce2-ae54e022189f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.049995698s +STEP: Saw pod success +Jun 18 11:35:04.303: INFO: Pod "pod-1b2cd48c-91bd-11e9-bce2-ae54e022189f" satisfied condition "success or failure" +Jun 18 11:35:04.401: INFO: Trying to get logs from node 10.72.74.149 pod pod-1b2cd48c-91bd-11e9-bce2-ae54e022189f container test-container: +STEP: delete the pod +Jun 18 11:35:04.480: INFO: Waiting for pod pod-1b2cd48c-91bd-11e9-bce2-ae54e022189f to disappear +Jun 18 11:35:04.494: INFO: Pod pod-1b2cd48c-91bd-11e9-bce2-ae54e022189f no longer exists +[AfterEach] [sig-storage] EmptyDir volumes /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 18:42:46.082: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-statefulset-vhs7b" for this suite. -May 13 18:42:52.134: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 18:42:52.177: INFO: namespace: e2e-tests-statefulset-vhs7b, resource: bindings, ignored listing per whitelist -May 13 18:42:52.379: INFO: namespace e2e-tests-statefulset-vhs7b deletion completed in 6.283998639s +Jun 18 11:35:04.494: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "e2e-tests-emptydir-5wsxk" for this suite. +Jun 18 11:35:12.565: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 11:35:12.751: INFO: namespace: e2e-tests-emptydir-5wsxk, resource: bindings, ignored listing per whitelist +Jun 18 11:35:13.046: INFO: namespace e2e-tests-emptydir-5wsxk deletion completed in 8.528593681s -• [SLOW TEST:89.934 seconds] -[sig-apps] StatefulSet -/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22 - [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 - Scaling should happen in predictable order and halt if any stateful pod is unhealthy [Conformance] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 +• [SLOW TEST:13.272 seconds] +[sig-storage] EmptyDir volumes +/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:40 + should support (root,0666,tmpfs) [NodeConformance] [Conformance] + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ -SS +SSSSSSSSSSSSS ------------------------------ -[sig-storage] ConfigMap - should be consumable from pods in volume with mappings and Item mode set [NodeConformance] [Conformance] +[k8s.io] Pods + should be submitted and removed [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -[BeforeEach] [sig-storage] ConfigMap +[BeforeEach] [k8s.io] Pods /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client -May 13 18:42:52.379: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -STEP: Building a namespace api object, basename configmap -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-configmap-nd2ws +Jun 18 11:35:13.049: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 +STEP: Building a namespace api object, basename pods +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-pods-lzfb8 STEP: Waiting for a default service account to be provisioned in namespace -[It] should be consumable from pods in volume with mappings and Item mode set [NodeConformance] [Conformance] +[BeforeEach] [k8s.io] Pods + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:132 +[It] should be submitted and removed [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -STEP: Creating configMap with name configmap-test-volume-map-ea42d42e-75ae-11e9-8f67-2632f168be36 -STEP: Creating a pod to test consume configMaps -May 13 18:42:52.675: INFO: Waiting up to 5m0s for pod "pod-configmaps-ea4439f4-75ae-11e9-8f67-2632f168be36" in namespace "e2e-tests-configmap-nd2ws" to be "success or failure" -May 13 18:42:52.683: INFO: Pod "pod-configmaps-ea4439f4-75ae-11e9-8f67-2632f168be36": Phase="Pending", Reason="", readiness=false. Elapsed: 7.247409ms -May 13 18:42:54.691: INFO: Pod "pod-configmaps-ea4439f4-75ae-11e9-8f67-2632f168be36": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.015426653s +STEP: creating the pod +STEP: setting up watch +STEP: submitting the pod to kubernetes +STEP: verifying the pod is in kubernetes +STEP: verifying pod creation was observed +Jun 18 11:35:19.701: INFO: running pod: &v1.Pod{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"pod-submit-remove-231d6bba-91bd-11e9-bce2-ae54e022189f", GenerateName:"", Namespace:"e2e-tests-pods-lzfb8", SelfLink:"/api/v1/namespaces/e2e-tests-pods-lzfb8/pods/pod-submit-remove-231d6bba-91bd-11e9-bce2-ae54e022189f", UID:"23232966-91bd-11e9-bf44-fa6f350b29f0", ResourceVersion:"88488", Generation:0, CreationTimestamp:v1.Time{Time:time.Time{wall:0x0, ext:63696454513, loc:(*time.Location)(0x7b33b80)}}, DeletionTimestamp:(*v1.Time)(nil), DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"name":"foo", "time":"548800715"}, Annotations:map[string]string{"kubernetes.io/psp":"e2e-test-privileged-psp"}, OwnerReferences:[]v1.OwnerReference(nil), Initializers:(*v1.Initializers)(nil), Finalizers:[]string(nil), ClusterName:""}, Spec:v1.PodSpec{Volumes:[]v1.Volume{v1.Volume{Name:"default-token-7l7c2", VolumeSource:v1.VolumeSource{HostPath:(*v1.HostPathVolumeSource)(nil), EmptyDir:(*v1.EmptyDirVolumeSource)(nil), GCEPersistentDisk:(*v1.GCEPersistentDiskVolumeSource)(nil), AWSElasticBlockStore:(*v1.AWSElasticBlockStoreVolumeSource)(nil), GitRepo:(*v1.GitRepoVolumeSource)(nil), Secret:(*v1.SecretVolumeSource)(0xc001382ac0), NFS:(*v1.NFSVolumeSource)(nil), ISCSI:(*v1.ISCSIVolumeSource)(nil), Glusterfs:(*v1.GlusterfsVolumeSource)(nil), PersistentVolumeClaim:(*v1.PersistentVolumeClaimVolumeSource)(nil), RBD:(*v1.RBDVolumeSource)(nil), FlexVolume:(*v1.FlexVolumeSource)(nil), Cinder:(*v1.CinderVolumeSource)(nil), CephFS:(*v1.CephFSVolumeSource)(nil), Flocker:(*v1.FlockerVolumeSource)(nil), DownwardAPI:(*v1.DownwardAPIVolumeSource)(nil), FC:(*v1.FCVolumeSource)(nil), AzureFile:(*v1.AzureFileVolumeSource)(nil), ConfigMap:(*v1.ConfigMapVolumeSource)(nil), VsphereVolume:(*v1.VsphereVirtualDiskVolumeSource)(nil), Quobyte:(*v1.QuobyteVolumeSource)(nil), AzureDisk:(*v1.AzureDiskVolumeSource)(nil), PhotonPersistentDisk:(*v1.PhotonPersistentDiskVolumeSource)(nil), Projected:(*v1.ProjectedVolumeSource)(nil), PortworxVolume:(*v1.PortworxVolumeSource)(nil), ScaleIO:(*v1.ScaleIOVolumeSource)(nil), StorageOS:(*v1.StorageOSVolumeSource)(nil)}}}, InitContainers:[]v1.Container(nil), Containers:[]v1.Container{v1.Container{Name:"nginx", Image:"docker.io/library/nginx:1.14-alpine", Command:[]string(nil), Args:[]string(nil), WorkingDir:"", Ports:[]v1.ContainerPort(nil), EnvFrom:[]v1.EnvFromSource(nil), Env:[]v1.EnvVar(nil), Resources:v1.ResourceRequirements{Limits:v1.ResourceList(nil), Requests:v1.ResourceList(nil)}, VolumeMounts:[]v1.VolumeMount{v1.VolumeMount{Name:"default-token-7l7c2", ReadOnly:true, MountPath:"/var/run/secrets/kubernetes.io/serviceaccount", SubPath:"", MountPropagation:(*v1.MountPropagationMode)(nil)}}, VolumeDevices:[]v1.VolumeDevice(nil), LivenessProbe:(*v1.Probe)(nil), ReadinessProbe:(*v1.Probe)(nil), Lifecycle:(*v1.Lifecycle)(nil), TerminationMessagePath:"/dev/termination-log", TerminationMessagePolicy:"File", ImagePullPolicy:"IfNotPresent", SecurityContext:(*v1.SecurityContext)(nil), Stdin:false, StdinOnce:false, TTY:false}}, RestartPolicy:"Always", TerminationGracePeriodSeconds:(*int64)(0xc000fde338), ActiveDeadlineSeconds:(*int64)(nil), DNSPolicy:"ClusterFirst", NodeSelector:map[string]string(nil), ServiceAccountName:"default", DeprecatedServiceAccount:"default", AutomountServiceAccountToken:(*bool)(nil), NodeName:"10.72.74.143", HostNetwork:false, HostPID:false, HostIPC:false, ShareProcessNamespace:(*bool)(nil), SecurityContext:(*v1.PodSecurityContext)(0xc0001e2fc0), ImagePullSecrets:[]v1.LocalObjectReference(nil), Hostname:"", Subdomain:"", Affinity:(*v1.Affinity)(nil), SchedulerName:"default-scheduler", Tolerations:[]v1.Toleration{v1.Toleration{Key:"node.kubernetes.io/not-ready", Operator:"Exists", Value:"", Effect:"NoExecute", TolerationSeconds:(*int64)(0xc000fde380)}, v1.Toleration{Key:"node.kubernetes.io/unreachable", Operator:"Exists", Value:"", Effect:"NoExecute", TolerationSeconds:(*int64)(0xc000fde3a0)}}, HostAliases:[]v1.HostAlias(nil), PriorityClassName:"", Priority:(*int32)(0xc000fde3a8), DNSConfig:(*v1.PodDNSConfig)(nil), ReadinessGates:[]v1.PodReadinessGate(nil), RuntimeClassName:(*string)(nil), EnableServiceLinks:(*bool)(0xc000fde3ac)}, Status:v1.PodStatus{Phase:"Running", Conditions:[]v1.PodCondition{v1.PodCondition{Type:"Initialized", Status:"True", LastProbeTime:v1.Time{Time:time.Time{wall:0x0, ext:0, loc:(*time.Location)(nil)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63696454513, loc:(*time.Location)(0x7b33b80)}}, Reason:"", Message:""}, v1.PodCondition{Type:"Ready", Status:"True", LastProbeTime:v1.Time{Time:time.Time{wall:0x0, ext:0, loc:(*time.Location)(nil)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63696454518, loc:(*time.Location)(0x7b33b80)}}, Reason:"", Message:""}, v1.PodCondition{Type:"ContainersReady", Status:"True", LastProbeTime:v1.Time{Time:time.Time{wall:0x0, ext:0, loc:(*time.Location)(nil)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63696454518, loc:(*time.Location)(0x7b33b80)}}, Reason:"", Message:""}, v1.PodCondition{Type:"PodScheduled", Status:"True", LastProbeTime:v1.Time{Time:time.Time{wall:0x0, ext:0, loc:(*time.Location)(nil)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63696454513, loc:(*time.Location)(0x7b33b80)}}, Reason:"", Message:""}}, Message:"", Reason:"", NominatedNodeName:"", HostIP:"10.72.74.143", PodIP:"172.30.58.140", StartTime:(*v1.Time)(0xc001f7d700), InitContainerStatuses:[]v1.ContainerStatus(nil), ContainerStatuses:[]v1.ContainerStatus{v1.ContainerStatus{Name:"nginx", State:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(nil), Running:(*v1.ContainerStateRunning)(0xc001f7d720), Terminated:(*v1.ContainerStateTerminated)(nil)}, LastTerminationState:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(nil), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(nil)}, Ready:true, RestartCount:0, Image:"docker.io/library/nginx:1.14-alpine", ImageID:"docker.io/library/nginx@sha256:485b610fefec7ff6c463ced9623314a04ed67e3945b9c08d7e53a47f6d108dc7", ContainerID:"containerd://f00e194769cead829e4e05962add82a982710bf6ef5b60bb89c3902c42d7cfcb"}}, QOSClass:"BestEffort"}} +STEP: deleting the pod gracefully +STEP: verifying the kubelet observed the termination notice +Jun 18 11:35:24.787: INFO: no pod exists with the name we were looking for, assuming the termination request was observed and completed +STEP: verifying pod deletion was observed +[AfterEach] [k8s.io] Pods + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 +Jun 18 11:35:24.801: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "e2e-tests-pods-lzfb8" for this suite. +Jun 18 11:35:30.898: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 11:35:31.263: INFO: namespace: e2e-tests-pods-lzfb8, resource: bindings, ignored listing per whitelist +Jun 18 11:35:31.422: INFO: namespace e2e-tests-pods-lzfb8 deletion completed in 6.60600513s + +• [SLOW TEST:18.373 seconds] +[k8s.io] Pods +/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 + should be submitted and removed [NodeConformance] [Conformance] + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 +------------------------------ +SSS +------------------------------ +[sig-node] Downward API + should provide container's limits.cpu/memory and requests.cpu/memory as env vars [NodeConformance] [Conformance] + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 +[BeforeEach] [sig-node] Downward API + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 +STEP: Creating a kubernetes client +Jun 18 11:35:31.422: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 +STEP: Building a namespace api object, basename downward-api +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-downward-api-75d9j +STEP: Waiting for a default service account to be provisioned in namespace +[It] should provide container's limits.cpu/memory and requests.cpu/memory as env vars [NodeConformance] [Conformance] + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 +STEP: Creating a pod to test downward api env vars +Jun 18 11:35:31.897: INFO: Waiting up to 5m0s for pod "downward-api-2e09838e-91bd-11e9-bce2-ae54e022189f" in namespace "e2e-tests-downward-api-75d9j" to be "success or failure" +Jun 18 11:35:31.912: INFO: Pod "downward-api-2e09838e-91bd-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 14.631039ms +Jun 18 11:35:33.929: INFO: Pod "downward-api-2e09838e-91bd-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 2.031273594s +Jun 18 11:35:35.951: INFO: Pod "downward-api-2e09838e-91bd-11e9-bce2-ae54e022189f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.053079691s STEP: Saw pod success -May 13 18:42:54.691: INFO: Pod "pod-configmaps-ea4439f4-75ae-11e9-8f67-2632f168be36" satisfied condition "success or failure" -May 13 18:42:54.698: INFO: Trying to get logs from node 10.170.219.153 pod pod-configmaps-ea4439f4-75ae-11e9-8f67-2632f168be36 container configmap-volume-test: +Jun 18 11:35:35.951: INFO: Pod "downward-api-2e09838e-91bd-11e9-bce2-ae54e022189f" satisfied condition "success or failure" +Jun 18 11:35:35.966: INFO: Trying to get logs from node 10.72.74.149 pod downward-api-2e09838e-91bd-11e9-bce2-ae54e022189f container dapi-container: STEP: delete the pod -May 13 18:42:54.744: INFO: Waiting for pod pod-configmaps-ea4439f4-75ae-11e9-8f67-2632f168be36 to disappear -May 13 18:42:54.751: INFO: Pod pod-configmaps-ea4439f4-75ae-11e9-8f67-2632f168be36 no longer exists -[AfterEach] [sig-storage] ConfigMap +Jun 18 11:35:36.059: INFO: Waiting for pod downward-api-2e09838e-91bd-11e9-bce2-ae54e022189f to disappear +Jun 18 11:35:36.073: INFO: Pod downward-api-2e09838e-91bd-11e9-bce2-ae54e022189f no longer exists +[AfterEach] [sig-node] Downward API /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 18:42:54.751: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-configmap-nd2ws" for this suite. -May 13 18:43:02.788: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 18:43:02.822: INFO: namespace: e2e-tests-configmap-nd2ws, resource: bindings, ignored listing per whitelist -May 13 18:43:03.124: INFO: namespace e2e-tests-configmap-nd2ws deletion completed in 8.364655139s +Jun 18 11:35:36.073: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "e2e-tests-downward-api-75d9j" for this suite. +Jun 18 11:35:42.152: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 11:35:42.257: INFO: namespace: e2e-tests-downward-api-75d9j, resource: bindings, ignored listing per whitelist +Jun 18 11:35:42.801: INFO: namespace e2e-tests-downward-api-75d9j deletion completed in 6.701250959s -• [SLOW TEST:10.745 seconds] -[sig-storage] ConfigMap -/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:33 - should be consumable from pods in volume with mappings and Item mode set [NodeConformance] [Conformance] +• [SLOW TEST:11.378 seconds] +[sig-node] Downward API +/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downward_api.go:38 + should provide container's limits.cpu/memory and requests.cpu/memory as env vars [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ -SSSSSS +S ------------------------------ -[sig-cli] Kubectl client [k8s.io] Kubectl patch - should add annotations for pods in rc [Conformance] +[sig-storage] Subpath Atomic writer volumes + should support subpaths with configmap pod with mountPath of existing file [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -[BeforeEach] [sig-cli] Kubectl client +[BeforeEach] [sig-storage] Subpath /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client -May 13 18:43:03.124: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -STEP: Building a namespace api object, basename kubectl -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-kubectl-tqt6c +Jun 18 11:35:42.801: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 +STEP: Building a namespace api object, basename subpath +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-subpath-sv2dm STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [sig-cli] Kubectl client - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:243 -[It] should add annotations for pods in rc [Conformance] +[BeforeEach] Atomic writer volumes + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:38 +STEP: Setting up data +[It] should support subpaths with configmap pod with mountPath of existing file [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -STEP: creating Redis RC -May 13 18:43:03.406: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 create -f - --namespace=e2e-tests-kubectl-tqt6c' -May 13 18:43:03.878: INFO: stderr: "" -May 13 18:43:03.878: INFO: stdout: "replicationcontroller/redis-master created\n" -STEP: Waiting for Redis master to start. -May 13 18:43:04.886: INFO: Selector matched 1 pods for map[app:redis] -May 13 18:43:04.886: INFO: Found 0 / 1 -May 13 18:43:05.886: INFO: Selector matched 1 pods for map[app:redis] -May 13 18:43:05.887: INFO: Found 1 / 1 -May 13 18:43:05.887: INFO: WaitFor completed with timeout 5m0s. Pods found = 1 out of 1 -STEP: patching all pods -May 13 18:43:05.894: INFO: Selector matched 1 pods for map[app:redis] -May 13 18:43:05.894: INFO: ForEach: Found 1 pods from the filter. Now looping through them. -May 13 18:43:05.894: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 patch pod redis-master-nsh4l --namespace=e2e-tests-kubectl-tqt6c -p {"metadata":{"annotations":{"x":"y"}}}' -May 13 18:43:06.024: INFO: stderr: "" -May 13 18:43:06.024: INFO: stdout: "pod/redis-master-nsh4l patched\n" -STEP: checking annotations -May 13 18:43:06.034: INFO: Selector matched 1 pods for map[app:redis] -May 13 18:43:06.034: INFO: ForEach: Found 1 pods from the filter. Now looping through them. -[AfterEach] [sig-cli] Kubectl client +STEP: Creating pod pod-subpath-test-configmap-vf74 +STEP: Creating a pod to test atomic-volume-subpath +Jun 18 11:35:43.332: INFO: Waiting up to 5m0s for pod "pod-subpath-test-configmap-vf74" in namespace "e2e-tests-subpath-sv2dm" to be "success or failure" +Jun 18 11:35:43.345: INFO: Pod "pod-subpath-test-configmap-vf74": Phase="Pending", Reason="", readiness=false. Elapsed: 13.061065ms +Jun 18 11:35:45.360: INFO: Pod "pod-subpath-test-configmap-vf74": Phase="Pending", Reason="", readiness=false. Elapsed: 2.028183825s +Jun 18 11:35:47.399: INFO: Pod "pod-subpath-test-configmap-vf74": Phase="Running", Reason="", readiness=false. Elapsed: 4.066802388s +Jun 18 11:35:49.414: INFO: Pod "pod-subpath-test-configmap-vf74": Phase="Running", Reason="", readiness=false. Elapsed: 6.08196526s +Jun 18 11:35:51.445: INFO: Pod "pod-subpath-test-configmap-vf74": Phase="Running", Reason="", readiness=false. Elapsed: 8.113189468s +Jun 18 11:35:53.517: INFO: Pod "pod-subpath-test-configmap-vf74": Phase="Running", Reason="", readiness=false. Elapsed: 10.185651286s +Jun 18 11:35:55.532: INFO: Pod "pod-subpath-test-configmap-vf74": Phase="Running", Reason="", readiness=false. Elapsed: 12.200305368s +Jun 18 11:35:57.548: INFO: Pod "pod-subpath-test-configmap-vf74": Phase="Running", Reason="", readiness=false. Elapsed: 14.215898756s +Jun 18 11:35:59.563: INFO: Pod "pod-subpath-test-configmap-vf74": Phase="Running", Reason="", readiness=false. Elapsed: 16.231264019s +Jun 18 11:36:01.596: INFO: Pod "pod-subpath-test-configmap-vf74": Phase="Running", Reason="", readiness=false. Elapsed: 18.263674965s +Jun 18 11:36:03.612: INFO: Pod "pod-subpath-test-configmap-vf74": Phase="Running", Reason="", readiness=false. Elapsed: 20.280467994s +Jun 18 11:36:05.627: INFO: Pod "pod-subpath-test-configmap-vf74": Phase="Running", Reason="", readiness=false. Elapsed: 22.295303171s +Jun 18 11:36:07.642: INFO: Pod "pod-subpath-test-configmap-vf74": Phase="Succeeded", Reason="", readiness=false. Elapsed: 24.310376023s +STEP: Saw pod success +Jun 18 11:36:07.642: INFO: Pod "pod-subpath-test-configmap-vf74" satisfied condition "success or failure" +Jun 18 11:36:07.657: INFO: Trying to get logs from node 10.72.74.149 pod pod-subpath-test-configmap-vf74 container test-container-subpath-configmap-vf74: +STEP: delete the pod +Jun 18 11:36:07.760: INFO: Waiting for pod pod-subpath-test-configmap-vf74 to disappear +Jun 18 11:36:07.774: INFO: Pod pod-subpath-test-configmap-vf74 no longer exists +STEP: Deleting pod pod-subpath-test-configmap-vf74 +Jun 18 11:36:07.775: INFO: Deleting pod "pod-subpath-test-configmap-vf74" in namespace "e2e-tests-subpath-sv2dm" +[AfterEach] [sig-storage] Subpath /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 18:43:06.034: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-kubectl-tqt6c" for this suite. -May 13 18:43:30.069: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 18:43:30.202: INFO: namespace: e2e-tests-kubectl-tqt6c, resource: bindings, ignored listing per whitelist -May 13 18:43:30.366: INFO: namespace e2e-tests-kubectl-tqt6c deletion completed in 24.324937167s +Jun 18 11:36:07.790: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "e2e-tests-subpath-sv2dm" for this suite. +Jun 18 11:36:15.866: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 11:36:16.056: INFO: namespace: e2e-tests-subpath-sv2dm, resource: bindings, ignored listing per whitelist +Jun 18 11:36:16.396: INFO: namespace e2e-tests-subpath-sv2dm deletion completed in 8.584928496s -• [SLOW TEST:27.242 seconds] -[sig-cli] Kubectl client -/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:22 - [k8s.io] Kubectl patch - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 - should add annotations for pods in rc [Conformance] +• [SLOW TEST:33.596 seconds] +[sig-storage] Subpath +/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:22 + Atomic writer volumes + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:34 + should support subpaths with configmap pod with mountPath of existing file [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ S ------------------------------ -[sig-storage] Projected secret - should be consumable in multiple volumes in a pod [NodeConformance] [Conformance] +[sig-api-machinery] Garbage collector + should delete pods created by rc when not orphaning [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -[BeforeEach] [sig-storage] Projected secret +[BeforeEach] [sig-api-machinery] Garbage collector /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client -May 13 18:43:30.366: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -STEP: Building a namespace api object, basename projected -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-projected-gcgz5 +Jun 18 11:36:16.397: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 +STEP: Building a namespace api object, basename gc +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-gc-v5rcx STEP: Waiting for a default service account to be provisioned in namespace -[It] should be consumable in multiple volumes in a pod [NodeConformance] [Conformance] +[It] should delete pods created by rc when not orphaning [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -STEP: Creating secret with name projected-secret-test-00e6ecc5-75af-11e9-8f67-2632f168be36 -STEP: Creating a pod to test consume secrets -May 13 18:43:30.666: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-00e8e9ab-75af-11e9-8f67-2632f168be36" in namespace "e2e-tests-projected-gcgz5" to be "success or failure" -May 13 18:43:30.675: INFO: Pod "pod-projected-secrets-00e8e9ab-75af-11e9-8f67-2632f168be36": Phase="Pending", Reason="", readiness=false. Elapsed: 8.682008ms -May 13 18:43:32.684: INFO: Pod "pod-projected-secrets-00e8e9ab-75af-11e9-8f67-2632f168be36": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.017266587s -STEP: Saw pod success -May 13 18:43:32.684: INFO: Pod "pod-projected-secrets-00e8e9ab-75af-11e9-8f67-2632f168be36" satisfied condition "success or failure" -May 13 18:43:32.691: INFO: Trying to get logs from node 10.170.219.177 pod pod-projected-secrets-00e8e9ab-75af-11e9-8f67-2632f168be36 container secret-volume-test: -STEP: delete the pod -May 13 18:43:32.733: INFO: Waiting for pod pod-projected-secrets-00e8e9ab-75af-11e9-8f67-2632f168be36 to disappear -May 13 18:43:32.740: INFO: Pod pod-projected-secrets-00e8e9ab-75af-11e9-8f67-2632f168be36 no longer exists -[AfterEach] [sig-storage] Projected secret +STEP: create the rc +STEP: delete the rc +STEP: wait for all pods to be garbage collected +STEP: Gathering metrics +W0618 11:36:26.974053 17 metrics_grabber.go:81] Master node is not registered. Grabbing metrics from Scheduler, ControllerManager and ClusterAutoscaler is disabled. +Jun 18 11:36:26.974: INFO: For apiserver_request_count: +For apiserver_request_latencies_summary: +For etcd_helper_cache_entry_count: +For etcd_helper_cache_hit_count: +For etcd_helper_cache_miss_count: +For etcd_request_cache_add_latencies_summary: +For etcd_request_cache_get_latencies_summary: +For etcd_request_latencies_summary: +For garbage_collector_attempt_to_delete_queue_latency: +For garbage_collector_attempt_to_delete_work_duration: +For garbage_collector_attempt_to_orphan_queue_latency: +For garbage_collector_attempt_to_orphan_work_duration: +For garbage_collector_dirty_processing_latency_microseconds: +For garbage_collector_event_processing_latency_microseconds: +For garbage_collector_graph_changes_queue_latency: +For garbage_collector_graph_changes_work_duration: +For garbage_collector_orphan_processing_latency_microseconds: +For namespace_queue_latency: +For namespace_queue_latency_sum: +For namespace_queue_latency_count: +For namespace_retries: +For namespace_work_duration: +For namespace_work_duration_sum: +For namespace_work_duration_count: +For function_duration_seconds: +For errors_total: +For evicted_pods_total: + +[AfterEach] [sig-api-machinery] Garbage collector /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 18:43:32.741: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-projected-gcgz5" for this suite. -May 13 18:43:38.776: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 18:43:38.893: INFO: namespace: e2e-tests-projected-gcgz5, resource: bindings, ignored listing per whitelist -May 13 18:43:39.035: INFO: namespace e2e-tests-projected-gcgz5 deletion completed in 6.287030363s +Jun 18 11:36:26.974: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "e2e-tests-gc-v5rcx" for this suite. +Jun 18 11:36:35.036: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 11:36:35.554: INFO: namespace: e2e-tests-gc-v5rcx, resource: bindings, ignored listing per whitelist +Jun 18 11:36:35.554: INFO: namespace e2e-tests-gc-v5rcx deletion completed in 8.565757905s -• [SLOW TEST:8.669 seconds] -[sig-storage] Projected secret -/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_secret.go:34 - should be consumable in multiple volumes in a pod [NodeConformance] [Conformance] +• [SLOW TEST:19.157 seconds] +[sig-api-machinery] Garbage collector +/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:22 + should delete pods created by rc when not orphaning [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +SSSSSSS ------------------------------ -[sig-cli] Kubectl client [k8s.io] Guestbook application - should create and stop a working application [Conformance] +[sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] + Burst scaling should run to completion even with unhealthy pods [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -[BeforeEach] [sig-cli] Kubectl client +[BeforeEach] [sig-apps] StatefulSet /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client -May 13 18:43:39.036: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -STEP: Building a namespace api object, basename kubectl -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-kubectl-mxmlx +Jun 18 11:36:35.554: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 +STEP: Building a namespace api object, basename statefulset +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-statefulset-mg5bt STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [sig-cli] Kubectl client - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:243 -[It] should create and stop a working application [Conformance] +[BeforeEach] [sig-apps] StatefulSet + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:59 +[BeforeEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:74 +STEP: Creating service test in namespace e2e-tests-statefulset-mg5bt +[It] Burst scaling should run to completion even with unhealthy pods [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -STEP: creating all guestbook components -May 13 18:43:39.296: INFO: apiVersion: v1 -kind: Service -metadata: - name: redis-slave - labels: - app: redis - role: slave - tier: backend -spec: - ports: - - port: 6379 - selector: - app: redis - role: slave - tier: backend +STEP: Creating stateful set ss in namespace e2e-tests-statefulset-mg5bt +STEP: Waiting until all stateful set ss replicas will be running in namespace e2e-tests-statefulset-mg5bt +Jun 18 11:36:36.044: INFO: Found 0 stateful pods, waiting for 1 +Jun 18 11:36:46.090: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true +STEP: Confirming that stateful set scale up will not halt with unhealthy stateful pod +Jun 18 11:36:46.105: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 exec --namespace=e2e-tests-statefulset-mg5bt ss-0 -- /bin/sh -c mv -v /usr/share/nginx/html/index.html /tmp/ || true' +Jun 18 11:36:46.541: INFO: stderr: "" +Jun 18 11:36:46.541: INFO: stdout: "'/usr/share/nginx/html/index.html' -> '/tmp/index.html'\n" +Jun 18 11:36:46.541: INFO: stdout of mv -v /usr/share/nginx/html/index.html /tmp/ || true on ss-0: '/usr/share/nginx/html/index.html' -> '/tmp/index.html' + +Jun 18 11:36:46.557: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=true +Jun 18 11:36:56.596: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=false +Jun 18 11:36:56.597: INFO: Waiting for statefulset status.replicas updated to 0 +Jun 18 11:36:56.658: INFO: POD NODE PHASE GRACE CONDITIONS +Jun 18 11:36:56.658: INFO: ss-0 10.72.74.149 Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:36:36 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:36:47 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:36:47 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:36:36 +0000 UTC }] +Jun 18 11:36:56.658: INFO: ss-1 Pending [] +Jun 18 11:36:56.658: INFO: +Jun 18 11:36:56.658: INFO: StatefulSet ss has not reached scale 3, at 2 +Jun 18 11:36:57.674: INFO: Verifying statefulset ss doesn't scale past 3 for another 8.98416794s +Jun 18 11:36:58.690: INFO: Verifying statefulset ss doesn't scale past 3 for another 7.968292944s +Jun 18 11:36:59.705: INFO: Verifying statefulset ss doesn't scale past 3 for another 6.952603706s +Jun 18 11:37:00.721: INFO: Verifying statefulset ss doesn't scale past 3 for another 5.936666834s +Jun 18 11:37:01.738: INFO: Verifying statefulset ss doesn't scale past 3 for another 4.921247951s +Jun 18 11:37:02.755: INFO: Verifying statefulset ss doesn't scale past 3 for another 3.903874215s +Jun 18 11:37:03.774: INFO: Verifying statefulset ss doesn't scale past 3 for another 2.886692689s +Jun 18 11:37:04.790: INFO: Verifying statefulset ss doesn't scale past 3 for another 1.868200041s +Jun 18 11:37:05.807: INFO: Verifying statefulset ss doesn't scale past 3 for another 851.961558ms +STEP: Scaling up stateful set ss to 3 replicas and waiting until all of them will be running in namespace e2e-tests-statefulset-mg5bt +Jun 18 11:37:06.843: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 exec --namespace=e2e-tests-statefulset-mg5bt ss-0 -- /bin/sh -c mv -v /tmp/index.html /usr/share/nginx/html/ || true' +Jun 18 11:37:07.511: INFO: stderr: "" +Jun 18 11:37:07.511: INFO: stdout: "'/tmp/index.html' -> '/usr/share/nginx/html/index.html'\n" +Jun 18 11:37:07.511: INFO: stdout of mv -v /tmp/index.html /usr/share/nginx/html/ || true on ss-0: '/tmp/index.html' -> '/usr/share/nginx/html/index.html' + +Jun 18 11:37:07.511: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 exec --namespace=e2e-tests-statefulset-mg5bt ss-1 -- /bin/sh -c mv -v /tmp/index.html /usr/share/nginx/html/ || true' +Jun 18 11:37:09.021: INFO: stderr: "mv: can't rename '/tmp/index.html': No such file or directory\n" +Jun 18 11:37:09.021: INFO: stdout: "'/tmp/index.html' -> '/usr/share/nginx/html/index.html'\n" +Jun 18 11:37:09.021: INFO: stdout of mv -v /tmp/index.html /usr/share/nginx/html/ || true on ss-1: '/tmp/index.html' -> '/usr/share/nginx/html/index.html' + +Jun 18 11:37:09.021: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 exec --namespace=e2e-tests-statefulset-mg5bt ss-2 -- /bin/sh -c mv -v /tmp/index.html /usr/share/nginx/html/ || true' +Jun 18 11:37:09.494: INFO: stderr: "mv: can't rename '/tmp/index.html': No such file or directory\n" +Jun 18 11:37:09.494: INFO: stdout: "'/tmp/index.html' -> '/usr/share/nginx/html/index.html'\n" +Jun 18 11:37:09.494: INFO: stdout of mv -v /tmp/index.html /usr/share/nginx/html/ || true on ss-2: '/tmp/index.html' -> '/usr/share/nginx/html/index.html' + +Jun 18 11:37:09.510: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true +Jun 18 11:37:09.510: INFO: Waiting for pod ss-1 to enter Running - Ready=true, currently Running - Ready=true +Jun 18 11:37:09.510: INFO: Waiting for pod ss-2 to enter Running - Ready=true, currently Running - Ready=true +STEP: Scale down will not halt with unhealthy stateful pod +Jun 18 11:37:09.525: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 exec --namespace=e2e-tests-statefulset-mg5bt ss-0 -- /bin/sh -c mv -v /usr/share/nginx/html/index.html /tmp/ || true' +Jun 18 11:37:09.928: INFO: stderr: "" +Jun 18 11:37:09.928: INFO: stdout: "'/usr/share/nginx/html/index.html' -> '/tmp/index.html'\n" +Jun 18 11:37:09.928: INFO: stdout of mv -v /usr/share/nginx/html/index.html /tmp/ || true on ss-0: '/usr/share/nginx/html/index.html' -> '/tmp/index.html' + +Jun 18 11:37:09.929: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 exec --namespace=e2e-tests-statefulset-mg5bt ss-1 -- /bin/sh -c mv -v /usr/share/nginx/html/index.html /tmp/ || true' +Jun 18 11:37:10.310: INFO: stderr: "" +Jun 18 11:37:10.310: INFO: stdout: "'/usr/share/nginx/html/index.html' -> '/tmp/index.html'\n" +Jun 18 11:37:10.310: INFO: stdout of mv -v /usr/share/nginx/html/index.html /tmp/ || true on ss-1: '/usr/share/nginx/html/index.html' -> '/tmp/index.html' + +Jun 18 11:37:10.310: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 exec --namespace=e2e-tests-statefulset-mg5bt ss-2 -- /bin/sh -c mv -v /usr/share/nginx/html/index.html /tmp/ || true' +Jun 18 11:37:10.719: INFO: stderr: "" +Jun 18 11:37:10.719: INFO: stdout: "'/usr/share/nginx/html/index.html' -> '/tmp/index.html'\n" +Jun 18 11:37:10.719: INFO: stdout of mv -v /usr/share/nginx/html/index.html /tmp/ || true on ss-2: '/usr/share/nginx/html/index.html' -> '/tmp/index.html' + +Jun 18 11:37:10.719: INFO: Waiting for statefulset status.replicas updated to 0 +Jun 18 11:37:10.734: INFO: Waiting for stateful set status.readyReplicas to become 0, currently 1 +Jun 18 11:37:20.784: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=false +Jun 18 11:37:20.784: INFO: Waiting for pod ss-1 to enter Running - Ready=false, currently Running - Ready=false +Jun 18 11:37:20.784: INFO: Waiting for pod ss-2 to enter Running - Ready=false, currently Running - Ready=false +Jun 18 11:37:20.831: INFO: POD NODE PHASE GRACE CONDITIONS +Jun 18 11:37:20.831: INFO: ss-0 10.72.74.149 Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:36:36 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:37:10 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:37:10 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:36:36 +0000 UTC }] +Jun 18 11:37:20.831: INFO: ss-1 10.72.74.143 Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:36:56 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:37:10 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:37:10 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:36:56 +0000 UTC }] +Jun 18 11:37:20.831: INFO: ss-2 10.72.74.144 Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:36:56 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:37:11 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:37:11 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:36:56 +0000 UTC }] +Jun 18 11:37:20.831: INFO: +Jun 18 11:37:20.831: INFO: StatefulSet ss has not reached scale 0, at 3 +Jun 18 11:37:21.885: INFO: POD NODE PHASE GRACE CONDITIONS +Jun 18 11:37:21.885: INFO: ss-0 10.72.74.149 Running 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:36:36 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:37:10 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:37:10 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:36:36 +0000 UTC }] +Jun 18 11:37:21.885: INFO: ss-1 10.72.74.143 Running 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:36:56 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:37:10 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:37:10 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:36:56 +0000 UTC }] +Jun 18 11:37:21.885: INFO: ss-2 10.72.74.144 Running 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:36:56 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:37:11 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:37:11 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:36:56 +0000 UTC }] +Jun 18 11:37:21.885: INFO: +Jun 18 11:37:21.885: INFO: StatefulSet ss has not reached scale 0, at 3 +Jun 18 11:37:22.901: INFO: POD NODE PHASE GRACE CONDITIONS +Jun 18 11:37:22.901: INFO: ss-0 10.72.74.149 Running 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:36:36 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:37:10 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:37:10 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:36:36 +0000 UTC }] +Jun 18 11:37:22.901: INFO: ss-1 10.72.74.143 Pending 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:36:56 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:37:10 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:37:10 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:36:56 +0000 UTC }] +Jun 18 11:37:22.901: INFO: ss-2 10.72.74.144 Running 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:36:56 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:37:11 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:37:11 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:36:56 +0000 UTC }] +Jun 18 11:37:22.901: INFO: +Jun 18 11:37:22.901: INFO: StatefulSet ss has not reached scale 0, at 3 +Jun 18 11:37:23.916: INFO: POD NODE PHASE GRACE CONDITIONS +Jun 18 11:37:23.916: INFO: ss-0 10.72.74.149 Running 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:36:36 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:37:10 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:37:10 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:36:36 +0000 UTC }] +Jun 18 11:37:23.916: INFO: ss-1 10.72.74.143 Pending 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:36:56 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:37:10 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:37:10 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:36:56 +0000 UTC }] +Jun 18 11:37:23.916: INFO: ss-2 10.72.74.144 Running 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:36:56 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:37:11 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:37:11 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:36:56 +0000 UTC }] +Jun 18 11:37:23.916: INFO: +Jun 18 11:37:23.916: INFO: StatefulSet ss has not reached scale 0, at 3 +Jun 18 11:37:24.933: INFO: POD NODE PHASE GRACE CONDITIONS +Jun 18 11:37:24.933: INFO: ss-0 10.72.74.149 Running 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:36:36 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:37:10 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:37:10 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:36:36 +0000 UTC }] +Jun 18 11:37:24.933: INFO: ss-1 10.72.74.143 Pending 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:36:56 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:37:10 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:37:10 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:36:56 +0000 UTC }] +Jun 18 11:37:24.933: INFO: ss-2 10.72.74.144 Running 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:36:56 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:37:11 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:37:11 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:36:56 +0000 UTC }] +Jun 18 11:37:24.933: INFO: +Jun 18 11:37:24.933: INFO: StatefulSet ss has not reached scale 0, at 3 +Jun 18 11:37:25.949: INFO: POD NODE PHASE GRACE CONDITIONS +Jun 18 11:37:25.949: INFO: ss-0 10.72.74.149 Running 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:36:36 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:37:10 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:37:10 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:36:36 +0000 UTC }] +Jun 18 11:37:25.949: INFO: ss-1 10.72.74.143 Pending 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:36:56 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:37:10 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:37:10 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:36:56 +0000 UTC }] +Jun 18 11:37:25.949: INFO: ss-2 10.72.74.144 Running 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:36:56 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:37:11 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:37:11 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:36:56 +0000 UTC }] +Jun 18 11:37:25.949: INFO: +Jun 18 11:37:25.949: INFO: StatefulSet ss has not reached scale 0, at 3 +Jun 18 11:37:26.964: INFO: POD NODE PHASE GRACE CONDITIONS +Jun 18 11:37:26.964: INFO: ss-0 10.72.74.149 Running 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:36:36 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:37:10 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:37:10 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:36:36 +0000 UTC }] +Jun 18 11:37:26.964: INFO: ss-2 10.72.74.144 Running 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:36:56 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:37:11 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:37:11 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:36:56 +0000 UTC }] +Jun 18 11:37:26.964: INFO: +Jun 18 11:37:26.964: INFO: StatefulSet ss has not reached scale 0, at 2 +Jun 18 11:37:27.979: INFO: POD NODE PHASE GRACE CONDITIONS +Jun 18 11:37:27.979: INFO: ss-0 10.72.74.149 Running 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:36:36 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:37:10 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:37:10 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:36:36 +0000 UTC }] +Jun 18 11:37:27.979: INFO: ss-2 10.72.74.144 Running 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:36:56 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:37:11 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:37:11 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:36:56 +0000 UTC }] +Jun 18 11:37:27.979: INFO: +Jun 18 11:37:27.979: INFO: StatefulSet ss has not reached scale 0, at 2 +Jun 18 11:37:29.458: INFO: POD NODE PHASE GRACE CONDITIONS +Jun 18 11:37:29.458: INFO: ss-2 10.72.74.144 Running 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:36:56 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:37:11 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:37:11 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:36:56 +0000 UTC }] +Jun 18 11:37:29.458: INFO: +Jun 18 11:37:29.458: INFO: StatefulSet ss has not reached scale 0, at 1 +Jun 18 11:37:30.499: INFO: POD NODE PHASE GRACE CONDITIONS +Jun 18 11:37:30.499: INFO: ss-2 10.72.74.144 Running 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:36:56 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:37:11 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:37:11 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:36:56 +0000 UTC }] +Jun 18 11:37:30.499: INFO: +Jun 18 11:37:30.499: INFO: StatefulSet ss has not reached scale 0, at 1 +STEP: Scaling down stateful set ss to 0 replicas and waiting until none of pods will run in namespacee2e-tests-statefulset-mg5bt +Jun 18 11:37:31.533: INFO: Scaling statefulset ss to 0 +Jun 18 11:37:31.575: INFO: Waiting for statefulset status.replicas updated to 0 +[AfterEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:85 +Jun 18 11:37:31.589: INFO: Deleting all statefulset in ns e2e-tests-statefulset-mg5bt +Jun 18 11:37:31.603: INFO: Scaling statefulset ss to 0 +Jun 18 11:37:31.647: INFO: Waiting for statefulset status.replicas updated to 0 +Jun 18 11:37:31.661: INFO: Deleting statefulset ss +[AfterEach] [sig-apps] StatefulSet + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 +Jun 18 11:37:31.721: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "e2e-tests-statefulset-mg5bt" for this suite. +Jun 18 11:37:39.794: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 11:37:41.238: INFO: namespace: e2e-tests-statefulset-mg5bt, resource: bindings, ignored listing per whitelist +Jun 18 11:37:41.383: INFO: namespace e2e-tests-statefulset-mg5bt deletion completed in 9.643137461s -May 13 18:43:39.296: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 create -f - --namespace=e2e-tests-kubectl-mxmlx' -May 13 18:43:39.512: INFO: stderr: "" -May 13 18:43:39.512: INFO: stdout: "service/redis-slave created\n" -May 13 18:43:39.512: INFO: apiVersion: v1 -kind: Service -metadata: - name: redis-master - labels: - app: redis - role: master - tier: backend -spec: - ports: - - port: 6379 - targetPort: 6379 - selector: - app: redis - role: master - tier: backend - -May 13 18:43:39.512: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 create -f - --namespace=e2e-tests-kubectl-mxmlx' -May 13 18:43:39.830: INFO: stderr: "" -May 13 18:43:39.830: INFO: stdout: "service/redis-master created\n" -May 13 18:43:39.830: INFO: apiVersion: v1 -kind: Service -metadata: - name: frontend - labels: - app: guestbook - tier: frontend -spec: - # if your cluster supports it, uncomment the following to automatically create - # an external load-balanced IP for the frontend service. - # type: LoadBalancer - ports: - - port: 80 - selector: - app: guestbook - tier: frontend - -May 13 18:43:39.830: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 create -f - --namespace=e2e-tests-kubectl-mxmlx' -May 13 18:43:40.018: INFO: stderr: "" -May 13 18:43:40.018: INFO: stdout: "service/frontend created\n" -May 13 18:43:40.018: INFO: apiVersion: extensions/v1beta1 -kind: Deployment -metadata: - name: frontend -spec: - replicas: 3 - template: - metadata: - labels: - app: guestbook - tier: frontend - spec: - containers: - - name: php-redis - image: gcr.io/google-samples/gb-frontend:v6 - resources: - requests: - cpu: 100m - memory: 100Mi - env: - - name: GET_HOSTS_FROM - value: dns - # If your cluster config does not include a dns service, then to - # instead access environment variables to find service host - # info, comment out the 'value: dns' line above, and uncomment the - # line below: - # value: env - ports: - - containerPort: 80 - -May 13 18:43:40.019: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 create -f - --namespace=e2e-tests-kubectl-mxmlx' -May 13 18:43:40.251: INFO: stderr: "" -May 13 18:43:40.251: INFO: stdout: "deployment.extensions/frontend created\n" -May 13 18:43:40.251: INFO: apiVersion: extensions/v1beta1 -kind: Deployment -metadata: - name: redis-master -spec: - replicas: 1 - template: - metadata: - labels: - app: redis - role: master - tier: backend - spec: - containers: - - name: master - image: gcr.io/kubernetes-e2e-test-images/redis:1.0 - resources: - requests: - cpu: 100m - memory: 100Mi - ports: - - containerPort: 6379 - -May 13 18:43:40.251: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 create -f - --namespace=e2e-tests-kubectl-mxmlx' -May 13 18:43:40.531: INFO: stderr: "" -May 13 18:43:40.531: INFO: stdout: "deployment.extensions/redis-master created\n" -May 13 18:43:40.532: INFO: apiVersion: extensions/v1beta1 -kind: Deployment -metadata: - name: redis-slave -spec: - replicas: 2 - template: - metadata: - labels: - app: redis - role: slave - tier: backend - spec: - containers: - - name: slave - image: gcr.io/google-samples/gb-redisslave:v3 - resources: - requests: - cpu: 100m - memory: 100Mi - env: - - name: GET_HOSTS_FROM - value: dns - # If your cluster config does not include a dns service, then to - # instead access an environment variable to find the master - # service's host, comment out the 'value: dns' line above, and - # uncomment the line below: - # value: env - ports: - - containerPort: 6379 - -May 13 18:43:40.532: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 create -f - --namespace=e2e-tests-kubectl-mxmlx' -May 13 18:43:40.786: INFO: stderr: "" -May 13 18:43:40.786: INFO: stdout: "deployment.extensions/redis-slave created\n" -STEP: validating guestbook app -May 13 18:43:40.786: INFO: Waiting for all frontend pods to be Running. -May 13 18:43:45.837: INFO: Waiting for frontend to serve content. -May 13 18:43:45.865: INFO: Trying to add a new entry to the guestbook. -May 13 18:43:45.891: INFO: Verifying that added entry can be retrieved. -May 13 18:43:45.913: INFO: Failed to get response from guestbook. err: , response: {"data": ""} -May 13 18:43:50.952: INFO: Failed to get response from guestbook. err: , response: {"data": ""} -May 13 18:43:55.977: INFO: Failed to get response from guestbook. err: , response: {"data": ""} -May 13 18:44:01.012: INFO: Failed to get response from guestbook. err: , response: {"data": ""} -May 13 18:44:06.039: INFO: Failed to get response from guestbook. err: , response: {"data": ""} -May 13 18:44:11.083: INFO: Failed to get response from guestbook. err: , response: {"data": ""} -May 13 18:44:16.112: INFO: Failed to get response from guestbook. err: , response: {"data": ""} -May 13 18:44:21.148: INFO: Failed to get response from guestbook. err: , response: {"data": ""} -May 13 18:44:26.170: INFO: Failed to get response from guestbook. err: , response: {"data": ""} -May 13 18:44:31.206: INFO: Failed to get response from guestbook. err: , response: {"data": ""} -May 13 18:44:36.231: INFO: Failed to get response from guestbook. err: , response: {"data": ""} -May 13 18:44:41.270: INFO: Failed to get response from guestbook. err: , response: {"data": ""} -STEP: using delete to clean up resources -May 13 18:44:46.295: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 delete --grace-period=0 --force -f - --namespace=e2e-tests-kubectl-mxmlx' -May 13 18:44:46.474: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" -May 13 18:44:46.474: INFO: stdout: "service \"redis-slave\" force deleted\n" -STEP: using delete to clean up resources -May 13 18:44:46.474: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 delete --grace-period=0 --force -f - --namespace=e2e-tests-kubectl-mxmlx' -May 13 18:44:46.625: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" -May 13 18:44:46.625: INFO: stdout: "service \"redis-master\" force deleted\n" -STEP: using delete to clean up resources -May 13 18:44:46.625: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 delete --grace-period=0 --force -f - --namespace=e2e-tests-kubectl-mxmlx' -May 13 18:44:46.752: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" -May 13 18:44:46.752: INFO: stdout: "service \"frontend\" force deleted\n" -STEP: using delete to clean up resources -May 13 18:44:46.752: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 delete --grace-period=0 --force -f - --namespace=e2e-tests-kubectl-mxmlx' -May 13 18:44:46.854: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" -May 13 18:44:46.854: INFO: stdout: "deployment.extensions \"frontend\" force deleted\n" -STEP: using delete to clean up resources -May 13 18:44:46.854: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 delete --grace-period=0 --force -f - --namespace=e2e-tests-kubectl-mxmlx' -May 13 18:44:46.972: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" -May 13 18:44:46.972: INFO: stdout: "deployment.extensions \"redis-master\" force deleted\n" -STEP: using delete to clean up resources -May 13 18:44:46.972: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 delete --grace-period=0 --force -f - --namespace=e2e-tests-kubectl-mxmlx' -May 13 18:44:47.094: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" -May 13 18:44:47.094: INFO: stdout: "deployment.extensions \"redis-slave\" force deleted\n" -[AfterEach] [sig-cli] Kubectl client - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 18:44:47.094: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-kubectl-mxmlx" for this suite. -May 13 18:45:27.132: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 18:45:27.335: INFO: namespace: e2e-tests-kubectl-mxmlx, resource: bindings, ignored listing per whitelist -May 13 18:45:27.533: INFO: namespace e2e-tests-kubectl-mxmlx deletion completed in 40.430973147s - -• [SLOW TEST:108.498 seconds] -[sig-cli] Kubectl client -/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:22 - [k8s.io] Guestbook application +• [SLOW TEST:65.829 seconds] +[sig-apps] StatefulSet +/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22 + [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 - should create and stop a working application [Conformance] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------- -S ------------------------------- -[sig-network] Networking Granular Checks: Pods - should function for intra-pod communication: http [NodeConformance] [Conformance] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -[BeforeEach] [sig-network] Networking - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 -STEP: Creating a kubernetes client -May 13 18:45:27.534: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -STEP: Building a namespace api object, basename pod-network-test -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-pod-network-test-nlztd -STEP: Waiting for a default service account to be provisioned in namespace -[It] should function for intra-pod communication: http [NodeConformance] [Conformance] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -STEP: Performing setup for networking test in namespace e2e-tests-pod-network-test-nlztd -STEP: creating a selector -STEP: Creating the service pods in kubernetes -May 13 18:45:27.816: INFO: Waiting up to 10m0s for all (but 0) nodes to be schedulable -STEP: Creating test pods -May 13 18:45:48.018: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s 'http://172.30.227.86:8080/dial?request=hostName&protocol=http&host=172.30.19.77&port=8080&tries=1'] Namespace:e2e-tests-pod-network-test-nlztd PodName:host-test-container-pod ContainerName:hostexec Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} -May 13 18:45:48.019: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -May 13 18:45:48.188: INFO: Waiting for endpoints: map[] -May 13 18:45:48.196: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s 'http://172.30.227.86:8080/dial?request=hostName&protocol=http&host=172.30.227.85&port=8080&tries=1'] Namespace:e2e-tests-pod-network-test-nlztd PodName:host-test-container-pod ContainerName:hostexec Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} -May 13 18:45:48.196: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -May 13 18:45:48.352: INFO: Waiting for endpoints: map[] -May 13 18:45:48.359: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s 'http://172.30.227.86:8080/dial?request=hostName&protocol=http&host=172.30.63.47&port=8080&tries=1'] Namespace:e2e-tests-pod-network-test-nlztd PodName:host-test-container-pod ContainerName:hostexec Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} -May 13 18:45:48.359: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -May 13 18:45:48.563: INFO: Waiting for endpoints: map[] -[AfterEach] [sig-network] Networking - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 18:45:48.563: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-pod-network-test-nlztd" for this suite. -May 13 18:46:12.601: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 18:46:12.787: INFO: namespace: e2e-tests-pod-network-test-nlztd, resource: bindings, ignored listing per whitelist -May 13 18:46:12.835: INFO: namespace e2e-tests-pod-network-test-nlztd deletion completed in 24.263113791s - -• [SLOW TEST:45.301 seconds] -[sig-network] Networking -/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:25 - Granular Checks: Pods - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:28 - should function for intra-pod communication: http [NodeConformance] [Conformance] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------- -[sig-storage] Projected downwardAPI - should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -[BeforeEach] [sig-storage] Projected downwardAPI - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 -STEP: Creating a kubernetes client -May 13 18:46:12.835: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -STEP: Building a namespace api object, basename projected -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-projected-nf25l -STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [sig-storage] Projected downwardAPI - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:39 -[It] should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -STEP: Creating a pod to test downward API volume plugin -May 13 18:46:13.136: INFO: Waiting up to 5m0s for pod "downwardapi-volume-61bfd266-75af-11e9-8f67-2632f168be36" in namespace "e2e-tests-projected-nf25l" to be "success or failure" -May 13 18:46:13.145: INFO: Pod "downwardapi-volume-61bfd266-75af-11e9-8f67-2632f168be36": Phase="Pending", Reason="", readiness=false. Elapsed: 8.508564ms -May 13 18:46:15.153: INFO: Pod "downwardapi-volume-61bfd266-75af-11e9-8f67-2632f168be36": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.016655135s -STEP: Saw pod success -May 13 18:46:15.153: INFO: Pod "downwardapi-volume-61bfd266-75af-11e9-8f67-2632f168be36" satisfied condition "success or failure" -May 13 18:46:15.161: INFO: Trying to get logs from node 10.170.219.153 pod downwardapi-volume-61bfd266-75af-11e9-8f67-2632f168be36 container client-container: -STEP: delete the pod -May 13 18:46:15.199: INFO: Waiting for pod downwardapi-volume-61bfd266-75af-11e9-8f67-2632f168be36 to disappear -May 13 18:46:15.205: INFO: Pod downwardapi-volume-61bfd266-75af-11e9-8f67-2632f168be36 no longer exists -[AfterEach] [sig-storage] Projected downwardAPI - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 18:46:15.205: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-projected-nf25l" for this suite. -May 13 18:46:21.259: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 18:46:21.392: INFO: namespace: e2e-tests-projected-nf25l, resource: bindings, ignored listing per whitelist -May 13 18:46:21.577: INFO: namespace e2e-tests-projected-nf25l deletion completed in 6.364091739s - -• [SLOW TEST:8.741 seconds] -[sig-storage] Projected downwardAPI -/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:33 - should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------- -SS ------------------------------- -[k8s.io] Kubelet when scheduling a busybox command that always fails in a pod - should have an terminated reason [NodeConformance] [Conformance] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -[BeforeEach] [k8s.io] Kubelet - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 -STEP: Creating a kubernetes client -May 13 18:46:21.577: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -STEP: Building a namespace api object, basename kubelet-test -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-kubelet-test-tj94q -STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [k8s.io] Kubelet - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:37 -[BeforeEach] when scheduling a busybox command that always fails in a pod - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:81 -[It] should have an terminated reason [NodeConformance] [Conformance] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -[AfterEach] [k8s.io] Kubelet - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 18:46:26.057: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-kubelet-test-tj94q" for this suite. -May 13 18:46:32.106: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 18:46:32.468: INFO: namespace: e2e-tests-kubelet-test-tj94q, resource: bindings, ignored listing per whitelist -May 13 18:46:32.517: INFO: namespace e2e-tests-kubelet-test-tj94q deletion completed in 6.451985052s - -• [SLOW TEST:10.940 seconds] -[k8s.io] Kubelet -/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 - when scheduling a busybox command that always fails in a pod - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:78 - should have an terminated reason [NodeConformance] [Conformance] + Burst scaling should run to completion even with unhealthy pods [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ -SSSSSSSSS ------------------------------- -[sig-api-machinery] Garbage collector - should delete pods created by rc when not orphaning [Conformance] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -[BeforeEach] [sig-api-machinery] Garbage collector - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 -STEP: Creating a kubernetes client -May 13 18:46:32.518: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -STEP: Building a namespace api object, basename gc -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-gc-jgmpr -STEP: Waiting for a default service account to be provisioned in namespace -[It] should delete pods created by rc when not orphaning [Conformance] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -STEP: create the rc -STEP: delete the rc -STEP: wait for all pods to be garbage collected -STEP: Gathering metrics -W0513 18:46:42.929095 16 metrics_grabber.go:81] Master node is not registered. Grabbing metrics from Scheduler, ControllerManager and ClusterAutoscaler is disabled. -May 13 18:46:42.929: INFO: For apiserver_request_count: -For apiserver_request_latencies_summary: -For etcd_helper_cache_entry_count: -For etcd_helper_cache_hit_count: -For etcd_helper_cache_miss_count: -For etcd_request_cache_add_latencies_summary: -For etcd_request_cache_get_latencies_summary: -For etcd_request_latencies_summary: -For garbage_collector_attempt_to_delete_queue_latency: -For garbage_collector_attempt_to_delete_work_duration: -For garbage_collector_attempt_to_orphan_queue_latency: -For garbage_collector_attempt_to_orphan_work_duration: -For garbage_collector_dirty_processing_latency_microseconds: -For garbage_collector_event_processing_latency_microseconds: -For garbage_collector_graph_changes_queue_latency: -For garbage_collector_graph_changes_work_duration: -For garbage_collector_orphan_processing_latency_microseconds: -For namespace_queue_latency: -For namespace_queue_latency_sum: -For namespace_queue_latency_count: -For namespace_retries: -For namespace_work_duration: -For namespace_work_duration_sum: -For namespace_work_duration_count: -For function_duration_seconds: -For errors_total: -For evicted_pods_total: - -[AfterEach] [sig-api-machinery] Garbage collector - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 18:46:42.929: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-gc-jgmpr" for this suite. -May 13 18:46:48.967: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 18:46:49.065: INFO: namespace: e2e-tests-gc-jgmpr, resource: bindings, ignored listing per whitelist -May 13 18:46:49.410: INFO: namespace e2e-tests-gc-jgmpr deletion completed in 6.472655659s - -• [SLOW TEST:16.891 seconds] -[sig-api-machinery] Garbage collector -/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:22 - should delete pods created by rc when not orphaning [Conformance] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------- -SSSSSSSSS +SSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[k8s.io] KubeletManagedEtcHosts - should test kubelet managed /etc/hosts file [NodeConformance] [Conformance] +[sig-storage] EmptyDir wrapper volumes + should not conflict [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -[BeforeEach] [k8s.io] KubeletManagedEtcHosts +[BeforeEach] [sig-storage] EmptyDir wrapper volumes /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client -May 13 18:46:49.410: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -STEP: Building a namespace api object, basename e2e-kubelet-etc-hosts -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-e2e-kubelet-etc-hosts-dzjcb +Jun 18 11:37:41.386: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 +STEP: Building a namespace api object, basename emptydir-wrapper +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-emptydir-wrapper-h9sfn STEP: Waiting for a default service account to be provisioned in namespace -[It] should test kubelet managed /etc/hosts file [NodeConformance] [Conformance] +[It] should not conflict [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -STEP: Setting up the test -STEP: Creating hostNetwork=false pod -STEP: Creating hostNetwork=true pod -STEP: Running the test -STEP: Verifying /etc/hosts of container is kubelet-managed for pod with hostNetwork=false -May 13 18:46:55.792: INFO: ExecWithOptions {Command:[cat /etc/hosts] Namespace:e2e-tests-e2e-kubelet-etc-hosts-dzjcb PodName:test-pod ContainerName:busybox-1 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} -May 13 18:46:55.792: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -May 13 18:46:55.973: INFO: Exec stderr: "" -May 13 18:46:55.973: INFO: ExecWithOptions {Command:[cat /etc/hosts-original] Namespace:e2e-tests-e2e-kubelet-etc-hosts-dzjcb PodName:test-pod ContainerName:busybox-1 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} -May 13 18:46:55.973: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -May 13 18:46:56.168: INFO: Exec stderr: "" -May 13 18:46:56.168: INFO: ExecWithOptions {Command:[cat /etc/hosts] Namespace:e2e-tests-e2e-kubelet-etc-hosts-dzjcb PodName:test-pod ContainerName:busybox-2 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} -May 13 18:46:56.168: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -May 13 18:46:56.344: INFO: Exec stderr: "" -May 13 18:46:56.344: INFO: ExecWithOptions {Command:[cat /etc/hosts-original] Namespace:e2e-tests-e2e-kubelet-etc-hosts-dzjcb PodName:test-pod ContainerName:busybox-2 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} -May 13 18:46:56.344: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -May 13 18:46:56.524: INFO: Exec stderr: "" -STEP: Verifying /etc/hosts of container is not kubelet-managed since container specifies /etc/hosts mount -May 13 18:46:56.524: INFO: ExecWithOptions {Command:[cat /etc/hosts] Namespace:e2e-tests-e2e-kubelet-etc-hosts-dzjcb PodName:test-pod ContainerName:busybox-3 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} -May 13 18:46:56.524: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -May 13 18:46:56.772: INFO: Exec stderr: "" -May 13 18:46:56.772: INFO: ExecWithOptions {Command:[cat /etc/hosts-original] Namespace:e2e-tests-e2e-kubelet-etc-hosts-dzjcb PodName:test-pod ContainerName:busybox-3 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} -May 13 18:46:56.772: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -May 13 18:46:56.932: INFO: Exec stderr: "" -STEP: Verifying /etc/hosts content of container is not kubelet-managed for pod with hostNetwork=true -May 13 18:46:56.932: INFO: ExecWithOptions {Command:[cat /etc/hosts] Namespace:e2e-tests-e2e-kubelet-etc-hosts-dzjcb PodName:test-host-network-pod ContainerName:busybox-1 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} -May 13 18:46:56.932: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -May 13 18:46:57.087: INFO: Exec stderr: "" -May 13 18:46:57.087: INFO: ExecWithOptions {Command:[cat /etc/hosts-original] Namespace:e2e-tests-e2e-kubelet-etc-hosts-dzjcb PodName:test-host-network-pod ContainerName:busybox-1 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} -May 13 18:46:57.087: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -May 13 18:46:57.285: INFO: Exec stderr: "" -May 13 18:46:57.285: INFO: ExecWithOptions {Command:[cat /etc/hosts] Namespace:e2e-tests-e2e-kubelet-etc-hosts-dzjcb PodName:test-host-network-pod ContainerName:busybox-2 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} -May 13 18:46:57.285: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -May 13 18:46:57.469: INFO: Exec stderr: "" -May 13 18:46:57.469: INFO: ExecWithOptions {Command:[cat /etc/hosts-original] Namespace:e2e-tests-e2e-kubelet-etc-hosts-dzjcb PodName:test-host-network-pod ContainerName:busybox-2 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} -May 13 18:46:57.469: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -May 13 18:46:57.720: INFO: Exec stderr: "" -[AfterEach] [k8s.io] KubeletManagedEtcHosts +STEP: Cleaning up the secret +STEP: Cleaning up the configmap +STEP: Cleaning up the pod +[AfterEach] [sig-storage] EmptyDir wrapper volumes /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 18:46:57.720: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-e2e-kubelet-etc-hosts-dzjcb" for this suite. -May 13 18:47:49.756: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 18:47:50.060: INFO: namespace: e2e-tests-e2e-kubelet-etc-hosts-dzjcb, resource: bindings, ignored listing per whitelist -May 13 18:47:50.075: INFO: namespace e2e-tests-e2e-kubelet-etc-hosts-dzjcb deletion completed in 52.347386727s +Jun 18 11:37:48.142: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "e2e-tests-emptydir-wrapper-h9sfn" for this suite. +Jun 18 11:37:56.216: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 11:37:56.298: INFO: namespace: e2e-tests-emptydir-wrapper-h9sfn, resource: bindings, ignored listing per whitelist +Jun 18 11:37:56.782: INFO: namespace e2e-tests-emptydir-wrapper-h9sfn deletion completed in 8.617274941s -• [SLOW TEST:60.665 seconds] -[k8s.io] KubeletManagedEtcHosts -/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 - should test kubelet managed /etc/hosts file [NodeConformance] [Conformance] +• [SLOW TEST:15.397 seconds] +[sig-storage] EmptyDir wrapper volumes +/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:22 + should not conflict [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ -S +SSSSSSSSSSSSSSS ------------------------------ -[sig-api-machinery] Secrets - should be consumable from pods in env vars [NodeConformance] [Conformance] +[sig-storage] Subpath Atomic writer volumes + should support subpaths with secret pod [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -[BeforeEach] [sig-api-machinery] Secrets +[BeforeEach] [sig-storage] Subpath /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client -May 13 18:47:50.075: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -STEP: Building a namespace api object, basename secrets -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-secrets-cmr2d +Jun 18 11:37:56.784: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 +STEP: Building a namespace api object, basename subpath +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-subpath-w9hsb STEP: Waiting for a default service account to be provisioned in namespace -[It] should be consumable from pods in env vars [NodeConformance] [Conformance] +[BeforeEach] Atomic writer volumes + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:38 +STEP: Setting up data +[It] should support subpaths with secret pod [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -STEP: Creating secret with name secret-test-9bce5bd7-75af-11e9-8f67-2632f168be36 -STEP: Creating a pod to test consume secrets -May 13 18:47:50.547: INFO: Waiting up to 5m0s for pod "pod-secrets-9bcfc224-75af-11e9-8f67-2632f168be36" in namespace "e2e-tests-secrets-cmr2d" to be "success or failure" -May 13 18:47:50.555: INFO: Pod "pod-secrets-9bcfc224-75af-11e9-8f67-2632f168be36": Phase="Pending", Reason="", readiness=false. Elapsed: 7.611558ms -May 13 18:47:52.565: INFO: Pod "pod-secrets-9bcfc224-75af-11e9-8f67-2632f168be36": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.017507039s +STEP: Creating pod pod-subpath-test-secret-wxjh +STEP: Creating a pod to test atomic-volume-subpath +Jun 18 11:37:57.290: INFO: Waiting up to 5m0s for pod "pod-subpath-test-secret-wxjh" in namespace "e2e-tests-subpath-w9hsb" to be "success or failure" +Jun 18 11:37:57.307: INFO: Pod "pod-subpath-test-secret-wxjh": Phase="Pending", Reason="", readiness=false. Elapsed: 16.891579ms +Jun 18 11:37:59.322: INFO: Pod "pod-subpath-test-secret-wxjh": Phase="Pending", Reason="", readiness=false. Elapsed: 2.031406857s +Jun 18 11:38:01.337: INFO: Pod "pod-subpath-test-secret-wxjh": Phase="Pending", Reason="", readiness=false. Elapsed: 4.046894145s +Jun 18 11:38:03.373: INFO: Pod "pod-subpath-test-secret-wxjh": Phase="Pending", Reason="", readiness=false. Elapsed: 6.082617315s +Jun 18 11:38:05.391: INFO: Pod "pod-subpath-test-secret-wxjh": Phase="Running", Reason="", readiness=false. Elapsed: 8.100470668s +Jun 18 11:38:07.406: INFO: Pod "pod-subpath-test-secret-wxjh": Phase="Running", Reason="", readiness=false. Elapsed: 10.115247887s +Jun 18 11:38:09.421: INFO: Pod "pod-subpath-test-secret-wxjh": Phase="Running", Reason="", readiness=false. Elapsed: 12.130341346s +Jun 18 11:38:11.435: INFO: Pod "pod-subpath-test-secret-wxjh": Phase="Running", Reason="", readiness=false. Elapsed: 14.144695589s +Jun 18 11:38:13.471: INFO: Pod "pod-subpath-test-secret-wxjh": Phase="Running", Reason="", readiness=false. Elapsed: 16.180015785s +Jun 18 11:38:15.485: INFO: Pod "pod-subpath-test-secret-wxjh": Phase="Running", Reason="", readiness=false. Elapsed: 18.194863623s +Jun 18 11:38:17.853: INFO: Pod "pod-subpath-test-secret-wxjh": Phase="Running", Reason="", readiness=false. Elapsed: 20.562042445s +Jun 18 11:38:19.885: INFO: Pod "pod-subpath-test-secret-wxjh": Phase="Running", Reason="", readiness=false. Elapsed: 22.593965861s +Jun 18 11:38:21.904: INFO: Pod "pod-subpath-test-secret-wxjh": Phase="Running", Reason="", readiness=false. Elapsed: 24.613684127s +Jun 18 11:38:23.985: INFO: Pod "pod-subpath-test-secret-wxjh": Phase="Running", Reason="", readiness=false. Elapsed: 26.694293432s +Jun 18 11:38:26.004: INFO: Pod "pod-subpath-test-secret-wxjh": Phase="Succeeded", Reason="", readiness=false. Elapsed: 28.712971408s STEP: Saw pod success -May 13 18:47:52.565: INFO: Pod "pod-secrets-9bcfc224-75af-11e9-8f67-2632f168be36" satisfied condition "success or failure" -May 13 18:47:52.572: INFO: Trying to get logs from node 10.170.219.177 pod pod-secrets-9bcfc224-75af-11e9-8f67-2632f168be36 container secret-env-test: +Jun 18 11:38:26.004: INFO: Pod "pod-subpath-test-secret-wxjh" satisfied condition "success or failure" +Jun 18 11:38:26.018: INFO: Trying to get logs from node 10.72.74.143 pod pod-subpath-test-secret-wxjh container test-container-subpath-secret-wxjh: STEP: delete the pod -May 13 18:47:52.610: INFO: Waiting for pod pod-secrets-9bcfc224-75af-11e9-8f67-2632f168be36 to disappear -May 13 18:47:52.618: INFO: Pod pod-secrets-9bcfc224-75af-11e9-8f67-2632f168be36 no longer exists -[AfterEach] [sig-api-machinery] Secrets +Jun 18 11:38:26.097: INFO: Waiting for pod pod-subpath-test-secret-wxjh to disappear +Jun 18 11:38:26.119: INFO: Pod pod-subpath-test-secret-wxjh no longer exists +STEP: Deleting pod pod-subpath-test-secret-wxjh +Jun 18 11:38:26.119: INFO: Deleting pod "pod-subpath-test-secret-wxjh" in namespace "e2e-tests-subpath-w9hsb" +[AfterEach] [sig-storage] Subpath /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 18:47:52.618: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-secrets-cmr2d" for this suite. -May 13 18:47:58.654: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 18:47:58.784: INFO: namespace: e2e-tests-secrets-cmr2d, resource: bindings, ignored listing per whitelist -May 13 18:47:58.939: INFO: namespace e2e-tests-secrets-cmr2d deletion completed in 6.314523453s +Jun 18 11:38:26.136: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "e2e-tests-subpath-w9hsb" for this suite. +Jun 18 11:38:32.202: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 11:38:32.570: INFO: namespace: e2e-tests-subpath-w9hsb, resource: bindings, ignored listing per whitelist +Jun 18 11:38:32.681: INFO: namespace e2e-tests-subpath-w9hsb deletion completed in 6.526698349s -• [SLOW TEST:8.864 seconds] -[sig-api-machinery] Secrets -/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/secrets.go:32 - should be consumable from pods in env vars [NodeConformance] [Conformance] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 +• [SLOW TEST:35.898 seconds] +[sig-storage] Subpath +/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:22 + Atomic writer volumes + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:34 + should support subpaths with secret pod [Conformance] + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ -SSSSSS +S ------------------------------ [sig-apps] Daemon set [Serial] - should update pod when spec was updated and update strategy is RollingUpdate [Conformance] + should run and stop simple daemon [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-apps] Daemon set [Serial] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client -May 13 18:47:58.940: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 +Jun 18 11:38:32.681: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 STEP: Building a namespace api object, basename daemonsets -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-daemonsets-nd57p +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-daemonsets-frc4s STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] Daemon set [Serial] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:102 -[It] should update pod when spec was updated and update strategy is RollingUpdate [Conformance] +[It] should run and stop simple daemon [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -May 13 18:47:59.363: INFO: Creating simple daemon set daemon-set +STEP: Creating simple DaemonSet "daemon-set" STEP: Check that daemon pods launch on every node of the cluster. -May 13 18:47:59.385: INFO: Number of nodes with available pods: 0 -May 13 18:47:59.385: INFO: Node 10.170.219.151 is running more than one daemon pod -May 13 18:48:00.413: INFO: Number of nodes with available pods: 0 -May 13 18:48:00.413: INFO: Node 10.170.219.151 is running more than one daemon pod -May 13 18:48:01.402: INFO: Number of nodes with available pods: 2 -May 13 18:48:01.402: INFO: Node 10.170.219.153 is running more than one daemon pod -May 13 18:48:02.401: INFO: Number of nodes with available pods: 3 -May 13 18:48:02.401: INFO: Number of running nodes: 3, number of available pods: 3 -STEP: Update daemon pods image. -STEP: Check that daemon pods images are updated. -May 13 18:48:02.539: INFO: Wrong image for pod: daemon-set-28d9j. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:02.539: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:02.539: INFO: Wrong image for pod: daemon-set-zglt4. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:03.554: INFO: Wrong image for pod: daemon-set-28d9j. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:03.554: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:03.554: INFO: Wrong image for pod: daemon-set-zglt4. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:04.554: INFO: Wrong image for pod: daemon-set-28d9j. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:04.555: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:04.555: INFO: Wrong image for pod: daemon-set-zglt4. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:05.555: INFO: Wrong image for pod: daemon-set-28d9j. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:05.555: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:05.555: INFO: Wrong image for pod: daemon-set-zglt4. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:06.555: INFO: Wrong image for pod: daemon-set-28d9j. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:06.555: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:06.555: INFO: Wrong image for pod: daemon-set-zglt4. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:07.555: INFO: Wrong image for pod: daemon-set-28d9j. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:07.555: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:07.555: INFO: Wrong image for pod: daemon-set-zglt4. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:08.554: INFO: Wrong image for pod: daemon-set-28d9j. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:08.554: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:08.554: INFO: Wrong image for pod: daemon-set-zglt4. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:09.554: INFO: Wrong image for pod: daemon-set-28d9j. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:09.554: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:09.554: INFO: Wrong image for pod: daemon-set-zglt4. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:10.567: INFO: Wrong image for pod: daemon-set-28d9j. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:10.568: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:10.568: INFO: Wrong image for pod: daemon-set-zglt4. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:11.555: INFO: Wrong image for pod: daemon-set-28d9j. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:11.555: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:11.555: INFO: Wrong image for pod: daemon-set-zglt4. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:12.555: INFO: Wrong image for pod: daemon-set-28d9j. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:12.555: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:12.555: INFO: Wrong image for pod: daemon-set-zglt4. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:13.555: INFO: Wrong image for pod: daemon-set-28d9j. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:13.555: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:13.555: INFO: Wrong image for pod: daemon-set-zglt4. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:14.555: INFO: Wrong image for pod: daemon-set-28d9j. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:14.555: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:14.555: INFO: Wrong image for pod: daemon-set-zglt4. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:15.554: INFO: Wrong image for pod: daemon-set-28d9j. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:15.554: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:15.554: INFO: Wrong image for pod: daemon-set-zglt4. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:16.554: INFO: Wrong image for pod: daemon-set-28d9j. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:16.554: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:16.554: INFO: Wrong image for pod: daemon-set-zglt4. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:17.554: INFO: Wrong image for pod: daemon-set-28d9j. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:17.554: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:17.554: INFO: Wrong image for pod: daemon-set-zglt4. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:18.555: INFO: Wrong image for pod: daemon-set-28d9j. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:18.555: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:18.555: INFO: Wrong image for pod: daemon-set-zglt4. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:19.555: INFO: Wrong image for pod: daemon-set-28d9j. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:19.555: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:19.555: INFO: Wrong image for pod: daemon-set-zglt4. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:20.554: INFO: Wrong image for pod: daemon-set-28d9j. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:20.554: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:20.554: INFO: Wrong image for pod: daemon-set-zglt4. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:21.566: INFO: Wrong image for pod: daemon-set-28d9j. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:21.566: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:21.566: INFO: Wrong image for pod: daemon-set-zglt4. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:22.554: INFO: Wrong image for pod: daemon-set-28d9j. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:22.554: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:22.554: INFO: Wrong image for pod: daemon-set-zglt4. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:23.620: INFO: Wrong image for pod: daemon-set-28d9j. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:23.620: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:23.620: INFO: Wrong image for pod: daemon-set-zglt4. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:24.556: INFO: Wrong image for pod: daemon-set-28d9j. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:24.556: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:24.556: INFO: Wrong image for pod: daemon-set-zglt4. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:25.555: INFO: Wrong image for pod: daemon-set-28d9j. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:25.555: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:25.555: INFO: Wrong image for pod: daemon-set-zglt4. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:26.555: INFO: Wrong image for pod: daemon-set-28d9j. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:26.555: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:26.555: INFO: Wrong image for pod: daemon-set-zglt4. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:27.555: INFO: Wrong image for pod: daemon-set-28d9j. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:27.555: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:27.555: INFO: Wrong image for pod: daemon-set-zglt4. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:28.555: INFO: Wrong image for pod: daemon-set-28d9j. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:28.555: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:28.555: INFO: Wrong image for pod: daemon-set-zglt4. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:29.555: INFO: Wrong image for pod: daemon-set-28d9j. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:29.555: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:29.555: INFO: Wrong image for pod: daemon-set-zglt4. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:30.554: INFO: Wrong image for pod: daemon-set-28d9j. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:30.554: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:30.554: INFO: Wrong image for pod: daemon-set-zglt4. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:31.554: INFO: Wrong image for pod: daemon-set-28d9j. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:31.554: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:31.554: INFO: Wrong image for pod: daemon-set-zglt4. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:32.827: INFO: Wrong image for pod: daemon-set-28d9j. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:32.827: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:32.827: INFO: Wrong image for pod: daemon-set-zglt4. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:33.556: INFO: Wrong image for pod: daemon-set-28d9j. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:33.556: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:33.556: INFO: Wrong image for pod: daemon-set-zglt4. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:34.554: INFO: Wrong image for pod: daemon-set-28d9j. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:34.554: INFO: Pod daemon-set-28d9j is not available -May 13 18:48:34.554: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:34.554: INFO: Wrong image for pod: daemon-set-zglt4. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:35.627: INFO: Wrong image for pod: daemon-set-28d9j. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:35.627: INFO: Pod daemon-set-28d9j is not available -May 13 18:48:35.627: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:35.627: INFO: Wrong image for pod: daemon-set-zglt4. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:36.554: INFO: Wrong image for pod: daemon-set-28d9j. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:36.555: INFO: Pod daemon-set-28d9j is not available -May 13 18:48:36.555: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:36.555: INFO: Wrong image for pod: daemon-set-zglt4. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:37.628: INFO: Wrong image for pod: daemon-set-28d9j. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:37.628: INFO: Pod daemon-set-28d9j is not available -May 13 18:48:37.628: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:37.628: INFO: Wrong image for pod: daemon-set-zglt4. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:38.628: INFO: Wrong image for pod: daemon-set-28d9j. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:38.628: INFO: Pod daemon-set-28d9j is not available -May 13 18:48:38.628: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:38.628: INFO: Wrong image for pod: daemon-set-zglt4. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:39.553: INFO: Wrong image for pod: daemon-set-28d9j. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:39.554: INFO: Pod daemon-set-28d9j is not available -May 13 18:48:39.554: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:39.554: INFO: Wrong image for pod: daemon-set-zglt4. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:40.620: INFO: Wrong image for pod: daemon-set-28d9j. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:40.620: INFO: Pod daemon-set-28d9j is not available -May 13 18:48:40.620: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:40.620: INFO: Wrong image for pod: daemon-set-zglt4. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:41.554: INFO: Wrong image for pod: daemon-set-28d9j. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:41.555: INFO: Pod daemon-set-28d9j is not available -May 13 18:48:41.555: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:41.555: INFO: Wrong image for pod: daemon-set-zglt4. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:42.554: INFO: Wrong image for pod: daemon-set-28d9j. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:42.554: INFO: Pod daemon-set-28d9j is not available -May 13 18:48:42.554: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:42.554: INFO: Wrong image for pod: daemon-set-zglt4. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:43.567: INFO: Wrong image for pod: daemon-set-28d9j. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:43.567: INFO: Pod daemon-set-28d9j is not available -May 13 18:48:43.567: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:43.567: INFO: Wrong image for pod: daemon-set-zglt4. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:44.620: INFO: Wrong image for pod: daemon-set-28d9j. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:44.620: INFO: Pod daemon-set-28d9j is not available -May 13 18:48:44.620: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:44.620: INFO: Wrong image for pod: daemon-set-zglt4. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:45.555: INFO: Wrong image for pod: daemon-set-28d9j. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:45.555: INFO: Pod daemon-set-28d9j is not available -May 13 18:48:45.555: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:45.555: INFO: Wrong image for pod: daemon-set-zglt4. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:46.555: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:46.555: INFO: Pod daemon-set-d7dmf is not available -May 13 18:48:46.555: INFO: Wrong image for pod: daemon-set-zglt4. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:47.554: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:47.554: INFO: Wrong image for pod: daemon-set-zglt4. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:48.554: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:48.554: INFO: Wrong image for pod: daemon-set-zglt4. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:49.555: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:49.555: INFO: Wrong image for pod: daemon-set-zglt4. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:50.620: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:50.620: INFO: Wrong image for pod: daemon-set-zglt4. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:51.554: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:51.554: INFO: Wrong image for pod: daemon-set-zglt4. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:52.555: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:52.555: INFO: Wrong image for pod: daemon-set-zglt4. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:53.555: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:53.555: INFO: Wrong image for pod: daemon-set-zglt4. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:54.566: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:54.566: INFO: Wrong image for pod: daemon-set-zglt4. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:55.684: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:55.684: INFO: Wrong image for pod: daemon-set-zglt4. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:56.555: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:56.555: INFO: Wrong image for pod: daemon-set-zglt4. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:57.555: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:57.555: INFO: Wrong image for pod: daemon-set-zglt4. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:58.554: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:58.554: INFO: Wrong image for pod: daemon-set-zglt4. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:59.555: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:48:59.555: INFO: Wrong image for pod: daemon-set-zglt4. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:49:00.554: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:49:00.554: INFO: Wrong image for pod: daemon-set-zglt4. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:49:01.554: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:49:01.554: INFO: Wrong image for pod: daemon-set-zglt4. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:49:02.554: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:49:02.554: INFO: Wrong image for pod: daemon-set-zglt4. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:49:03.554: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:49:03.554: INFO: Wrong image for pod: daemon-set-zglt4. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:49:04.554: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:49:04.554: INFO: Wrong image for pod: daemon-set-zglt4. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:49:05.567: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:49:05.567: INFO: Wrong image for pod: daemon-set-zglt4. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:49:06.554: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:49:06.555: INFO: Wrong image for pod: daemon-set-zglt4. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:49:07.554: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:49:07.554: INFO: Wrong image for pod: daemon-set-zglt4. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:49:08.554: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:49:08.555: INFO: Wrong image for pod: daemon-set-zglt4. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:49:09.554: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:49:09.554: INFO: Wrong image for pod: daemon-set-zglt4. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:49:10.554: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:49:10.554: INFO: Wrong image for pod: daemon-set-zglt4. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:49:11.554: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:49:11.555: INFO: Wrong image for pod: daemon-set-zglt4. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:49:12.554: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:49:12.554: INFO: Wrong image for pod: daemon-set-zglt4. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:49:13.620: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:49:13.620: INFO: Wrong image for pod: daemon-set-zglt4. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:49:14.554: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:49:14.554: INFO: Wrong image for pod: daemon-set-zglt4. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:49:15.555: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:49:15.555: INFO: Wrong image for pod: daemon-set-zglt4. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:49:16.583: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:49:16.583: INFO: Wrong image for pod: daemon-set-zglt4. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:49:17.554: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:49:17.554: INFO: Wrong image for pod: daemon-set-zglt4. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:49:18.555: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:49:18.555: INFO: Wrong image for pod: daemon-set-zglt4. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:49:19.555: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:49:19.555: INFO: Wrong image for pod: daemon-set-zglt4. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:49:19.555: INFO: Pod daemon-set-zglt4 is not available -May 13 18:49:20.554: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:49:20.555: INFO: Wrong image for pod: daemon-set-zglt4. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:49:20.555: INFO: Pod daemon-set-zglt4 is not available -May 13 18:49:21.555: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:49:21.555: INFO: Wrong image for pod: daemon-set-zglt4. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:49:21.555: INFO: Pod daemon-set-zglt4 is not available -May 13 18:49:22.557: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:49:22.557: INFO: Pod daemon-set-swv6c is not available -May 13 18:49:23.555: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:49:23.555: INFO: Pod daemon-set-swv6c is not available -May 13 18:49:24.554: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:49:25.555: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:49:26.555: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:49:27.570: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:49:28.556: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:49:29.556: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:49:30.620: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:49:31.554: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:49:32.554: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:49:33.554: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:49:34.555: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:49:35.554: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:49:36.583: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:49:37.554: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:49:38.567: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:49:39.555: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:49:40.554: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:49:41.554: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:49:42.554: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:49:44.172: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:49:44.555: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:49:45.555: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:49:46.555: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:49:47.554: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:49:48.554: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:49:49.567: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:49:50.554: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:49:51.554: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:49:52.554: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:49:53.554: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:49:54.554: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:49:55.554: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:49:55.554: INFO: Pod daemon-set-4cxm2 is not available -May 13 18:49:56.554: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:49:56.554: INFO: Pod daemon-set-4cxm2 is not available -May 13 18:49:57.554: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:49:57.554: INFO: Pod daemon-set-4cxm2 is not available -May 13 18:49:58.554: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:49:58.554: INFO: Pod daemon-set-4cxm2 is not available -May 13 18:49:59.554: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:49:59.554: INFO: Pod daemon-set-4cxm2 is not available -May 13 18:50:00.620: INFO: Wrong image for pod: daemon-set-4cxm2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. -May 13 18:50:00.620: INFO: Pod daemon-set-4cxm2 is not available -May 13 18:50:01.620: INFO: Pod daemon-set-42dj5 is not available -STEP: Check that daemon pods are still running on every node of the cluster. -May 13 18:50:01.642: INFO: Number of nodes with available pods: 2 -May 13 18:50:01.642: INFO: Node 10.170.219.177 is running more than one daemon pod -May 13 18:50:02.660: INFO: Number of nodes with available pods: 2 -May 13 18:50:02.660: INFO: Node 10.170.219.177 is running more than one daemon pod -May 13 18:50:03.728: INFO: Number of nodes with available pods: 3 -May 13 18:50:03.728: INFO: Number of running nodes: 3, number of available pods: 3 +Jun 18 11:38:33.280: INFO: Number of nodes with available pods: 0 +Jun 18 11:38:33.280: INFO: Node 10.72.74.143 is running more than one daemon pod +Jun 18 11:38:34.399: INFO: Number of nodes with available pods: 0 +Jun 18 11:38:34.399: INFO: Node 10.72.74.143 is running more than one daemon pod +Jun 18 11:38:35.323: INFO: Number of nodes with available pods: 0 +Jun 18 11:38:35.323: INFO: Node 10.72.74.143 is running more than one daemon pod +Jun 18 11:38:36.315: INFO: Number of nodes with available pods: 1 +Jun 18 11:38:36.315: INFO: Node 10.72.74.143 is running more than one daemon pod +Jun 18 11:38:37.314: INFO: Number of nodes with available pods: 3 +Jun 18 11:38:37.314: INFO: Number of running nodes: 3, number of available pods: 3 +STEP: Stop a daemon pod, check that the daemon pod is revived. +Jun 18 11:38:37.392: INFO: Number of nodes with available pods: 2 +Jun 18 11:38:37.392: INFO: Node 10.72.74.143 is running more than one daemon pod +Jun 18 11:38:38.504: INFO: Number of nodes with available pods: 2 +Jun 18 11:38:38.504: INFO: Node 10.72.74.143 is running more than one daemon pod +Jun 18 11:38:39.501: INFO: Number of nodes with available pods: 2 +Jun 18 11:38:39.501: INFO: Node 10.72.74.143 is running more than one daemon pod +Jun 18 11:38:40.427: INFO: Number of nodes with available pods: 2 +Jun 18 11:38:40.427: INFO: Node 10.72.74.143 is running more than one daemon pod +Jun 18 11:38:41.499: INFO: Number of nodes with available pods: 2 +Jun 18 11:38:41.499: INFO: Node 10.72.74.143 is running more than one daemon pod +Jun 18 11:38:42.428: INFO: Number of nodes with available pods: 2 +Jun 18 11:38:42.428: INFO: Node 10.72.74.143 is running more than one daemon pod +Jun 18 11:38:43.427: INFO: Number of nodes with available pods: 2 +Jun 18 11:38:43.427: INFO: Node 10.72.74.143 is running more than one daemon pod +Jun 18 11:38:44.443: INFO: Number of nodes with available pods: 2 +Jun 18 11:38:44.443: INFO: Node 10.72.74.143 is running more than one daemon pod +Jun 18 11:38:45.427: INFO: Number of nodes with available pods: 2 +Jun 18 11:38:45.427: INFO: Node 10.72.74.143 is running more than one daemon pod +Jun 18 11:38:46.424: INFO: Number of nodes with available pods: 2 +Jun 18 11:38:46.424: INFO: Node 10.72.74.143 is running more than one daemon pod +Jun 18 11:38:47.426: INFO: Number of nodes with available pods: 2 +Jun 18 11:38:47.426: INFO: Node 10.72.74.143 is running more than one daemon pod +Jun 18 11:38:48.499: INFO: Number of nodes with available pods: 2 +Jun 18 11:38:48.499: INFO: Node 10.72.74.143 is running more than one daemon pod +Jun 18 11:38:49.500: INFO: Number of nodes with available pods: 2 +Jun 18 11:38:49.500: INFO: Node 10.72.74.143 is running more than one daemon pod +Jun 18 11:38:50.425: INFO: Number of nodes with available pods: 2 +Jun 18 11:38:50.426: INFO: Node 10.72.74.143 is running more than one daemon pod +Jun 18 11:38:51.499: INFO: Number of nodes with available pods: 2 +Jun 18 11:38:51.499: INFO: Node 10.72.74.143 is running more than one daemon pod +Jun 18 11:38:52.425: INFO: Number of nodes with available pods: 2 +Jun 18 11:38:52.425: INFO: Node 10.72.74.143 is running more than one daemon pod +Jun 18 11:38:53.427: INFO: Number of nodes with available pods: 2 +Jun 18 11:38:53.427: INFO: Node 10.72.74.143 is running more than one daemon pod +Jun 18 11:38:54.484: INFO: Number of nodes with available pods: 2 +Jun 18 11:38:54.484: INFO: Node 10.72.74.143 is running more than one daemon pod +Jun 18 11:38:56.210: INFO: Number of nodes with available pods: 2 +Jun 18 11:38:56.210: INFO: Node 10.72.74.143 is running more than one daemon pod +Jun 18 11:38:56.425: INFO: Number of nodes with available pods: 2 +Jun 18 11:38:56.425: INFO: Node 10.72.74.143 is running more than one daemon pod +Jun 18 11:38:57.671: INFO: Number of nodes with available pods: 2 +Jun 18 11:38:57.672: INFO: Node 10.72.74.143 is running more than one daemon pod +Jun 18 11:38:58.499: INFO: Number of nodes with available pods: 2 +Jun 18 11:38:58.499: INFO: Node 10.72.74.143 is running more than one daemon pod +Jun 18 11:38:59.426: INFO: Number of nodes with available pods: 2 +Jun 18 11:38:59.426: INFO: Node 10.72.74.143 is running more than one daemon pod +Jun 18 11:39:00.500: INFO: Number of nodes with available pods: 2 +Jun 18 11:39:00.500: INFO: Node 10.72.74.143 is running more than one daemon pod +Jun 18 11:39:01.426: INFO: Number of nodes with available pods: 2 +Jun 18 11:39:01.426: INFO: Node 10.72.74.143 is running more than one daemon pod +Jun 18 11:39:02.426: INFO: Number of nodes with available pods: 2 +Jun 18 11:39:02.426: INFO: Node 10.72.74.143 is running more than one daemon pod +Jun 18 11:39:03.426: INFO: Number of nodes with available pods: 2 +Jun 18 11:39:03.426: INFO: Node 10.72.74.143 is running more than one daemon pod +Jun 18 11:39:04.426: INFO: Number of nodes with available pods: 2 +Jun 18 11:39:04.426: INFO: Node 10.72.74.143 is running more than one daemon pod +Jun 18 11:39:05.445: INFO: Number of nodes with available pods: 2 +Jun 18 11:39:05.445: INFO: Node 10.72.74.143 is running more than one daemon pod +Jun 18 11:39:06.499: INFO: Number of nodes with available pods: 2 +Jun 18 11:39:06.500: INFO: Node 10.72.74.143 is running more than one daemon pod +Jun 18 11:39:07.427: INFO: Number of nodes with available pods: 2 +Jun 18 11:39:07.427: INFO: Node 10.72.74.143 is running more than one daemon pod +Jun 18 11:39:08.427: INFO: Number of nodes with available pods: 2 +Jun 18 11:39:08.427: INFO: Node 10.72.74.143 is running more than one daemon pod +Jun 18 11:39:09.484: INFO: Number of nodes with available pods: 2 +Jun 18 11:39:09.484: INFO: Node 10.72.74.143 is running more than one daemon pod +Jun 18 11:39:10.427: INFO: Number of nodes with available pods: 2 +Jun 18 11:39:10.427: INFO: Node 10.72.74.143 is running more than one daemon pod +Jun 18 11:39:11.426: INFO: Number of nodes with available pods: 2 +Jun 18 11:39:11.426: INFO: Node 10.72.74.143 is running more than one daemon pod +Jun 18 11:39:12.427: INFO: Number of nodes with available pods: 2 +Jun 18 11:39:12.427: INFO: Node 10.72.74.143 is running more than one daemon pod +Jun 18 11:39:13.426: INFO: Number of nodes with available pods: 3 +Jun 18 11:39:13.426: INFO: Number of running nodes: 3, number of available pods: 3 [AfterEach] [sig-apps] Daemon set [Serial] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:68 STEP: Deleting DaemonSet "daemon-set" -STEP: deleting DaemonSet.extensions daemon-set in namespace e2e-tests-daemonsets-nd57p, will wait for the garbage collector to delete the pods -May 13 18:50:03.831: INFO: Deleting DaemonSet.extensions daemon-set took: 11.870457ms -May 13 18:50:03.931: INFO: Terminating DaemonSet.extensions daemon-set pods took: 100.23188ms -May 13 18:50:16.350: INFO: Number of nodes with available pods: 0 -May 13 18:50:16.350: INFO: Number of running nodes: 0, number of available pods: 0 -May 13 18:50:16.358: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"selfLink":"/apis/apps/v1/namespaces/e2e-tests-daemonsets-nd57p/daemonsets","resourceVersion":"33510"},"items":null} - -May 13 18:50:16.366: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"selfLink":"/api/v1/namespaces/e2e-tests-daemonsets-nd57p/pods","resourceVersion":"33510"},"items":null} - -[AfterEach] [sig-apps] Daemon set [Serial] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 18:50:16.390: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-daemonsets-nd57p" for this suite. -May 13 18:50:24.426: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 18:50:24.623: INFO: namespace: e2e-tests-daemonsets-nd57p, resource: bindings, ignored listing per whitelist -May 13 18:50:24.774: INFO: namespace e2e-tests-daemonsets-nd57p deletion completed in 8.37761987s - -• [SLOW TEST:145.834 seconds] -[sig-apps] Daemon set [Serial] -/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22 - should update pod when spec was updated and update strategy is RollingUpdate [Conformance] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------- -SSSSSS ------------------------------- -[sig-cli] Kubectl client [k8s.io] Kubectl cluster-info - should check if Kubernetes master services is included in cluster-info [Conformance] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -[BeforeEach] [sig-cli] Kubectl client - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 -STEP: Creating a kubernetes client -May 13 18:50:24.775: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -STEP: Building a namespace api object, basename kubectl -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-kubectl-nhhkm -STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [sig-cli] Kubectl client - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:243 -[It] should check if Kubernetes master services is included in cluster-info [Conformance] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -STEP: validating cluster-info -May 13 18:50:25.120: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 cluster-info' -May 13 18:50:25.242: INFO: stderr: "" -May 13 18:50:25.242: INFO: stdout: "\x1b[0;32mKubernetes master\x1b[0m is running at \x1b[0;33mhttps://172.21.0.1:443\x1b[0m\n\x1b[0;32mCoreDNS\x1b[0m is running at \x1b[0;33mhttps://172.21.0.1:443/api/v1/namespaces/kube-system/services/kube-dns:dns/proxy\x1b[0m\n\x1b[0;32mkubernetes-dashboard\x1b[0m is running at \x1b[0;33mhttps://172.21.0.1:443/api/v1/namespaces/kube-system/services/https:kubernetes-dashboard:/proxy\x1b[0m\n\x1b[0;32mMetrics-server\x1b[0m is running at \x1b[0;33mhttps://172.21.0.1:443/api/v1/namespaces/kube-system/services/https:metrics-server:/proxy\x1b[0m\n\nTo further debug and diagnose cluster problems, use 'kubectl cluster-info dump'.\n" -[AfterEach] [sig-cli] Kubectl client - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 18:50:25.242: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-kubectl-nhhkm" for this suite. -May 13 18:50:31.282: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 18:50:31.481: INFO: namespace: e2e-tests-kubectl-nhhkm, resource: bindings, ignored listing per whitelist -May 13 18:50:31.539: INFO: namespace e2e-tests-kubectl-nhhkm deletion completed in 6.289865859s - -• [SLOW TEST:6.764 seconds] -[sig-cli] Kubectl client -/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:22 - [k8s.io] Kubectl cluster-info - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 - should check if Kubernetes master services is included in cluster-info [Conformance] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------- -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------- -[sig-cli] Kubectl client [k8s.io] Update Demo - should scale a replication controller [Conformance] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -[BeforeEach] [sig-cli] Kubectl client - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 -STEP: Creating a kubernetes client -May 13 18:50:31.540: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -STEP: Building a namespace api object, basename kubectl -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-kubectl-ktwh4 -STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [sig-cli] Kubectl client - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:243 -[BeforeEach] [k8s.io] Update Demo - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:295 -[It] should scale a replication controller [Conformance] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -STEP: creating a replication controller -May 13 18:50:31.798: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 create -f - --namespace=e2e-tests-kubectl-ktwh4' -May 13 18:50:31.983: INFO: stderr: "" -May 13 18:50:31.983: INFO: stdout: "replicationcontroller/update-demo-nautilus created\n" -STEP: waiting for all containers in name=update-demo pods to come up. -May 13 18:50:31.983: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=e2e-tests-kubectl-ktwh4' -May 13 18:50:32.091: INFO: stderr: "" -May 13 18:50:32.091: INFO: stdout: "update-demo-nautilus-qlj7j update-demo-nautilus-rbt2q " -May 13 18:50:32.091: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 get pods update-demo-nautilus-qlj7j -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=e2e-tests-kubectl-ktwh4' -May 13 18:50:32.187: INFO: stderr: "" -May 13 18:50:32.187: INFO: stdout: "" -May 13 18:50:32.187: INFO: update-demo-nautilus-qlj7j is created but not running -May 13 18:50:37.187: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=e2e-tests-kubectl-ktwh4' -May 13 18:50:37.351: INFO: stderr: "" -May 13 18:50:37.351: INFO: stdout: "update-demo-nautilus-qlj7j update-demo-nautilus-rbt2q " -May 13 18:50:37.351: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 get pods update-demo-nautilus-qlj7j -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=e2e-tests-kubectl-ktwh4' -May 13 18:50:37.442: INFO: stderr: "" -May 13 18:50:37.442: INFO: stdout: "true" -May 13 18:50:37.442: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 get pods update-demo-nautilus-qlj7j -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=e2e-tests-kubectl-ktwh4' -May 13 18:50:37.538: INFO: stderr: "" -May 13 18:50:37.539: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0" -May 13 18:50:37.539: INFO: validating pod update-demo-nautilus-qlj7j -May 13 18:50:37.554: INFO: got data: { - "image": "nautilus.jpg" -} - -May 13 18:50:37.554: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . -May 13 18:50:37.554: INFO: update-demo-nautilus-qlj7j is verified up and running -May 13 18:50:37.554: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 get pods update-demo-nautilus-rbt2q -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=e2e-tests-kubectl-ktwh4' -May 13 18:50:37.664: INFO: stderr: "" -May 13 18:50:37.664: INFO: stdout: "true" -May 13 18:50:37.664: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 get pods update-demo-nautilus-rbt2q -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=e2e-tests-kubectl-ktwh4' -May 13 18:50:37.768: INFO: stderr: "" -May 13 18:50:37.769: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0" -May 13 18:50:37.769: INFO: validating pod update-demo-nautilus-rbt2q -May 13 18:50:37.783: INFO: got data: { - "image": "nautilus.jpg" -} - -May 13 18:50:37.783: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . -May 13 18:50:37.783: INFO: update-demo-nautilus-rbt2q is verified up and running -STEP: scaling down the replication controller -May 13 18:50:37.785: INFO: scanned /root for discovery docs: -May 13 18:50:37.785: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 scale rc update-demo-nautilus --replicas=1 --timeout=5m --namespace=e2e-tests-kubectl-ktwh4' -May 13 18:50:38.922: INFO: stderr: "" -May 13 18:50:38.922: INFO: stdout: "replicationcontroller/update-demo-nautilus scaled\n" -STEP: waiting for all containers in name=update-demo pods to come up. -May 13 18:50:38.922: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=e2e-tests-kubectl-ktwh4' -May 13 18:50:39.063: INFO: stderr: "" -May 13 18:50:39.063: INFO: stdout: "update-demo-nautilus-qlj7j update-demo-nautilus-rbt2q " -STEP: Replicas for name=update-demo: expected=1 actual=2 -May 13 18:50:44.063: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=e2e-tests-kubectl-ktwh4' -May 13 18:50:44.223: INFO: stderr: "" -May 13 18:50:44.223: INFO: stdout: "update-demo-nautilus-rbt2q " -May 13 18:50:44.223: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 get pods update-demo-nautilus-rbt2q -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=e2e-tests-kubectl-ktwh4' -May 13 18:50:44.331: INFO: stderr: "" -May 13 18:50:44.331: INFO: stdout: "true" -May 13 18:50:44.331: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 get pods update-demo-nautilus-rbt2q -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=e2e-tests-kubectl-ktwh4' -May 13 18:50:44.417: INFO: stderr: "" -May 13 18:50:44.417: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0" -May 13 18:50:44.417: INFO: validating pod update-demo-nautilus-rbt2q -May 13 18:50:44.429: INFO: got data: { - "image": "nautilus.jpg" -} - -May 13 18:50:44.430: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . -May 13 18:50:44.430: INFO: update-demo-nautilus-rbt2q is verified up and running -STEP: scaling up the replication controller -May 13 18:50:44.431: INFO: scanned /root for discovery docs: -May 13 18:50:44.432: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 scale rc update-demo-nautilus --replicas=2 --timeout=5m --namespace=e2e-tests-kubectl-ktwh4' -May 13 18:50:45.711: INFO: stderr: "" -May 13 18:50:45.711: INFO: stdout: "replicationcontroller/update-demo-nautilus scaled\n" -STEP: waiting for all containers in name=update-demo pods to come up. -May 13 18:50:45.711: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=e2e-tests-kubectl-ktwh4' -May 13 18:50:45.822: INFO: stderr: "" -May 13 18:50:45.822: INFO: stdout: "update-demo-nautilus-9l767 update-demo-nautilus-rbt2q " -May 13 18:50:45.822: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 get pods update-demo-nautilus-9l767 -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=e2e-tests-kubectl-ktwh4' -May 13 18:50:45.931: INFO: stderr: "" -May 13 18:50:45.931: INFO: stdout: "" -May 13 18:50:45.931: INFO: update-demo-nautilus-9l767 is created but not running -May 13 18:50:50.931: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=e2e-tests-kubectl-ktwh4' -May 13 18:50:51.031: INFO: stderr: "" -May 13 18:50:51.031: INFO: stdout: "update-demo-nautilus-9l767 update-demo-nautilus-rbt2q " -May 13 18:50:51.031: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 get pods update-demo-nautilus-9l767 -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=e2e-tests-kubectl-ktwh4' -May 13 18:50:51.151: INFO: stderr: "" -May 13 18:50:51.151: INFO: stdout: "true" -May 13 18:50:51.151: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 get pods update-demo-nautilus-9l767 -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=e2e-tests-kubectl-ktwh4' -May 13 18:50:51.255: INFO: stderr: "" -May 13 18:50:51.255: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0" -May 13 18:50:51.255: INFO: validating pod update-demo-nautilus-9l767 -May 13 18:50:51.271: INFO: got data: { - "image": "nautilus.jpg" -} - -May 13 18:50:51.271: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . -May 13 18:50:51.271: INFO: update-demo-nautilus-9l767 is verified up and running -May 13 18:50:51.271: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 get pods update-demo-nautilus-rbt2q -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=e2e-tests-kubectl-ktwh4' -May 13 18:50:51.365: INFO: stderr: "" -May 13 18:50:51.365: INFO: stdout: "true" -May 13 18:50:51.365: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 get pods update-demo-nautilus-rbt2q -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=e2e-tests-kubectl-ktwh4' -May 13 18:50:51.472: INFO: stderr: "" -May 13 18:50:51.472: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0" -May 13 18:50:51.472: INFO: validating pod update-demo-nautilus-rbt2q -May 13 18:50:51.483: INFO: got data: { - "image": "nautilus.jpg" -} - -May 13 18:50:51.483: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . -May 13 18:50:51.483: INFO: update-demo-nautilus-rbt2q is verified up and running -STEP: using delete to clean up resources -May 13 18:50:51.483: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 delete --grace-period=0 --force -f - --namespace=e2e-tests-kubectl-ktwh4' -May 13 18:50:51.597: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" -May 13 18:50:51.597: INFO: stdout: "replicationcontroller \"update-demo-nautilus\" force deleted\n" -May 13 18:50:51.598: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 get rc,svc -l name=update-demo --no-headers --namespace=e2e-tests-kubectl-ktwh4' -May 13 18:50:51.712: INFO: stderr: "No resources found.\n" -May 13 18:50:51.712: INFO: stdout: "" -May 13 18:50:51.712: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 get pods -l name=update-demo --namespace=e2e-tests-kubectl-ktwh4 -o go-template={{ range .items }}{{ if not .metadata.deletionTimestamp }}{{ .metadata.name }}{{ "\n" }}{{ end }}{{ end }}' -May 13 18:50:51.821: INFO: stderr: "" -May 13 18:50:51.821: INFO: stdout: "" -[AfterEach] [sig-cli] Kubectl client - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 18:50:51.821: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-kubectl-ktwh4" for this suite. -May 13 18:50:57.869: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 18:50:58.006: INFO: namespace: e2e-tests-kubectl-ktwh4, resource: bindings, ignored listing per whitelist -May 13 18:50:58.105: INFO: namespace e2e-tests-kubectl-ktwh4 deletion completed in 6.274579966s - -• [SLOW TEST:26.566 seconds] -[sig-cli] Kubectl client -/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:22 - [k8s.io] Update Demo - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 - should scale a replication controller [Conformance] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 +STEP: deleting DaemonSet.extensions daemon-set in namespace e2e-tests-daemonsets-frc4s, will wait for the garbage collector to delete the pods +Jun 18 11:39:13.529: INFO: Deleting DaemonSet.extensions daemon-set took: 28.035819ms +Jun 18 11:39:13.729: INFO: Terminating DaemonSet.extensions daemon-set pods took: 200.410306ms +Jun 18 11:39:56.063: INFO: Number of nodes with available pods: 0 +Jun 18 11:39:56.063: INFO: Number of running nodes: 0, number of available pods: 0 +Jun 18 11:39:56.081: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"selfLink":"/apis/apps/v1/namespaces/e2e-tests-daemonsets-frc4s/daemonsets","resourceVersion":"89552"},"items":null} + +Jun 18 11:39:56.095: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"selfLink":"/api/v1/namespaces/e2e-tests-daemonsets-frc4s/pods","resourceVersion":"89552"},"items":null} + +[AfterEach] [sig-apps] Daemon set [Serial] + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 +Jun 18 11:39:56.156: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "e2e-tests-daemonsets-frc4s" for this suite. +Jun 18 11:40:04.223: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 11:40:04.635: INFO: namespace: e2e-tests-daemonsets-frc4s, resource: bindings, ignored listing per whitelist +Jun 18 11:40:04.747: INFO: namespace e2e-tests-daemonsets-frc4s deletion completed in 8.576766657s + +• [SLOW TEST:92.066 seconds] +[sig-apps] Daemon set [Serial] +/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22 + should run and stop simple daemon [Conformance] + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ -SSSSS +SSSSSS ------------------------------ -[sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] - should perform rolling updates and roll backs of template modifications [Conformance] +[k8s.io] Probing container + should have monotonically increasing restart count [Slow][NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -[BeforeEach] [sig-apps] StatefulSet +[BeforeEach] [k8s.io] Probing container /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client -May 13 18:50:58.105: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -STEP: Building a namespace api object, basename statefulset -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-statefulset-cdqz9 +Jun 18 11:40:04.748: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 +STEP: Building a namespace api object, basename container-probe +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-container-probe-9cb5v STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [sig-apps] StatefulSet - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:59 -[BeforeEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:74 -STEP: Creating service test in namespace e2e-tests-statefulset-cdqz9 -[It] should perform rolling updates and roll backs of template modifications [Conformance] +[BeforeEach] [k8s.io] Probing container + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/container_probe.go:48 +[It] should have monotonically increasing restart count [Slow][NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -STEP: Creating a new StatefulSet -May 13 18:50:58.447: INFO: Found 0 stateful pods, waiting for 3 -May 13 18:51:08.469: INFO: Waiting for pod ss2-0 to enter Running - Ready=true, currently Running - Ready=true -May 13 18:51:08.469: INFO: Waiting for pod ss2-1 to enter Running - Ready=true, currently Running - Ready=true -May 13 18:51:08.469: INFO: Waiting for pod ss2-2 to enter Running - Ready=true, currently Running - Ready=true -May 13 18:51:08.490: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 exec --namespace=e2e-tests-statefulset-cdqz9 ss2-1 -- /bin/sh -c mv -v /usr/share/nginx/html/index.html /tmp/ || true' -May 13 18:51:08.785: INFO: stderr: "" -May 13 18:51:08.786: INFO: stdout: "'/usr/share/nginx/html/index.html' -> '/tmp/index.html'\n" -May 13 18:51:08.786: INFO: stdout of mv -v /usr/share/nginx/html/index.html /tmp/ || true on ss2-1: '/usr/share/nginx/html/index.html' -> '/tmp/index.html' - -STEP: Updating StatefulSet template: update image from docker.io/library/nginx:1.14-alpine to docker.io/library/nginx:1.15-alpine -May 13 18:51:18.848: INFO: Updating stateful set ss2 -STEP: Creating a new revision -STEP: Updating Pods in reverse ordinal order -May 13 18:51:28.928: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 exec --namespace=e2e-tests-statefulset-cdqz9 ss2-1 -- /bin/sh -c mv -v /tmp/index.html /usr/share/nginx/html/ || true' -May 13 18:51:29.173: INFO: stderr: "" -May 13 18:51:29.173: INFO: stdout: "'/tmp/index.html' -> '/usr/share/nginx/html/index.html'\n" -May 13 18:51:29.173: INFO: stdout of mv -v /tmp/index.html /usr/share/nginx/html/ || true on ss2-1: '/tmp/index.html' -> '/usr/share/nginx/html/index.html' - -May 13 18:51:39.320: INFO: Waiting for StatefulSet e2e-tests-statefulset-cdqz9/ss2 to complete update -May 13 18:51:39.320: INFO: Waiting for Pod e2e-tests-statefulset-cdqz9/ss2-0 to have revision ss2-c79899b9 update revision ss2-787997d666 -May 13 18:51:39.320: INFO: Waiting for Pod e2e-tests-statefulset-cdqz9/ss2-1 to have revision ss2-c79899b9 update revision ss2-787997d666 -May 13 18:51:39.320: INFO: Waiting for Pod e2e-tests-statefulset-cdqz9/ss2-2 to have revision ss2-c79899b9 update revision ss2-787997d666 -May 13 18:51:49.346: INFO: Waiting for StatefulSet e2e-tests-statefulset-cdqz9/ss2 to complete update -May 13 18:51:49.346: INFO: Waiting for Pod e2e-tests-statefulset-cdqz9/ss2-0 to have revision ss2-c79899b9 update revision ss2-787997d666 -STEP: Rolling back to a previous revision -May 13 18:51:59.335: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 exec --namespace=e2e-tests-statefulset-cdqz9 ss2-1 -- /bin/sh -c mv -v /usr/share/nginx/html/index.html /tmp/ || true' -May 13 18:51:59.719: INFO: stderr: "" -May 13 18:51:59.719: INFO: stdout: "'/usr/share/nginx/html/index.html' -> '/tmp/index.html'\n" -May 13 18:51:59.719: INFO: stdout of mv -v /usr/share/nginx/html/index.html /tmp/ || true on ss2-1: '/usr/share/nginx/html/index.html' -> '/tmp/index.html' - -May 13 18:52:09.819: INFO: Updating stateful set ss2 -STEP: Rolling back update in reverse ordinal order -May 13 18:52:09.840: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 exec --namespace=e2e-tests-statefulset-cdqz9 ss2-1 -- /bin/sh -c mv -v /tmp/index.html /usr/share/nginx/html/ || true' -May 13 18:52:10.117: INFO: stderr: "" -May 13 18:52:10.117: INFO: stdout: "'/tmp/index.html' -> '/usr/share/nginx/html/index.html'\n" -May 13 18:52:10.117: INFO: stdout of mv -v /tmp/index.html /usr/share/nginx/html/ || true on ss2-1: '/tmp/index.html' -> '/usr/share/nginx/html/index.html' - -May 13 18:52:30.173: INFO: Waiting for StatefulSet e2e-tests-statefulset-cdqz9/ss2 to complete update -May 13 18:52:30.173: INFO: Waiting for Pod e2e-tests-statefulset-cdqz9/ss2-0 to have revision ss2-787997d666 update revision ss2-c79899b9 -May 13 18:52:40.227: INFO: Waiting for StatefulSet e2e-tests-statefulset-cdqz9/ss2 to complete update -May 13 18:52:40.227: INFO: Waiting for Pod e2e-tests-statefulset-cdqz9/ss2-0 to have revision ss2-787997d666 update revision ss2-c79899b9 -[AfterEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:85 -May 13 18:52:50.187: INFO: Deleting all statefulset in ns e2e-tests-statefulset-cdqz9 -May 13 18:52:50.207: INFO: Scaling statefulset ss2 to 0 -May 13 18:53:20.260: INFO: Waiting for statefulset status.replicas updated to 0 -May 13 18:53:20.266: INFO: Deleting statefulset ss2 -[AfterEach] [sig-apps] StatefulSet +STEP: Creating pod liveness-http in namespace e2e-tests-container-probe-9cb5v +Jun 18 11:40:11.302: INFO: Started pod liveness-http in namespace e2e-tests-container-probe-9cb5v +STEP: checking the pod's current state and verifying that restartCount is present +Jun 18 11:40:11.319: INFO: Initial restart count of pod liveness-http is 0 +Jun 18 11:40:27.501: INFO: Restart count of pod e2e-tests-container-probe-9cb5v/liveness-http is now 1 (16.182143884s elapsed) +Jun 18 11:40:46.535: INFO: Restart count of pod e2e-tests-container-probe-9cb5v/liveness-http is now 2 (35.215712929s elapsed) +Jun 18 11:41:07.242: INFO: Restart count of pod e2e-tests-container-probe-9cb5v/liveness-http is now 3 (55.92263753s elapsed) +Jun 18 11:41:27.427: INFO: Restart count of pod e2e-tests-container-probe-9cb5v/liveness-http is now 4 (1m16.10776734s elapsed) +Jun 18 11:42:32.185: INFO: Restart count of pod e2e-tests-container-probe-9cb5v/liveness-http is now 5 (2m20.86648334s elapsed) +STEP: deleting the pod +[AfterEach] [k8s.io] Probing container /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 18:53:20.320: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-statefulset-cdqz9" for this suite. -May 13 18:53:28.355: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 18:53:28.708: INFO: namespace: e2e-tests-statefulset-cdqz9, resource: bindings, ignored listing per whitelist -May 13 18:53:28.778: INFO: namespace e2e-tests-statefulset-cdqz9 deletion completed in 8.449909689s +Jun 18 11:42:32.235: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "e2e-tests-container-probe-9cb5v" for this suite. +Jun 18 11:42:38.314: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 11:42:38.487: INFO: namespace: e2e-tests-container-probe-9cb5v, resource: bindings, ignored listing per whitelist +Jun 18 11:42:38.844: INFO: namespace e2e-tests-container-probe-9cb5v deletion completed in 6.585980418s -• [SLOW TEST:150.672 seconds] -[sig-apps] StatefulSet -/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22 - [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 - should perform rolling updates and roll backs of template modifications [Conformance] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 +• [SLOW TEST:154.096 seconds] +[k8s.io] Probing container +/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 + should have monotonically increasing restart count [Slow][NodeConformance] [Conformance] + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ -SSS +SSSS ------------------------------ -[sig-apps] ReplicaSet - should serve a basic image on each replica with a public image [Conformance] +[sig-storage] ConfigMap + should be consumable from pods in volume as non-root [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -[BeforeEach] [sig-apps] ReplicaSet +[BeforeEach] [sig-storage] ConfigMap /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client -May 13 18:53:28.778: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -STEP: Building a namespace api object, basename replicaset -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-replicaset-mks6t +Jun 18 11:42:38.844: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 +STEP: Building a namespace api object, basename configmap +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-configmap-bbl59 STEP: Waiting for a default service account to be provisioned in namespace -[It] should serve a basic image on each replica with a public image [Conformance] +[It] should be consumable from pods in volume as non-root [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -May 13 18:53:29.123: INFO: Creating ReplicaSet my-hostname-basic-65a0a4b3-75b0-11e9-8f67-2632f168be36 -May 13 18:53:29.142: INFO: Pod name my-hostname-basic-65a0a4b3-75b0-11e9-8f67-2632f168be36: Found 0 pods out of 1 -May 13 18:53:34.164: INFO: Pod name my-hostname-basic-65a0a4b3-75b0-11e9-8f67-2632f168be36: Found 1 pods out of 1 -May 13 18:53:34.164: INFO: Ensuring a pod for ReplicaSet "my-hostname-basic-65a0a4b3-75b0-11e9-8f67-2632f168be36" is running -May 13 18:53:34.177: INFO: Pod "my-hostname-basic-65a0a4b3-75b0-11e9-8f67-2632f168be36-x2j86" is running (conditions: [{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2019-05-13 18:53:29 +0000 UTC Reason: Message:} {Type:Ready Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2019-05-13 18:53:30 +0000 UTC Reason: Message:} {Type:ContainersReady Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2019-05-13 18:53:30 +0000 UTC Reason: Message:} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2019-05-13 18:53:29 +0000 UTC Reason: Message:}]) -May 13 18:53:34.177: INFO: Trying to dial the pod -May 13 18:53:39.242: INFO: Controller my-hostname-basic-65a0a4b3-75b0-11e9-8f67-2632f168be36: Got expected result from replica 1 [my-hostname-basic-65a0a4b3-75b0-11e9-8f67-2632f168be36-x2j86]: "my-hostname-basic-65a0a4b3-75b0-11e9-8f67-2632f168be36-x2j86", 1 of 1 required successes so far -[AfterEach] [sig-apps] ReplicaSet +STEP: Creating configMap with name configmap-test-volume-2ccc87bb-91be-11e9-bce2-ae54e022189f +STEP: Creating a pod to test consume configMaps +Jun 18 11:42:39.333: INFO: Waiting up to 5m0s for pod "pod-configmaps-2cce9b75-91be-11e9-bce2-ae54e022189f" in namespace "e2e-tests-configmap-bbl59" to be "success or failure" +Jun 18 11:42:39.384: INFO: Pod "pod-configmaps-2cce9b75-91be-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 51.1557ms +Jun 18 11:42:41.416: INFO: Pod "pod-configmaps-2cce9b75-91be-11e9-bce2-ae54e022189f": Phase="Running", Reason="", readiness=true. Elapsed: 2.08279105s +Jun 18 11:42:43.431: INFO: Pod "pod-configmaps-2cce9b75-91be-11e9-bce2-ae54e022189f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.098095355s +STEP: Saw pod success +Jun 18 11:42:43.431: INFO: Pod "pod-configmaps-2cce9b75-91be-11e9-bce2-ae54e022189f" satisfied condition "success or failure" +Jun 18 11:42:43.445: INFO: Trying to get logs from node 10.72.74.149 pod pod-configmaps-2cce9b75-91be-11e9-bce2-ae54e022189f container configmap-volume-test: +STEP: delete the pod +Jun 18 11:42:43.526: INFO: Waiting for pod pod-configmaps-2cce9b75-91be-11e9-bce2-ae54e022189f to disappear +Jun 18 11:42:43.542: INFO: Pod pod-configmaps-2cce9b75-91be-11e9-bce2-ae54e022189f no longer exists +[AfterEach] [sig-storage] ConfigMap /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 18:53:39.242: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-replicaset-mks6t" for this suite. -May 13 18:53:45.339: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 18:53:45.593: INFO: namespace: e2e-tests-replicaset-mks6t, resource: bindings, ignored listing per whitelist -May 13 18:53:45.606: INFO: namespace e2e-tests-replicaset-mks6t deletion completed in 6.352888264s +Jun 18 11:42:43.542: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "e2e-tests-configmap-bbl59" for this suite. +Jun 18 11:42:49.607: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 11:42:49.872: INFO: namespace: e2e-tests-configmap-bbl59, resource: bindings, ignored listing per whitelist +Jun 18 11:42:50.136: INFO: namespace e2e-tests-configmap-bbl59 deletion completed in 6.575749634s -• [SLOW TEST:16.828 seconds] -[sig-apps] ReplicaSet -/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22 - should serve a basic image on each replica with a public image [Conformance] +• [SLOW TEST:11.292 seconds] +[sig-storage] ConfigMap +/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:33 + should be consumable from pods in volume as non-root [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ -SSSSSSSS +SSSSSS ------------------------------ -[sig-node] Downward API - should provide pod name, namespace and IP address as env vars [NodeConformance] [Conformance] +[k8s.io] Pods + should contain environment variables for services [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -[BeforeEach] [sig-node] Downward API +[BeforeEach] [k8s.io] Pods /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client -May 13 18:53:45.607: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -STEP: Building a namespace api object, basename downward-api -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-downward-api-4mlpn +Jun 18 11:42:50.137: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 +STEP: Building a namespace api object, basename pods +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-pods-49ldv STEP: Waiting for a default service account to be provisioned in namespace -[It] should provide pod name, namespace and IP address as env vars [NodeConformance] [Conformance] +[BeforeEach] [k8s.io] Pods + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:132 +[It] should contain environment variables for services [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -STEP: Creating a pod to test downward api env vars -May 13 18:53:45.897: INFO: Waiting up to 5m0s for pod "downward-api-6f9db637-75b0-11e9-8f67-2632f168be36" in namespace "e2e-tests-downward-api-4mlpn" to be "success or failure" -May 13 18:53:45.904: INFO: Pod "downward-api-6f9db637-75b0-11e9-8f67-2632f168be36": Phase="Pending", Reason="", readiness=false. Elapsed: 7.231598ms -May 13 18:53:47.914: INFO: Pod "downward-api-6f9db637-75b0-11e9-8f67-2632f168be36": Phase="Running", Reason="", readiness=true. Elapsed: 2.017030187s -May 13 18:53:49.923: INFO: Pod "downward-api-6f9db637-75b0-11e9-8f67-2632f168be36": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.025552535s +Jun 18 11:42:54.799: INFO: Waiting up to 5m0s for pod "client-envvars-3606d9a3-91be-11e9-bce2-ae54e022189f" in namespace "e2e-tests-pods-49ldv" to be "success or failure" +Jun 18 11:42:54.814: INFO: Pod "client-envvars-3606d9a3-91be-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 15.003756ms +Jun 18 11:42:56.829: INFO: Pod "client-envvars-3606d9a3-91be-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 2.029770766s +Jun 18 11:42:58.843: INFO: Pod "client-envvars-3606d9a3-91be-11e9-bce2-ae54e022189f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.044526834s STEP: Saw pod success -May 13 18:53:49.923: INFO: Pod "downward-api-6f9db637-75b0-11e9-8f67-2632f168be36" satisfied condition "success or failure" -May 13 18:53:49.930: INFO: Trying to get logs from node 10.170.219.153 pod downward-api-6f9db637-75b0-11e9-8f67-2632f168be36 container dapi-container: +Jun 18 11:42:58.844: INFO: Pod "client-envvars-3606d9a3-91be-11e9-bce2-ae54e022189f" satisfied condition "success or failure" +Jun 18 11:42:58.858: INFO: Trying to get logs from node 10.72.74.144 pod client-envvars-3606d9a3-91be-11e9-bce2-ae54e022189f container env3cont: STEP: delete the pod -May 13 18:53:49.974: INFO: Waiting for pod downward-api-6f9db637-75b0-11e9-8f67-2632f168be36 to disappear -May 13 18:53:49.982: INFO: Pod downward-api-6f9db637-75b0-11e9-8f67-2632f168be36 no longer exists -[AfterEach] [sig-node] Downward API +Jun 18 11:42:58.937: INFO: Waiting for pod client-envvars-3606d9a3-91be-11e9-bce2-ae54e022189f to disappear +Jun 18 11:42:58.954: INFO: Pod client-envvars-3606d9a3-91be-11e9-bce2-ae54e022189f no longer exists +[AfterEach] [k8s.io] Pods /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 18:53:49.982: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-downward-api-4mlpn" for this suite. -May 13 18:53:56.029: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 18:53:56.188: INFO: namespace: e2e-tests-downward-api-4mlpn, resource: bindings, ignored listing per whitelist -May 13 18:53:56.261: INFO: namespace e2e-tests-downward-api-4mlpn deletion completed in 6.270758168s +Jun 18 11:42:58.954: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "e2e-tests-pods-49ldv" for this suite. +Jun 18 11:43:41.041: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 11:43:41.259: INFO: namespace: e2e-tests-pods-49ldv, resource: bindings, ignored listing per whitelist +Jun 18 11:43:41.540: INFO: namespace e2e-tests-pods-49ldv deletion completed in 42.55032617s -• [SLOW TEST:10.654 seconds] -[sig-node] Downward API -/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downward_api.go:38 - should provide pod name, namespace and IP address as env vars [NodeConformance] [Conformance] +• [SLOW TEST:51.403 seconds] +[k8s.io] Pods +/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 + should contain environment variables for services [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +SSSS ------------------------------ -[sig-network] Services - should serve a basic endpoint from pods [Conformance] +[sig-storage] Downward API volume + should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -[BeforeEach] [sig-network] Services +[BeforeEach] [sig-storage] Downward API volume /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client -May 13 18:53:56.262: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -STEP: Building a namespace api object, basename services -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-services-bbq75 +Jun 18 11:43:41.540: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 +STEP: Building a namespace api object, basename downward-api +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-downward-api-qdpzb STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [sig-network] Services - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:85 -[It] should serve a basic endpoint from pods [Conformance] +[BeforeEach] [sig-storage] Downward API volume + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:39 +[It] should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -STEP: creating service endpoint-test2 in namespace e2e-tests-services-bbq75 -STEP: waiting up to 3m0s for service endpoint-test2 in namespace e2e-tests-services-bbq75 to expose endpoints map[] -May 13 18:53:56.559: INFO: Get endpoints failed (5.246977ms elapsed, ignoring for 5s): endpoints "endpoint-test2" not found -May 13 18:53:57.565: INFO: successfully validated that service endpoint-test2 in namespace e2e-tests-services-bbq75 exposes endpoints map[] (1.011360677s elapsed) -STEP: Creating pod pod1 in namespace e2e-tests-services-bbq75 -STEP: waiting up to 3m0s for service endpoint-test2 in namespace e2e-tests-services-bbq75 to expose endpoints map[pod1:[80]] -May 13 18:53:59.623: INFO: successfully validated that service endpoint-test2 in namespace e2e-tests-services-bbq75 exposes endpoints map[pod1:[80]] (2.042324207s elapsed) -STEP: Creating pod pod2 in namespace e2e-tests-services-bbq75 -STEP: waiting up to 3m0s for service endpoint-test2 in namespace e2e-tests-services-bbq75 to expose endpoints map[pod1:[80] pod2:[80]] -May 13 18:54:01.841: INFO: successfully validated that service endpoint-test2 in namespace e2e-tests-services-bbq75 exposes endpoints map[pod1:[80] pod2:[80]] (2.207236477s elapsed) -STEP: Deleting pod pod1 in namespace e2e-tests-services-bbq75 -STEP: waiting up to 3m0s for service endpoint-test2 in namespace e2e-tests-services-bbq75 to expose endpoints map[pod2:[80]] -May 13 18:54:02.882: INFO: successfully validated that service endpoint-test2 in namespace e2e-tests-services-bbq75 exposes endpoints map[pod2:[80]] (1.026519591s elapsed) -STEP: Deleting pod pod2 in namespace e2e-tests-services-bbq75 -STEP: waiting up to 3m0s for service endpoint-test2 in namespace e2e-tests-services-bbq75 to expose endpoints map[] -May 13 18:54:03.907: INFO: successfully validated that service endpoint-test2 in namespace e2e-tests-services-bbq75 exposes endpoints map[] (1.011697819s elapsed) -[AfterEach] [sig-network] Services +STEP: Creating a pod to test downward API volume plugin +Jun 18 11:43:42.340: INFO: Waiting up to 5m0s for pod "downwardapi-volume-525cdba9-91be-11e9-bce2-ae54e022189f" in namespace "e2e-tests-downward-api-qdpzb" to be "success or failure" +Jun 18 11:43:42.358: INFO: Pod "downwardapi-volume-525cdba9-91be-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 18.119208ms +Jun 18 11:43:44.373: INFO: Pod "downwardapi-volume-525cdba9-91be-11e9-bce2-ae54e022189f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.033622535s +STEP: Saw pod success +Jun 18 11:43:44.373: INFO: Pod "downwardapi-volume-525cdba9-91be-11e9-bce2-ae54e022189f" satisfied condition "success or failure" +Jun 18 11:43:44.387: INFO: Trying to get logs from node 10.72.74.149 pod downwardapi-volume-525cdba9-91be-11e9-bce2-ae54e022189f container client-container: +STEP: delete the pod +Jun 18 11:43:44.557: INFO: Waiting for pod downwardapi-volume-525cdba9-91be-11e9-bce2-ae54e022189f to disappear +Jun 18 11:43:44.572: INFO: Pod downwardapi-volume-525cdba9-91be-11e9-bce2-ae54e022189f no longer exists +[AfterEach] [sig-storage] Downward API volume /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 18:54:03.963: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-services-bbq75" for this suite. -May 13 18:54:28.000: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 18:54:28.341: INFO: namespace: e2e-tests-services-bbq75, resource: bindings, ignored listing per whitelist -May 13 18:54:28.360: INFO: namespace e2e-tests-services-bbq75 deletion completed in 24.388848919s -[AfterEach] [sig-network] Services - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:90 +Jun 18 11:43:44.572: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "e2e-tests-downward-api-qdpzb" for this suite. +Jun 18 11:43:50.660: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 11:43:51.055: INFO: namespace: e2e-tests-downward-api-qdpzb, resource: bindings, ignored listing per whitelist +Jun 18 11:43:51.246: INFO: namespace e2e-tests-downward-api-qdpzb deletion completed in 6.654917712s -• [SLOW TEST:32.098 seconds] -[sig-network] Services -/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:22 - should serve a basic endpoint from pods [Conformance] +• [SLOW TEST:9.706 seconds] +[sig-storage] Downward API volume +/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:34 + should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +SSSSS ------------------------------ -[sig-apps] Deployment - RecreateDeployment should delete old pods and create new ones [Conformance] +[sig-apps] Daemon set [Serial] + should update pod when spec was updated and update strategy is RollingUpdate [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -[BeforeEach] [sig-apps] Deployment +[BeforeEach] [sig-apps] Daemon set [Serial] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client -May 13 18:54:28.363: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -STEP: Building a namespace api object, basename deployment -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-deployment-blp6j +Jun 18 11:43:51.246: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 +STEP: Building a namespace api object, basename daemonsets +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-daemonsets-sd8qh STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [sig-apps] Deployment - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:65 -[It] RecreateDeployment should delete old pods and create new ones [Conformance] +[BeforeEach] [sig-apps] Daemon set [Serial] + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:102 +[It] should update pod when spec was updated and update strategy is RollingUpdate [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -May 13 18:54:28.642: INFO: Creating deployment "test-recreate-deployment" -May 13 18:54:28.650: INFO: Waiting deployment "test-recreate-deployment" to be updated to revision 1 -May 13 18:54:28.662: INFO: new replicaset for deployment "test-recreate-deployment" is yet to be created -May 13 18:54:30.689: INFO: Waiting deployment "test-recreate-deployment" to complete -May 13 18:54:30.695: INFO: Triggering a new rollout for deployment "test-recreate-deployment" -May 13 18:54:30.706: INFO: Updating deployment test-recreate-deployment -May 13 18:54:30.706: INFO: Watching deployment "test-recreate-deployment" to verify that new pods will not run with olds pods -[AfterEach] [sig-apps] Deployment - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:59 -May 13 18:54:30.799: INFO: Deployment "test-recreate-deployment": -&Deployment{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-recreate-deployment,GenerateName:,Namespace:e2e-tests-deployment-blp6j,SelfLink:/apis/apps/v1/namespaces/e2e-tests-deployment-blp6j/deployments/test-recreate-deployment,UID:891b10aa-75b0-11e9-b786-da20024d205c,ResourceVersion:34719,Generation:2,CreationTimestamp:2019-05-13 18:54:28 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod-3,},Annotations:map[string]string{deployment.kubernetes.io/revision: 2,},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Spec:DeploymentSpec{Replicas:*1,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod-3,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod-3,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,},},Strategy:DeploymentStrategy{Type:Recreate,RollingUpdate:nil,},MinReadySeconds:0,RevisionHistoryLimit:*10,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:2,Replicas:1,UpdatedReplicas:1,AvailableReplicas:0,UnavailableReplicas:1,Conditions:[{Available False 2019-05-13 18:54:30 +0000 UTC 2019-05-13 18:54:30 +0000 UTC MinimumReplicasUnavailable Deployment does not have minimum availability.} {Progressing True 2019-05-13 18:54:30 +0000 UTC 2019-05-13 18:54:28 +0000 UTC ReplicaSetUpdated ReplicaSet "test-recreate-deployment-697fbf54bf" is progressing.}],ReadyReplicas:0,CollisionCount:nil,},} - -May 13 18:54:30.807: INFO: New ReplicaSet "test-recreate-deployment-697fbf54bf" of Deployment "test-recreate-deployment": -&ReplicaSet{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-recreate-deployment-697fbf54bf,GenerateName:,Namespace:e2e-tests-deployment-blp6j,SelfLink:/apis/apps/v1/namespaces/e2e-tests-deployment-blp6j/replicasets/test-recreate-deployment-697fbf54bf,UID:8a5c2507-75b0-11e9-83a8-7e0242378207,ResourceVersion:34717,Generation:1,CreationTimestamp:2019-05-13 18:54:30 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod-3,pod-template-hash: 697fbf54bf,},Annotations:map[string]string{deployment.kubernetes.io/desired-replicas: 1,deployment.kubernetes.io/max-replicas: 1,deployment.kubernetes.io/revision: 2,},OwnerReferences:[{apps/v1 Deployment test-recreate-deployment 891b10aa-75b0-11e9-b786-da20024d205c 0xc001076327 0xc001076328}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:ReplicaSetSpec{Replicas:*1,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod-3,pod-template-hash: 697fbf54bf,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod-3,pod-template-hash: 697fbf54bf,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,},},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:1,FullyLabeledReplicas:1,ObservedGeneration:1,ReadyReplicas:0,AvailableReplicas:0,Conditions:[],},} -May 13 18:54:30.807: INFO: All old ReplicaSets of Deployment "test-recreate-deployment": -May 13 18:54:30.807: INFO: &ReplicaSet{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-recreate-deployment-5dfdcc846d,GenerateName:,Namespace:e2e-tests-deployment-blp6j,SelfLink:/apis/apps/v1/namespaces/e2e-tests-deployment-blp6j/replicasets/test-recreate-deployment-5dfdcc846d,UID:891e5f35-75b0-11e9-83a8-7e0242378207,ResourceVersion:34707,Generation:2,CreationTimestamp:2019-05-13 18:54:28 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod-3,pod-template-hash: 5dfdcc846d,},Annotations:map[string]string{deployment.kubernetes.io/desired-replicas: 1,deployment.kubernetes.io/max-replicas: 1,deployment.kubernetes.io/revision: 1,},OwnerReferences:[{apps/v1 Deployment test-recreate-deployment 891b10aa-75b0-11e9-b786-da20024d205c 0xc001bc5ee7 0xc001bc5ee8}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:ReplicaSetSpec{Replicas:*0,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod-3,pod-template-hash: 5dfdcc846d,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod-3,pod-template-hash: 5dfdcc846d,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[],Containers:[{redis gcr.io/kubernetes-e2e-test-images/redis:1.0 [] [] [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,},},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:0,FullyLabeledReplicas:0,ObservedGeneration:2,ReadyReplicas:0,AvailableReplicas:0,Conditions:[],},} -May 13 18:54:30.816: INFO: Pod "test-recreate-deployment-697fbf54bf-wxctj" is not available: -&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-recreate-deployment-697fbf54bf-wxctj,GenerateName:test-recreate-deployment-697fbf54bf-,Namespace:e2e-tests-deployment-blp6j,SelfLink:/api/v1/namespaces/e2e-tests-deployment-blp6j/pods/test-recreate-deployment-697fbf54bf-wxctj,UID:8a5d3573-75b0-11e9-83a8-7e0242378207,ResourceVersion:34718,Generation:0,CreationTimestamp:2019-05-13 18:54:30 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod-3,pod-template-hash: 697fbf54bf,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet test-recreate-deployment-697fbf54bf 8a5c2507-75b0-11e9-83a8-7e0242378207 0xc000feaea7 0xc000feaea8}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-zpkl5 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-zpkl5,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-zpkl5 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.170.219.177,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc000feaf30} {node.kubernetes.io/unreachable Exists NoExecute 0xc000feaf50}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 18:54:30 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-05-13 18:54:30 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-05-13 18:54:30 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 18:54:30 +0000 UTC }],Message:,Reason:,HostIP:10.170.219.177,PodIP:,StartTime:2019-05-13 18:54:30 +0000 UTC,ContainerStatuses:[{nginx {ContainerStateWaiting{Reason:ContainerCreating,Message:,} nil nil} {nil nil nil} false 0 docker.io/library/nginx:1.14-alpine }],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} -[AfterEach] [sig-apps] Deployment +Jun 18 11:43:51.754: INFO: Creating simple daemon set daemon-set +STEP: Check that daemon pods launch on every node of the cluster. +Jun 18 11:43:51.797: INFO: Number of nodes with available pods: 0 +Jun 18 11:43:51.797: INFO: Node 10.72.74.143 is running more than one daemon pod +Jun 18 11:43:52.830: INFO: Number of nodes with available pods: 0 +Jun 18 11:43:52.830: INFO: Node 10.72.74.143 is running more than one daemon pod +Jun 18 11:43:53.830: INFO: Number of nodes with available pods: 2 +Jun 18 11:43:53.830: INFO: Node 10.72.74.149 is running more than one daemon pod +Jun 18 11:43:54.830: INFO: Number of nodes with available pods: 3 +Jun 18 11:43:54.830: INFO: Number of running nodes: 3, number of available pods: 3 +STEP: Update daemon pods image. +STEP: Check that daemon pods images are updated. +Jun 18 11:43:54.943: INFO: Wrong image for pod: daemon-set-c9vbb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:43:54.943: INFO: Wrong image for pod: daemon-set-dv9hc. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:43:54.943: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:43:55.984: INFO: Wrong image for pod: daemon-set-c9vbb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:43:55.984: INFO: Wrong image for pod: daemon-set-dv9hc. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:43:55.985: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:43:56.974: INFO: Wrong image for pod: daemon-set-c9vbb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:43:56.974: INFO: Wrong image for pod: daemon-set-dv9hc. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:43:56.974: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:43:58.034: INFO: Wrong image for pod: daemon-set-c9vbb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:43:58.034: INFO: Wrong image for pod: daemon-set-dv9hc. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:43:58.034: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:43:58.974: INFO: Wrong image for pod: daemon-set-c9vbb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:43:58.974: INFO: Wrong image for pod: daemon-set-dv9hc. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:43:58.974: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:43:59.984: INFO: Wrong image for pod: daemon-set-c9vbb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:43:59.985: INFO: Wrong image for pod: daemon-set-dv9hc. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:43:59.985: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:00.993: INFO: Wrong image for pod: daemon-set-c9vbb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:00.993: INFO: Wrong image for pod: daemon-set-dv9hc. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:00.993: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:02.285: INFO: Wrong image for pod: daemon-set-c9vbb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:02.285: INFO: Wrong image for pod: daemon-set-dv9hc. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:02.285: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:02.974: INFO: Wrong image for pod: daemon-set-c9vbb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:02.974: INFO: Wrong image for pod: daemon-set-dv9hc. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:02.974: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:03.985: INFO: Wrong image for pod: daemon-set-c9vbb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:03.985: INFO: Wrong image for pod: daemon-set-dv9hc. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:03.985: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:04.972: INFO: Wrong image for pod: daemon-set-c9vbb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:04.972: INFO: Wrong image for pod: daemon-set-dv9hc. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:04.972: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:05.972: INFO: Wrong image for pod: daemon-set-c9vbb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:05.972: INFO: Wrong image for pod: daemon-set-dv9hc. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:05.972: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:06.975: INFO: Wrong image for pod: daemon-set-c9vbb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:06.975: INFO: Wrong image for pod: daemon-set-dv9hc. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:06.975: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:07.972: INFO: Wrong image for pod: daemon-set-c9vbb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:07.972: INFO: Wrong image for pod: daemon-set-dv9hc. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:07.972: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:08.972: INFO: Wrong image for pod: daemon-set-c9vbb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:08.972: INFO: Wrong image for pod: daemon-set-dv9hc. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:08.972: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:09.972: INFO: Wrong image for pod: daemon-set-c9vbb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:09.972: INFO: Wrong image for pod: daemon-set-dv9hc. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:09.972: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:10.972: INFO: Wrong image for pod: daemon-set-c9vbb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:10.972: INFO: Wrong image for pod: daemon-set-dv9hc. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:10.972: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:11.989: INFO: Wrong image for pod: daemon-set-c9vbb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:11.989: INFO: Wrong image for pod: daemon-set-dv9hc. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:11.989: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:12.972: INFO: Wrong image for pod: daemon-set-c9vbb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:12.972: INFO: Wrong image for pod: daemon-set-dv9hc. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:12.972: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:13.973: INFO: Wrong image for pod: daemon-set-c9vbb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:13.973: INFO: Wrong image for pod: daemon-set-dv9hc. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:13.973: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:14.984: INFO: Wrong image for pod: daemon-set-c9vbb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:14.984: INFO: Wrong image for pod: daemon-set-dv9hc. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:14.984: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:15.984: INFO: Wrong image for pod: daemon-set-c9vbb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:15.985: INFO: Wrong image for pod: daemon-set-dv9hc. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:15.985: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:16.975: INFO: Wrong image for pod: daemon-set-c9vbb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:16.975: INFO: Wrong image for pod: daemon-set-dv9hc. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:16.975: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:17.985: INFO: Wrong image for pod: daemon-set-c9vbb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:17.985: INFO: Wrong image for pod: daemon-set-dv9hc. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:17.985: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:18.972: INFO: Wrong image for pod: daemon-set-c9vbb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:18.972: INFO: Wrong image for pod: daemon-set-dv9hc. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:18.972: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:19.985: INFO: Wrong image for pod: daemon-set-c9vbb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:19.985: INFO: Wrong image for pod: daemon-set-dv9hc. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:19.985: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:20.973: INFO: Wrong image for pod: daemon-set-c9vbb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:20.973: INFO: Wrong image for pod: daemon-set-dv9hc. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:20.973: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:21.979: INFO: Wrong image for pod: daemon-set-c9vbb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:21.979: INFO: Wrong image for pod: daemon-set-dv9hc. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:21.979: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:22.972: INFO: Wrong image for pod: daemon-set-c9vbb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:22.972: INFO: Wrong image for pod: daemon-set-dv9hc. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:22.972: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:23.973: INFO: Wrong image for pod: daemon-set-c9vbb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:23.973: INFO: Wrong image for pod: daemon-set-dv9hc. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:23.973: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:24.973: INFO: Wrong image for pod: daemon-set-c9vbb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:24.973: INFO: Wrong image for pod: daemon-set-dv9hc. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:24.973: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:25.973: INFO: Wrong image for pod: daemon-set-c9vbb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:25.973: INFO: Wrong image for pod: daemon-set-dv9hc. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:25.973: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:26.971: INFO: Wrong image for pod: daemon-set-c9vbb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:26.971: INFO: Wrong image for pod: daemon-set-dv9hc. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:26.971: INFO: Pod daemon-set-dv9hc is not available +Jun 18 11:44:26.971: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:27.974: INFO: Wrong image for pod: daemon-set-c9vbb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:27.974: INFO: Wrong image for pod: daemon-set-dv9hc. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:27.974: INFO: Pod daemon-set-dv9hc is not available +Jun 18 11:44:27.974: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:28.972: INFO: Wrong image for pod: daemon-set-c9vbb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:28.972: INFO: Wrong image for pod: daemon-set-dv9hc. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:28.972: INFO: Pod daemon-set-dv9hc is not available +Jun 18 11:44:28.972: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:29.984: INFO: Wrong image for pod: daemon-set-c9vbb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:29.985: INFO: Wrong image for pod: daemon-set-dv9hc. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:29.985: INFO: Pod daemon-set-dv9hc is not available +Jun 18 11:44:29.985: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:30.972: INFO: Wrong image for pod: daemon-set-c9vbb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:30.972: INFO: Wrong image for pod: daemon-set-dv9hc. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:30.972: INFO: Pod daemon-set-dv9hc is not available +Jun 18 11:44:30.972: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:31.985: INFO: Wrong image for pod: daemon-set-c9vbb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:31.985: INFO: Wrong image for pod: daemon-set-dv9hc. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:31.985: INFO: Pod daemon-set-dv9hc is not available +Jun 18 11:44:31.985: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:32.993: INFO: Wrong image for pod: daemon-set-c9vbb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:32.993: INFO: Wrong image for pod: daemon-set-dv9hc. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:32.994: INFO: Pod daemon-set-dv9hc is not available +Jun 18 11:44:32.994: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:33.979: INFO: Wrong image for pod: daemon-set-c9vbb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:33.979: INFO: Wrong image for pod: daemon-set-dv9hc. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:33.979: INFO: Pod daemon-set-dv9hc is not available +Jun 18 11:44:33.979: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:34.972: INFO: Wrong image for pod: daemon-set-c9vbb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:34.972: INFO: Wrong image for pod: daemon-set-dv9hc. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:34.972: INFO: Pod daemon-set-dv9hc is not available +Jun 18 11:44:34.972: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:35.985: INFO: Wrong image for pod: daemon-set-c9vbb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:35.985: INFO: Wrong image for pod: daemon-set-dv9hc. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:35.985: INFO: Pod daemon-set-dv9hc is not available +Jun 18 11:44:35.985: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:36.973: INFO: Wrong image for pod: daemon-set-c9vbb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:36.973: INFO: Pod daemon-set-nq2fm is not available +Jun 18 11:44:36.973: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:37.975: INFO: Wrong image for pod: daemon-set-c9vbb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:37.976: INFO: Pod daemon-set-nq2fm is not available +Jun 18 11:44:37.976: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:38.984: INFO: Wrong image for pod: daemon-set-c9vbb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:38.984: INFO: Pod daemon-set-nq2fm is not available +Jun 18 11:44:38.984: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:39.984: INFO: Wrong image for pod: daemon-set-c9vbb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:39.985: INFO: Pod daemon-set-nq2fm is not available +Jun 18 11:44:39.985: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:40.984: INFO: Wrong image for pod: daemon-set-c9vbb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:40.984: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:41.971: INFO: Wrong image for pod: daemon-set-c9vbb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:41.971: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:42.984: INFO: Wrong image for pod: daemon-set-c9vbb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:42.984: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:43.973: INFO: Wrong image for pod: daemon-set-c9vbb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:43.973: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:44.972: INFO: Wrong image for pod: daemon-set-c9vbb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:44.972: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:45.972: INFO: Wrong image for pod: daemon-set-c9vbb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:45.972: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:46.972: INFO: Wrong image for pod: daemon-set-c9vbb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:46.972: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:47.972: INFO: Wrong image for pod: daemon-set-c9vbb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:47.972: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:48.973: INFO: Wrong image for pod: daemon-set-c9vbb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:48.973: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:49.972: INFO: Wrong image for pod: daemon-set-c9vbb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:49.972: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:50.972: INFO: Wrong image for pod: daemon-set-c9vbb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:50.972: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:51.974: INFO: Wrong image for pod: daemon-set-c9vbb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:51.974: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:52.972: INFO: Wrong image for pod: daemon-set-c9vbb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:52.972: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:53.997: INFO: Wrong image for pod: daemon-set-c9vbb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:53.997: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:54.971: INFO: Wrong image for pod: daemon-set-c9vbb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:54.972: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:55.977: INFO: Wrong image for pod: daemon-set-c9vbb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:55.977: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:56.973: INFO: Wrong image for pod: daemon-set-c9vbb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:56.973: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:57.976: INFO: Wrong image for pod: daemon-set-c9vbb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:57.976: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:58.985: INFO: Wrong image for pod: daemon-set-c9vbb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:58.985: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:59.972: INFO: Wrong image for pod: daemon-set-c9vbb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:44:59.972: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:45:00.972: INFO: Wrong image for pod: daemon-set-c9vbb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:45:00.972: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:45:01.972: INFO: Wrong image for pod: daemon-set-c9vbb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:45:01.972: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:45:02.973: INFO: Wrong image for pod: daemon-set-c9vbb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:45:02.973: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:45:03.984: INFO: Wrong image for pod: daemon-set-c9vbb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:45:03.985: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:45:04.973: INFO: Wrong image for pod: daemon-set-c9vbb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:45:04.974: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:45:06.881: INFO: Wrong image for pod: daemon-set-c9vbb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:45:06.881: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:45:06.972: INFO: Wrong image for pod: daemon-set-c9vbb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:45:06.972: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:45:09.759: INFO: Wrong image for pod: daemon-set-c9vbb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:45:09.759: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:45:09.972: INFO: Wrong image for pod: daemon-set-c9vbb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:45:09.972: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:45:10.984: INFO: Wrong image for pod: daemon-set-c9vbb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:45:10.984: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:45:11.977: INFO: Wrong image for pod: daemon-set-c9vbb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:45:11.977: INFO: Pod daemon-set-c9vbb is not available +Jun 18 11:45:11.977: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:45:12.972: INFO: Wrong image for pod: daemon-set-c9vbb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:45:12.972: INFO: Pod daemon-set-c9vbb is not available +Jun 18 11:45:12.972: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:45:14.154: INFO: Wrong image for pod: daemon-set-c9vbb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:45:14.154: INFO: Pod daemon-set-c9vbb is not available +Jun 18 11:45:14.154: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:45:14.973: INFO: Wrong image for pod: daemon-set-c9vbb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:45:14.973: INFO: Pod daemon-set-c9vbb is not available +Jun 18 11:45:14.973: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:45:15.972: INFO: Wrong image for pod: daemon-set-c9vbb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:45:15.972: INFO: Pod daemon-set-c9vbb is not available +Jun 18 11:45:15.972: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:45:16.973: INFO: Wrong image for pod: daemon-set-c9vbb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:45:16.973: INFO: Pod daemon-set-c9vbb is not available +Jun 18 11:45:16.973: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:45:17.972: INFO: Wrong image for pod: daemon-set-c9vbb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:45:17.972: INFO: Pod daemon-set-c9vbb is not available +Jun 18 11:45:17.972: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:45:18.973: INFO: Wrong image for pod: daemon-set-c9vbb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:45:18.973: INFO: Pod daemon-set-c9vbb is not available +Jun 18 11:45:18.973: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:45:19.973: INFO: Wrong image for pod: daemon-set-c9vbb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:45:19.973: INFO: Pod daemon-set-c9vbb is not available +Jun 18 11:45:19.973: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:45:20.972: INFO: Wrong image for pod: daemon-set-c9vbb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:45:20.972: INFO: Pod daemon-set-c9vbb is not available +Jun 18 11:45:20.972: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:45:21.975: INFO: Pod daemon-set-t752l is not available +Jun 18 11:45:21.975: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:45:22.972: INFO: Pod daemon-set-t752l is not available +Jun 18 11:45:22.972: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:45:23.973: INFO: Pod daemon-set-t752l is not available +Jun 18 11:45:23.973: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:45:25.002: INFO: Pod daemon-set-t752l is not available +Jun 18 11:45:25.002: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:45:25.985: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:45:26.972: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:45:27.973: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:45:28.973: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:45:29.976: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:45:30.973: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:45:31.985: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:45:32.999: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:45:33.974: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:45:34.972: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:45:35.989: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:45:36.976: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:45:37.973: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:45:38.985: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:45:39.972: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:45:40.974: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:45:41.973: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:45:42.984: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:45:43.972: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:45:44.972: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:45:46.183: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:45:46.977: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:45:47.972: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:45:48.973: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:45:49.975: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:45:50.971: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:45:51.974: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:45:52.971: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:45:53.972: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:45:54.972: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:45:55.972: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:45:56.990: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:45:57.972: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:45:57.972: INFO: Pod daemon-set-zpfp6 is not available +Jun 18 11:45:58.984: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:45:58.984: INFO: Pod daemon-set-zpfp6 is not available +Jun 18 11:45:59.972: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:45:59.972: INFO: Pod daemon-set-zpfp6 is not available +Jun 18 11:46:00.973: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:46:00.973: INFO: Pod daemon-set-zpfp6 is not available +Jun 18 11:46:01.973: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:46:01.973: INFO: Pod daemon-set-zpfp6 is not available +Jun 18 11:46:02.973: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:46:02.973: INFO: Pod daemon-set-zpfp6 is not available +Jun 18 11:46:03.973: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:46:03.973: INFO: Pod daemon-set-zpfp6 is not available +Jun 18 11:46:04.973: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:46:04.973: INFO: Pod daemon-set-zpfp6 is not available +Jun 18 11:46:05.973: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:46:05.973: INFO: Pod daemon-set-zpfp6 is not available +Jun 18 11:46:06.972: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:46:06.972: INFO: Pod daemon-set-zpfp6 is not available +Jun 18 11:46:07.992: INFO: Wrong image for pod: daemon-set-zpfp6. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1. +Jun 18 11:46:07.992: INFO: Pod daemon-set-zpfp6 is not available +Jun 18 11:46:08.984: INFO: Pod daemon-set-969gc is not available +STEP: Check that daemon pods are still running on every node of the cluster. +Jun 18 11:46:09.041: INFO: Number of nodes with available pods: 2 +Jun 18 11:46:09.041: INFO: Node 10.72.74.149 is running more than one daemon pod +Jun 18 11:46:10.078: INFO: Number of nodes with available pods: 2 +Jun 18 11:46:10.078: INFO: Node 10.72.74.149 is running more than one daemon pod +Jun 18 11:46:11.078: INFO: Number of nodes with available pods: 2 +Jun 18 11:46:11.078: INFO: Node 10.72.74.149 is running more than one daemon pod +Jun 18 11:46:12.081: INFO: Number of nodes with available pods: 2 +Jun 18 11:46:12.082: INFO: Node 10.72.74.149 is running more than one daemon pod +Jun 18 11:46:13.077: INFO: Number of nodes with available pods: 3 +Jun 18 11:46:13.077: INFO: Number of running nodes: 3, number of available pods: 3 +[AfterEach] [sig-apps] Daemon set [Serial] + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:68 +STEP: Deleting DaemonSet "daemon-set" +STEP: deleting DaemonSet.extensions daemon-set in namespace e2e-tests-daemonsets-sd8qh, will wait for the garbage collector to delete the pods +Jun 18 11:46:13.237: INFO: Deleting DaemonSet.extensions daemon-set took: 27.514298ms +Jun 18 11:46:13.337: INFO: Terminating DaemonSet.extensions daemon-set pods took: 100.212922ms +Jun 18 11:46:26.077: INFO: Number of nodes with available pods: 0 +Jun 18 11:46:26.077: INFO: Number of running nodes: 0, number of available pods: 0 +Jun 18 11:46:26.090: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"selfLink":"/apis/apps/v1/namespaces/e2e-tests-daemonsets-sd8qh/daemonsets","resourceVersion":"90606"},"items":null} + +Jun 18 11:46:26.104: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"selfLink":"/api/v1/namespaces/e2e-tests-daemonsets-sd8qh/pods","resourceVersion":"90606"},"items":null} + +[AfterEach] [sig-apps] Daemon set [Serial] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 18:54:30.816: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-deployment-blp6j" for this suite. -May 13 18:54:36.851: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 18:54:37.119: INFO: namespace: e2e-tests-deployment-blp6j, resource: bindings, ignored listing per whitelist -May 13 18:54:37.122: INFO: namespace e2e-tests-deployment-blp6j deletion completed in 6.297523198s +Jun 18 11:46:26.166: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "e2e-tests-daemonsets-sd8qh" for this suite. +Jun 18 11:46:34.228: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 11:46:34.532: INFO: namespace: e2e-tests-daemonsets-sd8qh, resource: bindings, ignored listing per whitelist +Jun 18 11:46:34.709: INFO: namespace e2e-tests-daemonsets-sd8qh deletion completed in 8.528311597s -• [SLOW TEST:8.759 seconds] -[sig-apps] Deployment +• [SLOW TEST:163.463 seconds] +[sig-apps] Daemon set [Serial] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22 - RecreateDeployment should delete old pods and create new ones [Conformance] + should update pod when spec was updated and update strategy is RollingUpdate [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ -SS +SSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[k8s.io] Docker Containers - should be able to override the image's default arguments (docker cmd) [NodeConformance] [Conformance] +[sig-storage] Projected configMap + should be consumable from pods in volume with defaultMode set [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -[BeforeEach] [k8s.io] Docker Containers +[BeforeEach] [sig-storage] Projected configMap /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client -May 13 18:54:37.122: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -STEP: Building a namespace api object, basename containers -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-containers-mqxbm +Jun 18 11:46:34.712: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 +STEP: Building a namespace api object, basename projected +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-projected-f5gds STEP: Waiting for a default service account to be provisioned in namespace -[It] should be able to override the image's default arguments (docker cmd) [NodeConformance] [Conformance] +[It] should be consumable from pods in volume with defaultMode set [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -STEP: Creating a pod to test override arguments -May 13 18:54:38.250: INFO: Waiting up to 5m0s for pod "client-containers-8ed2395c-75b0-11e9-8f67-2632f168be36" in namespace "e2e-tests-containers-mqxbm" to be "success or failure" -May 13 18:54:38.257: INFO: Pod "client-containers-8ed2395c-75b0-11e9-8f67-2632f168be36": Phase="Pending", Reason="", readiness=false. Elapsed: 7.329415ms -May 13 18:54:40.265: INFO: Pod "client-containers-8ed2395c-75b0-11e9-8f67-2632f168be36": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.015559718s +STEP: Creating configMap with name projected-configmap-test-volume-b9623103-91be-11e9-bce2-ae54e022189f +STEP: Creating a pod to test consume configMaps +Jun 18 11:46:35.196: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-b96447a1-91be-11e9-bce2-ae54e022189f" in namespace "e2e-tests-projected-f5gds" to be "success or failure" +Jun 18 11:46:35.234: INFO: Pod "pod-projected-configmaps-b96447a1-91be-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 37.796569ms +Jun 18 11:46:37.284: INFO: Pod "pod-projected-configmaps-b96447a1-91be-11e9-bce2-ae54e022189f": Phase="Running", Reason="", readiness=true. Elapsed: 2.088452602s +Jun 18 11:46:39.299: INFO: Pod "pod-projected-configmaps-b96447a1-91be-11e9-bce2-ae54e022189f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.103025099s STEP: Saw pod success -May 13 18:54:40.265: INFO: Pod "client-containers-8ed2395c-75b0-11e9-8f67-2632f168be36" satisfied condition "success or failure" -May 13 18:54:40.273: INFO: Trying to get logs from node 10.170.219.153 pod client-containers-8ed2395c-75b0-11e9-8f67-2632f168be36 container test-container: +Jun 18 11:46:39.299: INFO: Pod "pod-projected-configmaps-b96447a1-91be-11e9-bce2-ae54e022189f" satisfied condition "success or failure" +Jun 18 11:46:39.315: INFO: Trying to get logs from node 10.72.74.143 pod pod-projected-configmaps-b96447a1-91be-11e9-bce2-ae54e022189f container projected-configmap-volume-test: STEP: delete the pod -May 13 18:54:40.310: INFO: Waiting for pod client-containers-8ed2395c-75b0-11e9-8f67-2632f168be36 to disappear -May 13 18:54:40.317: INFO: Pod client-containers-8ed2395c-75b0-11e9-8f67-2632f168be36 no longer exists -[AfterEach] [k8s.io] Docker Containers +Jun 18 11:46:39.424: INFO: Waiting for pod pod-projected-configmaps-b96447a1-91be-11e9-bce2-ae54e022189f to disappear +Jun 18 11:46:39.438: INFO: Pod pod-projected-configmaps-b96447a1-91be-11e9-bce2-ae54e022189f no longer exists +[AfterEach] [sig-storage] Projected configMap /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 18:54:40.317: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-containers-mqxbm" for this suite. -May 13 18:54:46.351: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 18:54:46.452: INFO: namespace: e2e-tests-containers-mqxbm, resource: bindings, ignored listing per whitelist -May 13 18:54:46.672: INFO: namespace e2e-tests-containers-mqxbm deletion completed in 6.347751879s +Jun 18 11:46:39.438: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "e2e-tests-projected-f5gds" for this suite. +Jun 18 11:46:45.520: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 11:46:46.157: INFO: namespace: e2e-tests-projected-f5gds, resource: bindings, ignored listing per whitelist +Jun 18 11:46:46.363: INFO: namespace e2e-tests-projected-f5gds deletion completed in 6.900672091s -• [SLOW TEST:9.550 seconds] -[k8s.io] Docker Containers -/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 - should be able to override the image's default arguments (docker cmd) [NodeConformance] [Conformance] +• [SLOW TEST:11.652 seconds] +[sig-storage] Projected configMap +/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_configmap.go:34 + should be consumable from pods in volume with defaultMode set [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ -SSSSSSSSSSS +SS ------------------------------ -[k8s.io] Variable Expansion - should allow substituting values in a container's args [NodeConformance] [Conformance] +[sig-network] Networking Granular Checks: Pods + should function for intra-pod communication: udp [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -[BeforeEach] [k8s.io] Variable Expansion +[BeforeEach] [sig-network] Networking /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client -May 13 18:54:46.672: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -STEP: Building a namespace api object, basename var-expansion -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-var-expansion-x8tm4 +Jun 18 11:46:46.363: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 +STEP: Building a namespace api object, basename pod-network-test +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-pod-network-test-cxzg9 STEP: Waiting for a default service account to be provisioned in namespace -[It] should allow substituting values in a container's args [NodeConformance] [Conformance] +[It] should function for intra-pod communication: udp [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -STEP: Creating a pod to test substitution in container's args -May 13 18:54:46.962: INFO: Waiting up to 5m0s for pod "var-expansion-94036e08-75b0-11e9-8f67-2632f168be36" in namespace "e2e-tests-var-expansion-x8tm4" to be "success or failure" -May 13 18:54:46.972: INFO: Pod "var-expansion-94036e08-75b0-11e9-8f67-2632f168be36": Phase="Pending", Reason="", readiness=false. Elapsed: 9.867252ms -May 13 18:54:48.980: INFO: Pod "var-expansion-94036e08-75b0-11e9-8f67-2632f168be36": Phase="Running", Reason="", readiness=true. Elapsed: 2.017912338s -May 13 18:54:50.988: INFO: Pod "var-expansion-94036e08-75b0-11e9-8f67-2632f168be36": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.02647884s -STEP: Saw pod success -May 13 18:54:50.988: INFO: Pod "var-expansion-94036e08-75b0-11e9-8f67-2632f168be36" satisfied condition "success or failure" -May 13 18:54:50.996: INFO: Trying to get logs from node 10.170.219.177 pod var-expansion-94036e08-75b0-11e9-8f67-2632f168be36 container dapi-container: -STEP: delete the pod -May 13 18:54:51.037: INFO: Waiting for pod var-expansion-94036e08-75b0-11e9-8f67-2632f168be36 to disappear -May 13 18:54:51.056: INFO: Pod var-expansion-94036e08-75b0-11e9-8f67-2632f168be36 no longer exists -[AfterEach] [k8s.io] Variable Expansion +STEP: Performing setup for networking test in namespace e2e-tests-pod-network-test-cxzg9 +STEP: creating a selector +STEP: Creating the service pods in kubernetes +Jun 18 11:46:46.891: INFO: Waiting up to 10m0s for all (but 0) nodes to be schedulable +STEP: Creating test pods +Jun 18 11:47:13.214: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s 'http://172.30.114.10:8080/dial?request=hostName&protocol=udp&host=172.30.39.16&port=8081&tries=1'] Namespace:e2e-tests-pod-network-test-cxzg9 PodName:host-test-container-pod ContainerName:hostexec Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} +Jun 18 11:47:13.214: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 +Jun 18 11:47:13.501: INFO: Waiting for endpoints: map[] +Jun 18 11:47:13.517: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s 'http://172.30.114.10:8080/dial?request=hostName&protocol=udp&host=172.30.58.150&port=8081&tries=1'] Namespace:e2e-tests-pod-network-test-cxzg9 PodName:host-test-container-pod ContainerName:hostexec Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} +Jun 18 11:47:13.517: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 +Jun 18 11:47:13.734: INFO: Waiting for endpoints: map[] +Jun 18 11:47:13.748: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s 'http://172.30.114.10:8080/dial?request=hostName&protocol=udp&host=172.30.114.9&port=8081&tries=1'] Namespace:e2e-tests-pod-network-test-cxzg9 PodName:host-test-container-pod ContainerName:hostexec Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} +Jun 18 11:47:13.748: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 +Jun 18 11:47:13.984: INFO: Waiting for endpoints: map[] +[AfterEach] [sig-network] Networking /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 18:54:51.056: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-var-expansion-x8tm4" for this suite. -May 13 18:54:57.092: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 18:54:57.171: INFO: namespace: e2e-tests-var-expansion-x8tm4, resource: bindings, ignored listing per whitelist -May 13 18:54:57.460: INFO: namespace e2e-tests-var-expansion-x8tm4 deletion completed in 6.396015784s +Jun 18 11:47:13.984: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "e2e-tests-pod-network-test-cxzg9" for this suite. +Jun 18 11:47:38.053: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 11:47:38.372: INFO: namespace: e2e-tests-pod-network-test-cxzg9, resource: bindings, ignored listing per whitelist +Jun 18 11:47:38.549: INFO: namespace e2e-tests-pod-network-test-cxzg9 deletion completed in 24.542866067s -• [SLOW TEST:10.788 seconds] -[k8s.io] Variable Expansion -/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 - should allow substituting values in a container's args [NodeConformance] [Conformance] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 +• [SLOW TEST:52.186 seconds] +[sig-network] Networking +/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:25 + Granular Checks: Pods + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:28 + should function for intra-pod communication: udp [NodeConformance] [Conformance] + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ -SSSSSSSSSSSSSSSSSSS +SSS ------------------------------ -[sig-storage] Projected downwardAPI - should set mode on item file [NodeConformance] [Conformance] +[sig-storage] Projected secret + should be consumable in multiple volumes in a pod [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -[BeforeEach] [sig-storage] Projected downwardAPI +[BeforeEach] [sig-storage] Projected secret /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client -May 13 18:54:57.460: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 +Jun 18 11:47:38.549: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 STEP: Building a namespace api object, basename projected -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-projected-vtsw7 +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-projected-pw2p9 STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [sig-storage] Projected downwardAPI - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:39 -[It] should set mode on item file [NodeConformance] [Conformance] +[It] should be consumable in multiple volumes in a pod [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -STEP: Creating a pod to test downward API volume plugin -May 13 18:54:57.748: INFO: Waiting up to 5m0s for pod "downwardapi-volume-9a715bda-75b0-11e9-8f67-2632f168be36" in namespace "e2e-tests-projected-vtsw7" to be "success or failure" -May 13 18:54:57.756: INFO: Pod "downwardapi-volume-9a715bda-75b0-11e9-8f67-2632f168be36": Phase="Pending", Reason="", readiness=false. Elapsed: 8.026421ms -May 13 18:54:59.767: INFO: Pod "downwardapi-volume-9a715bda-75b0-11e9-8f67-2632f168be36": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.018563247s +STEP: Creating secret with name projected-secret-test-df70f185-91be-11e9-bce2-ae54e022189f +STEP: Creating a pod to test consume secrets +Jun 18 11:47:39.044: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-df7325e7-91be-11e9-bce2-ae54e022189f" in namespace "e2e-tests-projected-pw2p9" to be "success or failure" +Jun 18 11:47:39.058: INFO: Pod "pod-projected-secrets-df7325e7-91be-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 13.543801ms +Jun 18 11:47:41.074: INFO: Pod "pod-projected-secrets-df7325e7-91be-11e9-bce2-ae54e022189f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.030086259s STEP: Saw pod success -May 13 18:54:59.767: INFO: Pod "downwardapi-volume-9a715bda-75b0-11e9-8f67-2632f168be36" satisfied condition "success or failure" -May 13 18:54:59.775: INFO: Trying to get logs from node 10.170.219.153 pod downwardapi-volume-9a715bda-75b0-11e9-8f67-2632f168be36 container client-container: +Jun 18 11:47:41.074: INFO: Pod "pod-projected-secrets-df7325e7-91be-11e9-bce2-ae54e022189f" satisfied condition "success or failure" +Jun 18 11:47:41.089: INFO: Trying to get logs from node 10.72.74.143 pod pod-projected-secrets-df7325e7-91be-11e9-bce2-ae54e022189f container secret-volume-test: STEP: delete the pod -May 13 18:54:59.858: INFO: Waiting for pod downwardapi-volume-9a715bda-75b0-11e9-8f67-2632f168be36 to disappear -May 13 18:54:59.869: INFO: Pod downwardapi-volume-9a715bda-75b0-11e9-8f67-2632f168be36 no longer exists -[AfterEach] [sig-storage] Projected downwardAPI +Jun 18 11:47:41.160: INFO: Waiting for pod pod-projected-secrets-df7325e7-91be-11e9-bce2-ae54e022189f to disappear +Jun 18 11:47:41.175: INFO: Pod pod-projected-secrets-df7325e7-91be-11e9-bce2-ae54e022189f no longer exists +[AfterEach] [sig-storage] Projected secret /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 18:54:59.869: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-projected-vtsw7" for this suite. -May 13 18:55:05.906: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 18:55:06.145: INFO: namespace: e2e-tests-projected-vtsw7, resource: bindings, ignored listing per whitelist -May 13 18:55:06.150: INFO: namespace e2e-tests-projected-vtsw7 deletion completed in 6.272991958s +Jun 18 11:47:41.175: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "e2e-tests-projected-pw2p9" for this suite. +Jun 18 11:47:49.250: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 11:47:49.399: INFO: namespace: e2e-tests-projected-pw2p9, resource: bindings, ignored listing per whitelist +Jun 18 11:47:49.757: INFO: namespace e2e-tests-projected-pw2p9 deletion completed in 8.557412971s -• [SLOW TEST:8.690 seconds] -[sig-storage] Projected downwardAPI -/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:33 - should set mode on item file [NodeConformance] [Conformance] +• [SLOW TEST:11.208 seconds] +[sig-storage] Projected secret +/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_secret.go:34 + should be consumable in multiple volumes in a pod [NodeConformance] [Conformance] + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 +------------------------------ +SSSSSSSSSSSSSSSSSSSSSSSSSSSS +------------------------------ +[sig-api-machinery] Namespaces [Serial] + should ensure that all services are removed when a namespace is deleted [Conformance] + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 +[BeforeEach] [sig-api-machinery] Namespaces [Serial] + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 +STEP: Creating a kubernetes client +Jun 18 11:47:49.757: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 +STEP: Building a namespace api object, basename namespaces +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-namespaces-cddjl +STEP: Waiting for a default service account to be provisioned in namespace +[It] should ensure that all services are removed when a namespace is deleted [Conformance] + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 +STEP: Creating a test namespace +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-nsdeletetest-942rq +STEP: Waiting for a default service account to be provisioned in namespace +STEP: Creating a service in the namespace +STEP: Deleting the namespace +STEP: Waiting for the namespace to be removed. +STEP: Recreating the namespace +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-nsdeletetest-856jq +STEP: Verifying there is no service in the namespace +[AfterEach] [sig-api-machinery] Namespaces [Serial] + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 +Jun 18 11:47:57.924: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "e2e-tests-namespaces-cddjl" for this suite. +Jun 18 11:48:04.018: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 11:48:04.677: INFO: namespace: e2e-tests-namespaces-cddjl, resource: bindings, ignored listing per whitelist +Jun 18 11:48:04.715: INFO: namespace e2e-tests-namespaces-cddjl deletion completed in 6.772836112s +STEP: Destroying namespace "e2e-tests-nsdeletetest-942rq" for this suite. +Jun 18 11:48:04.731: INFO: Namespace e2e-tests-nsdeletetest-942rq was already deleted +STEP: Destroying namespace "e2e-tests-nsdeletetest-856jq" for this suite. +Jun 18 11:48:10.780: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 11:48:11.007: INFO: namespace: e2e-tests-nsdeletetest-856jq, resource: bindings, ignored listing per whitelist +Jun 18 11:48:11.597: INFO: namespace e2e-tests-nsdeletetest-856jq deletion completed in 6.866103474s + +• [SLOW TEST:21.840 seconds] +[sig-api-machinery] Namespaces [Serial] +/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:22 + should ensure that all services are removed when a namespace is deleted [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ -S +SSSSS ------------------------------ -[sig-storage] Projected downwardAPI - should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance] +[sig-apps] Deployment + deployment should delete old replica sets [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -[BeforeEach] [sig-storage] Projected downwardAPI +[BeforeEach] [sig-apps] Deployment /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client -May 13 18:55:06.150: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -STEP: Building a namespace api object, basename projected -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-projected-b7v4m +Jun 18 11:48:11.597: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 +STEP: Building a namespace api object, basename deployment +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-deployment-fnr4d STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [sig-storage] Projected downwardAPI - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:39 -[It] should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance] +[BeforeEach] [sig-apps] Deployment + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:65 +[It] deployment should delete old replica sets [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -STEP: Creating a pod to test downward API volume plugin -May 13 18:55:06.427: INFO: Waiting up to 5m0s for pod "downwardapi-volume-9f9daf81-75b0-11e9-8f67-2632f168be36" in namespace "e2e-tests-projected-b7v4m" to be "success or failure" -May 13 18:55:06.434: INFO: Pod "downwardapi-volume-9f9daf81-75b0-11e9-8f67-2632f168be36": Phase="Pending", Reason="", readiness=false. Elapsed: 7.034518ms -May 13 18:55:08.443: INFO: Pod "downwardapi-volume-9f9daf81-75b0-11e9-8f67-2632f168be36": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.016118363s -STEP: Saw pod success -May 13 18:55:08.443: INFO: Pod "downwardapi-volume-9f9daf81-75b0-11e9-8f67-2632f168be36" satisfied condition "success or failure" -May 13 18:55:08.452: INFO: Trying to get logs from node 10.170.219.153 pod downwardapi-volume-9f9daf81-75b0-11e9-8f67-2632f168be36 container client-container: -STEP: delete the pod -May 13 18:55:08.492: INFO: Waiting for pod downwardapi-volume-9f9daf81-75b0-11e9-8f67-2632f168be36 to disappear -May 13 18:55:08.499: INFO: Pod downwardapi-volume-9f9daf81-75b0-11e9-8f67-2632f168be36 no longer exists -[AfterEach] [sig-storage] Projected downwardAPI +Jun 18 11:48:12.118: INFO: Pod name cleanup-pod: Found 0 pods out of 1 +Jun 18 11:48:17.153: INFO: Pod name cleanup-pod: Found 1 pods out of 1 +STEP: ensuring each pod is running +Jun 18 11:48:17.153: INFO: Creating deployment test-cleanup-deployment +STEP: Waiting for deployment test-cleanup-deployment history to be cleaned up +[AfterEach] [sig-apps] Deployment + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:59 +Jun 18 11:48:17.232: INFO: Deployment "test-cleanup-deployment": +&Deployment{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-cleanup-deployment,GenerateName:,Namespace:e2e-tests-deployment-fnr4d,SelfLink:/apis/apps/v1/namespaces/e2e-tests-deployment-fnr4d/deployments/test-cleanup-deployment,UID:f63253cd-91be-11e9-bf44-fa6f350b29f0,ResourceVersion:91136,Generation:1,CreationTimestamp:2019-06-18 11:48:17 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: cleanup-pod,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Spec:DeploymentSpec{Replicas:*1,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: cleanup-pod,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: cleanup-pod,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[],Containers:[{redis gcr.io/kubernetes-e2e-test-images/redis:1.0 [] [] [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,},},Strategy:DeploymentStrategy{Type:RollingUpdate,RollingUpdate:&RollingUpdateDeployment{MaxUnavailable:25%!,(MISSING)MaxSurge:25%!,(MISSING)},},MinReadySeconds:0,RevisionHistoryLimit:*0,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:0,Replicas:0,UpdatedReplicas:0,AvailableReplicas:0,UnavailableReplicas:0,Conditions:[],ReadyReplicas:0,CollisionCount:nil,},} + +Jun 18 11:48:17.246: INFO: New ReplicaSet of Deployment "test-cleanup-deployment" is nil. +Jun 18 11:48:17.246: INFO: All old ReplicaSets of Deployment "test-cleanup-deployment": +Jun 18 11:48:17.246: INFO: &ReplicaSet{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-cleanup-controller,GenerateName:,Namespace:e2e-tests-deployment-fnr4d,SelfLink:/apis/apps/v1/namespaces/e2e-tests-deployment-fnr4d/replicasets/test-cleanup-controller,UID:f329e69f-91be-11e9-bf44-fa6f350b29f0,ResourceVersion:91137,Generation:1,CreationTimestamp:2019-06-18 11:48:12 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: cleanup-pod,pod: nginx,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 Deployment test-cleanup-deployment f63253cd-91be-11e9-bf44-fa6f350b29f0 0xc0017351f7 0xc0017351f8}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:ReplicaSetSpec{Replicas:*1,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: cleanup-pod,pod: nginx,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: cleanup-pod,pod: nginx,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,},},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:1,FullyLabeledReplicas:1,ObservedGeneration:1,ReadyReplicas:1,AvailableReplicas:1,Conditions:[],},} +Jun 18 11:48:17.263: INFO: Pod "test-cleanup-controller-pslsj" is available: +&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-cleanup-controller-pslsj,GenerateName:test-cleanup-controller-,Namespace:e2e-tests-deployment-fnr4d,SelfLink:/api/v1/namespaces/e2e-tests-deployment-fnr4d/pods/test-cleanup-controller-pslsj,UID:f32fda23-91be-11e9-bf44-fa6f350b29f0,ResourceVersion:91132,Generation:0,CreationTimestamp:2019-06-18 11:48:12 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: cleanup-pod,pod: nginx,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet test-cleanup-controller f329e69f-91be-11e9-bf44-fa6f350b29f0 0xc0021a4587 0xc0021a4588}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-kmqbr {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-kmqbr,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-kmqbr true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.72.74.144,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc0021a4600} {node.kubernetes.io/unreachable Exists NoExecute 0xc0021a4620}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Running,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:48:12 +0000 UTC } {Ready True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:48:14 +0000 UTC } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:48:14 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:48:12 +0000 UTC }],Message:,Reason:,HostIP:10.72.74.144,PodIP:172.30.114.11,StartTime:2019-06-18 11:48:12 +0000 UTC,ContainerStatuses:[{nginx {nil ContainerStateRunning{StartedAt:2019-06-18 11:48:13 +0000 UTC,} nil} {nil nil nil} true 0 docker.io/library/nginx:1.14-alpine docker.io/library/nginx@sha256:485b610fefec7ff6c463ced9623314a04ed67e3945b9c08d7e53a47f6d108dc7 containerd://6fda0b030ff10f427b12682f8bf2ca5a738d8c9e4a6d8482ca92c59f4486b470}],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} +[AfterEach] [sig-apps] Deployment /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 18:55:08.499: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-projected-b7v4m" for this suite. -May 13 18:55:14.548: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 18:55:14.660: INFO: namespace: e2e-tests-projected-b7v4m, resource: bindings, ignored listing per whitelist -May 13 18:55:14.877: INFO: namespace e2e-tests-projected-b7v4m deletion completed in 6.369425s +Jun 18 11:48:17.263: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "e2e-tests-deployment-fnr4d" for this suite. +Jun 18 11:48:23.352: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 11:48:23.561: INFO: namespace: e2e-tests-deployment-fnr4d, resource: bindings, ignored listing per whitelist +Jun 18 11:48:24.055: INFO: namespace e2e-tests-deployment-fnr4d deletion completed in 6.769347211s -• [SLOW TEST:8.727 seconds] -[sig-storage] Projected downwardAPI -/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:33 - should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance] +• [SLOW TEST:12.458 seconds] +[sig-apps] Deployment +/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22 + deployment should delete old replica sets [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ -[sig-storage] Projected configMap - should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance] +SSSSSSSSSSSSSSSSSS +------------------------------ +[sig-api-machinery] Watchers + should observe add, update, and delete watch notifications on configmaps [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -[BeforeEach] [sig-storage] Projected configMap +[BeforeEach] [sig-api-machinery] Watchers /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client -May 13 18:55:14.877: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -STEP: Building a namespace api object, basename projected -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-projected-d68wz +Jun 18 11:48:24.055: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 +STEP: Building a namespace api object, basename watch +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-watch-qbwmb STEP: Waiting for a default service account to be provisioned in namespace -[It] should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance] +[It] should observe add, update, and delete watch notifications on configmaps [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -STEP: Creating configMap with name projected-configmap-test-volume-a4d285cc-75b0-11e9-8f67-2632f168be36 -STEP: Creating a pod to test consume configMaps -May 13 18:55:15.170: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-a4d3da3a-75b0-11e9-8f67-2632f168be36" in namespace "e2e-tests-projected-d68wz" to be "success or failure" -May 13 18:55:15.177: INFO: Pod "pod-projected-configmaps-a4d3da3a-75b0-11e9-8f67-2632f168be36": Phase="Pending", Reason="", readiness=false. Elapsed: 7.448996ms -May 13 18:55:17.185: INFO: Pod "pod-projected-configmaps-a4d3da3a-75b0-11e9-8f67-2632f168be36": Phase="Running", Reason="", readiness=true. Elapsed: 2.015376699s -May 13 18:55:19.194: INFO: Pod "pod-projected-configmaps-a4d3da3a-75b0-11e9-8f67-2632f168be36": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.023993986s -STEP: Saw pod success -May 13 18:55:19.194: INFO: Pod "pod-projected-configmaps-a4d3da3a-75b0-11e9-8f67-2632f168be36" satisfied condition "success or failure" -May 13 18:55:19.203: INFO: Trying to get logs from node 10.170.219.153 pod pod-projected-configmaps-a4d3da3a-75b0-11e9-8f67-2632f168be36 container projected-configmap-volume-test: -STEP: delete the pod -May 13 18:55:19.241: INFO: Waiting for pod pod-projected-configmaps-a4d3da3a-75b0-11e9-8f67-2632f168be36 to disappear -May 13 18:55:19.320: INFO: Pod pod-projected-configmaps-a4d3da3a-75b0-11e9-8f67-2632f168be36 no longer exists -[AfterEach] [sig-storage] Projected configMap +STEP: creating a watch on configmaps with label A +STEP: creating a watch on configmaps with label B +STEP: creating a watch on configmaps with label A or B +STEP: creating a configmap with label A and ensuring the correct watchers observe the notification +Jun 18 11:48:24.545: INFO: Got : ADDED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-configmap-a,GenerateName:,Namespace:e2e-tests-watch-qbwmb,SelfLink:/api/v1/namespaces/e2e-tests-watch-qbwmb/configmaps/e2e-watch-test-configmap-a,UID:fa94e668-91be-11e9-bf44-fa6f350b29f0,ResourceVersion:91209,Generation:0,CreationTimestamp:2019-06-18 11:48:24 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: multiple-watchers-A,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Data:map[string]string{},BinaryData:map[string][]byte{},} +Jun 18 11:48:24.545: INFO: Got : ADDED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-configmap-a,GenerateName:,Namespace:e2e-tests-watch-qbwmb,SelfLink:/api/v1/namespaces/e2e-tests-watch-qbwmb/configmaps/e2e-watch-test-configmap-a,UID:fa94e668-91be-11e9-bf44-fa6f350b29f0,ResourceVersion:91209,Generation:0,CreationTimestamp:2019-06-18 11:48:24 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: multiple-watchers-A,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Data:map[string]string{},BinaryData:map[string][]byte{},} +STEP: modifying configmap A and ensuring the correct watchers observe the notification +Jun 18 11:48:34.590: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-configmap-a,GenerateName:,Namespace:e2e-tests-watch-qbwmb,SelfLink:/api/v1/namespaces/e2e-tests-watch-qbwmb/configmaps/e2e-watch-test-configmap-a,UID:fa94e668-91be-11e9-bf44-fa6f350b29f0,ResourceVersion:91226,Generation:0,CreationTimestamp:2019-06-18 11:48:24 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: multiple-watchers-A,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Data:map[string]string{mutation: 1,},BinaryData:map[string][]byte{},} +Jun 18 11:48:34.591: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-configmap-a,GenerateName:,Namespace:e2e-tests-watch-qbwmb,SelfLink:/api/v1/namespaces/e2e-tests-watch-qbwmb/configmaps/e2e-watch-test-configmap-a,UID:fa94e668-91be-11e9-bf44-fa6f350b29f0,ResourceVersion:91226,Generation:0,CreationTimestamp:2019-06-18 11:48:24 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: multiple-watchers-A,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Data:map[string]string{mutation: 1,},BinaryData:map[string][]byte{},} +STEP: modifying configmap A again and ensuring the correct watchers observe the notification +Jun 18 11:48:44.700: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-configmap-a,GenerateName:,Namespace:e2e-tests-watch-qbwmb,SelfLink:/api/v1/namespaces/e2e-tests-watch-qbwmb/configmaps/e2e-watch-test-configmap-a,UID:fa94e668-91be-11e9-bf44-fa6f350b29f0,ResourceVersion:91244,Generation:0,CreationTimestamp:2019-06-18 11:48:24 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: multiple-watchers-A,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},} +Jun 18 11:48:44.700: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-configmap-a,GenerateName:,Namespace:e2e-tests-watch-qbwmb,SelfLink:/api/v1/namespaces/e2e-tests-watch-qbwmb/configmaps/e2e-watch-test-configmap-a,UID:fa94e668-91be-11e9-bf44-fa6f350b29f0,ResourceVersion:91244,Generation:0,CreationTimestamp:2019-06-18 11:48:24 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: multiple-watchers-A,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},} +STEP: deleting configmap A and ensuring the correct watchers observe the notification +Jun 18 11:48:54.743: INFO: Got : DELETED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-configmap-a,GenerateName:,Namespace:e2e-tests-watch-qbwmb,SelfLink:/api/v1/namespaces/e2e-tests-watch-qbwmb/configmaps/e2e-watch-test-configmap-a,UID:fa94e668-91be-11e9-bf44-fa6f350b29f0,ResourceVersion:91261,Generation:0,CreationTimestamp:2019-06-18 11:48:24 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: multiple-watchers-A,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},} +Jun 18 11:48:54.743: INFO: Got : DELETED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-configmap-a,GenerateName:,Namespace:e2e-tests-watch-qbwmb,SelfLink:/api/v1/namespaces/e2e-tests-watch-qbwmb/configmaps/e2e-watch-test-configmap-a,UID:fa94e668-91be-11e9-bf44-fa6f350b29f0,ResourceVersion:91261,Generation:0,CreationTimestamp:2019-06-18 11:48:24 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: multiple-watchers-A,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},} +STEP: creating a configmap with label B and ensuring the correct watchers observe the notification +Jun 18 11:49:05.117: INFO: Got : ADDED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-configmap-b,GenerateName:,Namespace:e2e-tests-watch-qbwmb,SelfLink:/api/v1/namespaces/e2e-tests-watch-qbwmb/configmaps/e2e-watch-test-configmap-b,UID:12c23309-91bf-11e9-bf44-fa6f350b29f0,ResourceVersion:91279,Generation:0,CreationTimestamp:2019-06-18 11:49:05 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: multiple-watchers-B,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Data:map[string]string{},BinaryData:map[string][]byte{},} +Jun 18 11:49:05.117: INFO: Got : ADDED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-configmap-b,GenerateName:,Namespace:e2e-tests-watch-qbwmb,SelfLink:/api/v1/namespaces/e2e-tests-watch-qbwmb/configmaps/e2e-watch-test-configmap-b,UID:12c23309-91bf-11e9-bf44-fa6f350b29f0,ResourceVersion:91279,Generation:0,CreationTimestamp:2019-06-18 11:49:05 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: multiple-watchers-B,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Data:map[string]string{},BinaryData:map[string][]byte{},} +STEP: deleting configmap B and ensuring the correct watchers observe the notification +Jun 18 11:49:15.351: INFO: Got : DELETED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-configmap-b,GenerateName:,Namespace:e2e-tests-watch-qbwmb,SelfLink:/api/v1/namespaces/e2e-tests-watch-qbwmb/configmaps/e2e-watch-test-configmap-b,UID:12c23309-91bf-11e9-bf44-fa6f350b29f0,ResourceVersion:91297,Generation:0,CreationTimestamp:2019-06-18 11:49:05 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: multiple-watchers-B,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Data:map[string]string{},BinaryData:map[string][]byte{},} +Jun 18 11:49:15.351: INFO: Got : DELETED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-configmap-b,GenerateName:,Namespace:e2e-tests-watch-qbwmb,SelfLink:/api/v1/namespaces/e2e-tests-watch-qbwmb/configmaps/e2e-watch-test-configmap-b,UID:12c23309-91bf-11e9-bf44-fa6f350b29f0,ResourceVersion:91297,Generation:0,CreationTimestamp:2019-06-18 11:49:05 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: multiple-watchers-B,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Data:map[string]string{},BinaryData:map[string][]byte{},} +[AfterEach] [sig-api-machinery] Watchers /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 18:55:19.320: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-projected-d68wz" for this suite. -May 13 18:55:25.368: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 18:55:25.486: INFO: namespace: e2e-tests-projected-d68wz, resource: bindings, ignored listing per whitelist -May 13 18:55:25.644: INFO: namespace e2e-tests-projected-d68wz deletion completed in 6.31621283s +Jun 18 11:49:25.352: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "e2e-tests-watch-qbwmb" for this suite. +Jun 18 11:49:31.436: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 11:49:32.722: INFO: namespace: e2e-tests-watch-qbwmb, resource: bindings, ignored listing per whitelist +Jun 18 11:49:33.099: INFO: namespace e2e-tests-watch-qbwmb deletion completed in 7.711628267s -• [SLOW TEST:10.767 seconds] -[sig-storage] Projected configMap -/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_configmap.go:34 - should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance] +• [SLOW TEST:69.044 seconds] +[sig-api-machinery] Watchers +/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:22 + should observe add, update, and delete watch notifications on configmaps [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSSSSS ------------------------------ -[sig-network] Networking Granular Checks: Pods - should function for intra-pod communication: udp [NodeConformance] [Conformance] +[sig-storage] Projected configMap + optional updates should be reflected in volume [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -[BeforeEach] [sig-network] Networking +[BeforeEach] [sig-storage] Projected configMap /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client -May 13 18:55:25.644: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -STEP: Building a namespace api object, basename pod-network-test -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-pod-network-test-kt5mg +Jun 18 11:49:33.099: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 +STEP: Building a namespace api object, basename projected +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-projected-mdjpd STEP: Waiting for a default service account to be provisioned in namespace -[It] should function for intra-pod communication: udp [NodeConformance] [Conformance] +[It] optional updates should be reflected in volume [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -STEP: Performing setup for networking test in namespace e2e-tests-pod-network-test-kt5mg -STEP: creating a selector -STEP: Creating the service pods in kubernetes -May 13 18:55:25.916: INFO: Waiting up to 10m0s for all (but 0) nodes to be schedulable -STEP: Creating test pods -May 13 18:55:42.141: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s 'http://172.30.227.101:8080/dial?request=hostName&protocol=udp&host=172.30.19.82&port=8081&tries=1'] Namespace:e2e-tests-pod-network-test-kt5mg PodName:host-test-container-pod ContainerName:hostexec Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} -May 13 18:55:42.141: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -May 13 18:55:42.329: INFO: Waiting for endpoints: map[] -May 13 18:55:42.337: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s 'http://172.30.227.101:8080/dial?request=hostName&protocol=udp&host=172.30.227.100&port=8081&tries=1'] Namespace:e2e-tests-pod-network-test-kt5mg PodName:host-test-container-pod ContainerName:hostexec Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} -May 13 18:55:42.337: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -May 13 18:55:42.523: INFO: Waiting for endpoints: map[] -May 13 18:55:42.531: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s 'http://172.30.227.101:8080/dial?request=hostName&protocol=udp&host=172.30.63.63&port=8081&tries=1'] Namespace:e2e-tests-pod-network-test-kt5mg PodName:host-test-container-pod ContainerName:hostexec Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} -May 13 18:55:42.531: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -May 13 18:55:42.686: INFO: Waiting for endpoints: map[] -[AfterEach] [sig-network] Networking +STEP: Creating configMap with name cm-test-opt-del-23bca456-91bf-11e9-bce2-ae54e022189f +STEP: Creating configMap with name cm-test-opt-upd-23bca4aa-91bf-11e9-bce2-ae54e022189f +STEP: Creating the pod +STEP: Deleting configmap cm-test-opt-del-23bca456-91bf-11e9-bce2-ae54e022189f +STEP: Updating configmap cm-test-opt-upd-23bca4aa-91bf-11e9-bce2-ae54e022189f +STEP: Creating configMap with name cm-test-opt-create-23bca4d7-91bf-11e9-bce2-ae54e022189f +STEP: waiting to observe update in volume +[AfterEach] [sig-storage] Projected configMap /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 18:55:42.686: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-pod-network-test-kt5mg" for this suite. -May 13 18:56:06.723: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 18:56:07.133: INFO: namespace: e2e-tests-pod-network-test-kt5mg, resource: bindings, ignored listing per whitelist -May 13 18:56:07.254: INFO: namespace e2e-tests-pod-network-test-kt5mg deletion completed in 24.559613399s +Jun 18 11:51:02.255: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "e2e-tests-projected-mdjpd" for this suite. +Jun 18 11:51:26.327: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 11:51:26.666: INFO: namespace: e2e-tests-projected-mdjpd, resource: bindings, ignored listing per whitelist +Jun 18 11:51:26.838: INFO: namespace e2e-tests-projected-mdjpd deletion completed in 24.561696136s -• [SLOW TEST:41.610 seconds] -[sig-network] Networking -/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:25 - Granular Checks: Pods - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:28 - should function for intra-pod communication: udp [NodeConformance] [Conformance] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 +• [SLOW TEST:113.739 seconds] +[sig-storage] Projected configMap +/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_configmap.go:34 + optional updates should be reflected in volume [NodeConformance] [Conformance] + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ -SSSS +SSSSSSSSSSSSSSS ------------------------------ [sig-storage] EmptyDir volumes - should support (non-root,0777,default) [NodeConformance] [Conformance] + should support (root,0644,tmpfs) [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-storage] EmptyDir volumes /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client -May 13 18:56:07.255: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 +Jun 18 11:51:26.839: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 STEP: Building a namespace api object, basename emptydir -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-emptydir-rpsd6 +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-emptydir-j4br4 STEP: Waiting for a default service account to be provisioned in namespace -[It] should support (non-root,0777,default) [NodeConformance] [Conformance] +[It] should support (root,0644,tmpfs) [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -STEP: Creating a pod to test emptydir 0777 on node default medium -May 13 18:56:07.628: INFO: Waiting up to 5m0s for pod "pod-c41817e4-75b0-11e9-8f67-2632f168be36" in namespace "e2e-tests-emptydir-rpsd6" to be "success or failure" -May 13 18:56:07.635: INFO: Pod "pod-c41817e4-75b0-11e9-8f67-2632f168be36": Phase="Pending", Reason="", readiness=false. Elapsed: 7.479152ms -May 13 18:56:09.657: INFO: Pod "pod-c41817e4-75b0-11e9-8f67-2632f168be36": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.02936211s +STEP: Creating a pod to test emptydir 0644 on tmpfs +Jun 18 11:51:27.516: INFO: Waiting up to 5m0s for pod "pod-67a1763c-91bf-11e9-bce2-ae54e022189f" in namespace "e2e-tests-emptydir-j4br4" to be "success or failure" +Jun 18 11:51:27.531: INFO: Pod "pod-67a1763c-91bf-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 14.883419ms +Jun 18 11:51:29.547: INFO: Pod "pod-67a1763c-91bf-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 2.0308532s +Jun 18 11:51:31.562: INFO: Pod "pod-67a1763c-91bf-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 4.045722555s +Jun 18 11:51:33.595: INFO: Pod "pod-67a1763c-91bf-11e9-bce2-ae54e022189f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.079131073s STEP: Saw pod success -May 13 18:56:09.657: INFO: Pod "pod-c41817e4-75b0-11e9-8f67-2632f168be36" satisfied condition "success or failure" -May 13 18:56:09.665: INFO: Trying to get logs from node 10.170.219.177 pod pod-c41817e4-75b0-11e9-8f67-2632f168be36 container test-container: +Jun 18 11:51:33.595: INFO: Pod "pod-67a1763c-91bf-11e9-bce2-ae54e022189f" satisfied condition "success or failure" +Jun 18 11:51:33.610: INFO: Trying to get logs from node 10.72.74.144 pod pod-67a1763c-91bf-11e9-bce2-ae54e022189f container test-container: STEP: delete the pod -May 13 18:56:09.703: INFO: Waiting for pod pod-c41817e4-75b0-11e9-8f67-2632f168be36 to disappear -May 13 18:56:09.720: INFO: Pod pod-c41817e4-75b0-11e9-8f67-2632f168be36 no longer exists +Jun 18 11:51:33.688: INFO: Waiting for pod pod-67a1763c-91bf-11e9-bce2-ae54e022189f to disappear +Jun 18 11:51:33.701: INFO: Pod pod-67a1763c-91bf-11e9-bce2-ae54e022189f no longer exists [AfterEach] [sig-storage] EmptyDir volumes /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 18:56:09.720: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-emptydir-rpsd6" for this suite. -May 13 18:56:15.761: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 18:56:15.852: INFO: namespace: e2e-tests-emptydir-rpsd6, resource: bindings, ignored listing per whitelist -May 13 18:56:16.001: INFO: namespace e2e-tests-emptydir-rpsd6 deletion completed in 6.27390006s +Jun 18 11:51:33.702: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "e2e-tests-emptydir-j4br4" for this suite. +Jun 18 11:51:39.773: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 11:51:40.112: INFO: namespace: e2e-tests-emptydir-j4br4, resource: bindings, ignored listing per whitelist +Jun 18 11:51:40.318: INFO: namespace e2e-tests-emptydir-j4br4 deletion completed in 6.595004758s -• [SLOW TEST:8.747 seconds] +• [SLOW TEST:13.480 seconds] [sig-storage] EmptyDir volumes /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:40 - should support (non-root,0777,default) [NodeConformance] [Conformance] + should support (root,0644,tmpfs) [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ -SSSSSSSS +SSSS ------------------------------ -[sig-storage] EmptyDir wrapper volumes - should not conflict [Conformance] +[sig-storage] Projected downwardAPI + should provide container's memory limit [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -[BeforeEach] [sig-storage] EmptyDir wrapper volumes +[BeforeEach] [sig-storage] Projected downwardAPI /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client -May 13 18:56:16.002: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -STEP: Building a namespace api object, basename emptydir-wrapper -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-emptydir-wrapper-xtjxg +Jun 18 11:51:40.320: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 +STEP: Building a namespace api object, basename projected +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-projected-mtbs9 STEP: Waiting for a default service account to be provisioned in namespace -[It] should not conflict [Conformance] +[BeforeEach] [sig-storage] Projected downwardAPI + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:39 +[It] should provide container's memory limit [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -STEP: Cleaning up the secret -STEP: Cleaning up the configmap -STEP: Cleaning up the pod -[AfterEach] [sig-storage] EmptyDir wrapper volumes +STEP: Creating a pod to test downward API volume plugin +Jun 18 11:51:40.795: INFO: Waiting up to 5m0s for pod "downwardapi-volume-6f8b7834-91bf-11e9-bce2-ae54e022189f" in namespace "e2e-tests-projected-mtbs9" to be "success or failure" +Jun 18 11:51:40.810: INFO: Pod "downwardapi-volume-6f8b7834-91bf-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 14.505605ms +Jun 18 11:51:42.825: INFO: Pod "downwardapi-volume-6f8b7834-91bf-11e9-bce2-ae54e022189f": Phase="Running", Reason="", readiness=true. Elapsed: 2.029532817s +Jun 18 11:51:44.857: INFO: Pod "downwardapi-volume-6f8b7834-91bf-11e9-bce2-ae54e022189f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.061563737s +STEP: Saw pod success +Jun 18 11:51:44.857: INFO: Pod "downwardapi-volume-6f8b7834-91bf-11e9-bce2-ae54e022189f" satisfied condition "success or failure" +Jun 18 11:51:44.871: INFO: Trying to get logs from node 10.72.74.149 pod downwardapi-volume-6f8b7834-91bf-11e9-bce2-ae54e022189f container client-container: +STEP: delete the pod +Jun 18 11:51:44.951: INFO: Waiting for pod downwardapi-volume-6f8b7834-91bf-11e9-bce2-ae54e022189f to disappear +Jun 18 11:51:44.965: INFO: Pod downwardapi-volume-6f8b7834-91bf-11e9-bce2-ae54e022189f no longer exists +[AfterEach] [sig-storage] Projected downwardAPI /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 18:56:20.527: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-emptydir-wrapper-xtjxg" for this suite. -May 13 18:56:26.564: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 18:56:26.628: INFO: namespace: e2e-tests-emptydir-wrapper-xtjxg, resource: bindings, ignored listing per whitelist -May 13 18:56:26.818: INFO: namespace e2e-tests-emptydir-wrapper-xtjxg deletion completed in 6.283524234s +Jun 18 11:51:44.965: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "e2e-tests-projected-mtbs9" for this suite. +Jun 18 11:51:53.041: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 11:51:53.295: INFO: namespace: e2e-tests-projected-mtbs9, resource: bindings, ignored listing per whitelist +Jun 18 11:51:53.833: INFO: namespace e2e-tests-projected-mtbs9 deletion completed in 8.85033182s -• [SLOW TEST:10.816 seconds] -[sig-storage] EmptyDir wrapper volumes -/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:22 - should not conflict [Conformance] +• [SLOW TEST:13.513 seconds] +[sig-storage] Projected downwardAPI +/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:33 + should provide container's memory limit [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ -S +SSSSSS ------------------------------ -[k8s.io] Kubelet when scheduling a busybox command in a pod - should print the output to logs [NodeConformance] [Conformance] +[sig-cli] Kubectl client [k8s.io] Kubectl patch + should add annotations for pods in rc [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -[BeforeEach] [k8s.io] Kubelet +[BeforeEach] [sig-cli] Kubectl client /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client -May 13 18:56:26.818: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -STEP: Building a namespace api object, basename kubelet-test -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-kubelet-test-hghcz +Jun 18 11:51:53.834: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 +STEP: Building a namespace api object, basename kubectl +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-kubectl-z65vx STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [k8s.io] Kubelet - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:37 -[It] should print the output to logs [NodeConformance] [Conformance] +[BeforeEach] [sig-cli] Kubectl client + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:243 +[It] should add annotations for pods in rc [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -[AfterEach] [k8s.io] Kubelet +STEP: creating Redis RC +Jun 18 11:51:54.275: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 create -f - --namespace=e2e-tests-kubectl-z65vx' +Jun 18 11:51:54.860: INFO: stderr: "" +Jun 18 11:51:54.861: INFO: stdout: "replicationcontroller/redis-master created\n" +STEP: Waiting for Redis master to start. +Jun 18 11:51:55.892: INFO: Selector matched 1 pods for map[app:redis] +Jun 18 11:51:55.892: INFO: Found 0 / 1 +Jun 18 11:51:56.875: INFO: Selector matched 1 pods for map[app:redis] +Jun 18 11:51:56.875: INFO: Found 1 / 1 +Jun 18 11:51:56.875: INFO: WaitFor completed with timeout 5m0s. Pods found = 1 out of 1 +STEP: patching all pods +Jun 18 11:51:56.890: INFO: Selector matched 1 pods for map[app:redis] +Jun 18 11:51:56.890: INFO: ForEach: Found 1 pods from the filter. Now looping through them. +Jun 18 11:51:56.890: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 patch pod redis-master-zd72d --namespace=e2e-tests-kubectl-z65vx -p {"metadata":{"annotations":{"x":"y"}}}' +Jun 18 11:51:57.037: INFO: stderr: "" +Jun 18 11:51:57.037: INFO: stdout: "pod/redis-master-zd72d patched\n" +STEP: checking annotations +Jun 18 11:51:57.052: INFO: Selector matched 1 pods for map[app:redis] +Jun 18 11:51:57.052: INFO: ForEach: Found 1 pods from the filter. Now looping through them. +[AfterEach] [sig-cli] Kubectl client /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 18:56:29.151: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-kubelet-test-hghcz" for this suite. -May 13 18:57:13.188: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 18:57:13.763: INFO: namespace: e2e-tests-kubelet-test-hghcz, resource: bindings, ignored listing per whitelist -May 13 18:57:13.785: INFO: namespace e2e-tests-kubelet-test-hghcz deletion completed in 44.62663585s +Jun 18 11:51:57.052: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "e2e-tests-kubectl-z65vx" for this suite. +Jun 18 11:52:21.116: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 11:52:21.527: INFO: namespace: e2e-tests-kubectl-z65vx, resource: bindings, ignored listing per whitelist +Jun 18 11:52:21.704: INFO: namespace e2e-tests-kubectl-z65vx deletion completed in 24.633980974s -• [SLOW TEST:46.967 seconds] -[k8s.io] Kubelet -/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 - when scheduling a busybox command in a pod - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:40 - should print the output to logs [NodeConformance] [Conformance] +• [SLOW TEST:27.870 seconds] +[sig-cli] Kubectl client +/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:22 + [k8s.io] Kubectl patch + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 + should add annotations for pods in rc [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ -SSSSS +SSSSSSS ------------------------------ -[sig-api-machinery] Watchers - should observe add, update, and delete watch notifications on configmaps [Conformance] +[sig-network] Services + should serve a basic endpoint from pods [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -[BeforeEach] [sig-api-machinery] Watchers +[BeforeEach] [sig-network] Services /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client -May 13 18:57:13.785: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -STEP: Building a namespace api object, basename watch -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-watch-9mnwf +Jun 18 11:52:21.706: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 +STEP: Building a namespace api object, basename services +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-services-x68nm STEP: Waiting for a default service account to be provisioned in namespace -[It] should observe add, update, and delete watch notifications on configmaps [Conformance] +[BeforeEach] [sig-network] Services + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:85 +[It] should serve a basic endpoint from pods [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -STEP: creating a watch on configmaps with label A -STEP: creating a watch on configmaps with label B -STEP: creating a watch on configmaps with label A or B -STEP: creating a configmap with label A and ensuring the correct watchers observe the notification -May 13 18:57:14.102: INFO: Got : ADDED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-configmap-a,GenerateName:,Namespace:e2e-tests-watch-9mnwf,SelfLink:/api/v1/namespaces/e2e-tests-watch-9mnwf/configmaps/e2e-watch-test-configmap-a,UID:ebb89a30-75b0-11e9-b786-da20024d205c,ResourceVersion:35432,Generation:0,CreationTimestamp:2019-05-13 18:57:14 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: multiple-watchers-A,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Data:map[string]string{},BinaryData:map[string][]byte{},} -May 13 18:57:14.102: INFO: Got : ADDED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-configmap-a,GenerateName:,Namespace:e2e-tests-watch-9mnwf,SelfLink:/api/v1/namespaces/e2e-tests-watch-9mnwf/configmaps/e2e-watch-test-configmap-a,UID:ebb89a30-75b0-11e9-b786-da20024d205c,ResourceVersion:35432,Generation:0,CreationTimestamp:2019-05-13 18:57:14 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: multiple-watchers-A,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Data:map[string]string{},BinaryData:map[string][]byte{},} -STEP: modifying configmap A and ensuring the correct watchers observe the notification -May 13 18:57:24.230: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-configmap-a,GenerateName:,Namespace:e2e-tests-watch-9mnwf,SelfLink:/api/v1/namespaces/e2e-tests-watch-9mnwf/configmaps/e2e-watch-test-configmap-a,UID:ebb89a30-75b0-11e9-b786-da20024d205c,ResourceVersion:35449,Generation:0,CreationTimestamp:2019-05-13 18:57:14 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: multiple-watchers-A,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Data:map[string]string{mutation: 1,},BinaryData:map[string][]byte{},} -May 13 18:57:24.230: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-configmap-a,GenerateName:,Namespace:e2e-tests-watch-9mnwf,SelfLink:/api/v1/namespaces/e2e-tests-watch-9mnwf/configmaps/e2e-watch-test-configmap-a,UID:ebb89a30-75b0-11e9-b786-da20024d205c,ResourceVersion:35449,Generation:0,CreationTimestamp:2019-05-13 18:57:14 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: multiple-watchers-A,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Data:map[string]string{mutation: 1,},BinaryData:map[string][]byte{},} -STEP: modifying configmap A again and ensuring the correct watchers observe the notification -May 13 18:57:34.261: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-configmap-a,GenerateName:,Namespace:e2e-tests-watch-9mnwf,SelfLink:/api/v1/namespaces/e2e-tests-watch-9mnwf/configmaps/e2e-watch-test-configmap-a,UID:ebb89a30-75b0-11e9-b786-da20024d205c,ResourceVersion:35466,Generation:0,CreationTimestamp:2019-05-13 18:57:14 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: multiple-watchers-A,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},} -May 13 18:57:34.261: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-configmap-a,GenerateName:,Namespace:e2e-tests-watch-9mnwf,SelfLink:/api/v1/namespaces/e2e-tests-watch-9mnwf/configmaps/e2e-watch-test-configmap-a,UID:ebb89a30-75b0-11e9-b786-da20024d205c,ResourceVersion:35466,Generation:0,CreationTimestamp:2019-05-13 18:57:14 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: multiple-watchers-A,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},} -STEP: deleting configmap A and ensuring the correct watchers observe the notification -May 13 18:57:44.292: INFO: Got : DELETED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-configmap-a,GenerateName:,Namespace:e2e-tests-watch-9mnwf,SelfLink:/api/v1/namespaces/e2e-tests-watch-9mnwf/configmaps/e2e-watch-test-configmap-a,UID:ebb89a30-75b0-11e9-b786-da20024d205c,ResourceVersion:35482,Generation:0,CreationTimestamp:2019-05-13 18:57:14 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: multiple-watchers-A,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},} -May 13 18:57:44.292: INFO: Got : DELETED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-configmap-a,GenerateName:,Namespace:e2e-tests-watch-9mnwf,SelfLink:/api/v1/namespaces/e2e-tests-watch-9mnwf/configmaps/e2e-watch-test-configmap-a,UID:ebb89a30-75b0-11e9-b786-da20024d205c,ResourceVersion:35482,Generation:0,CreationTimestamp:2019-05-13 18:57:14 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: multiple-watchers-A,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},} -STEP: creating a configmap with label B and ensuring the correct watchers observe the notification -May 13 18:57:54.320: INFO: Got : ADDED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-configmap-b,GenerateName:,Namespace:e2e-tests-watch-9mnwf,SelfLink:/api/v1/namespaces/e2e-tests-watch-9mnwf/configmaps/e2e-watch-test-configmap-b,UID:03b0d925-75b1-11e9-b786-da20024d205c,ResourceVersion:35535,Generation:0,CreationTimestamp:2019-05-13 18:57:54 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: multiple-watchers-B,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Data:map[string]string{},BinaryData:map[string][]byte{},} -May 13 18:57:54.320: INFO: Got : ADDED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-configmap-b,GenerateName:,Namespace:e2e-tests-watch-9mnwf,SelfLink:/api/v1/namespaces/e2e-tests-watch-9mnwf/configmaps/e2e-watch-test-configmap-b,UID:03b0d925-75b1-11e9-b786-da20024d205c,ResourceVersion:35535,Generation:0,CreationTimestamp:2019-05-13 18:57:54 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: multiple-watchers-B,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Data:map[string]string{},BinaryData:map[string][]byte{},} -STEP: deleting configmap B and ensuring the correct watchers observe the notification -May 13 18:58:04.350: INFO: Got : DELETED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-configmap-b,GenerateName:,Namespace:e2e-tests-watch-9mnwf,SelfLink:/api/v1/namespaces/e2e-tests-watch-9mnwf/configmaps/e2e-watch-test-configmap-b,UID:03b0d925-75b1-11e9-b786-da20024d205c,ResourceVersion:35552,Generation:0,CreationTimestamp:2019-05-13 18:57:54 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: multiple-watchers-B,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Data:map[string]string{},BinaryData:map[string][]byte{},} -May 13 18:58:04.350: INFO: Got : DELETED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-configmap-b,GenerateName:,Namespace:e2e-tests-watch-9mnwf,SelfLink:/api/v1/namespaces/e2e-tests-watch-9mnwf/configmaps/e2e-watch-test-configmap-b,UID:03b0d925-75b1-11e9-b786-da20024d205c,ResourceVersion:35552,Generation:0,CreationTimestamp:2019-05-13 18:57:54 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: multiple-watchers-B,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Data:map[string]string{},BinaryData:map[string][]byte{},} -[AfterEach] [sig-api-machinery] Watchers +STEP: creating service endpoint-test2 in namespace e2e-tests-services-x68nm +STEP: waiting up to 3m0s for service endpoint-test2 in namespace e2e-tests-services-x68nm to expose endpoints map[] +Jun 18 11:52:22.198: INFO: Get endpoints failed (11.828849ms elapsed, ignoring for 5s): endpoints "endpoint-test2" not found +Jun 18 11:52:23.210: INFO: successfully validated that service endpoint-test2 in namespace e2e-tests-services-x68nm exposes endpoints map[] (1.024407469s elapsed) +STEP: Creating pod pod1 in namespace e2e-tests-services-x68nm +STEP: waiting up to 3m0s for service endpoint-test2 in namespace e2e-tests-services-x68nm to expose endpoints map[pod1:[80]] +Jun 18 11:52:26.384: INFO: successfully validated that service endpoint-test2 in namespace e2e-tests-services-x68nm exposes endpoints map[pod1:[80]] (3.140390415s elapsed) +STEP: Creating pod pod2 in namespace e2e-tests-services-x68nm +STEP: waiting up to 3m0s for service endpoint-test2 in namespace e2e-tests-services-x68nm to expose endpoints map[pod1:[80] pod2:[80]] +Jun 18 11:52:29.605: INFO: successfully validated that service endpoint-test2 in namespace e2e-tests-services-x68nm exposes endpoints map[pod1:[80] pod2:[80]] (3.200879832s elapsed) +STEP: Deleting pod pod1 in namespace e2e-tests-services-x68nm +STEP: waiting up to 3m0s for service endpoint-test2 in namespace e2e-tests-services-x68nm to expose endpoints map[pod2:[80]] +Jun 18 11:52:29.699: INFO: successfully validated that service endpoint-test2 in namespace e2e-tests-services-x68nm exposes endpoints map[pod2:[80]] (69.673439ms elapsed) +STEP: Deleting pod pod2 in namespace e2e-tests-services-x68nm +STEP: waiting up to 3m0s for service endpoint-test2 in namespace e2e-tests-services-x68nm to expose endpoints map[] +Jun 18 11:52:29.734: INFO: successfully validated that service endpoint-test2 in namespace e2e-tests-services-x68nm exposes endpoints map[] (12.499935ms elapsed) +[AfterEach] [sig-network] Services /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 18:58:14.351: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-watch-9mnwf" for this suite. -May 13 18:58:20.400: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 18:58:20.710: INFO: namespace: e2e-tests-watch-9mnwf, resource: bindings, ignored listing per whitelist -May 13 18:58:20.754: INFO: namespace e2e-tests-watch-9mnwf deletion completed in 6.382440495s +Jun 18 11:52:29.799: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "e2e-tests-services-x68nm" for this suite. +Jun 18 11:52:37.868: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 11:52:38.007: INFO: namespace: e2e-tests-services-x68nm, resource: bindings, ignored listing per whitelist +Jun 18 11:52:39.542: INFO: namespace e2e-tests-services-x68nm deletion completed in 9.724294154s +[AfterEach] [sig-network] Services + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:90 -• [SLOW TEST:66.969 seconds] -[sig-api-machinery] Watchers -/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:22 - should observe add, update, and delete watch notifications on configmaps [Conformance] +• [SLOW TEST:17.837 seconds] +[sig-network] Services +/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:22 + should serve a basic endpoint from pods [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ -SSSS +SSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-auth] ServiceAccounts - should mount an API token into pods [Conformance] +[sig-storage] Downward API volume + should provide podname only [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -[BeforeEach] [sig-auth] ServiceAccounts +[BeforeEach] [sig-storage] Downward API volume /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client -May 13 18:58:20.756: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -STEP: Building a namespace api object, basename svcaccounts -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-svcaccounts-cthmt +Jun 18 11:52:39.546: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 +STEP: Building a namespace api object, basename downward-api +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-downward-api-jcczq STEP: Waiting for a default service account to be provisioned in namespace -[It] should mount an API token into pods [Conformance] +[BeforeEach] [sig-storage] Downward API volume + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:39 +[It] should provide podname only [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -STEP: getting the auto-created API token -STEP: Creating a pod to test consume service account token -May 13 18:58:21.562: INFO: Waiting up to 5m0s for pod "pod-service-account-13ecad26-75b1-11e9-8f67-2632f168be36-x46qz" in namespace "e2e-tests-svcaccounts-cthmt" to be "success or failure" -May 13 18:58:21.570: INFO: Pod "pod-service-account-13ecad26-75b1-11e9-8f67-2632f168be36-x46qz": Phase="Pending", Reason="", readiness=false. Elapsed: 8.052351ms -May 13 18:58:23.578: INFO: Pod "pod-service-account-13ecad26-75b1-11e9-8f67-2632f168be36-x46qz": Phase="Pending", Reason="", readiness=false. Elapsed: 2.01599754s -May 13 18:58:25.620: INFO: Pod "pod-service-account-13ecad26-75b1-11e9-8f67-2632f168be36-x46qz": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.057318835s -STEP: Saw pod success -May 13 18:58:25.620: INFO: Pod "pod-service-account-13ecad26-75b1-11e9-8f67-2632f168be36-x46qz" satisfied condition "success or failure" -May 13 18:58:25.628: INFO: Trying to get logs from node 10.170.219.153 pod pod-service-account-13ecad26-75b1-11e9-8f67-2632f168be36-x46qz container token-test: -STEP: delete the pod -May 13 18:58:25.668: INFO: Waiting for pod pod-service-account-13ecad26-75b1-11e9-8f67-2632f168be36-x46qz to disappear -May 13 18:58:25.675: INFO: Pod pod-service-account-13ecad26-75b1-11e9-8f67-2632f168be36-x46qz no longer exists -STEP: Creating a pod to test consume service account root CA -May 13 18:58:25.684: INFO: Waiting up to 5m0s for pod "pod-service-account-13ecad26-75b1-11e9-8f67-2632f168be36-th6ck" in namespace "e2e-tests-svcaccounts-cthmt" to be "success or failure" -May 13 18:58:25.692: INFO: Pod "pod-service-account-13ecad26-75b1-11e9-8f67-2632f168be36-th6ck": Phase="Pending", Reason="", readiness=false. Elapsed: 8.179352ms -May 13 18:58:27.701: INFO: Pod "pod-service-account-13ecad26-75b1-11e9-8f67-2632f168be36-th6ck": Phase="Pending", Reason="", readiness=false. Elapsed: 2.016547039s -May 13 18:58:29.709: INFO: Pod "pod-service-account-13ecad26-75b1-11e9-8f67-2632f168be36-th6ck": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.025384567s -STEP: Saw pod success -May 13 18:58:29.709: INFO: Pod "pod-service-account-13ecad26-75b1-11e9-8f67-2632f168be36-th6ck" satisfied condition "success or failure" -May 13 18:58:29.717: INFO: Trying to get logs from node 10.170.219.153 pod pod-service-account-13ecad26-75b1-11e9-8f67-2632f168be36-th6ck container root-ca-test: -STEP: delete the pod -May 13 18:58:29.760: INFO: Waiting for pod pod-service-account-13ecad26-75b1-11e9-8f67-2632f168be36-th6ck to disappear -May 13 18:58:29.768: INFO: Pod pod-service-account-13ecad26-75b1-11e9-8f67-2632f168be36-th6ck no longer exists -STEP: Creating a pod to test consume service account namespace -May 13 18:58:29.820: INFO: Waiting up to 5m0s for pod "pod-service-account-13ecad26-75b1-11e9-8f67-2632f168be36-x6797" in namespace "e2e-tests-svcaccounts-cthmt" to be "success or failure" -May 13 18:58:29.831: INFO: Pod "pod-service-account-13ecad26-75b1-11e9-8f67-2632f168be36-x6797": Phase="Pending", Reason="", readiness=false. Elapsed: 11.699904ms -May 13 18:58:31.839: INFO: Pod "pod-service-account-13ecad26-75b1-11e9-8f67-2632f168be36-x6797": Phase="Pending", Reason="", readiness=false. Elapsed: 2.01969841s -May 13 18:58:33.848: INFO: Pod "pod-service-account-13ecad26-75b1-11e9-8f67-2632f168be36-x6797": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.0280618s +STEP: Creating a pod to test downward API volume plugin +Jun 18 11:52:40.118: INFO: Waiting up to 5m0s for pod "downwardapi-volume-92e74509-91bf-11e9-bce2-ae54e022189f" in namespace "e2e-tests-downward-api-jcczq" to be "success or failure" +Jun 18 11:52:40.135: INFO: Pod "downwardapi-volume-92e74509-91bf-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 16.545017ms +Jun 18 11:52:42.150: INFO: Pod "downwardapi-volume-92e74509-91bf-11e9-bce2-ae54e022189f": Phase="Running", Reason="", readiness=true. Elapsed: 2.031397567s +Jun 18 11:52:44.165: INFO: Pod "downwardapi-volume-92e74509-91bf-11e9-bce2-ae54e022189f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.046114687s STEP: Saw pod success -May 13 18:58:33.848: INFO: Pod "pod-service-account-13ecad26-75b1-11e9-8f67-2632f168be36-x6797" satisfied condition "success or failure" -May 13 18:58:33.856: INFO: Trying to get logs from node 10.170.219.153 pod pod-service-account-13ecad26-75b1-11e9-8f67-2632f168be36-x6797 container namespace-test: +Jun 18 11:52:44.165: INFO: Pod "downwardapi-volume-92e74509-91bf-11e9-bce2-ae54e022189f" satisfied condition "success or failure" +Jun 18 11:52:44.179: INFO: Trying to get logs from node 10.72.74.143 pod downwardapi-volume-92e74509-91bf-11e9-bce2-ae54e022189f container client-container: STEP: delete the pod -May 13 18:58:33.896: INFO: Waiting for pod pod-service-account-13ecad26-75b1-11e9-8f67-2632f168be36-x6797 to disappear -May 13 18:58:33.903: INFO: Pod pod-service-account-13ecad26-75b1-11e9-8f67-2632f168be36-x6797 no longer exists -[AfterEach] [sig-auth] ServiceAccounts +Jun 18 11:52:44.261: INFO: Waiting for pod downwardapi-volume-92e74509-91bf-11e9-bce2-ae54e022189f to disappear +Jun 18 11:52:44.278: INFO: Pod downwardapi-volume-92e74509-91bf-11e9-bce2-ae54e022189f no longer exists +[AfterEach] [sig-storage] Downward API volume /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 18:58:33.903: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-svcaccounts-cthmt" for this suite. -May 13 18:58:39.939: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 18:58:40.211: INFO: namespace: e2e-tests-svcaccounts-cthmt, resource: bindings, ignored listing per whitelist -May 13 18:58:40.264: INFO: namespace e2e-tests-svcaccounts-cthmt deletion completed in 6.3540025s +Jun 18 11:52:44.278: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "e2e-tests-downward-api-jcczq" for this suite. +Jun 18 11:52:50.367: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 11:52:50.465: INFO: namespace: e2e-tests-downward-api-jcczq, resource: bindings, ignored listing per whitelist +Jun 18 11:52:51.009: INFO: namespace e2e-tests-downward-api-jcczq deletion completed in 6.711858518s -• [SLOW TEST:19.509 seconds] -[sig-auth] ServiceAccounts -/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/auth/framework.go:22 - should mount an API token into pods [Conformance] +• [SLOW TEST:11.464 seconds] +[sig-storage] Downward API volume +/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:34 + should provide podname only [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ -SSSSSSSSSSSSSSSSS +[sig-cli] Kubectl client [k8s.io] Proxy server + should support proxy with --port 0 [Conformance] + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 +[BeforeEach] [sig-cli] Kubectl client + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 +STEP: Creating a kubernetes client +Jun 18 11:52:51.010: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 +STEP: Building a namespace api object, basename kubectl +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-kubectl-wxs7n +STEP: Waiting for a default service account to be provisioned in namespace +[BeforeEach] [sig-cli] Kubectl client + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:243 +[It] should support proxy with --port 0 [Conformance] + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 +STEP: starting the proxy server +Jun 18 11:52:51.483: INFO: Asynchronously running '/usr/local/bin/kubectl kubectl --kubeconfig=/tmp/kubeconfig-953583206 proxy -p 0 --disable-filter' +STEP: curling proxy /api/ output +[AfterEach] [sig-cli] Kubectl client + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 +Jun 18 11:52:51.608: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "e2e-tests-kubectl-wxs7n" for this suite. +Jun 18 11:52:57.671: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 11:52:58.132: INFO: namespace: e2e-tests-kubectl-wxs7n, resource: bindings, ignored listing per whitelist +Jun 18 11:52:58.147: INFO: namespace e2e-tests-kubectl-wxs7n deletion completed in 6.524445614s + +• [SLOW TEST:7.137 seconds] +[sig-cli] Kubectl client +/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:22 + [k8s.io] Proxy server + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 + should support proxy with --port 0 [Conformance] + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ -[k8s.io] Container Lifecycle Hook when create a pod with lifecycle hook - should execute poststart exec hook properly [NodeConformance] [Conformance] +S +------------------------------ +[sig-storage] Downward API volume + should update annotations on modification [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -[BeforeEach] [k8s.io] Container Lifecycle Hook +[BeforeEach] [sig-storage] Downward API volume /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client -May 13 18:58:40.265: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -STEP: Building a namespace api object, basename container-lifecycle-hook -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-container-lifecycle-hook-z79gb +Jun 18 11:52:58.147: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 +STEP: Building a namespace api object, basename downward-api +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-downward-api-nm5fq STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] when create a pod with lifecycle hook - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:61 -STEP: create the container to handle the HTTPGet hook request. -[It] should execute poststart exec hook properly [NodeConformance] [Conformance] +[BeforeEach] [sig-storage] Downward API volume + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:39 +[It] should update annotations on modification [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -STEP: create the pod with lifecycle hook -STEP: check poststart hook -STEP: delete the pod with lifecycle hook -May 13 18:58:44.648: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear -May 13 18:58:44.655: INFO: Pod pod-with-poststart-exec-hook still exists -May 13 18:58:46.655: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear -May 13 18:58:46.676: INFO: Pod pod-with-poststart-exec-hook still exists -May 13 18:58:48.655: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear -May 13 18:58:48.663: INFO: Pod pod-with-poststart-exec-hook still exists -May 13 18:58:50.655: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear -May 13 18:58:50.664: INFO: Pod pod-with-poststart-exec-hook still exists -May 13 18:58:52.655: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear -May 13 18:58:52.663: INFO: Pod pod-with-poststart-exec-hook still exists -May 13 18:58:54.655: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear -May 13 18:58:54.663: INFO: Pod pod-with-poststart-exec-hook still exists -May 13 18:58:56.655: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear -May 13 18:58:56.663: INFO: Pod pod-with-poststart-exec-hook still exists -May 13 18:58:58.655: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear -May 13 18:58:58.675: INFO: Pod pod-with-poststart-exec-hook still exists -May 13 18:59:00.655: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear -May 13 18:59:00.663: INFO: Pod pod-with-poststart-exec-hook still exists -May 13 18:59:02.655: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear -May 13 18:59:02.663: INFO: Pod pod-with-poststart-exec-hook still exists -May 13 18:59:04.655: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear -May 13 18:59:04.664: INFO: Pod pod-with-poststart-exec-hook no longer exists -[AfterEach] [k8s.io] Container Lifecycle Hook +STEP: Creating the pod +Jun 18 11:53:01.238: INFO: Successfully updated pod "annotationupdate9deb3049-91bf-11e9-bce2-ae54e022189f" +[AfterEach] [sig-storage] Downward API volume /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 18:59:04.664: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-container-lifecycle-hook-z79gb" for this suite. -May 13 18:59:28.702: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 18:59:28.783: INFO: namespace: e2e-tests-container-lifecycle-hook-z79gb, resource: bindings, ignored listing per whitelist -May 13 18:59:28.967: INFO: namespace e2e-tests-container-lifecycle-hook-z79gb deletion completed in 24.295316995s +Jun 18 11:53:05.346: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "e2e-tests-downward-api-nm5fq" for this suite. +Jun 18 11:53:29.511: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 11:53:30.031: INFO: namespace: e2e-tests-downward-api-nm5fq, resource: bindings, ignored listing per whitelist +Jun 18 11:53:30.541: INFO: namespace e2e-tests-downward-api-nm5fq deletion completed in 25.176288634s -• [SLOW TEST:48.703 seconds] -[k8s.io] Container Lifecycle Hook -/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 - when create a pod with lifecycle hook - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:40 - should execute poststart exec hook properly [NodeConformance] [Conformance] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 +• [SLOW TEST:32.395 seconds] +[sig-storage] Downward API volume +/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:34 + should update annotations on modification [NodeConformance] [Conformance] + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ -[sig-apps] Daemon set [Serial] - should rollback without unnecessary restarts [Conformance] +SSSSSSSSSSSSSSSSSS +------------------------------ +[sig-apps] ReplicationController + should serve a basic image on each replica with a public image [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -[BeforeEach] [sig-apps] Daemon set [Serial] +[BeforeEach] [sig-apps] ReplicationController /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client -May 13 18:59:28.968: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -STEP: Building a namespace api object, basename daemonsets -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-daemonsets-9nsft +Jun 18 11:53:30.545: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 +STEP: Building a namespace api object, basename replication-controller +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-replication-controller-lhwk9 STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [sig-apps] Daemon set [Serial] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:102 -[It] should rollback without unnecessary restarts [Conformance] +[It] should serve a basic image on each replica with a public image [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -May 13 18:59:29.280: INFO: Requires at least 2 nodes (not -1) -[AfterEach] [sig-apps] Daemon set [Serial] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:68 -May 13 18:59:29.294: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"selfLink":"/apis/apps/v1/namespaces/e2e-tests-daemonsets-9nsft/daemonsets","resourceVersion":"35878"},"items":null} - -May 13 18:59:29.301: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"selfLink":"/api/v1/namespaces/e2e-tests-daemonsets-9nsft/pods","resourceVersion":"35878"},"items":null} - -[AfterEach] [sig-apps] Daemon set [Serial] +STEP: Creating replication controller my-hostname-basic-b14d1127-91bf-11e9-bce2-ae54e022189f +Jun 18 11:53:31.119: INFO: Pod name my-hostname-basic-b14d1127-91bf-11e9-bce2-ae54e022189f: Found 0 pods out of 1 +Jun 18 11:53:36.159: INFO: Pod name my-hostname-basic-b14d1127-91bf-11e9-bce2-ae54e022189f: Found 1 pods out of 1 +Jun 18 11:53:36.159: INFO: Ensuring all pods for ReplicationController "my-hostname-basic-b14d1127-91bf-11e9-bce2-ae54e022189f" are running +Jun 18 11:53:36.173: INFO: Pod "my-hostname-basic-b14d1127-91bf-11e9-bce2-ae54e022189f-lnnf9" is running (conditions: [{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2019-06-18 11:53:31 +0000 UTC Reason: Message:} {Type:Ready Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2019-06-18 11:53:33 +0000 UTC Reason: Message:} {Type:ContainersReady Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2019-06-18 11:53:33 +0000 UTC Reason: Message:} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2019-06-18 11:53:31 +0000 UTC Reason: Message:}]) +Jun 18 11:53:36.173: INFO: Trying to dial the pod +Jun 18 11:53:41.816: INFO: Controller my-hostname-basic-b14d1127-91bf-11e9-bce2-ae54e022189f: Got expected result from replica 1 [my-hostname-basic-b14d1127-91bf-11e9-bce2-ae54e022189f-lnnf9]: "my-hostname-basic-b14d1127-91bf-11e9-bce2-ae54e022189f-lnnf9", 1 of 1 required successes so far +[AfterEach] [sig-apps] ReplicationController /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 18:59:29.323: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-daemonsets-9nsft" for this suite. -May 13 18:59:35.357: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 18:59:35.494: INFO: namespace: e2e-tests-daemonsets-9nsft, resource: bindings, ignored listing per whitelist -May 13 18:59:35.596: INFO: namespace e2e-tests-daemonsets-9nsft deletion completed in 6.26603023s +Jun 18 11:53:41.816: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "e2e-tests-replication-controller-lhwk9" for this suite. +Jun 18 11:53:47.900: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 11:53:47.935: INFO: namespace: e2e-tests-replication-controller-lhwk9, resource: bindings, ignored listing per whitelist +Jun 18 11:53:48.392: INFO: namespace e2e-tests-replication-controller-lhwk9 deletion completed in 6.556722122s -S [SKIPPING] [6.628 seconds] -[sig-apps] Daemon set [Serial] +• [SLOW TEST:17.847 seconds] +[sig-apps] ReplicationController /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22 - should rollback without unnecessary restarts [Conformance] [It] + should serve a basic image on each replica with a public image [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 - - May 13 18:59:29.280: Requires at least 2 nodes (not -1) - - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/util.go:292 ------------------------------ -SSSSSSSSSSS +S ------------------------------ [sig-cli] Kubectl client [k8s.io] Kubectl run pod should create a pod from an image when restart is Never [Conformance] @@ -2139,9 +1908,9 @@ SSSSSSSSSSS [BeforeEach] [sig-cli] Kubectl client /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client -May 13 18:59:35.597: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 +Jun 18 11:53:48.393: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 STEP: Building a namespace api object, basename kubectl -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-kubectl-kh9ms +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-kubectl-dvrgz STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:243 @@ -2150,24 +1919,24 @@ STEP: Waiting for a default service account to be provisioned in namespace [It] should create a pod from an image when restart is Never [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: running the image docker.io/library/nginx:1.14-alpine -May 13 18:59:35.862: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 run e2e-test-nginx-pod --restart=Never --generator=run-pod/v1 --image=docker.io/library/nginx:1.14-alpine --namespace=e2e-tests-kubectl-kh9ms' -May 13 18:59:36.382: INFO: stderr: "" -May 13 18:59:36.382: INFO: stdout: "pod/e2e-test-nginx-pod created\n" +Jun 18 11:53:48.840: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 run e2e-test-nginx-pod --restart=Never --generator=run-pod/v1 --image=docker.io/library/nginx:1.14-alpine --namespace=e2e-tests-kubectl-dvrgz' +Jun 18 11:53:49.014: INFO: stderr: "" +Jun 18 11:53:49.014: INFO: stdout: "pod/e2e-test-nginx-pod created\n" STEP: verifying the pod e2e-test-nginx-pod was created [AfterEach] [k8s.io] Kubectl run pod /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1532 -May 13 18:59:37.118: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 delete pods e2e-test-nginx-pod --namespace=e2e-tests-kubectl-kh9ms' -May 13 18:59:39.658: INFO: stderr: "" -May 13 18:59:39.658: INFO: stdout: "pod \"e2e-test-nginx-pod\" deleted\n" +Jun 18 11:53:49.030: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 delete pods e2e-test-nginx-pod --namespace=e2e-tests-kubectl-dvrgz' +Jun 18 11:53:56.172: INFO: stderr: "" +Jun 18 11:53:56.172: INFO: stdout: "pod \"e2e-test-nginx-pod\" deleted\n" [AfterEach] [sig-cli] Kubectl client /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 18:59:39.658: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-kubectl-kh9ms" for this suite. -May 13 18:59:45.752: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 18:59:45.970: INFO: namespace: e2e-tests-kubectl-kh9ms, resource: bindings, ignored listing per whitelist -May 13 18:59:46.118: INFO: namespace e2e-tests-kubectl-kh9ms deletion completed in 6.452230691s +Jun 18 11:53:56.172: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "e2e-tests-kubectl-dvrgz" for this suite. +Jun 18 11:54:02.247: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 11:54:02.360: INFO: namespace: e2e-tests-kubectl-dvrgz, resource: bindings, ignored listing per whitelist +Jun 18 11:54:02.760: INFO: namespace e2e-tests-kubectl-dvrgz deletion completed in 6.567816546s -• [SLOW TEST:10.521 seconds] +• [SLOW TEST:14.367 seconds] [sig-cli] Kubectl client /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:22 [k8s.io] Kubectl run pod @@ -2175,2156 +1944,1188 @@ May 13 18:59:46.118: INFO: namespace e2e-tests-kubectl-kh9ms deletion completed should create a pod from an image when restart is Never [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ -SS ------------------------------- -[sig-storage] ConfigMap - optional updates should be reflected in volume [NodeConformance] [Conformance] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -[BeforeEach] [sig-storage] ConfigMap - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 -STEP: Creating a kubernetes client -May 13 18:59:46.118: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -STEP: Building a namespace api object, basename configmap -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-configmap-m8p6l -STEP: Waiting for a default service account to be provisioned in namespace -[It] optional updates should be reflected in volume [NodeConformance] [Conformance] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -STEP: Creating configMap with name cm-test-opt-del-46819088-75b1-11e9-8f67-2632f168be36 -STEP: Creating configMap with name cm-test-opt-upd-468190d9-75b1-11e9-8f67-2632f168be36 -STEP: Creating the pod -STEP: Deleting configmap cm-test-opt-del-46819088-75b1-11e9-8f67-2632f168be36 -STEP: Updating configmap cm-test-opt-upd-468190d9-75b1-11e9-8f67-2632f168be36 -STEP: Creating configMap with name cm-test-opt-create-468190fa-75b1-11e9-8f67-2632f168be36 -STEP: waiting to observe update in volume -[AfterEach] [sig-storage] ConfigMap - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 18:59:50.686: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-configmap-m8p6l" for this suite. -May 13 19:00:14.723: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 19:00:15.049: INFO: namespace: e2e-tests-configmap-m8p6l, resource: bindings, ignored listing per whitelist -May 13 19:00:15.091: INFO: namespace e2e-tests-configmap-m8p6l deletion completed in 24.396274324s - -• [SLOW TEST:28.973 seconds] -[sig-storage] ConfigMap -/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:33 - optional updates should be reflected in volume [NodeConformance] [Conformance] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------- -SSSSSSSS +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-storage] Projected configMap - optional updates should be reflected in volume [NodeConformance] [Conformance] +[sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] + Should recreate evicted statefulset [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -[BeforeEach] [sig-storage] Projected configMap +[BeforeEach] [sig-apps] StatefulSet /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client -May 13 19:00:15.091: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -STEP: Building a namespace api object, basename projected -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-projected-7xf6f +Jun 18 11:54:02.761: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 +STEP: Building a namespace api object, basename statefulset +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-statefulset-56kfg STEP: Waiting for a default service account to be provisioned in namespace -[It] optional updates should be reflected in volume [NodeConformance] [Conformance] +[BeforeEach] [sig-apps] StatefulSet + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:59 +[BeforeEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:74 +STEP: Creating service test in namespace e2e-tests-statefulset-56kfg +[It] Should recreate evicted statefulset [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -STEP: Creating configMap with name cm-test-opt-del-57c65024-75b1-11e9-8f67-2632f168be36 -STEP: Creating configMap with name cm-test-opt-upd-57c65073-75b1-11e9-8f67-2632f168be36 -STEP: Creating the pod -STEP: Deleting configmap cm-test-opt-del-57c65024-75b1-11e9-8f67-2632f168be36 -STEP: Updating configmap cm-test-opt-upd-57c65073-75b1-11e9-8f67-2632f168be36 -STEP: Creating configMap with name cm-test-opt-create-57c65093-75b1-11e9-8f67-2632f168be36 -STEP: waiting to observe update in volume -[AfterEach] [sig-storage] Projected configMap +STEP: Looking for a node to schedule stateful set and pod +STEP: Creating pod with conflicting port in namespace e2e-tests-statefulset-56kfg +STEP: Creating statefulset with conflicting port in namespace e2e-tests-statefulset-56kfg +STEP: Waiting until pod test-pod will start running in namespace e2e-tests-statefulset-56kfg +STEP: Waiting until stateful pod ss-0 will be recreated and deleted at least once in namespace e2e-tests-statefulset-56kfg +Jun 18 11:54:07.921: INFO: Observed stateful pod in namespace: e2e-tests-statefulset-56kfg, name: ss-0, uid: c5b85413-91bf-11e9-bf44-fa6f350b29f0, status phase: Failed. Waiting for statefulset controller to delete. +Jun 18 11:54:07.986: INFO: Observed stateful pod in namespace: e2e-tests-statefulset-56kfg, name: ss-0, uid: c5b85413-91bf-11e9-bf44-fa6f350b29f0, status phase: Failed. Waiting for statefulset controller to delete. +Jun 18 11:54:07.986: INFO: Observed delete event for stateful pod ss-0 in namespace e2e-tests-statefulset-56kfg +STEP: Removing pod with conflicting port in namespace e2e-tests-statefulset-56kfg +STEP: Waiting when stateful pod ss-0 will be recreated in namespace e2e-tests-statefulset-56kfg and will be in running state +[AfterEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:85 +Jun 18 11:54:18.182: INFO: Deleting all statefulset in ns e2e-tests-statefulset-56kfg +Jun 18 11:54:18.196: INFO: Scaling statefulset ss to 0 +Jun 18 11:54:28.277: INFO: Waiting for statefulset status.replicas updated to 0 +Jun 18 11:54:28.299: INFO: Deleting statefulset ss +[AfterEach] [sig-apps] StatefulSet /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 19:00:19.589: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-projected-7xf6f" for this suite. -May 13 19:00:43.651: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 19:00:44.271: INFO: namespace: e2e-tests-projected-7xf6f, resource: bindings, ignored listing per whitelist -May 13 19:00:44.315: INFO: namespace e2e-tests-projected-7xf6f deletion completed in 24.693817256s +Jun 18 11:54:28.357: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "e2e-tests-statefulset-56kfg" for this suite. +Jun 18 11:54:35.331: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 11:54:35.540: INFO: namespace: e2e-tests-statefulset-56kfg, resource: bindings, ignored listing per whitelist +Jun 18 11:54:39.082: INFO: namespace e2e-tests-statefulset-56kfg deletion completed in 10.707235979s -• [SLOW TEST:29.224 seconds] -[sig-storage] Projected configMap -/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_configmap.go:34 - optional updates should be reflected in volume [NodeConformance] [Conformance] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 +• [SLOW TEST:36.321 seconds] +[sig-apps] StatefulSet +/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22 + [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 + Should recreate evicted statefulset [Conformance] + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ -SSSS +SSSSSSS ------------------------------ -[sig-storage] EmptyDir volumes - should support (root,0777,tmpfs) [NodeConformance] [Conformance] +[sig-node] ConfigMap + should be consumable via environment variable [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -[BeforeEach] [sig-storage] EmptyDir volumes +[BeforeEach] [sig-node] ConfigMap /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client -May 13 19:00:44.316: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -STEP: Building a namespace api object, basename emptydir -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-emptydir-fspp7 +Jun 18 11:54:39.082: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 +STEP: Building a namespace api object, basename configmap +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-configmap-5dtpj STEP: Waiting for a default service account to be provisioned in namespace -[It] should support (root,0777,tmpfs) [NodeConformance] [Conformance] +[It] should be consumable via environment variable [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -STEP: Creating a pod to test emptydir 0777 on tmpfs -May 13 19:00:44.739: INFO: Waiting up to 5m0s for pod "pod-6943f5ce-75b1-11e9-8f67-2632f168be36" in namespace "e2e-tests-emptydir-fspp7" to be "success or failure" -May 13 19:00:44.748: INFO: Pod "pod-6943f5ce-75b1-11e9-8f67-2632f168be36": Phase="Pending", Reason="", readiness=false. Elapsed: 9.671139ms -May 13 19:00:46.846: INFO: Pod "pod-6943f5ce-75b1-11e9-8f67-2632f168be36": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.107523554s +STEP: Creating configMap e2e-tests-configmap-5dtpj/configmap-test-da2157e7-91bf-11e9-bce2-ae54e022189f +STEP: Creating a pod to test consume configMaps +Jun 18 11:54:39.630: INFO: Waiting up to 5m0s for pod "pod-configmaps-da236834-91bf-11e9-bce2-ae54e022189f" in namespace "e2e-tests-configmap-5dtpj" to be "success or failure" +Jun 18 11:54:39.644: INFO: Pod "pod-configmaps-da236834-91bf-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 14.554748ms +Jun 18 11:54:41.659: INFO: Pod "pod-configmaps-da236834-91bf-11e9-bce2-ae54e022189f": Phase="Running", Reason="", readiness=true. Elapsed: 2.029232063s +Jun 18 11:54:43.673: INFO: Pod "pod-configmaps-da236834-91bf-11e9-bce2-ae54e022189f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.043155517s STEP: Saw pod success -May 13 19:00:46.846: INFO: Pod "pod-6943f5ce-75b1-11e9-8f67-2632f168be36" satisfied condition "success or failure" -May 13 19:00:46.854: INFO: Trying to get logs from node 10.170.219.153 pod pod-6943f5ce-75b1-11e9-8f67-2632f168be36 container test-container: +Jun 18 11:54:43.673: INFO: Pod "pod-configmaps-da236834-91bf-11e9-bce2-ae54e022189f" satisfied condition "success or failure" +Jun 18 11:54:43.687: INFO: Trying to get logs from node 10.72.74.144 pod pod-configmaps-da236834-91bf-11e9-bce2-ae54e022189f container env-test: STEP: delete the pod -May 13 19:00:46.892: INFO: Waiting for pod pod-6943f5ce-75b1-11e9-8f67-2632f168be36 to disappear -May 13 19:00:46.900: INFO: Pod pod-6943f5ce-75b1-11e9-8f67-2632f168be36 no longer exists -[AfterEach] [sig-storage] EmptyDir volumes +Jun 18 11:54:43.785: INFO: Waiting for pod pod-configmaps-da236834-91bf-11e9-bce2-ae54e022189f to disappear +Jun 18 11:54:43.801: INFO: Pod pod-configmaps-da236834-91bf-11e9-bce2-ae54e022189f no longer exists +[AfterEach] [sig-node] ConfigMap /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 19:00:46.900: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-emptydir-fspp7" for this suite. -May 13 19:00:52.935: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 19:00:53.053: INFO: namespace: e2e-tests-emptydir-fspp7, resource: bindings, ignored listing per whitelist -May 13 19:00:53.227: INFO: namespace e2e-tests-emptydir-fspp7 deletion completed in 6.319583629s +Jun 18 11:54:43.801: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "e2e-tests-configmap-5dtpj" for this suite. +Jun 18 11:54:51.872: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 11:54:52.012: INFO: namespace: e2e-tests-configmap-5dtpj, resource: bindings, ignored listing per whitelist +Jun 18 11:54:52.416: INFO: namespace e2e-tests-configmap-5dtpj deletion completed in 8.595551754s -• [SLOW TEST:8.912 seconds] -[sig-storage] EmptyDir volumes -/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:40 - should support (root,0777,tmpfs) [NodeConformance] [Conformance] +• [SLOW TEST:13.334 seconds] +[sig-node] ConfigMap +/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/configmap.go:31 + should be consumable via environment variable [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ -SS +SSS ------------------------------ -[k8s.io] Variable Expansion - should allow substituting values in a container's command [NodeConformance] [Conformance] +[k8s.io] Pods + should support remote command execution over websockets [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -[BeforeEach] [k8s.io] Variable Expansion +[BeforeEach] [k8s.io] Pods /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client -May 13 19:00:53.227: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -STEP: Building a namespace api object, basename var-expansion -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-var-expansion-wfjxh +Jun 18 11:54:52.416: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 +STEP: Building a namespace api object, basename pods +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-pods-4hbw6 STEP: Waiting for a default service account to be provisioned in namespace -[It] should allow substituting values in a container's command [NodeConformance] [Conformance] +[BeforeEach] [k8s.io] Pods + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:132 +[It] should support remote command execution over websockets [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -STEP: Creating a pod to test substitution in container's command -May 13 19:00:53.512: INFO: Waiting up to 5m0s for pod "var-expansion-6e7e90d4-75b1-11e9-8f67-2632f168be36" in namespace "e2e-tests-var-expansion-wfjxh" to be "success or failure" -May 13 19:00:53.523: INFO: Pod "var-expansion-6e7e90d4-75b1-11e9-8f67-2632f168be36": Phase="Pending", Reason="", readiness=false. Elapsed: 10.519895ms -May 13 19:00:55.531: INFO: Pod "var-expansion-6e7e90d4-75b1-11e9-8f67-2632f168be36": Phase="Running", Reason="", readiness=true. Elapsed: 2.018798931s -May 13 19:00:57.539: INFO: Pod "var-expansion-6e7e90d4-75b1-11e9-8f67-2632f168be36": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.027058218s -STEP: Saw pod success -May 13 19:00:57.539: INFO: Pod "var-expansion-6e7e90d4-75b1-11e9-8f67-2632f168be36" satisfied condition "success or failure" -May 13 19:00:57.547: INFO: Trying to get logs from node 10.170.219.177 pod var-expansion-6e7e90d4-75b1-11e9-8f67-2632f168be36 container dapi-container: -STEP: delete the pod -May 13 19:00:57.644: INFO: Waiting for pod var-expansion-6e7e90d4-75b1-11e9-8f67-2632f168be36 to disappear -May 13 19:00:57.651: INFO: Pod var-expansion-6e7e90d4-75b1-11e9-8f67-2632f168be36 no longer exists -[AfterEach] [k8s.io] Variable Expansion +Jun 18 11:54:52.860: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 +STEP: creating the pod +STEP: submitting the pod to kubernetes +[AfterEach] [k8s.io] Pods /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 19:00:57.651: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-var-expansion-wfjxh" for this suite. -May 13 19:01:03.687: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 19:01:04.060: INFO: namespace: e2e-tests-var-expansion-wfjxh, resource: bindings, ignored listing per whitelist -May 13 19:01:04.128: INFO: namespace e2e-tests-var-expansion-wfjxh deletion completed in 6.468175173s +Jun 18 11:54:55.174: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "e2e-tests-pods-4hbw6" for this suite. +Jun 18 11:55:43.245: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 11:55:43.696: INFO: namespace: e2e-tests-pods-4hbw6, resource: bindings, ignored listing per whitelist +Jun 18 11:55:43.752: INFO: namespace e2e-tests-pods-4hbw6 deletion completed in 48.558932495s -• [SLOW TEST:10.900 seconds] -[k8s.io] Variable Expansion +• [SLOW TEST:51.336 seconds] +[k8s.io] Pods /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 - should allow substituting values in a container's command [NodeConformance] [Conformance] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------- -SSS ------------------------------- -[sig-cli] Kubectl client [k8s.io] Kubectl describe - should check if kubectl describe prints relevant information for rc and pods [Conformance] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -[BeforeEach] [sig-cli] Kubectl client - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 -STEP: Creating a kubernetes client -May 13 19:01:04.128: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -STEP: Building a namespace api object, basename kubectl -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-kubectl-dpvrq -STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [sig-cli] Kubectl client - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:243 -[It] should check if kubectl describe prints relevant information for rc and pods [Conformance] + should support remote command execution over websockets [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -May 13 19:01:04.423: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 version --client' -May 13 19:01:04.476: INFO: stderr: "" -May 13 19:01:04.476: INFO: stdout: "Client Version: version.Info{Major:\"1\", Minor:\"13\", GitVersion:\"v1.13.0\", GitCommit:\"ddf47ac13c1a9483ea035a79cd7c10005ff21a6d\", GitTreeState:\"clean\", BuildDate:\"2018-12-03T21:04:45Z\", GoVersion:\"go1.11.2\", Compiler:\"gc\", Platform:\"linux/amd64\"}\n" -May 13 19:01:04.489: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 create -f - --namespace=e2e-tests-kubectl-dpvrq' -May 13 19:01:04.757: INFO: stderr: "" -May 13 19:01:04.757: INFO: stdout: "replicationcontroller/redis-master created\n" -May 13 19:01:04.757: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 create -f - --namespace=e2e-tests-kubectl-dpvrq' -May 13 19:01:04.956: INFO: stderr: "" -May 13 19:01:04.956: INFO: stdout: "service/redis-master created\n" -STEP: Waiting for Redis master to start. -May 13 19:01:05.965: INFO: Selector matched 1 pods for map[app:redis] -May 13 19:01:05.965: INFO: Found 0 / 1 -May 13 19:01:06.964: INFO: Selector matched 1 pods for map[app:redis] -May 13 19:01:06.964: INFO: Found 1 / 1 -May 13 19:01:06.964: INFO: WaitFor completed with timeout 5m0s. Pods found = 1 out of 1 -May 13 19:01:06.972: INFO: Selector matched 1 pods for map[app:redis] -May 13 19:01:06.972: INFO: ForEach: Found 1 pods from the filter. Now looping through them. -May 13 19:01:06.972: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 describe pod redis-master-fbwpx --namespace=e2e-tests-kubectl-dpvrq' -May 13 19:01:07.088: INFO: stderr: "" -May 13 19:01:07.088: INFO: stdout: "Name: redis-master-fbwpx\nNamespace: e2e-tests-kubectl-dpvrq\nPriority: 0\nPriorityClassName: \nNode: 10.170.219.153/10.170.219.153\nStart Time: Mon, 13 May 2019 19:01:04 +0000\nLabels: app=redis\n role=master\nAnnotations: kubernetes.io/psp: e2e-test-privileged-psp\nStatus: Running\nIP: 172.30.63.15\nControlled By: ReplicationController/redis-master\nContainers:\n redis-master:\n Container ID: containerd://db8fb328abbc78c670154d0984428fe0bbd188ad2379e5cd29ec52a54899712c\n Image: gcr.io/kubernetes-e2e-test-images/redis:1.0\n Image ID: gcr.io/kubernetes-e2e-test-images/redis@sha256:af4748d1655c08dc54d4be5182135395db9ce87aba2d4699b26b14ae197c5830\n Port: 6379/TCP\n Host Port: 0/TCP\n State: Running\n Started: Mon, 13 May 2019 19:01:05 +0000\n Ready: True\n Restart Count: 0\n Environment: \n Mounts:\n /var/run/secrets/kubernetes.io/serviceaccount from default-token-x54mw (ro)\nConditions:\n Type Status\n Initialized True \n Ready True \n ContainersReady True \n PodScheduled True \nVolumes:\n default-token-x54mw:\n Type: Secret (a volume populated by a Secret)\n SecretName: default-token-x54mw\n Optional: false\nQoS Class: BestEffort\nNode-Selectors: \nTolerations: node.kubernetes.io/not-ready:NoExecute for 300s\n node.kubernetes.io/unreachable:NoExecute for 300s\nEvents:\n Type Reason Age From Message\n ---- ------ ---- ---- -------\n Normal Scheduled 3s default-scheduler Successfully assigned e2e-tests-kubectl-dpvrq/redis-master-fbwpx to 10.170.219.153\n Normal Pulled 2s kubelet, 10.170.219.153 Container image \"gcr.io/kubernetes-e2e-test-images/redis:1.0\" already present on machine\n Normal Created 2s kubelet, 10.170.219.153 Created container\n Normal Started 2s kubelet, 10.170.219.153 Started container\n" -May 13 19:01:07.089: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 describe rc redis-master --namespace=e2e-tests-kubectl-dpvrq' -May 13 19:01:07.331: INFO: stderr: "" -May 13 19:01:07.331: INFO: stdout: "Name: redis-master\nNamespace: e2e-tests-kubectl-dpvrq\nSelector: app=redis,role=master\nLabels: app=redis\n role=master\nAnnotations: \nReplicas: 1 current / 1 desired\nPods Status: 1 Running / 0 Waiting / 0 Succeeded / 0 Failed\nPod Template:\n Labels: app=redis\n role=master\n Containers:\n redis-master:\n Image: gcr.io/kubernetes-e2e-test-images/redis:1.0\n Port: 6379/TCP\n Host Port: 0/TCP\n Environment: \n Mounts: \n Volumes: \nEvents:\n Type Reason Age From Message\n ---- ------ ---- ---- -------\n Normal SuccessfulCreate 3s replication-controller Created pod: redis-master-fbwpx\n" -May 13 19:01:07.331: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 describe service redis-master --namespace=e2e-tests-kubectl-dpvrq' -May 13 19:01:07.469: INFO: stderr: "" -May 13 19:01:07.469: INFO: stdout: "Name: redis-master\nNamespace: e2e-tests-kubectl-dpvrq\nLabels: app=redis\n role=master\nAnnotations: \nSelector: app=redis,role=master\nType: ClusterIP\nIP: 172.21.170.97\nPort: 6379/TCP\nTargetPort: redis-server/TCP\nEndpoints: 172.30.63.15:6379\nSession Affinity: None\nEvents: \n" -May 13 19:01:07.477: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 describe node 10.170.219.151' -May 13 19:01:07.623: INFO: stderr: "" -May 13 19:01:07.623: INFO: stdout: "Name: 10.170.219.151\nRoles: \nLabels: arch=amd64\n beta.kubernetes.io/arch=amd64\n beta.kubernetes.io/instance-type=b3c.4x16.encrypted\n beta.kubernetes.io/os=linux\n failure-domain.beta.kubernetes.io/region=us-east\n failure-domain.beta.kubernetes.io/zone=wdc04\n ibm-cloud.kubernetes.io/encrypted-docker-data=true\n ibm-cloud.kubernetes.io/ha-worker=true\n ibm-cloud.kubernetes.io/iaas-provider=softlayer\n ibm-cloud.kubernetes.io/machine-type=b3c.4x16.encrypted\n ibm-cloud.kubernetes.io/os=UBUNTU_18_64\n ibm-cloud.kubernetes.io/sgx-enabled=false\n ibm-cloud.kubernetes.io/worker-pool-id=0ef1a5f1e33c44fe9c21650356d908ee-715ddc0\n ibm-cloud.kubernetes.io/worker-version=1.13.6_1521\n kubernetes.io/hostname=10.170.219.151\n privateVLAN=2615309\n publicVLAN=2615307\nAnnotations: node.alpha.kubernetes.io/ttl: 0\n volumes.kubernetes.io/controller-managed-attach-detach: true\nCreationTimestamp: Mon, 13 May 2019 16:27:05 +0000\nTaints: \nUnschedulable: false\nConditions:\n Type Status LastHeartbeatTime LastTransitionTime Reason Message\n ---- ------ ----------------- ------------------ ------ -------\n MemoryPressure False Mon, 13 May 2019 19:01:04 +0000 Mon, 13 May 2019 16:27:05 +0000 KubeletHasSufficientMemory kubelet has sufficient memory available\n DiskPressure False Mon, 13 May 2019 19:01:04 +0000 Mon, 13 May 2019 16:27:05 +0000 KubeletHasNoDiskPressure kubelet has no disk pressure\n PIDPressure False Mon, 13 May 2019 19:01:04 +0000 Mon, 13 May 2019 16:27:05 +0000 KubeletHasSufficientPID kubelet has sufficient PID available\n Ready True Mon, 13 May 2019 19:01:04 +0000 Mon, 13 May 2019 16:27:35 +0000 KubeletReady kubelet is posting ready status. AppArmor enabled\nAddresses:\n InternalIP: 10.170.219.151\n ExternalIP: 169.45.211.115\n Hostname: 10.170.219.151\nCapacity:\n cpu: 4\n ephemeral-storage: 102685624Ki\n hugepages-1Gi: 0\n hugepages-2Mi: 0\n memory: 16419940Ki\n pods: 110\nAllocatable:\n cpu: 3910m\n ephemeral-storage: 99892574949\n hugepages-1Gi: 0\n hugepages-2Mi: 0\n memory: 13627492Ki\n pods: 110\nSystem Info:\n Machine ID: b0d23c58eb874ec6b60fb1f5475b76c1\n System UUID: 8268FCE2-3882-6830-B330-6142D9A697FA\n Boot ID: 62457096-92c8-4668-8281-cd86fd53efd3\n Kernel Version: 4.15.0-47-generic\n OS Image: Ubuntu 18.04.2 LTS\n Operating System: linux\n Architecture: amd64\n Container Runtime Version: containerd://1.2.6\n Kubelet Version: v1.13.6+IKS\n Kube-Proxy Version: v1.13.6+IKS\nProviderID: ibm://d18c889395112a40d2f4e3065f237a7d///0ef1a5f1e33c44fe9c21650356d908ee/kube-wdc04-cr0ef1a5f1e33c44fe9c21650356d908ee-w3\nNon-terminated Pods: (14 in total)\n Namespace Name CPU Requests CPU Limits Memory Requests Memory Limits AGE\n --------- ---- ------------ ---------- --------------- ------------- ---\n heptio-sonobuoy sonobuoy-systemd-logs-daemon-set-8e5ce9bdf4a94ece-nlv42 0 (0%) 0 (0%) 0 (0%) 0 (0%) 20m\n kube-system calico-kube-controllers-7dd978d898-r8w74 10m (0%) 0 (0%) 25Mi (0%) 0 (0%) 162m\n kube-system calico-node-d9d6s 250m (6%) 0 (0%) 80Mi (0%) 0 (0%) 154m\n kube-system coredns-58d696879-rrkth 100m (2%) 0 (0%) 70Mi (0%) 400Mi (3%) 161m\n kube-system coredns-58d696879-zmj2g 100m (2%) 0 (0%) 70Mi (0%) 400Mi (3%) 153m\n kube-system coredns-autoscaler-64f9c5b4df-hzq57 20m (0%) 0 (0%) 10Mi (0%) 0 (0%) 161m\n kube-system ibm-file-plugin-7f6d8979bd-p5smn 50m (1%) 200m (5%) 100Mi (0%) 0 (0%) 160m\n kube-system ibm-keepalived-watcher-8wshf 5m (0%) 0 (0%) 10Mi (0%) 0 (0%) 154m\n kube-system ibm-kube-fluentd-6fw9c 25m (0%) 300m (7%) 150Mi (1%) 1600M (11%) 148m\n kube-system ibm-master-proxy-static-10.170.219.151 25m (0%) 300m (7%) 32M (0%) 512M (3%) 152m\n kube-system ibm-storage-watcher-845946d5b5-knxcm 50m (1%) 200m (5%) 100Mi (0%) 0 (0%) 160m\n kube-system kubernetes-dashboard-7996b848f4-mzbwj 50m (1%) 0 (0%) 100Mi (0%) 0 (0%) 157m\n kube-system public-cr0ef1a5f1e33c44fe9c21650356d908ee-alb1-8f59cf484-9v5pm 0 (0%) 0 (0%) 0 (0%) 0 (0%) 146m\n kube-system vpn-774cf5c6d4-6p6cr 5m (0%) 0 (0%) 5Mi (0%) 0 (0%) 157m\nAllocated resources:\n (Total limits may be over 100 percent, i.e., overcommitted.)\n Resource Requests Limits\n -------- -------- ------\n cpu 690m (17%) 1 (25%)\n memory 768530Ki (5%) 2881700Ki (21%)\n ephemeral-storage 0 (0%) 0 (0%)\nEvents: \n" -May 13 19:01:07.623: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 describe namespace e2e-tests-kubectl-dpvrq' -May 13 19:01:07.821: INFO: stderr: "" -May 13 19:01:07.821: INFO: stdout: "Name: e2e-tests-kubectl-dpvrq\nLabels: e2e-framework=kubectl\n e2e-run=ad7e8b22-75ae-11e9-8f67-2632f168be36\nAnnotations: \nStatus: Active\n\nNo resource quota.\n\nNo resource limits.\n" -[AfterEach] [sig-cli] Kubectl client - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 19:01:07.821: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-kubectl-dpvrq" for this suite. -May 13 19:01:31.858: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 19:01:32.046: INFO: namespace: e2e-tests-kubectl-dpvrq, resource: bindings, ignored listing per whitelist -May 13 19:01:32.249: INFO: namespace e2e-tests-kubectl-dpvrq deletion completed in 24.419041714s - -• [SLOW TEST:28.121 seconds] -[sig-cli] Kubectl client -/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:22 - [k8s.io] Kubectl describe - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 - should check if kubectl describe prints relevant information for rc and pods [Conformance] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSS +SSSSSSSSS ------------------------------ -[sig-apps] ReplicaSet - should adopt matching pods on creation and release no longer matching pods [Conformance] +[sig-node] Downward API + should provide default limits.cpu/memory from node allocatable [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -[BeforeEach] [sig-apps] ReplicaSet +[BeforeEach] [sig-node] Downward API /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client -May 13 19:01:32.250: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -STEP: Building a namespace api object, basename replicaset -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-replicaset-vvtcp +Jun 18 11:55:43.754: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 +STEP: Building a namespace api object, basename downward-api +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-downward-api-wh69j STEP: Waiting for a default service account to be provisioned in namespace -[It] should adopt matching pods on creation and release no longer matching pods [Conformance] +[It] should provide default limits.cpu/memory from node allocatable [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -STEP: Given a Pod with a 'name' label pod-adoption-release is created -STEP: When a replicaset with a matching selector is created -STEP: Then the orphan pod is adopted -STEP: When the matched label of one of its pods change -May 13 19:01:36.329: INFO: Pod name pod-adoption-release: Found 1 pods out of 1 -STEP: Then the pod is released -[AfterEach] [sig-apps] ReplicaSet +STEP: Creating a pod to test downward api env vars +Jun 18 11:55:44.238: INFO: Waiting up to 5m0s for pod "downward-api-00a5f64f-91c0-11e9-bce2-ae54e022189f" in namespace "e2e-tests-downward-api-wh69j" to be "success or failure" +Jun 18 11:55:44.256: INFO: Pod "downward-api-00a5f64f-91c0-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 17.595203ms +Jun 18 11:55:46.270: INFO: Pod "downward-api-00a5f64f-91c0-11e9-bce2-ae54e022189f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.032024221s +STEP: Saw pod success +Jun 18 11:55:46.270: INFO: Pod "downward-api-00a5f64f-91c0-11e9-bce2-ae54e022189f" satisfied condition "success or failure" +Jun 18 11:55:46.285: INFO: Trying to get logs from node 10.72.74.143 pod downward-api-00a5f64f-91c0-11e9-bce2-ae54e022189f container dapi-container: +STEP: delete the pod +Jun 18 11:55:46.366: INFO: Waiting for pod downward-api-00a5f64f-91c0-11e9-bce2-ae54e022189f to disappear +Jun 18 11:55:46.380: INFO: Pod downward-api-00a5f64f-91c0-11e9-bce2-ae54e022189f no longer exists +[AfterEach] [sig-node] Downward API /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 19:01:36.436: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-replicaset-vvtcp" for this suite. -May 13 19:02:00.472: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 19:02:00.555: INFO: namespace: e2e-tests-replicaset-vvtcp, resource: bindings, ignored listing per whitelist -May 13 19:02:00.710: INFO: namespace e2e-tests-replicaset-vvtcp deletion completed in 24.26539389s +Jun 18 11:55:46.380: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "e2e-tests-downward-api-wh69j" for this suite. +Jun 18 11:55:52.453: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 11:55:52.700: INFO: namespace: e2e-tests-downward-api-wh69j, resource: bindings, ignored listing per whitelist +Jun 18 11:55:52.950: INFO: namespace e2e-tests-downward-api-wh69j deletion completed in 6.550708025s -• [SLOW TEST:28.460 seconds] -[sig-apps] ReplicaSet -/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22 - should adopt matching pods on creation and release no longer matching pods [Conformance] +• [SLOW TEST:9.196 seconds] +[sig-node] Downward API +/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downward_api.go:38 + should provide default limits.cpu/memory from node allocatable [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSSSSSSSSSS ------------------------------ -[sig-apps] Deployment - deployment should support rollover [Conformance] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -[BeforeEach] [sig-apps] Deployment - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 -STEP: Creating a kubernetes client -May 13 19:02:00.711: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -STEP: Building a namespace api object, basename deployment -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-deployment-2fcc2 -STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [sig-apps] Deployment - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:65 -[It] deployment should support rollover [Conformance] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -May 13 19:02:01.161: INFO: Pod name rollover-pod: Found 0 pods out of 1 -May 13 19:02:06.182: INFO: Pod name rollover-pod: Found 1 pods out of 1 -STEP: ensuring each pod is running -May 13 19:02:06.182: INFO: Waiting for pods owned by replica set "test-rollover-controller" to become ready -May 13 19:02:08.191: INFO: Creating deployment "test-rollover-deployment" -May 13 19:02:08.203: INFO: Make sure deployment "test-rollover-deployment" performs scaling operations -May 13 19:02:10.214: INFO: Check revision of new replica set for deployment "test-rollover-deployment" -May 13 19:02:10.227: INFO: Ensure that both replica sets have 1 created replica -May 13 19:02:10.242: INFO: Rollover old replica sets for deployment "test-rollover-deployment" with new image update -May 13 19:02:10.254: INFO: Updating deployment test-rollover-deployment -May 13 19:02:10.254: INFO: Wait deployment "test-rollover-deployment" to be observed by the deployment controller -May 13 19:02:12.265: INFO: Wait for revision update of deployment "test-rollover-deployment" to 2 -May 13 19:02:12.281: INFO: Make sure deployment "test-rollover-deployment" is complete -May 13 19:02:12.294: INFO: all replica sets need to contain the pod-template-hash label -May 13 19:02:12.294: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:1, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63693370928, loc:(*time.Location)(0x7b33b80)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63693370928, loc:(*time.Location)(0x7b33b80)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63693370930, loc:(*time.Location)(0x7b33b80)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63693370928, loc:(*time.Location)(0x7b33b80)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-6b7f9d6597\" is progressing."}}, CollisionCount:(*int32)(nil)} -May 13 19:02:14.308: INFO: all replica sets need to contain the pod-template-hash label -May 13 19:02:14.308: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:2, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63693370928, loc:(*time.Location)(0x7b33b80)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63693370928, loc:(*time.Location)(0x7b33b80)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63693370933, loc:(*time.Location)(0x7b33b80)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63693370928, loc:(*time.Location)(0x7b33b80)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-6b7f9d6597\" is progressing."}}, CollisionCount:(*int32)(nil)} -May 13 19:02:16.324: INFO: all replica sets need to contain the pod-template-hash label -May 13 19:02:16.324: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:2, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63693370928, loc:(*time.Location)(0x7b33b80)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63693370928, loc:(*time.Location)(0x7b33b80)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63693370933, loc:(*time.Location)(0x7b33b80)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63693370928, loc:(*time.Location)(0x7b33b80)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-6b7f9d6597\" is progressing."}}, CollisionCount:(*int32)(nil)} -May 13 19:02:18.308: INFO: all replica sets need to contain the pod-template-hash label -May 13 19:02:18.308: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:2, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63693370928, loc:(*time.Location)(0x7b33b80)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63693370928, loc:(*time.Location)(0x7b33b80)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63693370933, loc:(*time.Location)(0x7b33b80)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63693370928, loc:(*time.Location)(0x7b33b80)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-6b7f9d6597\" is progressing."}}, CollisionCount:(*int32)(nil)} -May 13 19:02:20.308: INFO: all replica sets need to contain the pod-template-hash label -May 13 19:02:20.308: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:2, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63693370928, loc:(*time.Location)(0x7b33b80)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63693370928, loc:(*time.Location)(0x7b33b80)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63693370933, loc:(*time.Location)(0x7b33b80)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63693370928, loc:(*time.Location)(0x7b33b80)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-6b7f9d6597\" is progressing."}}, CollisionCount:(*int32)(nil)} -May 13 19:02:22.308: INFO: all replica sets need to contain the pod-template-hash label -May 13 19:02:22.309: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:2, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63693370928, loc:(*time.Location)(0x7b33b80)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63693370928, loc:(*time.Location)(0x7b33b80)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63693370933, loc:(*time.Location)(0x7b33b80)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63693370928, loc:(*time.Location)(0x7b33b80)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-6b7f9d6597\" is progressing."}}, CollisionCount:(*int32)(nil)} -May 13 19:02:24.333: INFO: -May 13 19:02:24.333: INFO: Ensure that both old replica sets have no replicas -[AfterEach] [sig-apps] Deployment - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:59 -May 13 19:02:24.356: INFO: Deployment "test-rollover-deployment": -&Deployment{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-rollover-deployment,GenerateName:,Namespace:e2e-tests-deployment-2fcc2,SelfLink:/apis/apps/v1/namespaces/e2e-tests-deployment-2fcc2/deployments/test-rollover-deployment,UID:9b0497b7-75b1-11e9-b786-da20024d205c,ResourceVersion:36591,Generation:2,CreationTimestamp:2019-05-13 19:02:08 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: rollover-pod,},Annotations:map[string]string{deployment.kubernetes.io/revision: 2,},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Spec:DeploymentSpec{Replicas:*1,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: rollover-pod,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: rollover-pod,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[],Containers:[{redis gcr.io/kubernetes-e2e-test-images/redis:1.0 [] [] [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,},},Strategy:DeploymentStrategy{Type:RollingUpdate,RollingUpdate:&RollingUpdateDeployment{MaxUnavailable:0,MaxSurge:1,},},MinReadySeconds:10,RevisionHistoryLimit:*10,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:2,Replicas:1,UpdatedReplicas:1,AvailableReplicas:1,UnavailableReplicas:0,Conditions:[{Available True 2019-05-13 19:02:08 +0000 UTC 2019-05-13 19:02:08 +0000 UTC MinimumReplicasAvailable Deployment has minimum availability.} {Progressing True 2019-05-13 19:02:23 +0000 UTC 2019-05-13 19:02:08 +0000 UTC NewReplicaSetAvailable ReplicaSet "test-rollover-deployment-6b7f9d6597" has successfully progressed.}],ReadyReplicas:1,CollisionCount:nil,},} - -May 13 19:02:24.364: INFO: New ReplicaSet "test-rollover-deployment-6b7f9d6597" of Deployment "test-rollover-deployment": -&ReplicaSet{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-rollover-deployment-6b7f9d6597,GenerateName:,Namespace:e2e-tests-deployment-2fcc2,SelfLink:/apis/apps/v1/namespaces/e2e-tests-deployment-2fcc2/replicasets/test-rollover-deployment-6b7f9d6597,UID:9c3f96c7-75b1-11e9-83a8-7e0242378207,ResourceVersion:36582,Generation:2,CreationTimestamp:2019-05-13 19:02:10 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: rollover-pod,pod-template-hash: 6b7f9d6597,},Annotations:map[string]string{deployment.kubernetes.io/desired-replicas: 1,deployment.kubernetes.io/max-replicas: 2,deployment.kubernetes.io/revision: 2,},OwnerReferences:[{apps/v1 Deployment test-rollover-deployment 9b0497b7-75b1-11e9-b786-da20024d205c 0xc000b999b7 0xc000b999b8}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:ReplicaSetSpec{Replicas:*1,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: rollover-pod,pod-template-hash: 6b7f9d6597,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: rollover-pod,pod-template-hash: 6b7f9d6597,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[],Containers:[{redis gcr.io/kubernetes-e2e-test-images/redis:1.0 [] [] [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,},},MinReadySeconds:10,},Status:ReplicaSetStatus{Replicas:1,FullyLabeledReplicas:1,ObservedGeneration:2,ReadyReplicas:1,AvailableReplicas:1,Conditions:[],},} -May 13 19:02:24.364: INFO: All old ReplicaSets of Deployment "test-rollover-deployment": -May 13 19:02:24.364: INFO: &ReplicaSet{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-rollover-controller,GenerateName:,Namespace:e2e-tests-deployment-2fcc2,SelfLink:/apis/apps/v1/namespaces/e2e-tests-deployment-2fcc2/replicasets/test-rollover-controller,UID:96d15d8f-75b1-11e9-b786-da20024d205c,ResourceVersion:36590,Generation:2,CreationTimestamp:2019-05-13 19:02:01 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: rollover-pod,pod: nginx,},Annotations:map[string]string{deployment.kubernetes.io/desired-replicas: 1,deployment.kubernetes.io/max-replicas: 2,},OwnerReferences:[{apps/v1 Deployment test-rollover-deployment 9b0497b7-75b1-11e9-b786-da20024d205c 0xc000b99827 0xc000b99828}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:ReplicaSetSpec{Replicas:*0,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: rollover-pod,pod: nginx,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: rollover-pod,pod: nginx,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,},},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:0,FullyLabeledReplicas:0,ObservedGeneration:2,ReadyReplicas:0,AvailableReplicas:0,Conditions:[],},} -May 13 19:02:24.364: INFO: &ReplicaSet{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-rollover-deployment-6586df867b,GenerateName:,Namespace:e2e-tests-deployment-2fcc2,SelfLink:/apis/apps/v1/namespaces/e2e-tests-deployment-2fcc2/replicasets/test-rollover-deployment-6586df867b,UID:9b09b286-75b1-11e9-83a8-7e0242378207,ResourceVersion:36543,Generation:2,CreationTimestamp:2019-05-13 19:02:08 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: rollover-pod,pod-template-hash: 6586df867b,},Annotations:map[string]string{deployment.kubernetes.io/desired-replicas: 1,deployment.kubernetes.io/max-replicas: 2,deployment.kubernetes.io/revision: 1,},OwnerReferences:[{apps/v1 Deployment test-rollover-deployment 9b0497b7-75b1-11e9-b786-da20024d205c 0xc000b998e7 0xc000b998e8}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:ReplicaSetSpec{Replicas:*0,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: rollover-pod,pod-template-hash: 6586df867b,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: rollover-pod,pod-template-hash: 6586df867b,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[],Containers:[{redis-slave gcr.io/google_samples/gb-redisslave:nonexistent [] [] [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,},},MinReadySeconds:10,},Status:ReplicaSetStatus{Replicas:0,FullyLabeledReplicas:0,ObservedGeneration:2,ReadyReplicas:0,AvailableReplicas:0,Conditions:[],},} -May 13 19:02:24.372: INFO: Pod "test-rollover-deployment-6b7f9d6597-csk2h" is available: -&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-rollover-deployment-6b7f9d6597-csk2h,GenerateName:test-rollover-deployment-6b7f9d6597-,Namespace:e2e-tests-deployment-2fcc2,SelfLink:/api/v1/namespaces/e2e-tests-deployment-2fcc2/pods/test-rollover-deployment-6b7f9d6597-csk2h,UID:9c46fce4-75b1-11e9-83a8-7e0242378207,ResourceVersion:36563,Generation:0,CreationTimestamp:2019-05-13 19:02:10 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: rollover-pod,pod-template-hash: 6b7f9d6597,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet test-rollover-deployment-6b7f9d6597 9c3f96c7-75b1-11e9-83a8-7e0242378207 0xc0013d0b77 0xc0013d0b78}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-zd9mx {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-zd9mx,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{redis gcr.io/kubernetes-e2e-test-images/redis:1.0 [] [] [] [] [] {map[] map[]} [{default-token-zd9mx true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.170.219.177,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc0013d0c00} {node.kubernetes.io/unreachable Exists NoExecute 0xc0013d0c20}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Running,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:02:10 +0000 UTC } {Ready True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:02:12 +0000 UTC } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:02:12 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:02:10 +0000 UTC }],Message:,Reason:,HostIP:10.170.219.177,PodIP:172.30.227.110,StartTime:2019-05-13 19:02:10 +0000 UTC,ContainerStatuses:[{redis {nil ContainerStateRunning{StartedAt:2019-05-13 19:02:12 +0000 UTC,} nil} {nil nil nil} true 0 gcr.io/kubernetes-e2e-test-images/redis:1.0 gcr.io/kubernetes-e2e-test-images/redis@sha256:af4748d1655c08dc54d4be5182135395db9ce87aba2d4699b26b14ae197c5830 containerd://0a4cd6e65077caab9430c6d8d7b75bdc3c155772e4bad9accb30615766b6ff81}],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} -[AfterEach] [sig-apps] Deployment - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 19:02:24.372: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-deployment-2fcc2" for this suite. -May 13 19:02:30.409: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 19:02:30.486: INFO: namespace: e2e-tests-deployment-2fcc2, resource: bindings, ignored listing per whitelist -May 13 19:02:30.645: INFO: namespace e2e-tests-deployment-2fcc2 deletion completed in 6.264249525s - -• [SLOW TEST:29.933 seconds] -[sig-apps] Deployment -/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22 - deployment should support rollover [Conformance] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------- -[sig-storage] EmptyDir wrapper volumes - should not cause race condition when used for configmaps [Serial] [Slow] [Conformance] +[k8s.io] [sig-node] PreStop + should call prestop when killing a pod [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -[BeforeEach] [sig-storage] EmptyDir wrapper volumes +[BeforeEach] [k8s.io] [sig-node] PreStop /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client -May 13 19:02:30.645: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -STEP: Building a namespace api object, basename emptydir-wrapper -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-emptydir-wrapper-l22w6 +Jun 18 11:55:52.950: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 +STEP: Building a namespace api object, basename prestop +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-prestop-267tt STEP: Waiting for a default service account to be provisioned in namespace -[It] should not cause race condition when used for configmaps [Serial] [Slow] [Conformance] +[It] should call prestop when killing a pod [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -STEP: Creating 50 configmaps -STEP: Creating RC which spawns configmap-volume pods -May 13 19:02:31.395: INFO: Pod name wrapped-volume-race-a8d4d8a6-75b1-11e9-8f67-2632f168be36: Found 0 pods out of 5 -May 13 19:02:36.409: INFO: Pod name wrapped-volume-race-a8d4d8a6-75b1-11e9-8f67-2632f168be36: Found 5 pods out of 5 -STEP: Ensuring each pod is running -STEP: deleting ReplicationController wrapped-volume-race-a8d4d8a6-75b1-11e9-8f67-2632f168be36 in namespace e2e-tests-emptydir-wrapper-l22w6, will wait for the garbage collector to delete the pods -May 13 19:05:31.416: INFO: Deleting ReplicationController wrapped-volume-race-a8d4d8a6-75b1-11e9-8f67-2632f168be36 took: 19.638259ms -May 13 19:05:31.516: INFO: Terminating ReplicationController wrapped-volume-race-a8d4d8a6-75b1-11e9-8f67-2632f168be36 pods took: 100.210253ms -STEP: Creating RC which spawns configmap-volume pods -May 13 19:06:14.961: INFO: Pod name wrapped-volume-race-2e137194-75b2-11e9-8f67-2632f168be36: Found 0 pods out of 5 -May 13 19:06:19.975: INFO: Pod name wrapped-volume-race-2e137194-75b2-11e9-8f67-2632f168be36: Found 5 pods out of 5 -STEP: Ensuring each pod is running -STEP: deleting ReplicationController wrapped-volume-race-2e137194-75b2-11e9-8f67-2632f168be36 in namespace e2e-tests-emptydir-wrapper-l22w6, will wait for the garbage collector to delete the pods -May 13 19:08:12.112: INFO: Deleting ReplicationController wrapped-volume-race-2e137194-75b2-11e9-8f67-2632f168be36 took: 21.487831ms -May 13 19:08:12.213: INFO: Terminating ReplicationController wrapped-volume-race-2e137194-75b2-11e9-8f67-2632f168be36 pods took: 100.243481ms -STEP: Creating RC which spawns configmap-volume pods -May 13 19:08:53.056: INFO: Pod name wrapped-volume-race-8c4f1ce3-75b2-11e9-8f67-2632f168be36: Found 0 pods out of 5 -May 13 19:08:58.071: INFO: Pod name wrapped-volume-race-8c4f1ce3-75b2-11e9-8f67-2632f168be36: Found 5 pods out of 5 -STEP: Ensuring each pod is running -STEP: deleting ReplicationController wrapped-volume-race-8c4f1ce3-75b2-11e9-8f67-2632f168be36 in namespace e2e-tests-emptydir-wrapper-l22w6, will wait for the garbage collector to delete the pods -May 13 19:10:50.258: INFO: Deleting ReplicationController wrapped-volume-race-8c4f1ce3-75b2-11e9-8f67-2632f168be36 took: 20.516259ms -May 13 19:10:50.358: INFO: Terminating ReplicationController wrapped-volume-race-8c4f1ce3-75b2-11e9-8f67-2632f168be36 pods took: 100.193356ms -STEP: Cleaning up the configMaps -[AfterEach] [sig-storage] EmptyDir wrapper volumes +STEP: Creating server pod server in namespace e2e-tests-prestop-267tt +STEP: Waiting for pods to come up. +STEP: Creating tester pod tester in namespace e2e-tests-prestop-267tt +STEP: Deleting pre-stop pod +Jun 18 11:56:08.648: INFO: Saw: { + "Hostname": "server", + "Sent": null, + "Received": { + "prestop": 1 + }, + "Errors": null, + "Log": [ + "default/nettest has 0 endpoints ([]), which is less than 8 as expected. Waiting for all endpoints to come up.", + "default/nettest has 0 endpoints ([]), which is less than 8 as expected. Waiting for all endpoints to come up.", + "default/nettest has 0 endpoints ([]), which is less than 8 as expected. Waiting for all endpoints to come up." + ], + "StillContactingPeers": true +} +STEP: Deleting the server pod +[AfterEach] [k8s.io] [sig-node] PreStop /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 19:11:33.777: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-emptydir-wrapper-l22w6" for this suite. -May 13 19:11:40.586: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 19:11:40.829: INFO: namespace: e2e-tests-emptydir-wrapper-l22w6, resource: bindings, ignored listing per whitelist -May 13 19:11:40.842: INFO: namespace e2e-tests-emptydir-wrapper-l22w6 deletion completed in 7.056672976s +Jun 18 11:56:08.684: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "e2e-tests-prestop-267tt" for this suite. +Jun 18 11:56:48.759: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 11:56:48.854: INFO: namespace: e2e-tests-prestop-267tt, resource: bindings, ignored listing per whitelist +Jun 18 11:56:49.279: INFO: namespace e2e-tests-prestop-267tt deletion completed in 40.569554785s -• [SLOW TEST:550.197 seconds] -[sig-storage] EmptyDir wrapper volumes -/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:22 - should not cause race condition when used for configmaps [Serial] [Slow] [Conformance] +• [SLOW TEST:56.329 seconds] +[k8s.io] [sig-node] PreStop +/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 + should call prestop when killing a pod [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ -SSSSSSSSSSS +S ------------------------------ -[sig-network] Proxy version v1 - should proxy through a service and a pod [Conformance] +[sig-storage] Downward API volume + should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -[BeforeEach] version v1 +[BeforeEach] [sig-storage] Downward API volume /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client -May 13 19:11:40.843: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -STEP: Building a namespace api object, basename proxy -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-proxy-nc6vk +Jun 18 11:56:49.279: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 +STEP: Building a namespace api object, basename downward-api +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-downward-api-749wd STEP: Waiting for a default service account to be provisioned in namespace -[It] should proxy through a service and a pod [Conformance] +[BeforeEach] [sig-storage] Downward API volume + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:39 +[It] should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -STEP: starting an echo server on multiple ports -STEP: creating replication controller proxy-service-fgbbn in namespace e2e-tests-proxy-nc6vk -I0513 19:11:41.145462 16 runners.go:184] Created replication controller with name: proxy-service-fgbbn, namespace: e2e-tests-proxy-nc6vk, replica count: 1 -I0513 19:11:42.195792 16 runners.go:184] proxy-service-fgbbn Pods: 1 out of 1 created, 0 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady -I0513 19:11:43.196066 16 runners.go:184] proxy-service-fgbbn Pods: 1 out of 1 created, 0 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady -I0513 19:11:44.196329 16 runners.go:184] proxy-service-fgbbn Pods: 1 out of 1 created, 0 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 1 runningButNotReady -I0513 19:11:45.196582 16 runners.go:184] proxy-service-fgbbn Pods: 1 out of 1 created, 0 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 1 runningButNotReady -I0513 19:11:46.196834 16 runners.go:184] proxy-service-fgbbn Pods: 1 out of 1 created, 0 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 1 runningButNotReady -I0513 19:11:47.197103 16 runners.go:184] proxy-service-fgbbn Pods: 1 out of 1 created, 0 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 1 runningButNotReady -I0513 19:11:48.197405 16 runners.go:184] proxy-service-fgbbn Pods: 1 out of 1 created, 0 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 1 runningButNotReady -I0513 19:11:49.197993 16 runners.go:184] proxy-service-fgbbn Pods: 1 out of 1 created, 0 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 1 runningButNotReady -I0513 19:11:50.198314 16 runners.go:184] proxy-service-fgbbn Pods: 1 out of 1 created, 0 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 1 runningButNotReady -I0513 19:11:51.198708 16 runners.go:184] proxy-service-fgbbn Pods: 1 out of 1 created, 0 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 1 runningButNotReady -I0513 19:11:52.199001 16 runners.go:184] proxy-service-fgbbn Pods: 1 out of 1 created, 0 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 1 runningButNotReady -I0513 19:11:53.199229 16 runners.go:184] proxy-service-fgbbn Pods: 1 out of 1 created, 1 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady -May 13 19:11:53.221: INFO: setup took 12.106728088s, starting test cases -STEP: running 16 cases, 20 attempts per case, 320 total attempts -May 13 19:11:53.320: INFO: (0) /api/v1/namespaces/e2e-tests-proxy-nc6vk/pods/http:proxy-service-fgbbn-5l2nv:1080/proxy/: +STEP: delete the pod +Jun 18 11:56:53.984: INFO: Waiting for pod downwardapi-volume-27b56a70-91c0-11e9-bce2-ae54e022189f to disappear +Jun 18 11:56:54.003: INFO: Pod downwardapi-volume-27b56a70-91c0-11e9-bce2-ae54e022189f no longer exists +[AfterEach] [sig-storage] Downward API volume /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 19:11:55.941: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-proxy-nc6vk" for this suite. -May 13 19:12:02.003: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 19:12:02.104: INFO: namespace: e2e-tests-proxy-nc6vk, resource: bindings, ignored listing per whitelist -May 13 19:12:02.241: INFO: namespace e2e-tests-proxy-nc6vk deletion completed in 6.289017561s +Jun 18 11:56:54.003: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "e2e-tests-downward-api-749wd" for this suite. +Jun 18 11:57:00.074: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 11:57:00.142: INFO: namespace: e2e-tests-downward-api-749wd, resource: bindings, ignored listing per whitelist +Jun 18 11:57:00.624: INFO: namespace e2e-tests-downward-api-749wd deletion completed in 6.599051909s -• [SLOW TEST:21.397 seconds] -[sig-network] Proxy -/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:22 - version v1 - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/proxy.go:56 - should proxy through a service and a pod [Conformance] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 +• [SLOW TEST:11.345 seconds] +[sig-storage] Downward API volume +/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:34 + should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance] + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ -SSSSS +SSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-network] Service endpoints latency - should not be very high [Conformance] +[sig-cli] Kubectl client [k8s.io] Kubectl version + should check is all data is printed [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -[BeforeEach] [sig-network] Service endpoints latency +[BeforeEach] [sig-cli] Kubectl client /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client -May 13 19:12:02.241: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -STEP: Building a namespace api object, basename svc-latency -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-svc-latency-lm64d +Jun 18 11:57:00.625: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 +STEP: Building a namespace api object, basename kubectl +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-kubectl-qc5bj STEP: Waiting for a default service account to be provisioned in namespace -[It] should not be very high [Conformance] +[BeforeEach] [sig-cli] Kubectl client + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:243 +[It] should check is all data is printed [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -STEP: creating replication controller svc-latency-rc in namespace e2e-tests-svc-latency-lm64d -I0513 19:12:02.549373 16 runners.go:184] Created replication controller with name: svc-latency-rc, namespace: e2e-tests-svc-latency-lm64d, replica count: 1 -I0513 19:12:03.599737 16 runners.go:184] svc-latency-rc Pods: 1 out of 1 created, 0 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady -I0513 19:12:04.599960 16 runners.go:184] svc-latency-rc Pods: 1 out of 1 created, 1 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady -May 13 19:12:04.722: INFO: Created: latency-svc-9wfh9 -May 13 19:12:04.754: INFO: Got endpoints: latency-svc-9wfh9 [54.430948ms] -May 13 19:12:04.840: INFO: Created: latency-svc-5rc9p -May 13 19:12:04.848: INFO: Got endpoints: latency-svc-5rc9p [93.182329ms] -May 13 19:12:04.854: INFO: Created: latency-svc-7l7gx -May 13 19:12:04.860: INFO: Got endpoints: latency-svc-7l7gx [105.120469ms] -May 13 19:12:04.867: INFO: Created: latency-svc-stjpd -May 13 19:12:04.873: INFO: Got endpoints: latency-svc-stjpd [118.943449ms] -May 13 19:12:04.879: INFO: Created: latency-svc-hnj9f -May 13 19:12:04.883: INFO: Got endpoints: latency-svc-hnj9f [128.62891ms] -May 13 19:12:04.891: INFO: Created: latency-svc-j92kd -May 13 19:12:04.896: INFO: Got endpoints: latency-svc-j92kd [141.22021ms] -May 13 19:12:04.902: INFO: Created: latency-svc-qwft2 -May 13 19:12:04.906: INFO: Got endpoints: latency-svc-qwft2 [151.824777ms] -May 13 19:12:04.914: INFO: Created: latency-svc-8tmv5 -May 13 19:12:04.920: INFO: Got endpoints: latency-svc-8tmv5 [165.236232ms] -May 13 19:12:04.926: INFO: Created: latency-svc-6hr4k -May 13 19:12:04.932: INFO: Got endpoints: latency-svc-6hr4k [177.608128ms] -May 13 19:12:04.938: INFO: Created: latency-svc-59m79 -May 13 19:12:04.944: INFO: Got endpoints: latency-svc-59m79 [189.479167ms] -May 13 19:12:04.949: INFO: Created: latency-svc-wz796 -May 13 19:12:04.955: INFO: Got endpoints: latency-svc-wz796 [200.047441ms] -May 13 19:12:04.961: INFO: Created: latency-svc-xv4pg -May 13 19:12:04.966: INFO: Got endpoints: latency-svc-xv4pg [211.718577ms] -May 13 19:12:04.973: INFO: Created: latency-svc-4c88h -May 13 19:12:04.980: INFO: Got endpoints: latency-svc-4c88h [225.314401ms] -May 13 19:12:04.989: INFO: Created: latency-svc-ll8tc -May 13 19:12:04.994: INFO: Got endpoints: latency-svc-ll8tc [240.074954ms] -May 13 19:12:05.003: INFO: Created: latency-svc-ck7md -May 13 19:12:05.010: INFO: Got endpoints: latency-svc-ck7md [255.017595ms] -May 13 19:12:05.018: INFO: Created: latency-svc-rfb5g -May 13 19:12:05.024: INFO: Got endpoints: latency-svc-rfb5g [269.920041ms] -May 13 19:12:05.031: INFO: Created: latency-svc-xzb6f -May 13 19:12:05.034: INFO: Got endpoints: latency-svc-xzb6f [186.552923ms] -May 13 19:12:05.042: INFO: Created: latency-svc-4s9ws -May 13 19:12:05.047: INFO: Got endpoints: latency-svc-4s9ws [187.156387ms] -May 13 19:12:05.055: INFO: Created: latency-svc-cndns -May 13 19:12:05.060: INFO: Got endpoints: latency-svc-cndns [186.69902ms] -May 13 19:12:05.077: INFO: Created: latency-svc-fc7gb -May 13 19:12:05.082: INFO: Got endpoints: latency-svc-fc7gb [198.976533ms] -May 13 19:12:05.092: INFO: Created: latency-svc-6hvvl -May 13 19:12:05.097: INFO: Got endpoints: latency-svc-6hvvl [201.201521ms] -May 13 19:12:05.104: INFO: Created: latency-svc-559tg -May 13 19:12:05.109: INFO: Got endpoints: latency-svc-559tg [203.181036ms] -May 13 19:12:05.116: INFO: Created: latency-svc-8q9qj -May 13 19:12:05.122: INFO: Got endpoints: latency-svc-8q9qj [201.928244ms] -May 13 19:12:05.127: INFO: Created: latency-svc-5tv89 -May 13 19:12:05.132: INFO: Got endpoints: latency-svc-5tv89 [200.418481ms] -May 13 19:12:05.138: INFO: Created: latency-svc-vp4ch -May 13 19:12:05.142: INFO: Got endpoints: latency-svc-vp4ch [197.680969ms] -May 13 19:12:05.149: INFO: Created: latency-svc-629f5 -May 13 19:12:05.156: INFO: Got endpoints: latency-svc-629f5 [201.015649ms] -May 13 19:12:05.161: INFO: Created: latency-svc-246v5 -May 13 19:12:05.166: INFO: Got endpoints: latency-svc-246v5 [199.891915ms] -May 13 19:12:05.171: INFO: Created: latency-svc-l5q5q -May 13 19:12:05.177: INFO: Got endpoints: latency-svc-l5q5q [196.98303ms] -May 13 19:12:05.182: INFO: Created: latency-svc-4zbgj -May 13 19:12:05.188: INFO: Got endpoints: latency-svc-4zbgj [193.03613ms] -May 13 19:12:05.194: INFO: Created: latency-svc-v4q7r -May 13 19:12:05.200: INFO: Got endpoints: latency-svc-v4q7r [190.876464ms] -May 13 19:12:05.207: INFO: Created: latency-svc-4c6l8 -May 13 19:12:05.212: INFO: Got endpoints: latency-svc-4c6l8 [187.863188ms] -May 13 19:12:05.219: INFO: Created: latency-svc-8ct5m -May 13 19:12:05.225: INFO: Got endpoints: latency-svc-8ct5m [190.934967ms] -May 13 19:12:05.230: INFO: Created: latency-svc-ljpk7 -May 13 19:12:05.235: INFO: Got endpoints: latency-svc-ljpk7 [188.468772ms] -May 13 19:12:05.242: INFO: Created: latency-svc-4ppw5 -May 13 19:12:05.247: INFO: Got endpoints: latency-svc-4ppw5 [186.89939ms] -May 13 19:12:05.254: INFO: Created: latency-svc-kv84f -May 13 19:12:05.260: INFO: Got endpoints: latency-svc-kv84f [177.532721ms] -May 13 19:12:05.268: INFO: Created: latency-svc-t4r2j -May 13 19:12:05.272: INFO: Got endpoints: latency-svc-t4r2j [175.455083ms] -May 13 19:12:05.277: INFO: Created: latency-svc-m8rkt -May 13 19:12:05.283: INFO: Got endpoints: latency-svc-m8rkt [173.032686ms] -May 13 19:12:05.289: INFO: Created: latency-svc-5d2px -May 13 19:12:05.295: INFO: Got endpoints: latency-svc-5d2px [173.750924ms] -May 13 19:12:05.300: INFO: Created: latency-svc-w2kf6 -May 13 19:12:05.308: INFO: Got endpoints: latency-svc-w2kf6 [175.963147ms] -May 13 19:12:05.312: INFO: Created: latency-svc-gcltr -May 13 19:12:05.317: INFO: Got endpoints: latency-svc-gcltr [175.346837ms] -May 13 19:12:05.322: INFO: Created: latency-svc-bm7fg -May 13 19:12:05.334: INFO: Created: latency-svc-q9nsc -May 13 19:12:05.349: INFO: Created: latency-svc-bh5b6 -May 13 19:12:05.353: INFO: Got endpoints: latency-svc-bm7fg [197.20046ms] -May 13 19:12:05.357: INFO: Created: latency-svc-t8shc -May 13 19:12:05.367: INFO: Created: latency-svc-kkk8q -May 13 19:12:05.379: INFO: Created: latency-svc-x7q7m -May 13 19:12:05.392: INFO: Created: latency-svc-p7mpm -May 13 19:12:05.397: INFO: Got endpoints: latency-svc-q9nsc [230.893183ms] -May 13 19:12:05.407: INFO: Created: latency-svc-w65bq -May 13 19:12:05.417: INFO: Created: latency-svc-k4rsc -May 13 19:12:05.429: INFO: Created: latency-svc-wx7ww -May 13 19:12:05.440: INFO: Created: latency-svc-dbh6l -May 13 19:12:05.469: INFO: Got endpoints: latency-svc-bh5b6 [292.124451ms] -May 13 19:12:05.476: INFO: Created: latency-svc-m82fz -May 13 19:12:05.488: INFO: Created: latency-svc-r9n9q -May 13 19:12:05.497: INFO: Got endpoints: latency-svc-t8shc [309.001007ms] -May 13 19:12:05.501: INFO: Created: latency-svc-fmgpp -May 13 19:12:05.513: INFO: Created: latency-svc-4gcj8 -May 13 19:12:05.528: INFO: Created: latency-svc-svv4h -May 13 19:12:05.539: INFO: Created: latency-svc-fglvj -May 13 19:12:05.547: INFO: Got endpoints: latency-svc-kkk8q [346.100178ms] -May 13 19:12:05.550: INFO: Created: latency-svc-l7wqp -May 13 19:12:05.562: INFO: Created: latency-svc-vnwb7 -May 13 19:12:05.573: INFO: Created: latency-svc-r5xnv -May 13 19:12:05.597: INFO: Got endpoints: latency-svc-x7q7m [385.232241ms] -May 13 19:12:05.618: INFO: Created: latency-svc-zqlxj -May 13 19:12:05.648: INFO: Got endpoints: latency-svc-p7mpm [422.454801ms] -May 13 19:12:05.669: INFO: Created: latency-svc-twncp -May 13 19:12:05.698: INFO: Got endpoints: latency-svc-w65bq [462.677975ms] -May 13 19:12:05.721: INFO: Created: latency-svc-h42fn -May 13 19:12:05.748: INFO: Got endpoints: latency-svc-k4rsc [500.653197ms] -May 13 19:12:05.769: INFO: Created: latency-svc-lk765 -May 13 19:12:05.797: INFO: Got endpoints: latency-svc-wx7ww [537.311488ms] -May 13 19:12:05.818: INFO: Created: latency-svc-wqp5k -May 13 19:12:05.847: INFO: Got endpoints: latency-svc-dbh6l [574.774522ms] -May 13 19:12:05.868: INFO: Created: latency-svc-h7h8t -May 13 19:12:05.898: INFO: Got endpoints: latency-svc-m82fz [614.824167ms] -May 13 19:12:05.920: INFO: Created: latency-svc-hxlgn -May 13 19:12:05.947: INFO: Got endpoints: latency-svc-r9n9q [651.890946ms] -May 13 19:12:05.971: INFO: Created: latency-svc-znqh9 -May 13 19:12:05.998: INFO: Got endpoints: latency-svc-fmgpp [689.368202ms] -May 13 19:12:06.020: INFO: Created: latency-svc-f84mr -May 13 19:12:06.047: INFO: Got endpoints: latency-svc-4gcj8 [730.19722ms] -May 13 19:12:06.068: INFO: Created: latency-svc-bhk8k -May 13 19:12:06.097: INFO: Got endpoints: latency-svc-svv4h [744.387609ms] -May 13 19:12:06.118: INFO: Created: latency-svc-9ckh6 -May 13 19:12:06.147: INFO: Got endpoints: latency-svc-fglvj [750.139233ms] -May 13 19:12:06.171: INFO: Created: latency-svc-xfv2z -May 13 19:12:06.198: INFO: Got endpoints: latency-svc-l7wqp [728.560479ms] -May 13 19:12:06.218: INFO: Created: latency-svc-c92jj -May 13 19:12:06.247: INFO: Got endpoints: latency-svc-vnwb7 [750.338318ms] -May 13 19:12:06.267: INFO: Created: latency-svc-vnnm5 -May 13 19:12:06.304: INFO: Got endpoints: latency-svc-r5xnv [757.381887ms] -May 13 19:12:06.325: INFO: Created: latency-svc-gfgrr -May 13 19:12:06.347: INFO: Got endpoints: latency-svc-zqlxj [749.833946ms] -May 13 19:12:06.370: INFO: Created: latency-svc-vthjm -May 13 19:12:06.397: INFO: Got endpoints: latency-svc-twncp [748.953994ms] -May 13 19:12:06.419: INFO: Created: latency-svc-fkl6w -May 13 19:12:06.447: INFO: Got endpoints: latency-svc-h42fn [749.352456ms] -May 13 19:12:06.469: INFO: Created: latency-svc-7d4sh -May 13 19:12:06.497: INFO: Got endpoints: latency-svc-lk765 [749.619808ms] -May 13 19:12:06.620: INFO: Created: latency-svc-kbpww -May 13 19:12:06.620: INFO: Got endpoints: latency-svc-h7h8t [772.306996ms] -May 13 19:12:06.620: INFO: Got endpoints: latency-svc-wqp5k [822.902659ms] -May 13 19:12:06.641: INFO: Created: latency-svc-66qj5 -May 13 19:12:06.647: INFO: Got endpoints: latency-svc-hxlgn [748.967596ms] -May 13 19:12:06.651: INFO: Created: latency-svc-zrdkc -May 13 19:12:06.669: INFO: Created: latency-svc-w5g8p -May 13 19:12:06.697: INFO: Got endpoints: latency-svc-znqh9 [750.094128ms] -May 13 19:12:06.723: INFO: Created: latency-svc-pvl28 -May 13 19:12:06.747: INFO: Got endpoints: latency-svc-f84mr [749.217811ms] -May 13 19:12:06.768: INFO: Created: latency-svc-g6bxr -May 13 19:12:06.797: INFO: Got endpoints: latency-svc-bhk8k [749.754345ms] -May 13 19:12:06.818: INFO: Created: latency-svc-w7mxq -May 13 19:12:06.846: INFO: Got endpoints: latency-svc-9ckh6 [749.121334ms] -May 13 19:12:06.867: INFO: Created: latency-svc-dppzq -May 13 19:12:06.897: INFO: Got endpoints: latency-svc-xfv2z [749.806697ms] -May 13 19:12:06.917: INFO: Created: latency-svc-4ddmw -May 13 19:12:06.947: INFO: Got endpoints: latency-svc-c92jj [749.168602ms] -May 13 19:12:06.967: INFO: Created: latency-svc-z4n28 -May 13 19:12:06.997: INFO: Got endpoints: latency-svc-vnnm5 [750.437879ms] -May 13 19:12:07.018: INFO: Created: latency-svc-pcs2s -May 13 19:12:07.055: INFO: Got endpoints: latency-svc-gfgrr [751.244567ms] -May 13 19:12:07.080: INFO: Created: latency-svc-zgvfm -May 13 19:12:07.098: INFO: Got endpoints: latency-svc-vthjm [750.672233ms] -May 13 19:12:07.118: INFO: Created: latency-svc-b9xg7 -May 13 19:12:07.147: INFO: Got endpoints: latency-svc-fkl6w [749.918235ms] -May 13 19:12:07.168: INFO: Created: latency-svc-l6q9r -May 13 19:12:07.198: INFO: Got endpoints: latency-svc-7d4sh [750.314986ms] -May 13 19:12:07.220: INFO: Created: latency-svc-srncs -May 13 19:12:07.248: INFO: Got endpoints: latency-svc-kbpww [750.139111ms] -May 13 19:12:07.270: INFO: Created: latency-svc-25lsz -May 13 19:12:07.297: INFO: Got endpoints: latency-svc-66qj5 [677.304818ms] -May 13 19:12:07.322: INFO: Created: latency-svc-4qmhg -May 13 19:12:07.349: INFO: Got endpoints: latency-svc-zrdkc [729.248662ms] -May 13 19:12:07.371: INFO: Created: latency-svc-zltst -May 13 19:12:07.398: INFO: Got endpoints: latency-svc-w5g8p [751.284444ms] -May 13 19:12:07.419: INFO: Created: latency-svc-bnhcn -May 13 19:12:07.447: INFO: Got endpoints: latency-svc-pvl28 [749.742941ms] -May 13 19:12:07.468: INFO: Created: latency-svc-7nk79 -May 13 19:12:07.504: INFO: Got endpoints: latency-svc-g6bxr [756.57464ms] -May 13 19:12:07.524: INFO: Created: latency-svc-2fns6 -May 13 19:12:07.547: INFO: Got endpoints: latency-svc-w7mxq [749.790882ms] -May 13 19:12:07.568: INFO: Created: latency-svc-qzw2j -May 13 19:12:07.597: INFO: Got endpoints: latency-svc-dppzq [750.695128ms] -May 13 19:12:07.618: INFO: Created: latency-svc-5n6nx -May 13 19:12:07.647: INFO: Got endpoints: latency-svc-4ddmw [750.358416ms] -May 13 19:12:07.677: INFO: Created: latency-svc-cvzqf -May 13 19:12:07.697: INFO: Got endpoints: latency-svc-z4n28 [750.548199ms] -May 13 19:12:07.719: INFO: Created: latency-svc-8s55l -May 13 19:12:07.747: INFO: Got endpoints: latency-svc-pcs2s [749.422726ms] -May 13 19:12:07.768: INFO: Created: latency-svc-lhrrp -May 13 19:12:07.797: INFO: Got endpoints: latency-svc-zgvfm [741.70654ms] -May 13 19:12:07.818: INFO: Created: latency-svc-2hddk -May 13 19:12:07.847: INFO: Got endpoints: latency-svc-b9xg7 [749.240659ms] -May 13 19:12:07.883: INFO: Created: latency-svc-df54r -May 13 19:12:07.897: INFO: Got endpoints: latency-svc-l6q9r [750.336588ms] -May 13 19:12:07.918: INFO: Created: latency-svc-xxt28 -May 13 19:12:07.948: INFO: Got endpoints: latency-svc-srncs [749.762218ms] -May 13 19:12:07.969: INFO: Created: latency-svc-c9zhr -May 13 19:12:07.998: INFO: Got endpoints: latency-svc-25lsz [750.257493ms] -May 13 19:12:08.019: INFO: Created: latency-svc-5grsf -May 13 19:12:08.047: INFO: Got endpoints: latency-svc-4qmhg [749.937568ms] -May 13 19:12:08.068: INFO: Created: latency-svc-j2854 -May 13 19:12:08.097: INFO: Got endpoints: latency-svc-zltst [748.103666ms] -May 13 19:12:08.119: INFO: Created: latency-svc-gdpmr -May 13 19:12:08.148: INFO: Got endpoints: latency-svc-bnhcn [749.653371ms] -May 13 19:12:08.169: INFO: Created: latency-svc-dvz5b -May 13 19:12:08.198: INFO: Got endpoints: latency-svc-7nk79 [750.352207ms] -May 13 19:12:08.218: INFO: Created: latency-svc-mcmc4 -May 13 19:12:08.247: INFO: Got endpoints: latency-svc-2fns6 [743.502942ms] -May 13 19:12:08.280: INFO: Created: latency-svc-h25jt -May 13 19:12:08.297: INFO: Got endpoints: latency-svc-qzw2j [750.164014ms] -May 13 19:12:08.317: INFO: Created: latency-svc-jnjqw -May 13 19:12:08.347: INFO: Got endpoints: latency-svc-5n6nx [749.918849ms] -May 13 19:12:08.367: INFO: Created: latency-svc-ddxwb -May 13 19:12:08.397: INFO: Got endpoints: latency-svc-cvzqf [750.119048ms] -May 13 19:12:08.418: INFO: Created: latency-svc-dwtvg -May 13 19:12:08.448: INFO: Got endpoints: latency-svc-8s55l [750.576956ms] -May 13 19:12:08.468: INFO: Created: latency-svc-z6hqb -May 13 19:12:08.497: INFO: Got endpoints: latency-svc-lhrrp [750.349953ms] -May 13 19:12:08.517: INFO: Created: latency-svc-jbsbk -May 13 19:12:08.548: INFO: Got endpoints: latency-svc-2hddk [750.570451ms] -May 13 19:12:08.582: INFO: Created: latency-svc-7bfp2 -May 13 19:12:08.597: INFO: Got endpoints: latency-svc-df54r [749.626827ms] -May 13 19:12:08.620: INFO: Created: latency-svc-8s67v -May 13 19:12:08.647: INFO: Got endpoints: latency-svc-xxt28 [749.503148ms] -May 13 19:12:08.667: INFO: Created: latency-svc-d29bv -May 13 19:12:08.697: INFO: Got endpoints: latency-svc-c9zhr [749.008811ms] -May 13 19:12:08.718: INFO: Created: latency-svc-zjp67 -May 13 19:12:08.747: INFO: Got endpoints: latency-svc-5grsf [749.182556ms] -May 13 19:12:08.767: INFO: Created: latency-svc-f7n8j -May 13 19:12:08.798: INFO: Got endpoints: latency-svc-j2854 [750.286529ms] -May 13 19:12:08.818: INFO: Created: latency-svc-2ktcq -May 13 19:12:08.847: INFO: Got endpoints: latency-svc-gdpmr [749.883903ms] -May 13 19:12:08.867: INFO: Created: latency-svc-g8992 -May 13 19:12:08.899: INFO: Got endpoints: latency-svc-dvz5b [751.585456ms] -May 13 19:12:08.920: INFO: Created: latency-svc-zl99z -May 13 19:12:08.947: INFO: Got endpoints: latency-svc-mcmc4 [749.415971ms] -May 13 19:12:08.969: INFO: Created: latency-svc-47qqt -May 13 19:12:09.000: INFO: Got endpoints: latency-svc-h25jt [752.467286ms] -May 13 19:12:09.020: INFO: Created: latency-svc-mlw4n -May 13 19:12:09.048: INFO: Got endpoints: latency-svc-jnjqw [750.253194ms] -May 13 19:12:09.071: INFO: Created: latency-svc-lfh9t -May 13 19:12:09.099: INFO: Got endpoints: latency-svc-ddxwb [751.655091ms] -May 13 19:12:09.122: INFO: Created: latency-svc-6r8v9 -May 13 19:12:09.148: INFO: Got endpoints: latency-svc-dwtvg [750.30322ms] -May 13 19:12:09.170: INFO: Created: latency-svc-jcgcs -May 13 19:12:09.197: INFO: Got endpoints: latency-svc-z6hqb [749.215224ms] -May 13 19:12:09.217: INFO: Created: latency-svc-5hrxs -May 13 19:12:09.248: INFO: Got endpoints: latency-svc-jbsbk [750.536001ms] -May 13 19:12:09.272: INFO: Created: latency-svc-7cvfp -May 13 19:12:09.297: INFO: Got endpoints: latency-svc-7bfp2 [738.07179ms] -May 13 19:12:09.318: INFO: Created: latency-svc-s6xnj -May 13 19:12:09.347: INFO: Got endpoints: latency-svc-8s67v [750.287653ms] -May 13 19:12:09.368: INFO: Created: latency-svc-vrjpt -May 13 19:12:09.398: INFO: Got endpoints: latency-svc-d29bv [751.616266ms] -May 13 19:12:09.421: INFO: Created: latency-svc-qsg6s -May 13 19:12:09.447: INFO: Got endpoints: latency-svc-zjp67 [750.519432ms] -May 13 19:12:09.520: INFO: Got endpoints: latency-svc-f7n8j [772.89061ms] -May 13 19:12:09.520: INFO: Created: latency-svc-pfjwf -May 13 19:12:09.543: INFO: Created: latency-svc-2tsdm -May 13 19:12:09.547: INFO: Got endpoints: latency-svc-2ktcq [749.374254ms] -May 13 19:12:09.567: INFO: Created: latency-svc-qlrgt -May 13 19:12:09.597: INFO: Got endpoints: latency-svc-g8992 [749.889159ms] -May 13 19:12:09.617: INFO: Created: latency-svc-lsj4b -May 13 19:12:09.647: INFO: Got endpoints: latency-svc-zl99z [747.850698ms] -May 13 19:12:09.668: INFO: Created: latency-svc-j96w6 -May 13 19:12:09.697: INFO: Got endpoints: latency-svc-47qqt [749.63925ms] -May 13 19:12:09.717: INFO: Created: latency-svc-26s65 -May 13 19:12:09.747: INFO: Got endpoints: latency-svc-mlw4n [747.311495ms] -May 13 19:12:09.767: INFO: Created: latency-svc-gdq8x -May 13 19:12:09.796: INFO: Got endpoints: latency-svc-lfh9t [748.889644ms] -May 13 19:12:09.818: INFO: Created: latency-svc-fgmbw -May 13 19:12:09.847: INFO: Got endpoints: latency-svc-6r8v9 [748.045828ms] -May 13 19:12:09.868: INFO: Created: latency-svc-g2rm9 -May 13 19:12:09.897: INFO: Got endpoints: latency-svc-jcgcs [749.159104ms] -May 13 19:12:09.918: INFO: Created: latency-svc-4246w -May 13 19:12:09.947: INFO: Got endpoints: latency-svc-5hrxs [749.495611ms] -May 13 19:12:09.968: INFO: Created: latency-svc-6nbvm -May 13 19:12:09.997: INFO: Got endpoints: latency-svc-7cvfp [749.093257ms] -May 13 19:12:10.018: INFO: Created: latency-svc-ft2pl -May 13 19:12:10.048: INFO: Got endpoints: latency-svc-s6xnj [750.664178ms] -May 13 19:12:10.070: INFO: Created: latency-svc-5cwhz -May 13 19:12:10.098: INFO: Got endpoints: latency-svc-vrjpt [750.383633ms] -May 13 19:12:10.119: INFO: Created: latency-svc-t67k7 -May 13 19:12:10.147: INFO: Got endpoints: latency-svc-qsg6s [748.139356ms] -May 13 19:12:10.167: INFO: Created: latency-svc-45dfz -May 13 19:12:10.197: INFO: Got endpoints: latency-svc-pfjwf [749.949964ms] -May 13 19:12:10.217: INFO: Created: latency-svc-kgz6j -May 13 19:12:10.247: INFO: Got endpoints: latency-svc-2tsdm [726.448266ms] -May 13 19:12:10.268: INFO: Created: latency-svc-25ptq -May 13 19:12:10.297: INFO: Got endpoints: latency-svc-qlrgt [750.135752ms] -May 13 19:12:10.318: INFO: Created: latency-svc-s6lgw -May 13 19:12:10.347: INFO: Got endpoints: latency-svc-lsj4b [749.642484ms] -May 13 19:12:10.368: INFO: Created: latency-svc-72hdd -May 13 19:12:10.397: INFO: Got endpoints: latency-svc-j96w6 [749.879655ms] -May 13 19:12:10.418: INFO: Created: latency-svc-nvlbd -May 13 19:12:10.447: INFO: Got endpoints: latency-svc-26s65 [750.260674ms] -May 13 19:12:10.468: INFO: Created: latency-svc-tx6br -May 13 19:12:10.497: INFO: Got endpoints: latency-svc-gdq8x [750.055439ms] -May 13 19:12:10.520: INFO: Created: latency-svc-s4prc -May 13 19:12:10.547: INFO: Got endpoints: latency-svc-fgmbw [750.110261ms] -May 13 19:12:10.567: INFO: Created: latency-svc-9xt8q -May 13 19:12:10.597: INFO: Got endpoints: latency-svc-g2rm9 [749.846539ms] -May 13 19:12:10.617: INFO: Created: latency-svc-njhjf -May 13 19:12:10.647: INFO: Got endpoints: latency-svc-4246w [750.112283ms] -May 13 19:12:10.667: INFO: Created: latency-svc-gqs9d -May 13 19:12:10.697: INFO: Got endpoints: latency-svc-6nbvm [750.078413ms] -May 13 19:12:10.717: INFO: Created: latency-svc-69gnk -May 13 19:12:10.747: INFO: Got endpoints: latency-svc-ft2pl [749.977468ms] -May 13 19:12:10.770: INFO: Created: latency-svc-q292n -May 13 19:12:10.797: INFO: Got endpoints: latency-svc-5cwhz [748.277529ms] -May 13 19:12:10.823: INFO: Created: latency-svc-gchln -May 13 19:12:10.847: INFO: Got endpoints: latency-svc-t67k7 [749.60842ms] -May 13 19:12:10.901: INFO: Created: latency-svc-wr85z -May 13 19:12:10.901: INFO: Got endpoints: latency-svc-45dfz [754.458337ms] -May 13 19:12:10.922: INFO: Created: latency-svc-jfz6g -May 13 19:12:10.947: INFO: Got endpoints: latency-svc-kgz6j [749.970926ms] -May 13 19:12:10.967: INFO: Created: latency-svc-vn7sm -May 13 19:12:10.997: INFO: Got endpoints: latency-svc-25ptq [750.614375ms] -May 13 19:12:11.017: INFO: Created: latency-svc-btdpb -May 13 19:12:11.047: INFO: Got endpoints: latency-svc-s6lgw [749.687666ms] -May 13 19:12:11.072: INFO: Created: latency-svc-fzqzj -May 13 19:12:11.098: INFO: Got endpoints: latency-svc-72hdd [750.855789ms] -May 13 19:12:11.120: INFO: Created: latency-svc-485q4 -May 13 19:12:11.148: INFO: Got endpoints: latency-svc-nvlbd [750.676929ms] -May 13 19:12:11.170: INFO: Created: latency-svc-p82gb -May 13 19:12:11.197: INFO: Got endpoints: latency-svc-tx6br [749.907462ms] -May 13 19:12:11.422: INFO: Got endpoints: latency-svc-9xt8q [875.357434ms] -May 13 19:12:11.422: INFO: Got endpoints: latency-svc-njhjf [825.297835ms] -May 13 19:12:11.422: INFO: Got endpoints: latency-svc-s4prc [925.098829ms] -May 13 19:12:11.520: INFO: Got endpoints: latency-svc-q292n [772.923033ms] -May 13 19:12:11.520: INFO: Got endpoints: latency-svc-69gnk [823.354169ms] -May 13 19:12:11.520: INFO: Got endpoints: latency-svc-gqs9d [873.241625ms] -May 13 19:12:11.520: INFO: Created: latency-svc-cg2gz -May 13 19:12:11.541: INFO: Created: latency-svc-nf75m -May 13 19:12:11.547: INFO: Got endpoints: latency-svc-gchln [750.661384ms] -May 13 19:12:11.553: INFO: Created: latency-svc-t7dvc -May 13 19:12:11.564: INFO: Created: latency-svc-nb7c4 -May 13 19:12:11.577: INFO: Created: latency-svc-64qwt -May 13 19:12:11.589: INFO: Created: latency-svc-zthhj -May 13 19:12:11.597: INFO: Got endpoints: latency-svc-wr85z [749.749073ms] -May 13 19:12:11.601: INFO: Created: latency-svc-jcz2b -May 13 19:12:11.613: INFO: Created: latency-svc-5h2hk -May 13 19:12:11.625: INFO: Created: latency-svc-z6t9l -May 13 19:12:11.647: INFO: Got endpoints: latency-svc-jfz6g [746.32362ms] -May 13 19:12:11.668: INFO: Created: latency-svc-qrktn -May 13 19:12:11.697: INFO: Got endpoints: latency-svc-vn7sm [749.571474ms] -May 13 19:12:11.718: INFO: Created: latency-svc-9z26w -May 13 19:12:11.748: INFO: Got endpoints: latency-svc-btdpb [750.27264ms] -May 13 19:12:11.771: INFO: Created: latency-svc-mxf9h -May 13 19:12:11.797: INFO: Got endpoints: latency-svc-fzqzj [749.705547ms] -May 13 19:12:11.818: INFO: Created: latency-svc-qmx9k -May 13 19:12:11.848: INFO: Got endpoints: latency-svc-485q4 [750.814847ms] -May 13 19:12:11.873: INFO: Created: latency-svc-d9g55 -May 13 19:12:11.897: INFO: Got endpoints: latency-svc-p82gb [749.228933ms] -May 13 19:12:11.917: INFO: Created: latency-svc-wzck9 -May 13 19:12:11.947: INFO: Got endpoints: latency-svc-cg2gz [749.378514ms] -May 13 19:12:11.968: INFO: Created: latency-svc-cnxh7 -May 13 19:12:11.997: INFO: Got endpoints: latency-svc-nf75m [574.251501ms] -May 13 19:12:12.023: INFO: Created: latency-svc-cpvkk -May 13 19:12:12.047: INFO: Got endpoints: latency-svc-t7dvc [624.453951ms] -May 13 19:12:12.067: INFO: Created: latency-svc-fknxm -May 13 19:12:12.097: INFO: Got endpoints: latency-svc-nb7c4 [674.629377ms] -May 13 19:12:12.117: INFO: Created: latency-svc-9654c -May 13 19:12:12.147: INFO: Got endpoints: latency-svc-64qwt [626.798036ms] -May 13 19:12:12.168: INFO: Created: latency-svc-jvjs4 -May 13 19:12:12.197: INFO: Got endpoints: latency-svc-zthhj [676.600975ms] -May 13 19:12:12.217: INFO: Created: latency-svc-6m8gd -May 13 19:12:12.247: INFO: Got endpoints: latency-svc-jcz2b [726.709805ms] -May 13 19:12:12.267: INFO: Created: latency-svc-278w7 -May 13 19:12:12.297: INFO: Got endpoints: latency-svc-5h2hk [749.453563ms] -May 13 19:12:12.318: INFO: Created: latency-svc-nsznp -May 13 19:12:12.347: INFO: Got endpoints: latency-svc-z6t9l [750.34854ms] -May 13 19:12:12.368: INFO: Created: latency-svc-h9kgl -May 13 19:12:12.398: INFO: Got endpoints: latency-svc-qrktn [749.975497ms] -May 13 19:12:12.420: INFO: Created: latency-svc-4bzhb -May 13 19:12:12.447: INFO: Got endpoints: latency-svc-9z26w [749.785348ms] -May 13 19:12:12.469: INFO: Created: latency-svc-xcvz8 -May 13 19:12:12.497: INFO: Got endpoints: latency-svc-mxf9h [749.234483ms] -May 13 19:12:12.518: INFO: Created: latency-svc-42f4h -May 13 19:12:12.547: INFO: Got endpoints: latency-svc-qmx9k [749.723623ms] -May 13 19:12:12.567: INFO: Created: latency-svc-pwwr8 -May 13 19:12:12.597: INFO: Got endpoints: latency-svc-d9g55 [748.742584ms] -May 13 19:12:12.618: INFO: Created: latency-svc-dq4j5 -May 13 19:12:12.647: INFO: Got endpoints: latency-svc-wzck9 [750.089914ms] -May 13 19:12:12.697: INFO: Got endpoints: latency-svc-cnxh7 [750.340779ms] -May 13 19:12:12.747: INFO: Got endpoints: latency-svc-cpvkk [750.168318ms] -May 13 19:12:12.820: INFO: Got endpoints: latency-svc-fknxm [772.72133ms] -May 13 19:12:12.847: INFO: Got endpoints: latency-svc-9654c [750.025009ms] -May 13 19:12:12.897: INFO: Got endpoints: latency-svc-jvjs4 [750.278003ms] -May 13 19:12:12.948: INFO: Got endpoints: latency-svc-6m8gd [750.511521ms] -May 13 19:12:12.997: INFO: Got endpoints: latency-svc-278w7 [750.12643ms] -May 13 19:12:13.049: INFO: Got endpoints: latency-svc-nsznp [751.465069ms] -May 13 19:12:13.120: INFO: Got endpoints: latency-svc-h9kgl [772.338833ms] -May 13 19:12:13.148: INFO: Got endpoints: latency-svc-4bzhb [750.827965ms] -May 13 19:12:13.197: INFO: Got endpoints: latency-svc-xcvz8 [750.420813ms] -May 13 19:12:13.248: INFO: Got endpoints: latency-svc-42f4h [750.824723ms] -May 13 19:12:13.298: INFO: Got endpoints: latency-svc-pwwr8 [751.08869ms] -May 13 19:12:13.347: INFO: Got endpoints: latency-svc-dq4j5 [749.796745ms] -May 13 19:12:13.347: INFO: Latencies: [93.182329ms 105.120469ms 118.943449ms 128.62891ms 141.22021ms 151.824777ms 165.236232ms 173.032686ms 173.750924ms 175.346837ms 175.455083ms 175.963147ms 177.532721ms 177.608128ms 186.552923ms 186.69902ms 186.89939ms 187.156387ms 187.863188ms 188.468772ms 189.479167ms 190.876464ms 190.934967ms 193.03613ms 196.98303ms 197.20046ms 197.680969ms 198.976533ms 199.891915ms 200.047441ms 200.418481ms 201.015649ms 201.201521ms 201.928244ms 203.181036ms 211.718577ms 225.314401ms 230.893183ms 240.074954ms 255.017595ms 269.920041ms 292.124451ms 309.001007ms 346.100178ms 385.232241ms 422.454801ms 462.677975ms 500.653197ms 537.311488ms 574.251501ms 574.774522ms 614.824167ms 624.453951ms 626.798036ms 651.890946ms 674.629377ms 676.600975ms 677.304818ms 689.368202ms 726.448266ms 726.709805ms 728.560479ms 729.248662ms 730.19722ms 738.07179ms 741.70654ms 743.502942ms 744.387609ms 746.32362ms 747.311495ms 747.850698ms 748.045828ms 748.103666ms 748.139356ms 748.277529ms 748.742584ms 748.889644ms 748.953994ms 748.967596ms 749.008811ms 749.093257ms 749.121334ms 749.159104ms 749.168602ms 749.182556ms 749.215224ms 749.217811ms 749.228933ms 749.234483ms 749.240659ms 749.352456ms 749.374254ms 749.378514ms 749.415971ms 749.422726ms 749.453563ms 749.495611ms 749.503148ms 749.571474ms 749.60842ms 749.619808ms 749.626827ms 749.63925ms 749.642484ms 749.653371ms 749.687666ms 749.705547ms 749.723623ms 749.742941ms 749.749073ms 749.754345ms 749.762218ms 749.785348ms 749.790882ms 749.796745ms 749.806697ms 749.833946ms 749.846539ms 749.879655ms 749.883903ms 749.889159ms 749.907462ms 749.918235ms 749.918849ms 749.937568ms 749.949964ms 749.970926ms 749.975497ms 749.977468ms 750.025009ms 750.055439ms 750.078413ms 750.089914ms 750.094128ms 750.110261ms 750.112283ms 750.119048ms 750.12643ms 750.135752ms 750.139111ms 750.139233ms 750.164014ms 750.168318ms 750.253194ms 750.257493ms 750.260674ms 750.27264ms 750.278003ms 750.286529ms 750.287653ms 750.30322ms 750.314986ms 750.336588ms 750.338318ms 750.340779ms 750.34854ms 750.349953ms 750.352207ms 750.358416ms 750.383633ms 750.420813ms 750.437879ms 750.511521ms 750.519432ms 750.536001ms 750.548199ms 750.570451ms 750.576956ms 750.614375ms 750.661384ms 750.664178ms 750.672233ms 750.676929ms 750.695128ms 750.814847ms 750.824723ms 750.827965ms 750.855789ms 751.08869ms 751.244567ms 751.284444ms 751.465069ms 751.585456ms 751.616266ms 751.655091ms 752.467286ms 754.458337ms 756.57464ms 757.381887ms 772.306996ms 772.338833ms 772.72133ms 772.89061ms 772.923033ms 822.902659ms 823.354169ms 825.297835ms 873.241625ms 875.357434ms 925.098829ms] -May 13 19:12:13.347: INFO: 50 %ile: 749.619808ms -May 13 19:12:13.347: INFO: 90 %ile: 751.284444ms -May 13 19:12:13.347: INFO: 99 %ile: 875.357434ms -May 13 19:12:13.347: INFO: Total sample count: 200 -[AfterEach] [sig-network] Service endpoints latency +Jun 18 11:57:01.066: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 version' +Jun 18 11:57:01.196: INFO: stderr: "" +Jun 18 11:57:01.196: INFO: stdout: "Client Version: version.Info{Major:\"1\", Minor:\"13\", GitVersion:\"v1.13.0\", GitCommit:\"ddf47ac13c1a9483ea035a79cd7c10005ff21a6d\", GitTreeState:\"clean\", BuildDate:\"2018-12-03T21:04:45Z\", GoVersion:\"go1.11.2\", Compiler:\"gc\", Platform:\"linux/amd64\"}\nServer Version: version.Info{Major:\"1\", Minor:\"13\", GitVersion:\"v1.13.7+IKS\", GitCommit:\"675df39b011fd4f4f54aa131d903bab685cde6b8\", GitTreeState:\"clean\", BuildDate:\"2019-06-10T19:51:36Z\", GoVersion:\"go1.11.5\", Compiler:\"gc\", Platform:\"linux/amd64\"}\n" +[AfterEach] [sig-cli] Kubectl client /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 19:12:13.348: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-svc-latency-lm64d" for this suite. -May 13 19:12:29.457: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 19:12:29.543: INFO: namespace: e2e-tests-svc-latency-lm64d, resource: bindings, ignored listing per whitelist -May 13 19:12:29.742: INFO: namespace e2e-tests-svc-latency-lm64d deletion completed in 16.314634404s +Jun 18 11:57:01.196: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "e2e-tests-kubectl-qc5bj" for this suite. +Jun 18 11:57:07.265: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 11:57:08.233: INFO: namespace: e2e-tests-kubectl-qc5bj, resource: bindings, ignored listing per whitelist +Jun 18 11:57:08.545: INFO: namespace e2e-tests-kubectl-qc5bj deletion completed in 7.330865465s -• [SLOW TEST:27.501 seconds] -[sig-network] Service endpoints latency -/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:22 - should not be very high [Conformance] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 +• [SLOW TEST:7.920 seconds] +[sig-cli] Kubectl client +/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:22 + [k8s.io] Kubectl version + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 + should check is all data is printed [Conformance] + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ -SSS +SSSSSS ------------------------------ -[sig-node] Downward API - should provide pod UID as env vars [NodeConformance] [Conformance] +[sig-storage] ConfigMap + should be consumable from pods in volume with mappings and Item mode set [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -[BeforeEach] [sig-node] Downward API +[BeforeEach] [sig-storage] ConfigMap /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client -May 13 19:12:29.742: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -STEP: Building a namespace api object, basename downward-api -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-downward-api-j58m5 +Jun 18 11:57:08.545: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 +STEP: Building a namespace api object, basename configmap +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-configmap-spkd5 STEP: Waiting for a default service account to be provisioned in namespace -[It] should provide pod UID as env vars [NodeConformance] [Conformance] +[It] should be consumable from pods in volume with mappings and Item mode set [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -STEP: Creating a pod to test downward api env vars -May 13 19:12:30.073: INFO: Waiting up to 5m0s for pod "downward-api-0dad88c0-75b3-11e9-8f67-2632f168be36" in namespace "e2e-tests-downward-api-j58m5" to be "success or failure" -May 13 19:12:30.081: INFO: Pod "downward-api-0dad88c0-75b3-11e9-8f67-2632f168be36": Phase="Pending", Reason="", readiness=false. Elapsed: 7.401613ms -May 13 19:12:32.089: INFO: Pod "downward-api-0dad88c0-75b3-11e9-8f67-2632f168be36": Phase="Pending", Reason="", readiness=false. Elapsed: 2.015270117s -May 13 19:12:34.098: INFO: Pod "downward-api-0dad88c0-75b3-11e9-8f67-2632f168be36": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.024297872s +STEP: Creating configMap with name configmap-test-volume-map-332b67b1-91c0-11e9-bce2-ae54e022189f +STEP: Creating a pod to test consume configMaps +Jun 18 11:57:09.013: INFO: Waiting up to 5m0s for pod "pod-configmaps-332d94b4-91c0-11e9-bce2-ae54e022189f" in namespace "e2e-tests-configmap-spkd5" to be "success or failure" +Jun 18 11:57:09.028: INFO: Pod "pod-configmaps-332d94b4-91c0-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 14.796586ms +Jun 18 11:57:11.043: INFO: Pod "pod-configmaps-332d94b4-91c0-11e9-bce2-ae54e022189f": Phase="Running", Reason="", readiness=true. Elapsed: 2.030098892s +Jun 18 11:57:13.058: INFO: Pod "pod-configmaps-332d94b4-91c0-11e9-bce2-ae54e022189f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.044933817s STEP: Saw pod success -May 13 19:12:34.098: INFO: Pod "downward-api-0dad88c0-75b3-11e9-8f67-2632f168be36" satisfied condition "success or failure" -May 13 19:12:34.107: INFO: Trying to get logs from node 10.170.219.153 pod downward-api-0dad88c0-75b3-11e9-8f67-2632f168be36 container dapi-container: +Jun 18 11:57:13.058: INFO: Pod "pod-configmaps-332d94b4-91c0-11e9-bce2-ae54e022189f" satisfied condition "success or failure" +Jun 18 11:57:13.073: INFO: Trying to get logs from node 10.72.74.144 pod pod-configmaps-332d94b4-91c0-11e9-bce2-ae54e022189f container configmap-volume-test: STEP: delete the pod -May 13 19:12:34.220: INFO: Waiting for pod downward-api-0dad88c0-75b3-11e9-8f67-2632f168be36 to disappear -May 13 19:12:34.228: INFO: Pod downward-api-0dad88c0-75b3-11e9-8f67-2632f168be36 no longer exists -[AfterEach] [sig-node] Downward API +Jun 18 11:57:13.157: INFO: Waiting for pod pod-configmaps-332d94b4-91c0-11e9-bce2-ae54e022189f to disappear +Jun 18 11:57:13.171: INFO: Pod pod-configmaps-332d94b4-91c0-11e9-bce2-ae54e022189f no longer exists +[AfterEach] [sig-storage] ConfigMap /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 19:12:34.228: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-downward-api-j58m5" for this suite. -May 13 19:12:40.264: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 19:12:40.303: INFO: namespace: e2e-tests-downward-api-j58m5, resource: bindings, ignored listing per whitelist -May 13 19:12:40.597: INFO: namespace e2e-tests-downward-api-j58m5 deletion completed in 6.361772947s +Jun 18 11:57:13.171: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "e2e-tests-configmap-spkd5" for this suite. +Jun 18 11:57:19.264: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 11:57:19.310: INFO: namespace: e2e-tests-configmap-spkd5, resource: bindings, ignored listing per whitelist +Jun 18 11:57:20.513: INFO: namespace e2e-tests-configmap-spkd5 deletion completed in 7.322777767s -• [SLOW TEST:10.855 seconds] -[sig-node] Downward API -/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downward_api.go:38 - should provide pod UID as env vars [NodeConformance] [Conformance] +• [SLOW TEST:11.968 seconds] +[sig-storage] ConfigMap +/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:33 + should be consumable from pods in volume with mappings and Item mode set [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ -[sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] - should perform canary updates and phased rolling updates of template modifications [Conformance] +SSSSSS +------------------------------ +[k8s.io] InitContainer [NodeConformance] + should not start app containers and fail the pod if init containers fail on a RestartNever pod [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -[BeforeEach] [sig-apps] StatefulSet +[BeforeEach] [k8s.io] InitContainer [NodeConformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client -May 13 19:12:40.598: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -STEP: Building a namespace api object, basename statefulset -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-statefulset-4npj8 +Jun 18 11:57:20.514: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 +STEP: Building a namespace api object, basename init-container +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-init-container-rn6tl STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [sig-apps] StatefulSet - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:59 -[BeforeEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:74 -STEP: Creating service test in namespace e2e-tests-statefulset-4npj8 -[It] should perform canary updates and phased rolling updates of template modifications [Conformance] +[BeforeEach] [k8s.io] InitContainer [NodeConformance] + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/init_container.go:43 +[It] should not start app containers and fail the pod if init containers fail on a RestartNever pod [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -STEP: Creating a new StaefulSet -May 13 19:12:40.894: INFO: Found 0 stateful pods, waiting for 3 -May 13 19:12:50.916: INFO: Waiting for pod ss2-0 to enter Running - Ready=true, currently Running - Ready=true -May 13 19:12:50.916: INFO: Waiting for pod ss2-1 to enter Running - Ready=true, currently Running - Ready=true -May 13 19:12:50.916: INFO: Waiting for pod ss2-2 to enter Running - Ready=true, currently Running - Ready=true -STEP: Updating stateful set template: update image from docker.io/library/nginx:1.14-alpine to docker.io/library/nginx:1.15-alpine -May 13 19:12:50.958: INFO: Updating stateful set ss2 -STEP: Creating a new revision -STEP: Not applying an update when the partition is greater than the number of replicas -STEP: Performing a canary update -May 13 19:13:01.019: INFO: Updating stateful set ss2 -May 13 19:13:01.031: INFO: Waiting for Pod e2e-tests-statefulset-4npj8/ss2-2 to have revision ss2-c79899b9 update revision ss2-787997d666 -STEP: Restoring Pods to the correct revision when they are deleted -May 13 19:13:11.127: INFO: Found 2 stateful pods, waiting for 3 -May 13 19:13:21.150: INFO: Waiting for pod ss2-0 to enter Running - Ready=true, currently Running - Ready=true -May 13 19:13:21.151: INFO: Waiting for pod ss2-1 to enter Running - Ready=true, currently Running - Ready=true -May 13 19:13:21.151: INFO: Waiting for pod ss2-2 to enter Running - Ready=true, currently Running - Ready=true -STEP: Performing a phased rolling update -May 13 19:13:21.227: INFO: Updating stateful set ss2 -May 13 19:13:21.239: INFO: Waiting for Pod e2e-tests-statefulset-4npj8/ss2-1 to have revision ss2-c79899b9 update revision ss2-787997d666 -May 13 19:13:31.287: INFO: Updating stateful set ss2 -May 13 19:13:31.300: INFO: Waiting for StatefulSet e2e-tests-statefulset-4npj8/ss2 to complete update -May 13 19:13:31.300: INFO: Waiting for Pod e2e-tests-statefulset-4npj8/ss2-0 to have revision ss2-c79899b9 update revision ss2-787997d666 -May 13 19:13:41.428: INFO: Waiting for StatefulSet e2e-tests-statefulset-4npj8/ss2 to complete update -[AfterEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:85 -May 13 19:13:51.316: INFO: Deleting all statefulset in ns e2e-tests-statefulset-4npj8 -May 13 19:13:51.333: INFO: Scaling statefulset ss2 to 0 -May 13 19:14:11.360: INFO: Waiting for statefulset status.replicas updated to 0 -May 13 19:14:11.365: INFO: Deleting statefulset ss2 -[AfterEach] [sig-apps] StatefulSet +STEP: creating the pod +Jun 18 11:57:20.967: INFO: PodSpec: initContainers in spec.initContainers +[AfterEach] [k8s.io] InitContainer [NodeConformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 19:14:11.390: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-statefulset-4npj8" for this suite. -May 13 19:14:19.427: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 19:14:19.552: INFO: namespace: e2e-tests-statefulset-4npj8, resource: bindings, ignored listing per whitelist -May 13 19:14:19.774: INFO: namespace e2e-tests-statefulset-4npj8 deletion completed in 8.376672674s +Jun 18 11:57:24.346: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "e2e-tests-init-container-rn6tl" for this suite. +Jun 18 11:57:30.463: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 11:57:30.698: INFO: namespace: e2e-tests-init-container-rn6tl, resource: bindings, ignored listing per whitelist +Jun 18 11:57:30.947: INFO: namespace e2e-tests-init-container-rn6tl deletion completed in 6.557255362s -• [SLOW TEST:99.177 seconds] -[sig-apps] StatefulSet -/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22 - [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 - should perform canary updates and phased rolling updates of template modifications [Conformance] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 +• [SLOW TEST:10.434 seconds] +[k8s.io] InitContainer [NodeConformance] +/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 + should not start app containers and fail the pod if init containers fail on a RestartNever pod [Conformance] + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ -SS +SSS ------------------------------ -[k8s.io] Probing container - should be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance] +[sig-storage] Projected downwardAPI + should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -[BeforeEach] [k8s.io] Probing container +[BeforeEach] [sig-storage] Projected downwardAPI /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client -May 13 19:14:19.776: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -STEP: Building a namespace api object, basename container-probe -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-container-probe-qjl2t +Jun 18 11:57:30.949: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 +STEP: Building a namespace api object, basename projected +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-projected-p6vrf STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [k8s.io] Probing container - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/container_probe.go:48 -[It] should be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance] +[BeforeEach] [sig-storage] Projected downwardAPI + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:39 +[It] should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -STEP: Creating pod liveness-exec in namespace e2e-tests-container-probe-qjl2t -May 13 19:14:24.100: INFO: Started pod liveness-exec in namespace e2e-tests-container-probe-qjl2t -STEP: checking the pod's current state and verifying that restartCount is present -May 13 19:14:24.109: INFO: Initial restart count of pod liveness-exec is 0 -May 13 19:15:09.040: INFO: Restart count of pod e2e-tests-container-probe-qjl2t/liveness-exec is now 1 (44.93137547s elapsed) -STEP: deleting the pod -[AfterEach] [k8s.io] Probing container +STEP: Creating a pod to test downward API volume plugin +Jun 18 11:57:31.519: INFO: Waiting up to 5m0s for pod "downwardapi-volume-40977b36-91c0-11e9-bce2-ae54e022189f" in namespace "e2e-tests-projected-p6vrf" to be "success or failure" +Jun 18 11:57:31.533: INFO: Pod "downwardapi-volume-40977b36-91c0-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 14.764601ms +Jun 18 11:57:33.549: INFO: Pod "downwardapi-volume-40977b36-91c0-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 2.030349171s +Jun 18 11:57:35.565: INFO: Pod "downwardapi-volume-40977b36-91c0-11e9-bce2-ae54e022189f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.046008441s +STEP: Saw pod success +Jun 18 11:57:35.565: INFO: Pod "downwardapi-volume-40977b36-91c0-11e9-bce2-ae54e022189f" satisfied condition "success or failure" +Jun 18 11:57:35.580: INFO: Trying to get logs from node 10.72.74.144 pod downwardapi-volume-40977b36-91c0-11e9-bce2-ae54e022189f container client-container: +STEP: delete the pod +Jun 18 11:57:35.659: INFO: Waiting for pod downwardapi-volume-40977b36-91c0-11e9-bce2-ae54e022189f to disappear +Jun 18 11:57:35.672: INFO: Pod downwardapi-volume-40977b36-91c0-11e9-bce2-ae54e022189f no longer exists +[AfterEach] [sig-storage] Projected downwardAPI /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 19:15:09.063: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-container-probe-qjl2t" for this suite. -May 13 19:15:15.160: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 19:15:15.316: INFO: namespace: e2e-tests-container-probe-qjl2t, resource: bindings, ignored listing per whitelist -May 13 19:15:15.460: INFO: namespace e2e-tests-container-probe-qjl2t deletion completed in 6.340251171s +Jun 18 11:57:35.672: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "e2e-tests-projected-p6vrf" for this suite. +Jun 18 11:57:41.783: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 11:57:42.109: INFO: namespace: e2e-tests-projected-p6vrf, resource: bindings, ignored listing per whitelist +Jun 18 11:57:42.257: INFO: namespace e2e-tests-projected-p6vrf deletion completed in 6.543140343s -• [SLOW TEST:55.684 seconds] -[k8s.io] Probing container -/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 - should be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance] +• [SLOW TEST:11.309 seconds] +[sig-storage] Projected downwardAPI +/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:33 + should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ -SSSSSSSSSSSSSSSSSSSSSS +SSSSSSSSSSSS ------------------------------ [sig-storage] Secrets - should be consumable in multiple volumes in a pod [NodeConformance] [Conformance] + should be consumable from pods in volume with mappings and Item Mode set [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-storage] Secrets /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client -May 13 19:15:15.460: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 +Jun 18 11:57:42.258: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 STEP: Building a namespace api object, basename secrets -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-secrets-lxbll +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-secrets-k4jfb STEP: Waiting for a default service account to be provisioned in namespace -[It] should be consumable in multiple volumes in a pod [NodeConformance] [Conformance] +[It] should be consumable from pods in volume with mappings and Item Mode set [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -STEP: Creating secret with name secret-test-706d22b1-75b3-11e9-8f67-2632f168be36 +STEP: Creating secret with name secret-test-map-4745231b-91c0-11e9-bce2-ae54e022189f STEP: Creating a pod to test consume secrets -May 13 19:15:15.753: INFO: Waiting up to 5m0s for pod "pod-secrets-706e72d7-75b3-11e9-8f67-2632f168be36" in namespace "e2e-tests-secrets-lxbll" to be "success or failure" -May 13 19:15:15.761: INFO: Pod "pod-secrets-706e72d7-75b3-11e9-8f67-2632f168be36": Phase="Pending", Reason="", readiness=false. Elapsed: 7.785607ms -May 13 19:15:17.769: INFO: Pod "pod-secrets-706e72d7-75b3-11e9-8f67-2632f168be36": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.016256283s +Jun 18 11:57:42.736: INFO: Waiting up to 5m0s for pod "pod-secrets-474735d8-91c0-11e9-bce2-ae54e022189f" in namespace "e2e-tests-secrets-k4jfb" to be "success or failure" +Jun 18 11:57:42.751: INFO: Pod "pod-secrets-474735d8-91c0-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 14.998026ms +Jun 18 11:57:44.798: INFO: Pod "pod-secrets-474735d8-91c0-11e9-bce2-ae54e022189f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.061461631s STEP: Saw pod success -May 13 19:15:17.769: INFO: Pod "pod-secrets-706e72d7-75b3-11e9-8f67-2632f168be36" satisfied condition "success or failure" -May 13 19:15:17.776: INFO: Trying to get logs from node 10.170.219.177 pod pod-secrets-706e72d7-75b3-11e9-8f67-2632f168be36 container secret-volume-test: +Jun 18 11:57:44.798: INFO: Pod "pod-secrets-474735d8-91c0-11e9-bce2-ae54e022189f" satisfied condition "success or failure" +Jun 18 11:57:44.812: INFO: Trying to get logs from node 10.72.74.144 pod pod-secrets-474735d8-91c0-11e9-bce2-ae54e022189f container secret-volume-test: STEP: delete the pod -May 13 19:15:17.818: INFO: Waiting for pod pod-secrets-706e72d7-75b3-11e9-8f67-2632f168be36 to disappear -May 13 19:15:17.827: INFO: Pod pod-secrets-706e72d7-75b3-11e9-8f67-2632f168be36 no longer exists +Jun 18 11:57:44.885: INFO: Waiting for pod pod-secrets-474735d8-91c0-11e9-bce2-ae54e022189f to disappear +Jun 18 11:57:44.900: INFO: Pod pod-secrets-474735d8-91c0-11e9-bce2-ae54e022189f no longer exists [AfterEach] [sig-storage] Secrets /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 19:15:17.827: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-secrets-lxbll" for this suite. -May 13 19:15:23.863: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 19:15:23.995: INFO: namespace: e2e-tests-secrets-lxbll, resource: bindings, ignored listing per whitelist -May 13 19:15:24.159: INFO: namespace e2e-tests-secrets-lxbll deletion completed in 6.324747219s +Jun 18 11:57:44.900: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "e2e-tests-secrets-k4jfb" for this suite. +Jun 18 11:57:50.969: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 11:57:51.152: INFO: namespace: e2e-tests-secrets-k4jfb, resource: bindings, ignored listing per whitelist +Jun 18 11:57:51.498: INFO: namespace e2e-tests-secrets-k4jfb deletion completed in 6.578493169s -• [SLOW TEST:8.699 seconds] +• [SLOW TEST:9.240 seconds] [sig-storage] Secrets /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/secrets_volume.go:34 - should be consumable in multiple volumes in a pod [NodeConformance] [Conformance] + should be consumable from pods in volume with mappings and Item Mode set [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ -SSSSSSSSSS +S ------------------------------ -[sig-network] DNS - should provide DNS for services [Conformance] +[sig-api-machinery] Garbage collector + should orphan pods created by rc if delete options say so [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -[BeforeEach] [sig-network] DNS +[BeforeEach] [sig-api-machinery] Garbage collector /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client -May 13 19:15:24.160: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -STEP: Building a namespace api object, basename dns -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-dns-xhpc9 +Jun 18 11:57:51.499: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 +STEP: Building a namespace api object, basename gc +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-gc-7zv6h STEP: Waiting for a default service account to be provisioned in namespace -[It] should provide DNS for services [Conformance] +[It] should orphan pods created by rc if delete options say so [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -STEP: Creating a test headless service -STEP: Running these commands on wheezy: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search dns-test-service A)" && test -n "$$check" && echo OK > /results/wheezy_udp@dns-test-service;check="$$(dig +tcp +noall +answer +search dns-test-service A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@dns-test-service;check="$$(dig +notcp +noall +answer +search dns-test-service.e2e-tests-dns-xhpc9 A)" && test -n "$$check" && echo OK > /results/wheezy_udp@dns-test-service.e2e-tests-dns-xhpc9;check="$$(dig +tcp +noall +answer +search dns-test-service.e2e-tests-dns-xhpc9 A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@dns-test-service.e2e-tests-dns-xhpc9;check="$$(dig +notcp +noall +answer +search dns-test-service.e2e-tests-dns-xhpc9.svc A)" && test -n "$$check" && echo OK > /results/wheezy_udp@dns-test-service.e2e-tests-dns-xhpc9.svc;check="$$(dig +tcp +noall +answer +search dns-test-service.e2e-tests-dns-xhpc9.svc A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@dns-test-service.e2e-tests-dns-xhpc9.svc;check="$$(dig +notcp +noall +answer +search _http._tcp.dns-test-service.e2e-tests-dns-xhpc9.svc SRV)" && test -n "$$check" && echo OK > /results/wheezy_udp@_http._tcp.dns-test-service.e2e-tests-dns-xhpc9.svc;check="$$(dig +tcp +noall +answer +search _http._tcp.dns-test-service.e2e-tests-dns-xhpc9.svc SRV)" && test -n "$$check" && echo OK > /results/wheezy_tcp@_http._tcp.dns-test-service.e2e-tests-dns-xhpc9.svc;check="$$(dig +notcp +noall +answer +search _http._tcp.test-service-2.e2e-tests-dns-xhpc9.svc SRV)" && test -n "$$check" && echo OK > /results/wheezy_udp@_http._tcp.test-service-2.e2e-tests-dns-xhpc9.svc;check="$$(dig +tcp +noall +answer +search _http._tcp.test-service-2.e2e-tests-dns-xhpc9.svc SRV)" && test -n "$$check" && echo OK > /results/wheezy_tcp@_http._tcp.test-service-2.e2e-tests-dns-xhpc9.svc;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".e2e-tests-dns-xhpc9.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@PodARecord;check="$$(dig +notcp +noall +answer +search 73.188.21.172.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/172.21.188.73_udp@PTR;check="$$(dig +tcp +noall +answer +search 73.188.21.172.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/172.21.188.73_tcp@PTR;sleep 1; done - -STEP: Running these commands on jessie: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search dns-test-service A)" && test -n "$$check" && echo OK > /results/jessie_udp@dns-test-service;check="$$(dig +tcp +noall +answer +search dns-test-service A)" && test -n "$$check" && echo OK > /results/jessie_tcp@dns-test-service;check="$$(dig +notcp +noall +answer +search dns-test-service.e2e-tests-dns-xhpc9 A)" && test -n "$$check" && echo OK > /results/jessie_udp@dns-test-service.e2e-tests-dns-xhpc9;check="$$(dig +tcp +noall +answer +search dns-test-service.e2e-tests-dns-xhpc9 A)" && test -n "$$check" && echo OK > /results/jessie_tcp@dns-test-service.e2e-tests-dns-xhpc9;check="$$(dig +notcp +noall +answer +search dns-test-service.e2e-tests-dns-xhpc9.svc A)" && test -n "$$check" && echo OK > /results/jessie_udp@dns-test-service.e2e-tests-dns-xhpc9.svc;check="$$(dig +tcp +noall +answer +search dns-test-service.e2e-tests-dns-xhpc9.svc A)" && test -n "$$check" && echo OK > /results/jessie_tcp@dns-test-service.e2e-tests-dns-xhpc9.svc;check="$$(dig +notcp +noall +answer +search _http._tcp.dns-test-service.e2e-tests-dns-xhpc9.svc SRV)" && test -n "$$check" && echo OK > /results/jessie_udp@_http._tcp.dns-test-service.e2e-tests-dns-xhpc9.svc;check="$$(dig +tcp +noall +answer +search _http._tcp.dns-test-service.e2e-tests-dns-xhpc9.svc SRV)" && test -n "$$check" && echo OK > /results/jessie_tcp@_http._tcp.dns-test-service.e2e-tests-dns-xhpc9.svc;check="$$(dig +notcp +noall +answer +search _http._tcp.test-service-2.e2e-tests-dns-xhpc9.svc SRV)" && test -n "$$check" && echo OK > /results/jessie_udp@_http._tcp.test-service-2.e2e-tests-dns-xhpc9.svc;check="$$(dig +tcp +noall +answer +search _http._tcp.test-service-2.e2e-tests-dns-xhpc9.svc SRV)" && test -n "$$check" && echo OK > /results/jessie_tcp@_http._tcp.test-service-2.e2e-tests-dns-xhpc9.svc;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".e2e-tests-dns-xhpc9.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_tcp@PodARecord;check="$$(dig +notcp +noall +answer +search 73.188.21.172.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/172.21.188.73_udp@PTR;check="$$(dig +tcp +noall +answer +search 73.188.21.172.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/172.21.188.73_tcp@PTR;sleep 1; done - -STEP: creating a pod to probe DNS -STEP: submitting the pod to kubernetes -STEP: retrieving the pod -STEP: looking for the results for each expected name from probers -May 13 19:15:28.541: INFO: Unable to read wheezy_udp@dns-test-service from pod e2e-tests-dns-xhpc9/dns-test-75a159c3-75b3-11e9-8f67-2632f168be36: the server could not find the requested resource (get pods dns-test-75a159c3-75b3-11e9-8f67-2632f168be36) -May 13 19:15:28.573: INFO: Unable to read wheezy_tcp@dns-test-service.e2e-tests-dns-xhpc9 from pod e2e-tests-dns-xhpc9/dns-test-75a159c3-75b3-11e9-8f67-2632f168be36: the server could not find the requested resource (get pods dns-test-75a159c3-75b3-11e9-8f67-2632f168be36) -May 13 19:15:28.606: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.e2e-tests-dns-xhpc9.svc from pod e2e-tests-dns-xhpc9/dns-test-75a159c3-75b3-11e9-8f67-2632f168be36: the server could not find the requested resource (get pods dns-test-75a159c3-75b3-11e9-8f67-2632f168be36) -May 13 19:15:28.621: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.e2e-tests-dns-xhpc9.svc from pod e2e-tests-dns-xhpc9/dns-test-75a159c3-75b3-11e9-8f67-2632f168be36: the server could not find the requested resource (get pods dns-test-75a159c3-75b3-11e9-8f67-2632f168be36) -May 13 19:15:28.691: INFO: Unable to read jessie_udp@dns-test-service from pod e2e-tests-dns-xhpc9/dns-test-75a159c3-75b3-11e9-8f67-2632f168be36: the server could not find the requested resource (get pods dns-test-75a159c3-75b3-11e9-8f67-2632f168be36) -May 13 19:15:28.701: INFO: Unable to read jessie_tcp@dns-test-service from pod e2e-tests-dns-xhpc9/dns-test-75a159c3-75b3-11e9-8f67-2632f168be36: the server could not find the requested resource (get pods dns-test-75a159c3-75b3-11e9-8f67-2632f168be36) -May 13 19:15:28.711: INFO: Unable to read jessie_udp@dns-test-service.e2e-tests-dns-xhpc9 from pod e2e-tests-dns-xhpc9/dns-test-75a159c3-75b3-11e9-8f67-2632f168be36: the server could not find the requested resource (get pods dns-test-75a159c3-75b3-11e9-8f67-2632f168be36) -May 13 19:15:28.722: INFO: Unable to read jessie_tcp@dns-test-service.e2e-tests-dns-xhpc9 from pod e2e-tests-dns-xhpc9/dns-test-75a159c3-75b3-11e9-8f67-2632f168be36: the server could not find the requested resource (get pods dns-test-75a159c3-75b3-11e9-8f67-2632f168be36) -May 13 19:15:28.732: INFO: Unable to read jessie_udp@dns-test-service.e2e-tests-dns-xhpc9.svc from pod e2e-tests-dns-xhpc9/dns-test-75a159c3-75b3-11e9-8f67-2632f168be36: the server could not find the requested resource (get pods dns-test-75a159c3-75b3-11e9-8f67-2632f168be36) -May 13 19:15:28.742: INFO: Unable to read jessie_tcp@dns-test-service.e2e-tests-dns-xhpc9.svc from pod e2e-tests-dns-xhpc9/dns-test-75a159c3-75b3-11e9-8f67-2632f168be36: the server could not find the requested resource (get pods dns-test-75a159c3-75b3-11e9-8f67-2632f168be36) -May 13 19:15:28.751: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.e2e-tests-dns-xhpc9.svc from pod e2e-tests-dns-xhpc9/dns-test-75a159c3-75b3-11e9-8f67-2632f168be36: the server could not find the requested resource (get pods dns-test-75a159c3-75b3-11e9-8f67-2632f168be36) -May 13 19:15:28.761: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.e2e-tests-dns-xhpc9.svc from pod e2e-tests-dns-xhpc9/dns-test-75a159c3-75b3-11e9-8f67-2632f168be36: the server could not find the requested resource (get pods dns-test-75a159c3-75b3-11e9-8f67-2632f168be36) -May 13 19:15:28.829: INFO: Lookups using e2e-tests-dns-xhpc9/dns-test-75a159c3-75b3-11e9-8f67-2632f168be36 failed for: [wheezy_udp@dns-test-service wheezy_tcp@dns-test-service.e2e-tests-dns-xhpc9 wheezy_udp@_http._tcp.dns-test-service.e2e-tests-dns-xhpc9.svc wheezy_tcp@_http._tcp.dns-test-service.e2e-tests-dns-xhpc9.svc jessie_udp@dns-test-service jessie_tcp@dns-test-service jessie_udp@dns-test-service.e2e-tests-dns-xhpc9 jessie_tcp@dns-test-service.e2e-tests-dns-xhpc9 jessie_udp@dns-test-service.e2e-tests-dns-xhpc9.svc jessie_tcp@dns-test-service.e2e-tests-dns-xhpc9.svc jessie_udp@_http._tcp.dns-test-service.e2e-tests-dns-xhpc9.svc jessie_tcp@_http._tcp.dns-test-service.e2e-tests-dns-xhpc9.svc] - -May 13 19:15:33.842: INFO: Unable to read wheezy_udp@dns-test-service from pod e2e-tests-dns-xhpc9/dns-test-75a159c3-75b3-11e9-8f67-2632f168be36: the server could not find the requested resource (get pods dns-test-75a159c3-75b3-11e9-8f67-2632f168be36) -May 13 19:15:33.875: INFO: Unable to read wheezy_tcp@dns-test-service.e2e-tests-dns-xhpc9 from pod e2e-tests-dns-xhpc9/dns-test-75a159c3-75b3-11e9-8f67-2632f168be36: the server could not find the requested resource (get pods dns-test-75a159c3-75b3-11e9-8f67-2632f168be36) -May 13 19:15:33.910: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.e2e-tests-dns-xhpc9.svc from pod e2e-tests-dns-xhpc9/dns-test-75a159c3-75b3-11e9-8f67-2632f168be36: the server could not find the requested resource (get pods dns-test-75a159c3-75b3-11e9-8f67-2632f168be36) -May 13 19:15:33.920: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.e2e-tests-dns-xhpc9.svc from pod e2e-tests-dns-xhpc9/dns-test-75a159c3-75b3-11e9-8f67-2632f168be36: the server could not find the requested resource (get pods dns-test-75a159c3-75b3-11e9-8f67-2632f168be36) -May 13 19:15:33.994: INFO: Unable to read jessie_udp@dns-test-service from pod e2e-tests-dns-xhpc9/dns-test-75a159c3-75b3-11e9-8f67-2632f168be36: the server could not find the requested resource (get pods dns-test-75a159c3-75b3-11e9-8f67-2632f168be36) -May 13 19:15:34.007: INFO: Unable to read jessie_tcp@dns-test-service from pod e2e-tests-dns-xhpc9/dns-test-75a159c3-75b3-11e9-8f67-2632f168be36: the server could not find the requested resource (get pods dns-test-75a159c3-75b3-11e9-8f67-2632f168be36) -May 13 19:15:34.018: INFO: Unable to read jessie_udp@dns-test-service.e2e-tests-dns-xhpc9 from pod e2e-tests-dns-xhpc9/dns-test-75a159c3-75b3-11e9-8f67-2632f168be36: the server could not find the requested resource (get pods dns-test-75a159c3-75b3-11e9-8f67-2632f168be36) -May 13 19:15:34.028: INFO: Unable to read jessie_tcp@dns-test-service.e2e-tests-dns-xhpc9 from pod e2e-tests-dns-xhpc9/dns-test-75a159c3-75b3-11e9-8f67-2632f168be36: the server could not find the requested resource (get pods dns-test-75a159c3-75b3-11e9-8f67-2632f168be36) -May 13 19:15:34.040: INFO: Unable to read jessie_udp@dns-test-service.e2e-tests-dns-xhpc9.svc from pod e2e-tests-dns-xhpc9/dns-test-75a159c3-75b3-11e9-8f67-2632f168be36: the server could not find the requested resource (get pods dns-test-75a159c3-75b3-11e9-8f67-2632f168be36) -May 13 19:15:34.051: INFO: Unable to read jessie_tcp@dns-test-service.e2e-tests-dns-xhpc9.svc from pod e2e-tests-dns-xhpc9/dns-test-75a159c3-75b3-11e9-8f67-2632f168be36: the server could not find the requested resource (get pods dns-test-75a159c3-75b3-11e9-8f67-2632f168be36) -May 13 19:15:34.061: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.e2e-tests-dns-xhpc9.svc from pod e2e-tests-dns-xhpc9/dns-test-75a159c3-75b3-11e9-8f67-2632f168be36: the server could not find the requested resource (get pods dns-test-75a159c3-75b3-11e9-8f67-2632f168be36) -May 13 19:15:34.071: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.e2e-tests-dns-xhpc9.svc from pod e2e-tests-dns-xhpc9/dns-test-75a159c3-75b3-11e9-8f67-2632f168be36: the server could not find the requested resource (get pods dns-test-75a159c3-75b3-11e9-8f67-2632f168be36) -May 13 19:15:34.139: INFO: Lookups using e2e-tests-dns-xhpc9/dns-test-75a159c3-75b3-11e9-8f67-2632f168be36 failed for: [wheezy_udp@dns-test-service wheezy_tcp@dns-test-service.e2e-tests-dns-xhpc9 wheezy_udp@_http._tcp.dns-test-service.e2e-tests-dns-xhpc9.svc wheezy_tcp@_http._tcp.dns-test-service.e2e-tests-dns-xhpc9.svc jessie_udp@dns-test-service jessie_tcp@dns-test-service jessie_udp@dns-test-service.e2e-tests-dns-xhpc9 jessie_tcp@dns-test-service.e2e-tests-dns-xhpc9 jessie_udp@dns-test-service.e2e-tests-dns-xhpc9.svc jessie_tcp@dns-test-service.e2e-tests-dns-xhpc9.svc jessie_udp@_http._tcp.dns-test-service.e2e-tests-dns-xhpc9.svc jessie_tcp@_http._tcp.dns-test-service.e2e-tests-dns-xhpc9.svc] - -May 13 19:15:38.853: INFO: Unable to read wheezy_udp@dns-test-service from pod e2e-tests-dns-xhpc9/dns-test-75a159c3-75b3-11e9-8f67-2632f168be36: the server could not find the requested resource (get pods dns-test-75a159c3-75b3-11e9-8f67-2632f168be36) -May 13 19:15:38.952: INFO: Unable to read wheezy_tcp@dns-test-service.e2e-tests-dns-xhpc9 from pod e2e-tests-dns-xhpc9/dns-test-75a159c3-75b3-11e9-8f67-2632f168be36: the server could not find the requested resource (get pods dns-test-75a159c3-75b3-11e9-8f67-2632f168be36) -May 13 19:15:38.984: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.e2e-tests-dns-xhpc9.svc from pod e2e-tests-dns-xhpc9/dns-test-75a159c3-75b3-11e9-8f67-2632f168be36: the server could not find the requested resource (get pods dns-test-75a159c3-75b3-11e9-8f67-2632f168be36) -May 13 19:15:38.994: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.e2e-tests-dns-xhpc9.svc from pod e2e-tests-dns-xhpc9/dns-test-75a159c3-75b3-11e9-8f67-2632f168be36: the server could not find the requested resource (get pods dns-test-75a159c3-75b3-11e9-8f67-2632f168be36) -May 13 19:15:39.072: INFO: Unable to read jessie_udp@dns-test-service from pod e2e-tests-dns-xhpc9/dns-test-75a159c3-75b3-11e9-8f67-2632f168be36: the server could not find the requested resource (get pods dns-test-75a159c3-75b3-11e9-8f67-2632f168be36) -May 13 19:15:39.082: INFO: Unable to read jessie_tcp@dns-test-service from pod e2e-tests-dns-xhpc9/dns-test-75a159c3-75b3-11e9-8f67-2632f168be36: the server could not find the requested resource (get pods dns-test-75a159c3-75b3-11e9-8f67-2632f168be36) -May 13 19:15:39.092: INFO: Unable to read jessie_udp@dns-test-service.e2e-tests-dns-xhpc9 from pod e2e-tests-dns-xhpc9/dns-test-75a159c3-75b3-11e9-8f67-2632f168be36: the server could not find the requested resource (get pods dns-test-75a159c3-75b3-11e9-8f67-2632f168be36) -May 13 19:15:39.104: INFO: Unable to read jessie_tcp@dns-test-service.e2e-tests-dns-xhpc9 from pod e2e-tests-dns-xhpc9/dns-test-75a159c3-75b3-11e9-8f67-2632f168be36: the server could not find the requested resource (get pods dns-test-75a159c3-75b3-11e9-8f67-2632f168be36) -May 13 19:15:39.114: INFO: Unable to read jessie_udp@dns-test-service.e2e-tests-dns-xhpc9.svc from pod e2e-tests-dns-xhpc9/dns-test-75a159c3-75b3-11e9-8f67-2632f168be36: the server could not find the requested resource (get pods dns-test-75a159c3-75b3-11e9-8f67-2632f168be36) -May 13 19:15:39.126: INFO: Unable to read jessie_tcp@dns-test-service.e2e-tests-dns-xhpc9.svc from pod e2e-tests-dns-xhpc9/dns-test-75a159c3-75b3-11e9-8f67-2632f168be36: the server could not find the requested resource (get pods dns-test-75a159c3-75b3-11e9-8f67-2632f168be36) -May 13 19:15:39.137: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.e2e-tests-dns-xhpc9.svc from pod e2e-tests-dns-xhpc9/dns-test-75a159c3-75b3-11e9-8f67-2632f168be36: the server could not find the requested resource (get pods dns-test-75a159c3-75b3-11e9-8f67-2632f168be36) -May 13 19:15:39.147: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.e2e-tests-dns-xhpc9.svc from pod e2e-tests-dns-xhpc9/dns-test-75a159c3-75b3-11e9-8f67-2632f168be36: the server could not find the requested resource (get pods dns-test-75a159c3-75b3-11e9-8f67-2632f168be36) -May 13 19:15:39.211: INFO: Lookups using e2e-tests-dns-xhpc9/dns-test-75a159c3-75b3-11e9-8f67-2632f168be36 failed for: [wheezy_udp@dns-test-service wheezy_tcp@dns-test-service.e2e-tests-dns-xhpc9 wheezy_udp@_http._tcp.dns-test-service.e2e-tests-dns-xhpc9.svc wheezy_tcp@_http._tcp.dns-test-service.e2e-tests-dns-xhpc9.svc jessie_udp@dns-test-service jessie_tcp@dns-test-service jessie_udp@dns-test-service.e2e-tests-dns-xhpc9 jessie_tcp@dns-test-service.e2e-tests-dns-xhpc9 jessie_udp@dns-test-service.e2e-tests-dns-xhpc9.svc jessie_tcp@dns-test-service.e2e-tests-dns-xhpc9.svc jessie_udp@_http._tcp.dns-test-service.e2e-tests-dns-xhpc9.svc jessie_tcp@_http._tcp.dns-test-service.e2e-tests-dns-xhpc9.svc] - -May 13 19:15:43.947: INFO: Unable to read wheezy_udp@dns-test-service from pod e2e-tests-dns-xhpc9/dns-test-75a159c3-75b3-11e9-8f67-2632f168be36: the server could not find the requested resource (get pods dns-test-75a159c3-75b3-11e9-8f67-2632f168be36) -May 13 19:15:43.982: INFO: Unable to read wheezy_tcp@dns-test-service.e2e-tests-dns-xhpc9 from pod e2e-tests-dns-xhpc9/dns-test-75a159c3-75b3-11e9-8f67-2632f168be36: the server could not find the requested resource (get pods dns-test-75a159c3-75b3-11e9-8f67-2632f168be36) -May 13 19:15:44.052: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.e2e-tests-dns-xhpc9.svc from pod e2e-tests-dns-xhpc9/dns-test-75a159c3-75b3-11e9-8f67-2632f168be36: the server could not find the requested resource (get pods dns-test-75a159c3-75b3-11e9-8f67-2632f168be36) -May 13 19:15:44.062: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.e2e-tests-dns-xhpc9.svc from pod e2e-tests-dns-xhpc9/dns-test-75a159c3-75b3-11e9-8f67-2632f168be36: the server could not find the requested resource (get pods dns-test-75a159c3-75b3-11e9-8f67-2632f168be36) -May 13 19:15:44.139: INFO: Unable to read jessie_udp@dns-test-service from pod e2e-tests-dns-xhpc9/dns-test-75a159c3-75b3-11e9-8f67-2632f168be36: the server could not find the requested resource (get pods dns-test-75a159c3-75b3-11e9-8f67-2632f168be36) -May 13 19:15:44.149: INFO: Unable to read jessie_tcp@dns-test-service from pod e2e-tests-dns-xhpc9/dns-test-75a159c3-75b3-11e9-8f67-2632f168be36: the server could not find the requested resource (get pods dns-test-75a159c3-75b3-11e9-8f67-2632f168be36) -May 13 19:15:44.160: INFO: Unable to read jessie_udp@dns-test-service.e2e-tests-dns-xhpc9 from pod e2e-tests-dns-xhpc9/dns-test-75a159c3-75b3-11e9-8f67-2632f168be36: the server could not find the requested resource (get pods dns-test-75a159c3-75b3-11e9-8f67-2632f168be36) -May 13 19:15:44.171: INFO: Unable to read jessie_tcp@dns-test-service.e2e-tests-dns-xhpc9 from pod e2e-tests-dns-xhpc9/dns-test-75a159c3-75b3-11e9-8f67-2632f168be36: the server could not find the requested resource (get pods dns-test-75a159c3-75b3-11e9-8f67-2632f168be36) -May 13 19:15:44.181: INFO: Unable to read jessie_udp@dns-test-service.e2e-tests-dns-xhpc9.svc from pod e2e-tests-dns-xhpc9/dns-test-75a159c3-75b3-11e9-8f67-2632f168be36: the server could not find the requested resource (get pods dns-test-75a159c3-75b3-11e9-8f67-2632f168be36) -May 13 19:15:44.191: INFO: Unable to read jessie_tcp@dns-test-service.e2e-tests-dns-xhpc9.svc from pod e2e-tests-dns-xhpc9/dns-test-75a159c3-75b3-11e9-8f67-2632f168be36: the server could not find the requested resource (get pods dns-test-75a159c3-75b3-11e9-8f67-2632f168be36) -May 13 19:15:44.203: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.e2e-tests-dns-xhpc9.svc from pod e2e-tests-dns-xhpc9/dns-test-75a159c3-75b3-11e9-8f67-2632f168be36: the server could not find the requested resource (get pods dns-test-75a159c3-75b3-11e9-8f67-2632f168be36) -May 13 19:15:44.217: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.e2e-tests-dns-xhpc9.svc from pod e2e-tests-dns-xhpc9/dns-test-75a159c3-75b3-11e9-8f67-2632f168be36: the server could not find the requested resource (get pods dns-test-75a159c3-75b3-11e9-8f67-2632f168be36) -May 13 19:15:44.295: INFO: Lookups using e2e-tests-dns-xhpc9/dns-test-75a159c3-75b3-11e9-8f67-2632f168be36 failed for: [wheezy_udp@dns-test-service wheezy_tcp@dns-test-service.e2e-tests-dns-xhpc9 wheezy_udp@_http._tcp.dns-test-service.e2e-tests-dns-xhpc9.svc wheezy_tcp@_http._tcp.dns-test-service.e2e-tests-dns-xhpc9.svc jessie_udp@dns-test-service jessie_tcp@dns-test-service jessie_udp@dns-test-service.e2e-tests-dns-xhpc9 jessie_tcp@dns-test-service.e2e-tests-dns-xhpc9 jessie_udp@dns-test-service.e2e-tests-dns-xhpc9.svc jessie_tcp@dns-test-service.e2e-tests-dns-xhpc9.svc jessie_udp@_http._tcp.dns-test-service.e2e-tests-dns-xhpc9.svc jessie_tcp@_http._tcp.dns-test-service.e2e-tests-dns-xhpc9.svc] - -May 13 19:15:48.841: INFO: Unable to read wheezy_udp@dns-test-service from pod e2e-tests-dns-xhpc9/dns-test-75a159c3-75b3-11e9-8f67-2632f168be36: the server could not find the requested resource (get pods dns-test-75a159c3-75b3-11e9-8f67-2632f168be36) -May 13 19:15:48.887: INFO: Unable to read wheezy_tcp@dns-test-service.e2e-tests-dns-xhpc9 from pod e2e-tests-dns-xhpc9/dns-test-75a159c3-75b3-11e9-8f67-2632f168be36: the server could not find the requested resource (get pods dns-test-75a159c3-75b3-11e9-8f67-2632f168be36) -May 13 19:15:48.919: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.e2e-tests-dns-xhpc9.svc from pod e2e-tests-dns-xhpc9/dns-test-75a159c3-75b3-11e9-8f67-2632f168be36: the server could not find the requested resource (get pods dns-test-75a159c3-75b3-11e9-8f67-2632f168be36) -May 13 19:15:48.932: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.e2e-tests-dns-xhpc9.svc from pod e2e-tests-dns-xhpc9/dns-test-75a159c3-75b3-11e9-8f67-2632f168be36: the server could not find the requested resource (get pods dns-test-75a159c3-75b3-11e9-8f67-2632f168be36) -May 13 19:15:49.008: INFO: Unable to read jessie_udp@dns-test-service from pod e2e-tests-dns-xhpc9/dns-test-75a159c3-75b3-11e9-8f67-2632f168be36: the server could not find the requested resource (get pods dns-test-75a159c3-75b3-11e9-8f67-2632f168be36) -May 13 19:15:49.019: INFO: Unable to read jessie_tcp@dns-test-service from pod e2e-tests-dns-xhpc9/dns-test-75a159c3-75b3-11e9-8f67-2632f168be36: the server could not find the requested resource (get pods dns-test-75a159c3-75b3-11e9-8f67-2632f168be36) -May 13 19:15:49.030: INFO: Unable to read jessie_udp@dns-test-service.e2e-tests-dns-xhpc9 from pod e2e-tests-dns-xhpc9/dns-test-75a159c3-75b3-11e9-8f67-2632f168be36: the server could not find the requested resource (get pods dns-test-75a159c3-75b3-11e9-8f67-2632f168be36) -May 13 19:15:49.041: INFO: Unable to read jessie_tcp@dns-test-service.e2e-tests-dns-xhpc9 from pod e2e-tests-dns-xhpc9/dns-test-75a159c3-75b3-11e9-8f67-2632f168be36: the server could not find the requested resource (get pods dns-test-75a159c3-75b3-11e9-8f67-2632f168be36) -May 13 19:15:49.052: INFO: Unable to read jessie_udp@dns-test-service.e2e-tests-dns-xhpc9.svc from pod e2e-tests-dns-xhpc9/dns-test-75a159c3-75b3-11e9-8f67-2632f168be36: the server could not find the requested resource (get pods dns-test-75a159c3-75b3-11e9-8f67-2632f168be36) -May 13 19:15:49.061: INFO: Unable to read jessie_tcp@dns-test-service.e2e-tests-dns-xhpc9.svc from pod e2e-tests-dns-xhpc9/dns-test-75a159c3-75b3-11e9-8f67-2632f168be36: the server could not find the requested resource (get pods dns-test-75a159c3-75b3-11e9-8f67-2632f168be36) -May 13 19:15:49.071: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.e2e-tests-dns-xhpc9.svc from pod e2e-tests-dns-xhpc9/dns-test-75a159c3-75b3-11e9-8f67-2632f168be36: the server could not find the requested resource (get pods dns-test-75a159c3-75b3-11e9-8f67-2632f168be36) -May 13 19:15:49.082: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.e2e-tests-dns-xhpc9.svc from pod e2e-tests-dns-xhpc9/dns-test-75a159c3-75b3-11e9-8f67-2632f168be36: the server could not find the requested resource (get pods dns-test-75a159c3-75b3-11e9-8f67-2632f168be36) -May 13 19:15:49.144: INFO: Lookups using e2e-tests-dns-xhpc9/dns-test-75a159c3-75b3-11e9-8f67-2632f168be36 failed for: [wheezy_udp@dns-test-service wheezy_tcp@dns-test-service.e2e-tests-dns-xhpc9 wheezy_udp@_http._tcp.dns-test-service.e2e-tests-dns-xhpc9.svc wheezy_tcp@_http._tcp.dns-test-service.e2e-tests-dns-xhpc9.svc jessie_udp@dns-test-service jessie_tcp@dns-test-service jessie_udp@dns-test-service.e2e-tests-dns-xhpc9 jessie_tcp@dns-test-service.e2e-tests-dns-xhpc9 jessie_udp@dns-test-service.e2e-tests-dns-xhpc9.svc jessie_tcp@dns-test-service.e2e-tests-dns-xhpc9.svc jessie_udp@_http._tcp.dns-test-service.e2e-tests-dns-xhpc9.svc jessie_tcp@_http._tcp.dns-test-service.e2e-tests-dns-xhpc9.svc] - -May 13 19:15:53.841: INFO: Unable to read wheezy_udp@dns-test-service from pod e2e-tests-dns-xhpc9/dns-test-75a159c3-75b3-11e9-8f67-2632f168be36: the server could not find the requested resource (get pods dns-test-75a159c3-75b3-11e9-8f67-2632f168be36) -May 13 19:15:53.874: INFO: Unable to read wheezy_tcp@dns-test-service.e2e-tests-dns-xhpc9 from pod e2e-tests-dns-xhpc9/dns-test-75a159c3-75b3-11e9-8f67-2632f168be36: the server could not find the requested resource (get pods dns-test-75a159c3-75b3-11e9-8f67-2632f168be36) -May 13 19:15:53.934: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.e2e-tests-dns-xhpc9.svc from pod e2e-tests-dns-xhpc9/dns-test-75a159c3-75b3-11e9-8f67-2632f168be36: the server could not find the requested resource (get pods dns-test-75a159c3-75b3-11e9-8f67-2632f168be36) -May 13 19:15:53.946: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.e2e-tests-dns-xhpc9.svc from pod e2e-tests-dns-xhpc9/dns-test-75a159c3-75b3-11e9-8f67-2632f168be36: the server could not find the requested resource (get pods dns-test-75a159c3-75b3-11e9-8f67-2632f168be36) -May 13 19:15:54.021: INFO: Unable to read jessie_udp@dns-test-service from pod e2e-tests-dns-xhpc9/dns-test-75a159c3-75b3-11e9-8f67-2632f168be36: the server could not find the requested resource (get pods dns-test-75a159c3-75b3-11e9-8f67-2632f168be36) -May 13 19:15:54.031: INFO: Unable to read jessie_tcp@dns-test-service from pod e2e-tests-dns-xhpc9/dns-test-75a159c3-75b3-11e9-8f67-2632f168be36: the server could not find the requested resource (get pods dns-test-75a159c3-75b3-11e9-8f67-2632f168be36) -May 13 19:15:54.042: INFO: Unable to read jessie_udp@dns-test-service.e2e-tests-dns-xhpc9 from pod e2e-tests-dns-xhpc9/dns-test-75a159c3-75b3-11e9-8f67-2632f168be36: the server could not find the requested resource (get pods dns-test-75a159c3-75b3-11e9-8f67-2632f168be36) -May 13 19:15:54.053: INFO: Unable to read jessie_tcp@dns-test-service.e2e-tests-dns-xhpc9 from pod e2e-tests-dns-xhpc9/dns-test-75a159c3-75b3-11e9-8f67-2632f168be36: the server could not find the requested resource (get pods dns-test-75a159c3-75b3-11e9-8f67-2632f168be36) -May 13 19:15:54.063: INFO: Unable to read jessie_udp@dns-test-service.e2e-tests-dns-xhpc9.svc from pod e2e-tests-dns-xhpc9/dns-test-75a159c3-75b3-11e9-8f67-2632f168be36: the server could not find the requested resource (get pods dns-test-75a159c3-75b3-11e9-8f67-2632f168be36) -May 13 19:15:54.073: INFO: Unable to read jessie_tcp@dns-test-service.e2e-tests-dns-xhpc9.svc from pod e2e-tests-dns-xhpc9/dns-test-75a159c3-75b3-11e9-8f67-2632f168be36: the server could not find the requested resource (get pods dns-test-75a159c3-75b3-11e9-8f67-2632f168be36) -May 13 19:15:54.083: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.e2e-tests-dns-xhpc9.svc from pod e2e-tests-dns-xhpc9/dns-test-75a159c3-75b3-11e9-8f67-2632f168be36: the server could not find the requested resource (get pods dns-test-75a159c3-75b3-11e9-8f67-2632f168be36) -May 13 19:15:54.095: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.e2e-tests-dns-xhpc9.svc from pod e2e-tests-dns-xhpc9/dns-test-75a159c3-75b3-11e9-8f67-2632f168be36: the server could not find the requested resource (get pods dns-test-75a159c3-75b3-11e9-8f67-2632f168be36) -May 13 19:15:54.162: INFO: Lookups using e2e-tests-dns-xhpc9/dns-test-75a159c3-75b3-11e9-8f67-2632f168be36 failed for: [wheezy_udp@dns-test-service wheezy_tcp@dns-test-service.e2e-tests-dns-xhpc9 wheezy_udp@_http._tcp.dns-test-service.e2e-tests-dns-xhpc9.svc wheezy_tcp@_http._tcp.dns-test-service.e2e-tests-dns-xhpc9.svc jessie_udp@dns-test-service jessie_tcp@dns-test-service jessie_udp@dns-test-service.e2e-tests-dns-xhpc9 jessie_tcp@dns-test-service.e2e-tests-dns-xhpc9 jessie_udp@dns-test-service.e2e-tests-dns-xhpc9.svc jessie_tcp@dns-test-service.e2e-tests-dns-xhpc9.svc jessie_udp@_http._tcp.dns-test-service.e2e-tests-dns-xhpc9.svc jessie_tcp@_http._tcp.dns-test-service.e2e-tests-dns-xhpc9.svc] - -May 13 19:15:59.227: INFO: DNS probes using e2e-tests-dns-xhpc9/dns-test-75a159c3-75b3-11e9-8f67-2632f168be36 succeeded +STEP: create the rc +STEP: delete the rc +STEP: wait for the rc to be deleted +STEP: wait for 30 seconds to see if the garbage collector mistakenly deletes the pods +STEP: Gathering metrics +W0618 11:58:33.200421 17 metrics_grabber.go:81] Master node is not registered. Grabbing metrics from Scheduler, ControllerManager and ClusterAutoscaler is disabled. +Jun 18 11:58:33.200: INFO: For apiserver_request_count: +For apiserver_request_latencies_summary: +For etcd_helper_cache_entry_count: +For etcd_helper_cache_hit_count: +For etcd_helper_cache_miss_count: +For etcd_request_cache_add_latencies_summary: +For etcd_request_cache_get_latencies_summary: +For etcd_request_latencies_summary: +For garbage_collector_attempt_to_delete_queue_latency: +For garbage_collector_attempt_to_delete_work_duration: +For garbage_collector_attempt_to_orphan_queue_latency: +For garbage_collector_attempt_to_orphan_work_duration: +For garbage_collector_dirty_processing_latency_microseconds: +For garbage_collector_event_processing_latency_microseconds: +For garbage_collector_graph_changes_queue_latency: +For garbage_collector_graph_changes_work_duration: +For garbage_collector_orphan_processing_latency_microseconds: +For namespace_queue_latency: +For namespace_queue_latency_sum: +For namespace_queue_latency_count: +For namespace_retries: +For namespace_work_duration: +For namespace_work_duration_sum: +For namespace_work_duration_count: +For function_duration_seconds: +For errors_total: +For evicted_pods_total: -STEP: deleting the pod -STEP: deleting the test service -STEP: deleting the test headless service -[AfterEach] [sig-network] DNS +[AfterEach] [sig-api-machinery] Garbage collector /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 19:15:59.324: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-dns-xhpc9" for this suite. -May 13 19:16:05.362: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 19:16:05.777: INFO: namespace: e2e-tests-dns-xhpc9, resource: bindings, ignored listing per whitelist -May 13 19:16:05.832: INFO: namespace e2e-tests-dns-xhpc9 deletion completed in 6.499923168s +Jun 18 11:58:33.200: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "e2e-tests-gc-7zv6h" for this suite. +Jun 18 11:58:41.264: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 11:58:41.645: INFO: namespace: e2e-tests-gc-7zv6h, resource: bindings, ignored listing per whitelist +Jun 18 11:58:41.822: INFO: namespace e2e-tests-gc-7zv6h deletion completed in 8.607733478s -• [SLOW TEST:41.671 seconds] -[sig-network] DNS -/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:22 - should provide DNS for services [Conformance] +• [SLOW TEST:50.323 seconds] +[sig-api-machinery] Garbage collector +/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:22 + should orphan pods created by rc if delete options say so [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ -S +SSS ------------------------------ -[sig-storage] ConfigMap - should be consumable from pods in volume with mappings as non-root [NodeConformance] [Conformance] +[k8s.io] Container Lifecycle Hook when create a pod with lifecycle hook + should execute poststart http hook properly [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -[BeforeEach] [sig-storage] ConfigMap +[BeforeEach] [k8s.io] Container Lifecycle Hook /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client -May 13 19:16:05.832: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -STEP: Building a namespace api object, basename configmap -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-configmap-tkq9d +Jun 18 11:58:41.822: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 +STEP: Building a namespace api object, basename container-lifecycle-hook +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-container-lifecycle-hook-ckrfx STEP: Waiting for a default service account to be provisioned in namespace -[It] should be consumable from pods in volume with mappings as non-root [NodeConformance] [Conformance] +[BeforeEach] when create a pod with lifecycle hook + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:61 +STEP: create the container to handle the HTTPGet hook request. +[It] should execute poststart http hook properly [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -STEP: Creating configMap with name configmap-test-volume-map-8e764299-75b3-11e9-8f67-2632f168be36 -STEP: Creating a pod to test consume configMaps -May 13 19:16:06.146: INFO: Waiting up to 5m0s for pod "pod-configmaps-8e779268-75b3-11e9-8f67-2632f168be36" in namespace "e2e-tests-configmap-tkq9d" to be "success or failure" -May 13 19:16:06.154: INFO: Pod "pod-configmaps-8e779268-75b3-11e9-8f67-2632f168be36": Phase="Pending", Reason="", readiness=false. Elapsed: 7.732265ms -May 13 19:16:08.162: INFO: Pod "pod-configmaps-8e779268-75b3-11e9-8f67-2632f168be36": Phase="Running", Reason="", readiness=true. Elapsed: 2.016099431s -May 13 19:16:10.182: INFO: Pod "pod-configmaps-8e779268-75b3-11e9-8f67-2632f168be36": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.0358075s -STEP: Saw pod success -May 13 19:16:10.182: INFO: Pod "pod-configmaps-8e779268-75b3-11e9-8f67-2632f168be36" satisfied condition "success or failure" -May 13 19:16:10.190: INFO: Trying to get logs from node 10.170.219.177 pod pod-configmaps-8e779268-75b3-11e9-8f67-2632f168be36 container configmap-volume-test: -STEP: delete the pod -May 13 19:16:10.236: INFO: Waiting for pod pod-configmaps-8e779268-75b3-11e9-8f67-2632f168be36 to disappear -May 13 19:16:10.243: INFO: Pod pod-configmaps-8e779268-75b3-11e9-8f67-2632f168be36 no longer exists -[AfterEach] [sig-storage] ConfigMap +STEP: create the pod with lifecycle hook +STEP: check poststart hook +STEP: delete the pod with lifecycle hook +Jun 18 11:58:47.158: INFO: Waiting for pod pod-with-poststart-http-hook to disappear +Jun 18 11:58:47.174: INFO: Pod pod-with-poststart-http-hook still exists +Jun 18 11:58:49.174: INFO: Waiting for pod pod-with-poststart-http-hook to disappear +Jun 18 11:58:49.190: INFO: Pod pod-with-poststart-http-hook still exists +Jun 18 11:58:51.175: INFO: Waiting for pod pod-with-poststart-http-hook to disappear +Jun 18 11:58:51.189: INFO: Pod pod-with-poststart-http-hook still exists +Jun 18 11:58:53.174: INFO: Waiting for pod pod-with-poststart-http-hook to disappear +Jun 18 11:58:53.190: INFO: Pod pod-with-poststart-http-hook still exists +Jun 18 11:58:55.174: INFO: Waiting for pod pod-with-poststart-http-hook to disappear +Jun 18 11:58:55.207: INFO: Pod pod-with-poststart-http-hook still exists +Jun 18 11:58:57.174: INFO: Waiting for pod pod-with-poststart-http-hook to disappear +Jun 18 11:58:57.190: INFO: Pod pod-with-poststart-http-hook still exists +Jun 18 11:58:59.174: INFO: Waiting for pod pod-with-poststart-http-hook to disappear +Jun 18 11:58:59.199: INFO: Pod pod-with-poststart-http-hook still exists +Jun 18 11:59:01.174: INFO: Waiting for pod pod-with-poststart-http-hook to disappear +Jun 18 11:59:01.196: INFO: Pod pod-with-poststart-http-hook no longer exists +[AfterEach] [k8s.io] Container Lifecycle Hook /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 19:16:10.243: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-configmap-tkq9d" for this suite. -May 13 19:16:16.279: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 19:16:16.460: INFO: namespace: e2e-tests-configmap-tkq9d, resource: bindings, ignored listing per whitelist -May 13 19:16:16.546: INFO: namespace e2e-tests-configmap-tkq9d deletion completed in 6.295552206s +Jun 18 11:59:01.196: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "e2e-tests-container-lifecycle-hook-ckrfx" for this suite. +Jun 18 11:59:25.333: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 11:59:25.472: INFO: namespace: e2e-tests-container-lifecycle-hook-ckrfx, resource: bindings, ignored listing per whitelist +Jun 18 11:59:25.871: INFO: namespace e2e-tests-container-lifecycle-hook-ckrfx deletion completed in 24.586517682s -• [SLOW TEST:10.715 seconds] -[sig-storage] ConfigMap -/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:33 - should be consumable from pods in volume with mappings as non-root [NodeConformance] [Conformance] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 +• [SLOW TEST:44.049 seconds] +[k8s.io] Container Lifecycle Hook +/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 + when create a pod with lifecycle hook + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:40 + should execute poststart http hook properly [NodeConformance] [Conformance] + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ -SSSSSSSSSSSSSSSSSS +SSSSS ------------------------------ -[sig-storage] HostPath - should give a volume the correct mode [NodeConformance] [Conformance] +[k8s.io] Container Lifecycle Hook when create a pod with lifecycle hook + should execute prestop http hook properly [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -[BeforeEach] [sig-storage] HostPath +[BeforeEach] [k8s.io] Container Lifecycle Hook /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client -May 13 19:16:16.549: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -STEP: Building a namespace api object, basename hostpath -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-hostpath-6kkbc +Jun 18 11:59:25.872: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 +STEP: Building a namespace api object, basename container-lifecycle-hook +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-container-lifecycle-hook-4sdnf STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [sig-storage] HostPath - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/host_path.go:37 -[It] should give a volume the correct mode [NodeConformance] [Conformance] +[BeforeEach] when create a pod with lifecycle hook + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:61 +STEP: create the container to handle the HTTPGet hook request. +[It] should execute prestop http hook properly [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -STEP: Creating a pod to test hostPath mode -May 13 19:16:16.840: INFO: Waiting up to 5m0s for pod "pod-host-path-test" in namespace "e2e-tests-hostpath-6kkbc" to be "success or failure" -May 13 19:16:16.850: INFO: Pod "pod-host-path-test": Phase="Pending", Reason="", readiness=false. Elapsed: 9.585843ms -May 13 19:16:18.858: INFO: Pod "pod-host-path-test": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.017797788s -STEP: Saw pod success -May 13 19:16:18.858: INFO: Pod "pod-host-path-test" satisfied condition "success or failure" -May 13 19:16:18.866: INFO: Trying to get logs from node 10.170.219.177 pod pod-host-path-test container test-container-1: -STEP: delete the pod -May 13 19:16:18.906: INFO: Waiting for pod pod-host-path-test to disappear -May 13 19:16:18.913: INFO: Pod pod-host-path-test no longer exists -[AfterEach] [sig-storage] HostPath +STEP: create the pod with lifecycle hook +STEP: delete the pod with lifecycle hook +Jun 18 11:59:30.521: INFO: Waiting for pod pod-with-prestop-http-hook to disappear +Jun 18 11:59:30.535: INFO: Pod pod-with-prestop-http-hook still exists +Jun 18 11:59:32.535: INFO: Waiting for pod pod-with-prestop-http-hook to disappear +Jun 18 11:59:32.550: INFO: Pod pod-with-prestop-http-hook still exists +Jun 18 11:59:34.535: INFO: Waiting for pod pod-with-prestop-http-hook to disappear +Jun 18 11:59:34.552: INFO: Pod pod-with-prestop-http-hook still exists +Jun 18 11:59:36.535: INFO: Waiting for pod pod-with-prestop-http-hook to disappear +Jun 18 11:59:36.550: INFO: Pod pod-with-prestop-http-hook no longer exists +STEP: check prestop hook +[AfterEach] [k8s.io] Container Lifecycle Hook /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 19:16:18.913: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-hostpath-6kkbc" for this suite. -May 13 19:16:24.953: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 19:16:25.175: INFO: namespace: e2e-tests-hostpath-6kkbc, resource: bindings, ignored listing per whitelist -May 13 19:16:25.264: INFO: namespace e2e-tests-hostpath-6kkbc deletion completed in 6.338408087s +Jun 18 11:59:36.613: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "e2e-tests-container-lifecycle-hook-4sdnf" for this suite. +Jun 18 12:00:00.739: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:00:00.845: INFO: namespace: e2e-tests-container-lifecycle-hook-4sdnf, resource: bindings, ignored listing per whitelist +Jun 18 12:00:01.251: INFO: namespace e2e-tests-container-lifecycle-hook-4sdnf deletion completed in 24.566359761s -• [SLOW TEST:8.715 seconds] -[sig-storage] HostPath -/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/host_path.go:34 - should give a volume the correct mode [NodeConformance] [Conformance] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 +• [SLOW TEST:35.379 seconds] +[k8s.io] Container Lifecycle Hook +/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 + when create a pod with lifecycle hook + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:40 + should execute prestop http hook properly [NodeConformance] [Conformance] + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ -SSSSS +SSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[k8s.io] Pods - should contain environment variables for services [NodeConformance] [Conformance] +[sig-node] Downward API + should provide host IP as an env var [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -[BeforeEach] [k8s.io] Pods +[BeforeEach] [sig-node] Downward API /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client -May 13 19:16:25.266: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -STEP: Building a namespace api object, basename pods -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-pods-25npw +Jun 18 12:00:01.252: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 +STEP: Building a namespace api object, basename downward-api +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-downward-api-ps68s STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [k8s.io] Pods - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:132 -[It] should contain environment variables for services [NodeConformance] [Conformance] +[It] should provide host IP as an env var [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -May 13 19:16:27.851: INFO: Waiting up to 5m0s for pod "client-envvars-9b6882c2-75b3-11e9-8f67-2632f168be36" in namespace "e2e-tests-pods-25npw" to be "success or failure" -May 13 19:16:27.859: INFO: Pod "client-envvars-9b6882c2-75b3-11e9-8f67-2632f168be36": Phase="Pending", Reason="", readiness=false. Elapsed: 7.721469ms -May 13 19:16:29.867: INFO: Pod "client-envvars-9b6882c2-75b3-11e9-8f67-2632f168be36": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.015713151s +STEP: Creating a pod to test downward api env vars +Jun 18 12:00:01.821: INFO: Waiting up to 5m0s for pod "downward-api-9a2e11b9-91c0-11e9-bce2-ae54e022189f" in namespace "e2e-tests-downward-api-ps68s" to be "success or failure" +Jun 18 12:00:01.835: INFO: Pod "downward-api-9a2e11b9-91c0-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 14.4461ms +Jun 18 12:00:03.850: INFO: Pod "downward-api-9a2e11b9-91c0-11e9-bce2-ae54e022189f": Phase="Running", Reason="", readiness=true. Elapsed: 2.029147638s +Jun 18 12:00:05.865: INFO: Pod "downward-api-9a2e11b9-91c0-11e9-bce2-ae54e022189f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.044732589s STEP: Saw pod success -May 13 19:16:29.867: INFO: Pod "client-envvars-9b6882c2-75b3-11e9-8f67-2632f168be36" satisfied condition "success or failure" -May 13 19:16:29.874: INFO: Trying to get logs from node 10.170.219.177 pod client-envvars-9b6882c2-75b3-11e9-8f67-2632f168be36 container env3cont: +Jun 18 12:00:05.865: INFO: Pod "downward-api-9a2e11b9-91c0-11e9-bce2-ae54e022189f" satisfied condition "success or failure" +Jun 18 12:00:05.880: INFO: Trying to get logs from node 10.72.74.149 pod downward-api-9a2e11b9-91c0-11e9-bce2-ae54e022189f container dapi-container: STEP: delete the pod -May 13 19:16:29.913: INFO: Waiting for pod client-envvars-9b6882c2-75b3-11e9-8f67-2632f168be36 to disappear -May 13 19:16:29.921: INFO: Pod client-envvars-9b6882c2-75b3-11e9-8f67-2632f168be36 no longer exists -[AfterEach] [k8s.io] Pods +Jun 18 12:00:05.954: INFO: Waiting for pod downward-api-9a2e11b9-91c0-11e9-bce2-ae54e022189f to disappear +Jun 18 12:00:05.968: INFO: Pod downward-api-9a2e11b9-91c0-11e9-bce2-ae54e022189f no longer exists +[AfterEach] [sig-node] Downward API /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 19:16:29.921: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-pods-25npw" for this suite. -May 13 19:17:09.959: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 19:17:10.121: INFO: namespace: e2e-tests-pods-25npw, resource: bindings, ignored listing per whitelist -May 13 19:17:10.341: INFO: namespace e2e-tests-pods-25npw deletion completed in 40.412397184s +Jun 18 12:00:05.968: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "e2e-tests-downward-api-ps68s" for this suite. +Jun 18 12:00:12.054: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:00:12.247: INFO: namespace: e2e-tests-downward-api-ps68s, resource: bindings, ignored listing per whitelist +Jun 18 12:00:12.596: INFO: namespace e2e-tests-downward-api-ps68s deletion completed in 6.60657604s -• [SLOW TEST:45.076 seconds] -[k8s.io] Pods -/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 - should contain environment variables for services [NodeConformance] [Conformance] +• [SLOW TEST:11.344 seconds] +[sig-node] Downward API +/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downward_api.go:38 + should provide host IP as an env var [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSSS ------------------------------ -[sig-storage] Downward API volume - should update annotations on modification [NodeConformance] [Conformance] +[sig-storage] Projected secret + should be consumable from pods in volume with mappings [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -[BeforeEach] [sig-storage] Downward API volume +[BeforeEach] [sig-storage] Projected secret /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client -May 13 19:17:10.342: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -STEP: Building a namespace api object, basename downward-api -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-downward-api-9jx28 +Jun 18 12:00:12.596: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 +STEP: Building a namespace api object, basename projected +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-projected-pnjr4 STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [sig-storage] Downward API volume - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:39 -[It] should update annotations on modification [NodeConformance] [Conformance] +[It] should be consumable from pods in volume with mappings [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -STEP: Creating the pod -May 13 19:17:13.199: INFO: Successfully updated pod "annotationupdateb4e75dfa-75b3-11e9-8f67-2632f168be36" -[AfterEach] [sig-storage] Downward API volume +STEP: Creating projection with secret that has name projected-secret-test-map-a0e1ca97-91c0-11e9-bce2-ae54e022189f +STEP: Creating a pod to test consume secrets +Jun 18 12:00:13.094: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-a0e3c2ae-91c0-11e9-bce2-ae54e022189f" in namespace "e2e-tests-projected-pnjr4" to be "success or failure" +Jun 18 12:00:13.113: INFO: Pod "pod-projected-secrets-a0e3c2ae-91c0-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 18.57701ms +Jun 18 12:00:15.129: INFO: Pod "pod-projected-secrets-a0e3c2ae-91c0-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 2.035149747s +Jun 18 12:00:17.144: INFO: Pod "pod-projected-secrets-a0e3c2ae-91c0-11e9-bce2-ae54e022189f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.050113616s +STEP: Saw pod success +Jun 18 12:00:17.144: INFO: Pod "pod-projected-secrets-a0e3c2ae-91c0-11e9-bce2-ae54e022189f" satisfied condition "success or failure" +Jun 18 12:00:17.160: INFO: Trying to get logs from node 10.72.74.149 pod pod-projected-secrets-a0e3c2ae-91c0-11e9-bce2-ae54e022189f container projected-secret-volume-test: +STEP: delete the pod +Jun 18 12:00:17.238: INFO: Waiting for pod pod-projected-secrets-a0e3c2ae-91c0-11e9-bce2-ae54e022189f to disappear +Jun 18 12:00:17.299: INFO: Pod pod-projected-secrets-a0e3c2ae-91c0-11e9-bce2-ae54e022189f no longer exists +[AfterEach] [sig-storage] Projected secret /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 19:17:17.260: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-downward-api-9jx28" for this suite. -May 13 19:17:35.301: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 19:17:35.500: INFO: namespace: e2e-tests-downward-api-9jx28, resource: bindings, ignored listing per whitelist -May 13 19:17:35.549: INFO: namespace e2e-tests-downward-api-9jx28 deletion completed in 18.280846578s +Jun 18 12:00:17.299: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "e2e-tests-projected-pnjr4" for this suite. +Jun 18 12:00:25.376: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:00:25.912: INFO: namespace: e2e-tests-projected-pnjr4, resource: bindings, ignored listing per whitelist +Jun 18 12:00:25.951: INFO: namespace e2e-tests-projected-pnjr4 deletion completed in 8.632905815s -• [SLOW TEST:25.207 seconds] -[sig-storage] Downward API volume -/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:34 - should update annotations on modification [NodeConformance] [Conformance] +• [SLOW TEST:13.355 seconds] +[sig-storage] Projected secret +/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_secret.go:34 + should be consumable from pods in volume with mappings [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ -SSSSSSSSSSS +SSSSS ------------------------------ -[sig-cli] Kubectl client [k8s.io] Proxy server - should support --unix-socket=/path [Conformance] +[sig-api-machinery] Garbage collector + should orphan RS created by deployment when deleteOptions.PropagationPolicy is Orphan [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -[BeforeEach] [sig-cli] Kubectl client +[BeforeEach] [sig-api-machinery] Garbage collector /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client -May 13 19:17:35.550: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -STEP: Building a namespace api object, basename kubectl -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-kubectl-qj5jf +Jun 18 12:00:25.952: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 +STEP: Building a namespace api object, basename gc +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-gc-bsxl6 STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [sig-cli] Kubectl client - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:243 -[It] should support --unix-socket=/path [Conformance] +[It] should orphan RS created by deployment when deleteOptions.PropagationPolicy is Orphan [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -STEP: Starting the proxy -May 13 19:17:35.819: INFO: Asynchronously running '/usr/local/bin/kubectl kubectl --kubeconfig=/tmp/kubeconfig-555138423 proxy --unix-socket=/tmp/kubectl-proxy-unix058550378/test' -STEP: retrieving proxy /api/ output -[AfterEach] [sig-cli] Kubectl client +STEP: create the deployment +STEP: Wait for the Deployment to create new ReplicaSet +STEP: delete the deployment +STEP: wait for 30 seconds to see if the garbage collector mistakenly deletes the rs +STEP: Gathering metrics +W0618 12:00:57.049622 17 metrics_grabber.go:81] Master node is not registered. Grabbing metrics from Scheduler, ControllerManager and ClusterAutoscaler is disabled. +Jun 18 12:00:57.049: INFO: For apiserver_request_count: +For apiserver_request_latencies_summary: +For etcd_helper_cache_entry_count: +For etcd_helper_cache_hit_count: +For etcd_helper_cache_miss_count: +For etcd_request_cache_add_latencies_summary: +For etcd_request_cache_get_latencies_summary: +For etcd_request_latencies_summary: +For garbage_collector_attempt_to_delete_queue_latency: +For garbage_collector_attempt_to_delete_work_duration: +For garbage_collector_attempt_to_orphan_queue_latency: +For garbage_collector_attempt_to_orphan_work_duration: +For garbage_collector_dirty_processing_latency_microseconds: +For garbage_collector_event_processing_latency_microseconds: +For garbage_collector_graph_changes_queue_latency: +For garbage_collector_graph_changes_work_duration: +For garbage_collector_orphan_processing_latency_microseconds: +For namespace_queue_latency: +For namespace_queue_latency_sum: +For namespace_queue_latency_count: +For namespace_retries: +For namespace_work_duration: +For namespace_work_duration_sum: +For namespace_work_duration_count: +For function_duration_seconds: +For errors_total: +For evicted_pods_total: + +[AfterEach] [sig-api-machinery] Garbage collector /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 19:17:35.882: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-kubectl-qj5jf" for this suite. -May 13 19:17:41.918: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 19:17:42.065: INFO: namespace: e2e-tests-kubectl-qj5jf, resource: bindings, ignored listing per whitelist -May 13 19:17:42.281: INFO: namespace e2e-tests-kubectl-qj5jf deletion completed in 6.390768653s +Jun 18 12:00:57.049: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "e2e-tests-gc-bsxl6" for this suite. +Jun 18 12:01:03.205: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:01:03.313: INFO: namespace: e2e-tests-gc-bsxl6, resource: bindings, ignored listing per whitelist +Jun 18 12:01:03.765: INFO: namespace e2e-tests-gc-bsxl6 deletion completed in 6.680265217s -• [SLOW TEST:6.731 seconds] -[sig-cli] Kubectl client -/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:22 - [k8s.io] Proxy server - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 - should support --unix-socket=/path [Conformance] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 +• [SLOW TEST:37.814 seconds] +[sig-api-machinery] Garbage collector +/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:22 + should orphan RS created by deployment when deleteOptions.PropagationPolicy is Orphan [Conformance] + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ -SSSSSS +SSSSSSSSSSSSSSSSSS ------------------------------ -[sig-cli] Kubectl client [k8s.io] Kubectl replace - should update a single-container pod's image [Conformance] +[sig-cli] Kubectl client [k8s.io] Kubectl run --rm job + should create a job from an image, then delete the job [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-cli] Kubectl client /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client -May 13 19:17:42.282: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 +Jun 18 12:01:03.769: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 STEP: Building a namespace api object, basename kubectl -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-kubectl-sxt2f +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-kubectl-4cfh7 STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:243 -[BeforeEach] [k8s.io] Kubectl replace - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1563 -[It] should update a single-container pod's image [Conformance] +[It] should create a job from an image, then delete the job [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -STEP: running the image docker.io/library/nginx:1.14-alpine -May 13 19:17:42.547: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 run e2e-test-nginx-pod --generator=run-pod/v1 --image=docker.io/library/nginx:1.14-alpine --labels=run=e2e-test-nginx-pod --namespace=e2e-tests-kubectl-sxt2f' -May 13 19:17:42.845: INFO: stderr: "" -May 13 19:17:42.845: INFO: stdout: "pod/e2e-test-nginx-pod created\n" -STEP: verifying the pod e2e-test-nginx-pod is running -STEP: verifying the pod e2e-test-nginx-pod was created -May 13 19:17:47.896: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 get pod e2e-test-nginx-pod --namespace=e2e-tests-kubectl-sxt2f -o json' -May 13 19:17:48.011: INFO: stderr: "" -May 13 19:17:48.011: INFO: stdout: "{\n \"apiVersion\": \"v1\",\n \"kind\": \"Pod\",\n \"metadata\": {\n \"annotations\": {\n \"kubernetes.io/psp\": \"e2e-test-privileged-psp\"\n },\n \"creationTimestamp\": \"2019-05-13T19:17:42Z\",\n \"labels\": {\n \"run\": \"e2e-test-nginx-pod\"\n },\n \"name\": \"e2e-test-nginx-pod\",\n \"namespace\": \"e2e-tests-kubectl-sxt2f\",\n \"resourceVersion\": \"40772\",\n \"selfLink\": \"/api/v1/namespaces/e2e-tests-kubectl-sxt2f/pods/e2e-test-nginx-pod\",\n \"uid\": \"c81a14bf-75b3-11e9-8e1a-4ad9c8c454e0\"\n },\n \"spec\": {\n \"containers\": [\n {\n \"image\": \"docker.io/library/nginx:1.14-alpine\",\n \"imagePullPolicy\": \"IfNotPresent\",\n \"name\": \"e2e-test-nginx-pod\",\n \"resources\": {},\n \"terminationMessagePath\": \"/dev/termination-log\",\n \"terminationMessagePolicy\": \"File\",\n \"volumeMounts\": [\n {\n \"mountPath\": \"/var/run/secrets/kubernetes.io/serviceaccount\",\n \"name\": \"default-token-fbl5r\",\n \"readOnly\": true\n }\n ]\n }\n ],\n \"dnsPolicy\": \"ClusterFirst\",\n \"enableServiceLinks\": true,\n \"nodeName\": \"10.170.219.177\",\n \"priority\": 0,\n \"restartPolicy\": \"Always\",\n \"schedulerName\": \"default-scheduler\",\n \"securityContext\": {},\n \"serviceAccount\": \"default\",\n \"serviceAccountName\": \"default\",\n \"terminationGracePeriodSeconds\": 30,\n \"tolerations\": [\n {\n \"effect\": \"NoExecute\",\n \"key\": \"node.kubernetes.io/not-ready\",\n \"operator\": \"Exists\",\n \"tolerationSeconds\": 300\n },\n {\n \"effect\": \"NoExecute\",\n \"key\": \"node.kubernetes.io/unreachable\",\n \"operator\": \"Exists\",\n \"tolerationSeconds\": 300\n }\n ],\n \"volumes\": [\n {\n \"name\": \"default-token-fbl5r\",\n \"secret\": {\n \"defaultMode\": 420,\n \"secretName\": \"default-token-fbl5r\"\n }\n }\n ]\n },\n \"status\": {\n \"conditions\": [\n {\n \"lastProbeTime\": null,\n \"lastTransitionTime\": \"2019-05-13T19:17:42Z\",\n \"status\": \"True\",\n \"type\": \"Initialized\"\n },\n {\n \"lastProbeTime\": null,\n \"lastTransitionTime\": \"2019-05-13T19:17:44Z\",\n \"status\": \"True\",\n \"type\": \"Ready\"\n },\n {\n \"lastProbeTime\": null,\n \"lastTransitionTime\": \"2019-05-13T19:17:44Z\",\n \"status\": \"True\",\n \"type\": \"ContainersReady\"\n },\n {\n \"lastProbeTime\": null,\n \"lastTransitionTime\": \"2019-05-13T19:17:42Z\",\n \"status\": \"True\",\n \"type\": \"PodScheduled\"\n }\n ],\n \"containerStatuses\": [\n {\n \"containerID\": \"containerd://9b14c725dec78eeb6cef878824ae32d03177b624f5042f29d13b6cfc7c683831\",\n \"image\": \"docker.io/library/nginx:1.14-alpine\",\n \"imageID\": \"docker.io/library/nginx@sha256:485b610fefec7ff6c463ced9623314a04ed67e3945b9c08d7e53a47f6d108dc7\",\n \"lastState\": {},\n \"name\": \"e2e-test-nginx-pod\",\n \"ready\": true,\n \"restartCount\": 0,\n \"state\": {\n \"running\": {\n \"startedAt\": \"2019-05-13T19:17:44Z\"\n }\n }\n }\n ],\n \"hostIP\": \"10.170.219.177\",\n \"phase\": \"Running\",\n \"podIP\": \"172.30.227.117\",\n \"qosClass\": \"BestEffort\",\n \"startTime\": \"2019-05-13T19:17:42Z\"\n }\n}\n" -STEP: replace the image in the pod -May 13 19:17:48.012: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 replace -f - --namespace=e2e-tests-kubectl-sxt2f' -May 13 19:17:48.265: INFO: stderr: "" -May 13 19:17:48.265: INFO: stdout: "pod/e2e-test-nginx-pod replaced\n" -STEP: verifying the pod e2e-test-nginx-pod has the right image docker.io/library/busybox:1.29 -[AfterEach] [k8s.io] Kubectl replace - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1568 -May 13 19:17:48.274: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 delete pods e2e-test-nginx-pod --namespace=e2e-tests-kubectl-sxt2f' -May 13 19:18:01.321: INFO: stderr: "" -May 13 19:18:01.321: INFO: stdout: "pod \"e2e-test-nginx-pod\" deleted\n" +STEP: executing a command with run --rm and attach with stdin +Jun 18 12:01:04.222: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 --namespace=e2e-tests-kubectl-4cfh7 run e2e-test-rm-busybox-job --image=docker.io/library/busybox:1.29 --rm=true --generator=job/v1 --restart=OnFailure --attach=true --stdin -- sh -c cat && echo 'stdin closed'' +Jun 18 12:01:06.622: INFO: stderr: "kubectl run --generator=job/v1 is DEPRECATED and will be removed in a future version. Use kubectl run --generator=run-pod/v1 or kubectl create instead.\nIf you don't see a command prompt, try pressing enter.\n" +Jun 18 12:01:06.622: INFO: stdout: "abcd1234stdin closed\njob.batch \"e2e-test-rm-busybox-job\" deleted\n" +STEP: verifying the job e2e-test-rm-busybox-job was deleted [AfterEach] [sig-cli] Kubectl client /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 19:18:01.322: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-kubectl-sxt2f" for this suite. -May 13 19:18:07.376: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 19:18:07.689: INFO: namespace: e2e-tests-kubectl-sxt2f, resource: bindings, ignored listing per whitelist -May 13 19:18:07.836: INFO: namespace e2e-tests-kubectl-sxt2f deletion completed in 6.491615668s +Jun 18 12:01:08.684: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "e2e-tests-kubectl-4cfh7" for this suite. +Jun 18 12:01:16.760: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:01:16.992: INFO: namespace: e2e-tests-kubectl-4cfh7, resource: bindings, ignored listing per whitelist +Jun 18 12:01:17.535: INFO: namespace e2e-tests-kubectl-4cfh7 deletion completed in 8.831526822s -• [SLOW TEST:25.555 seconds] +• [SLOW TEST:13.766 seconds] [sig-cli] Kubectl client /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:22 - [k8s.io] Kubectl replace + [k8s.io] Kubectl run --rm job /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 - should update a single-container pod's image [Conformance] + should create a job from an image, then delete the job [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ -S ------------------------------- -[k8s.io] Docker Containers - should use the image defaults if command and args are blank [NodeConformance] [Conformance] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -[BeforeEach] [k8s.io] Docker Containers - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 -STEP: Creating a kubernetes client -May 13 19:18:07.837: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -STEP: Building a namespace api object, basename containers -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-containers-bnrrz -STEP: Waiting for a default service account to be provisioned in namespace -[It] should use the image defaults if command and args are blank [NodeConformance] [Conformance] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -STEP: Creating a pod to test use defaults -May 13 19:18:08.136: INFO: Waiting up to 5m0s for pod "client-containers-d72dd42a-75b3-11e9-8f67-2632f168be36" in namespace "e2e-tests-containers-bnrrz" to be "success or failure" -May 13 19:18:08.144: INFO: Pod "client-containers-d72dd42a-75b3-11e9-8f67-2632f168be36": Phase="Pending", Reason="", readiness=false. Elapsed: 8.48713ms -May 13 19:18:10.152: INFO: Pod "client-containers-d72dd42a-75b3-11e9-8f67-2632f168be36": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.016220531s -STEP: Saw pod success -May 13 19:18:10.152: INFO: Pod "client-containers-d72dd42a-75b3-11e9-8f67-2632f168be36" satisfied condition "success or failure" -May 13 19:18:10.229: INFO: Trying to get logs from node 10.170.219.153 pod client-containers-d72dd42a-75b3-11e9-8f67-2632f168be36 container test-container: -STEP: delete the pod -May 13 19:18:10.273: INFO: Waiting for pod client-containers-d72dd42a-75b3-11e9-8f67-2632f168be36 to disappear -May 13 19:18:10.280: INFO: Pod client-containers-d72dd42a-75b3-11e9-8f67-2632f168be36 no longer exists -[AfterEach] [k8s.io] Docker Containers - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 19:18:10.280: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-containers-bnrrz" for this suite. -May 13 19:18:16.315: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 19:18:16.512: INFO: namespace: e2e-tests-containers-bnrrz, resource: bindings, ignored listing per whitelist -May 13 19:18:16.587: INFO: namespace e2e-tests-containers-bnrrz deletion completed in 6.299171012s - -• [SLOW TEST:8.749 seconds] -[k8s.io] Docker Containers -/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 - should use the image defaults if command and args are blank [NodeConformance] [Conformance] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------- -SSSSS +SSSSSSSSSSSSSS ------------------------------ -[k8s.io] Pods - should be submitted and removed [NodeConformance] [Conformance] +[sig-storage] ConfigMap + optional updates should be reflected in volume [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -[BeforeEach] [k8s.io] Pods +[BeforeEach] [sig-storage] ConfigMap /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client -May 13 19:18:16.587: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -STEP: Building a namespace api object, basename pods -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-pods-5nwsq +Jun 18 12:01:17.536: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 +STEP: Building a namespace api object, basename configmap +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-configmap-4n88m STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [k8s.io] Pods - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:132 -[It] should be submitted and removed [NodeConformance] [Conformance] +[It] optional updates should be reflected in volume [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -STEP: creating the pod -STEP: setting up watch -STEP: submitting the pod to kubernetes -STEP: verifying the pod is in kubernetes -STEP: verifying pod creation was observed -May 13 19:18:18.916: INFO: running pod: &v1.Pod{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"pod-submit-remove-dc629541-75b3-11e9-8f67-2632f168be36", GenerateName:"", Namespace:"e2e-tests-pods-5nwsq", SelfLink:"/api/v1/namespaces/e2e-tests-pods-5nwsq/pods/pod-submit-remove-dc629541-75b3-11e9-8f67-2632f168be36", UID:"dc657334-75b3-11e9-b786-da20024d205c", ResourceVersion:"40941", Generation:0, CreationTimestamp:v1.Time{Time:time.Time{wall:0x0, ext:63693371896, loc:(*time.Location)(0x7b33b80)}}, DeletionTimestamp:(*v1.Time)(nil), DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"name":"foo", "time":"855487199"}, Annotations:map[string]string{"kubernetes.io/psp":"e2e-test-privileged-psp"}, OwnerReferences:[]v1.OwnerReference(nil), Initializers:(*v1.Initializers)(nil), Finalizers:[]string(nil), ClusterName:""}, Spec:v1.PodSpec{Volumes:[]v1.Volume{v1.Volume{Name:"default-token-bqmmt", VolumeSource:v1.VolumeSource{HostPath:(*v1.HostPathVolumeSource)(nil), EmptyDir:(*v1.EmptyDirVolumeSource)(nil), GCEPersistentDisk:(*v1.GCEPersistentDiskVolumeSource)(nil), AWSElasticBlockStore:(*v1.AWSElasticBlockStoreVolumeSource)(nil), GitRepo:(*v1.GitRepoVolumeSource)(nil), Secret:(*v1.SecretVolumeSource)(0xc001219180), NFS:(*v1.NFSVolumeSource)(nil), ISCSI:(*v1.ISCSIVolumeSource)(nil), Glusterfs:(*v1.GlusterfsVolumeSource)(nil), PersistentVolumeClaim:(*v1.PersistentVolumeClaimVolumeSource)(nil), RBD:(*v1.RBDVolumeSource)(nil), FlexVolume:(*v1.FlexVolumeSource)(nil), Cinder:(*v1.CinderVolumeSource)(nil), CephFS:(*v1.CephFSVolumeSource)(nil), Flocker:(*v1.FlockerVolumeSource)(nil), DownwardAPI:(*v1.DownwardAPIVolumeSource)(nil), FC:(*v1.FCVolumeSource)(nil), AzureFile:(*v1.AzureFileVolumeSource)(nil), ConfigMap:(*v1.ConfigMapVolumeSource)(nil), VsphereVolume:(*v1.VsphereVirtualDiskVolumeSource)(nil), Quobyte:(*v1.QuobyteVolumeSource)(nil), AzureDisk:(*v1.AzureDiskVolumeSource)(nil), PhotonPersistentDisk:(*v1.PhotonPersistentDiskVolumeSource)(nil), Projected:(*v1.ProjectedVolumeSource)(nil), PortworxVolume:(*v1.PortworxVolumeSource)(nil), ScaleIO:(*v1.ScaleIOVolumeSource)(nil), StorageOS:(*v1.StorageOSVolumeSource)(nil)}}}, InitContainers:[]v1.Container(nil), Containers:[]v1.Container{v1.Container{Name:"nginx", Image:"docker.io/library/nginx:1.14-alpine", Command:[]string(nil), Args:[]string(nil), WorkingDir:"", Ports:[]v1.ContainerPort(nil), EnvFrom:[]v1.EnvFromSource(nil), Env:[]v1.EnvVar(nil), Resources:v1.ResourceRequirements{Limits:v1.ResourceList(nil), Requests:v1.ResourceList(nil)}, VolumeMounts:[]v1.VolumeMount{v1.VolumeMount{Name:"default-token-bqmmt", ReadOnly:true, MountPath:"/var/run/secrets/kubernetes.io/serviceaccount", SubPath:"", MountPropagation:(*v1.MountPropagationMode)(nil)}}, VolumeDevices:[]v1.VolumeDevice(nil), LivenessProbe:(*v1.Probe)(nil), ReadinessProbe:(*v1.Probe)(nil), Lifecycle:(*v1.Lifecycle)(nil), TerminationMessagePath:"/dev/termination-log", TerminationMessagePolicy:"File", ImagePullPolicy:"IfNotPresent", SecurityContext:(*v1.SecurityContext)(nil), Stdin:false, StdinOnce:false, TTY:false}}, RestartPolicy:"Always", TerminationGracePeriodSeconds:(*int64)(0xc0014c8818), ActiveDeadlineSeconds:(*int64)(nil), DNSPolicy:"ClusterFirst", NodeSelector:map[string]string(nil), ServiceAccountName:"default", DeprecatedServiceAccount:"default", AutomountServiceAccountToken:(*bool)(nil), NodeName:"10.170.219.177", HostNetwork:false, HostPID:false, HostIPC:false, ShareProcessNamespace:(*bool)(nil), SecurityContext:(*v1.PodSecurityContext)(0xc001a85aa0), ImagePullSecrets:[]v1.LocalObjectReference(nil), Hostname:"", Subdomain:"", Affinity:(*v1.Affinity)(nil), SchedulerName:"default-scheduler", Tolerations:[]v1.Toleration{v1.Toleration{Key:"node.kubernetes.io/not-ready", Operator:"Exists", Value:"", Effect:"NoExecute", TolerationSeconds:(*int64)(0xc0014c8860)}, v1.Toleration{Key:"node.kubernetes.io/unreachable", Operator:"Exists", Value:"", Effect:"NoExecute", TolerationSeconds:(*int64)(0xc0014c8bb0)}}, HostAliases:[]v1.HostAlias(nil), PriorityClassName:"", Priority:(*int32)(0xc0014c8bb8), DNSConfig:(*v1.PodDNSConfig)(nil), ReadinessGates:[]v1.PodReadinessGate(nil), RuntimeClassName:(*string)(nil), EnableServiceLinks:(*bool)(0xc0014c8bbc)}, Status:v1.PodStatus{Phase:"Running", Conditions:[]v1.PodCondition{v1.PodCondition{Type:"Initialized", Status:"True", LastProbeTime:v1.Time{Time:time.Time{wall:0x0, ext:0, loc:(*time.Location)(nil)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63693371896, loc:(*time.Location)(0x7b33b80)}}, Reason:"", Message:""}, v1.PodCondition{Type:"Ready", Status:"True", LastProbeTime:v1.Time{Time:time.Time{wall:0x0, ext:0, loc:(*time.Location)(nil)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63693371898, loc:(*time.Location)(0x7b33b80)}}, Reason:"", Message:""}, v1.PodCondition{Type:"ContainersReady", Status:"True", LastProbeTime:v1.Time{Time:time.Time{wall:0x0, ext:0, loc:(*time.Location)(nil)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63693371898, loc:(*time.Location)(0x7b33b80)}}, Reason:"", Message:""}, v1.PodCondition{Type:"PodScheduled", Status:"True", LastProbeTime:v1.Time{Time:time.Time{wall:0x0, ext:0, loc:(*time.Location)(nil)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63693371896, loc:(*time.Location)(0x7b33b80)}}, Reason:"", Message:""}}, Message:"", Reason:"", NominatedNodeName:"", HostIP:"10.170.219.177", PodIP:"172.30.227.123", StartTime:(*v1.Time)(0xc001427a20), InitContainerStatuses:[]v1.ContainerStatus(nil), ContainerStatuses:[]v1.ContainerStatus{v1.ContainerStatus{Name:"nginx", State:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(nil), Running:(*v1.ContainerStateRunning)(0xc001427a40), Terminated:(*v1.ContainerStateTerminated)(nil)}, LastTerminationState:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(nil), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(nil)}, Ready:true, RestartCount:0, Image:"docker.io/library/nginx:1.14-alpine", ImageID:"docker.io/library/nginx@sha256:485b610fefec7ff6c463ced9623314a04ed67e3945b9c08d7e53a47f6d108dc7", ContainerID:"containerd://663e0b11eff0a86565118312302ebac94275ab95a73e6e60ca2793cb4671828c"}}, QOSClass:"BestEffort"}} -STEP: deleting the pod gracefully -STEP: verifying the kubelet observed the termination notice -STEP: verifying pod deletion was observed -[AfterEach] [k8s.io] Pods +STEP: Creating configMap with name cm-test-opt-del-c7985325-91c0-11e9-bce2-ae54e022189f +STEP: Creating configMap with name cm-test-opt-upd-c798537e-91c0-11e9-bce2-ae54e022189f +STEP: Creating the pod +STEP: Deleting configmap cm-test-opt-del-c7985325-91c0-11e9-bce2-ae54e022189f +STEP: Updating configmap cm-test-opt-upd-c798537e-91c0-11e9-bce2-ae54e022189f +STEP: Creating configMap with name cm-test-opt-create-c79853a7-91c0-11e9-bce2-ae54e022189f +STEP: waiting to observe update in volume +[AfterEach] [sig-storage] ConfigMap /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 19:18:31.326: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-pods-5nwsq" for this suite. -May 13 19:18:37.364: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 19:18:37.558: INFO: namespace: e2e-tests-pods-5nwsq, resource: bindings, ignored listing per whitelist -May 13 19:18:37.660: INFO: namespace e2e-tests-pods-5nwsq deletion completed in 6.324462116s +Jun 18 12:02:31.953: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "e2e-tests-configmap-4n88m" for this suite. +Jun 18 12:02:48.036: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:02:48.813: INFO: namespace: e2e-tests-configmap-4n88m, resource: bindings, ignored listing per whitelist +Jun 18 12:02:48.955: INFO: namespace e2e-tests-configmap-4n88m deletion completed in 16.981179102s -• [SLOW TEST:21.073 seconds] -[k8s.io] Pods -/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 - should be submitted and removed [NodeConformance] [Conformance] +• [SLOW TEST:91.418 seconds] +[sig-storage] ConfigMap +/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:33 + optional updates should be reflected in volume [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSSSSSSS ------------------------------ -[sig-storage] Projected configMap - should be consumable from pods in volume with mappings [NodeConformance] [Conformance] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -[BeforeEach] [sig-storage] Projected configMap - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 -STEP: Creating a kubernetes client -May 13 19:18:37.662: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -STEP: Building a namespace api object, basename projected -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-projected-s5zqc -STEP: Waiting for a default service account to be provisioned in namespace -[It] should be consumable from pods in volume with mappings [NodeConformance] [Conformance] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -STEP: Creating configMap with name projected-configmap-test-volume-map-e8f4b05c-75b3-11e9-8f67-2632f168be36 -STEP: Creating a pod to test consume configMaps -May 13 19:18:37.970: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-e8f612b1-75b3-11e9-8f67-2632f168be36" in namespace "e2e-tests-projected-s5zqc" to be "success or failure" -May 13 19:18:37.978: INFO: Pod "pod-projected-configmaps-e8f612b1-75b3-11e9-8f67-2632f168be36": Phase="Pending", Reason="", readiness=false. Elapsed: 8.264931ms -May 13 19:18:39.987: INFO: Pod "pod-projected-configmaps-e8f612b1-75b3-11e9-8f67-2632f168be36": Phase="Running", Reason="", readiness=true. Elapsed: 2.017145147s -May 13 19:18:41.995: INFO: Pod "pod-projected-configmaps-e8f612b1-75b3-11e9-8f67-2632f168be36": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.025146902s -STEP: Saw pod success -May 13 19:18:41.995: INFO: Pod "pod-projected-configmaps-e8f612b1-75b3-11e9-8f67-2632f168be36" satisfied condition "success or failure" -May 13 19:18:42.020: INFO: Trying to get logs from node 10.170.219.153 pod pod-projected-configmaps-e8f612b1-75b3-11e9-8f67-2632f168be36 container projected-configmap-volume-test: -STEP: delete the pod -May 13 19:18:42.058: INFO: Waiting for pod pod-projected-configmaps-e8f612b1-75b3-11e9-8f67-2632f168be36 to disappear -May 13 19:18:42.066: INFO: Pod pod-projected-configmaps-e8f612b1-75b3-11e9-8f67-2632f168be36 no longer exists -[AfterEach] [sig-storage] Projected configMap - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 19:18:42.066: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-projected-s5zqc" for this suite. -May 13 19:18:48.102: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 19:18:48.234: INFO: namespace: e2e-tests-projected-s5zqc, resource: bindings, ignored listing per whitelist -May 13 19:18:48.349: INFO: namespace e2e-tests-projected-s5zqc deletion completed in 6.275239784s - -• [SLOW TEST:10.688 seconds] -[sig-storage] Projected configMap -/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_configmap.go:34 - should be consumable from pods in volume with mappings [NodeConformance] [Conformance] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------- -SSSS ------------------------------- -[sig-api-machinery] Garbage collector - should not be blocked by dependency circle [Conformance] +[sig-storage] Projected downwardAPI + should provide container's memory request [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -[BeforeEach] [sig-api-machinery] Garbage collector +[BeforeEach] [sig-storage] Projected downwardAPI /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client -May 13 19:18:48.351: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -STEP: Building a namespace api object, basename gc -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-gc-kvxxq +Jun 18 12:02:48.955: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 +STEP: Building a namespace api object, basename projected +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-projected-v642t STEP: Waiting for a default service account to be provisioned in namespace -[It] should not be blocked by dependency circle [Conformance] +[BeforeEach] [sig-storage] Projected downwardAPI + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:39 +[It] should provide container's memory request [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -May 13 19:18:48.775: INFO: pod1.ObjectMeta.OwnerReferences=[]v1.OwnerReference{v1.OwnerReference{APIVersion:"v1", Kind:"Pod", Name:"pod3", UID:"ef65f172-75b3-11e9-b786-da20024d205c", Controller:(*bool)(0xc0024a56c2), BlockOwnerDeletion:(*bool)(0xc0024a56c3)}} -May 13 19:18:48.792: INFO: pod2.ObjectMeta.OwnerReferences=[]v1.OwnerReference{v1.OwnerReference{APIVersion:"v1", Kind:"Pod", Name:"pod1", UID:"ef62af6d-75b3-11e9-b786-da20024d205c", Controller:(*bool)(0xc001958fbe), BlockOwnerDeletion:(*bool)(0xc001958fbf)}} -May 13 19:18:48.800: INFO: pod3.ObjectMeta.OwnerReferences=[]v1.OwnerReference{v1.OwnerReference{APIVersion:"v1", Kind:"Pod", Name:"pod2", UID:"ef6456a6-75b3-11e9-b786-da20024d205c", Controller:(*bool)(0xc0024a58d6), BlockOwnerDeletion:(*bool)(0xc0024a58d7)}} -[AfterEach] [sig-api-machinery] Garbage collector +STEP: Creating a pod to test downward API volume plugin +Jun 18 12:02:49.648: INFO: Waiting up to 5m0s for pod "downwardapi-volume-fe361889-91c0-11e9-bce2-ae54e022189f" in namespace "e2e-tests-projected-v642t" to be "success or failure" +Jun 18 12:02:49.662: INFO: Pod "downwardapi-volume-fe361889-91c0-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 14.230622ms +Jun 18 12:02:51.677: INFO: Pod "downwardapi-volume-fe361889-91c0-11e9-bce2-ae54e022189f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.028814137s +STEP: Saw pod success +Jun 18 12:02:51.677: INFO: Pod "downwardapi-volume-fe361889-91c0-11e9-bce2-ae54e022189f" satisfied condition "success or failure" +Jun 18 12:02:51.698: INFO: Trying to get logs from node 10.72.74.149 pod downwardapi-volume-fe361889-91c0-11e9-bce2-ae54e022189f container client-container: +STEP: delete the pod +Jun 18 12:02:51.770: INFO: Waiting for pod downwardapi-volume-fe361889-91c0-11e9-bce2-ae54e022189f to disappear +Jun 18 12:02:51.784: INFO: Pod downwardapi-volume-fe361889-91c0-11e9-bce2-ae54e022189f no longer exists +[AfterEach] [sig-storage] Projected downwardAPI /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 19:18:53.825: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-gc-kvxxq" for this suite. -May 13 19:18:59.920: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 19:19:00.034: INFO: namespace: e2e-tests-gc-kvxxq, resource: bindings, ignored listing per whitelist -May 13 19:19:00.202: INFO: namespace e2e-tests-gc-kvxxq deletion completed in 6.369142013s +Jun 18 12:02:51.784: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "e2e-tests-projected-v642t" for this suite. +Jun 18 12:02:57.854: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:02:57.976: INFO: namespace: e2e-tests-projected-v642t, resource: bindings, ignored listing per whitelist +Jun 18 12:02:58.343: INFO: namespace e2e-tests-projected-v642t deletion completed in 6.536820803s -• [SLOW TEST:11.851 seconds] -[sig-api-machinery] Garbage collector -/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:22 - should not be blocked by dependency circle [Conformance] +• [SLOW TEST:9.388 seconds] +[sig-storage] Projected downwardAPI +/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:33 + should provide container's memory request [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSS +S ------------------------------ -[k8s.io] [sig-node] PreStop - should call prestop when killing a pod [Conformance] +[sig-storage] EmptyDir volumes + should support (non-root,0777,tmpfs) [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -[BeforeEach] [k8s.io] [sig-node] PreStop +[BeforeEach] [sig-storage] EmptyDir volumes /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client -May 13 19:19:00.202: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -STEP: Building a namespace api object, basename prestop -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-prestop-ttdzd +Jun 18 12:02:58.345: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 +STEP: Building a namespace api object, basename emptydir +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-emptydir-8cd82 STEP: Waiting for a default service account to be provisioned in namespace -[It] should call prestop when killing a pod [Conformance] +[It] should support (non-root,0777,tmpfs) [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -STEP: Creating server pod server in namespace e2e-tests-prestop-ttdzd -STEP: Waiting for pods to come up. -STEP: Creating tester pod tester in namespace e2e-tests-prestop-ttdzd -STEP: Deleting pre-stop pod -May 13 19:19:09.585: INFO: Saw: { - "Hostname": "server", - "Sent": null, - "Received": { - "prestop": 1 - }, - "Errors": null, - "Log": [ - "default/nettest has 0 endpoints ([]), which is less than 8 as expected. Waiting for all endpoints to come up.", - "default/nettest has 0 endpoints ([]), which is less than 8 as expected. Waiting for all endpoints to come up." - ], - "StillContactingPeers": true -} -STEP: Deleting the server pod -[AfterEach] [k8s.io] [sig-node] PreStop +STEP: Creating a pod to test emptydir 0777 on tmpfs +Jun 18 12:02:58.802: INFO: Waiting up to 5m0s for pod "pod-03aaac9b-91c1-11e9-bce2-ae54e022189f" in namespace "e2e-tests-emptydir-8cd82" to be "success or failure" +Jun 18 12:02:58.816: INFO: Pod "pod-03aaac9b-91c1-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 14.060147ms +Jun 18 12:03:00.833: INFO: Pod "pod-03aaac9b-91c1-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 2.030584723s +Jun 18 12:03:02.848: INFO: Pod "pod-03aaac9b-91c1-11e9-bce2-ae54e022189f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.045749513s +STEP: Saw pod success +Jun 18 12:03:02.848: INFO: Pod "pod-03aaac9b-91c1-11e9-bce2-ae54e022189f" satisfied condition "success or failure" +Jun 18 12:03:02.885: INFO: Trying to get logs from node 10.72.74.149 pod pod-03aaac9b-91c1-11e9-bce2-ae54e022189f container test-container: +STEP: delete the pod +Jun 18 12:03:02.957: INFO: Waiting for pod pod-03aaac9b-91c1-11e9-bce2-ae54e022189f to disappear +Jun 18 12:03:02.972: INFO: Pod pod-03aaac9b-91c1-11e9-bce2-ae54e022189f no longer exists +[AfterEach] [sig-storage] EmptyDir volumes /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 19:19:09.599: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-prestop-ttdzd" for this suite. -May 13 19:19:49.635: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 19:19:49.726: INFO: namespace: e2e-tests-prestop-ttdzd, resource: bindings, ignored listing per whitelist -May 13 19:19:49.912: INFO: namespace e2e-tests-prestop-ttdzd deletion completed in 40.305565784s +Jun 18 12:03:02.972: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "e2e-tests-emptydir-8cd82" for this suite. +Jun 18 12:03:09.045: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:03:09.436: INFO: namespace: e2e-tests-emptydir-8cd82, resource: bindings, ignored listing per whitelist +Jun 18 12:03:09.623: INFO: namespace e2e-tests-emptydir-8cd82 deletion completed in 6.631004789s -• [SLOW TEST:49.710 seconds] -[k8s.io] [sig-node] PreStop -/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 - should call prestop when killing a pod [Conformance] +• [SLOW TEST:11.278 seconds] +[sig-storage] EmptyDir volumes +/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:40 + should support (non-root,0777,tmpfs) [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ -SSSSSSSSS +S ------------------------------ -[k8s.io] Probing container - should be restarted with a /healthz http liveness probe [NodeConformance] [Conformance] +[sig-network] DNS + should provide DNS for services [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -[BeforeEach] [k8s.io] Probing container +[BeforeEach] [sig-network] DNS /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client -May 13 19:19:49.913: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -STEP: Building a namespace api object, basename container-probe -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-container-probe-g862n +Jun 18 12:03:09.623: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 +STEP: Building a namespace api object, basename dns +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-dns-s9vcd STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [k8s.io] Probing container - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/container_probe.go:48 -[It] should be restarted with a /healthz http liveness probe [NodeConformance] [Conformance] +[It] should provide DNS for services [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -STEP: Creating pod liveness-http in namespace e2e-tests-container-probe-g862n -May 13 19:19:52.222: INFO: Started pod liveness-http in namespace e2e-tests-container-probe-g862n -STEP: checking the pod's current state and verifying that restartCount is present -May 13 19:19:52.231: INFO: Initial restart count of pod liveness-http is 0 -May 13 19:20:10.427: INFO: Restart count of pod e2e-tests-container-probe-g862n/liveness-http is now 1 (18.195996991s elapsed) +STEP: Creating a test headless service +STEP: Running these commands on wheezy: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search dns-test-service A)" && test -n "$$check" && echo OK > /results/wheezy_udp@dns-test-service;check="$$(dig +tcp +noall +answer +search dns-test-service A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@dns-test-service;check="$$(dig +notcp +noall +answer +search dns-test-service.e2e-tests-dns-s9vcd A)" && test -n "$$check" && echo OK > /results/wheezy_udp@dns-test-service.e2e-tests-dns-s9vcd;check="$$(dig +tcp +noall +answer +search dns-test-service.e2e-tests-dns-s9vcd A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@dns-test-service.e2e-tests-dns-s9vcd;check="$$(dig +notcp +noall +answer +search dns-test-service.e2e-tests-dns-s9vcd.svc A)" && test -n "$$check" && echo OK > /results/wheezy_udp@dns-test-service.e2e-tests-dns-s9vcd.svc;check="$$(dig +tcp +noall +answer +search dns-test-service.e2e-tests-dns-s9vcd.svc A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@dns-test-service.e2e-tests-dns-s9vcd.svc;check="$$(dig +notcp +noall +answer +search _http._tcp.dns-test-service.e2e-tests-dns-s9vcd.svc SRV)" && test -n "$$check" && echo OK > /results/wheezy_udp@_http._tcp.dns-test-service.e2e-tests-dns-s9vcd.svc;check="$$(dig +tcp +noall +answer +search _http._tcp.dns-test-service.e2e-tests-dns-s9vcd.svc SRV)" && test -n "$$check" && echo OK > /results/wheezy_tcp@_http._tcp.dns-test-service.e2e-tests-dns-s9vcd.svc;check="$$(dig +notcp +noall +answer +search _http._tcp.test-service-2.e2e-tests-dns-s9vcd.svc SRV)" && test -n "$$check" && echo OK > /results/wheezy_udp@_http._tcp.test-service-2.e2e-tests-dns-s9vcd.svc;check="$$(dig +tcp +noall +answer +search _http._tcp.test-service-2.e2e-tests-dns-s9vcd.svc SRV)" && test -n "$$check" && echo OK > /results/wheezy_tcp@_http._tcp.test-service-2.e2e-tests-dns-s9vcd.svc;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".e2e-tests-dns-s9vcd.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@PodARecord;check="$$(dig +notcp +noall +answer +search 233.65.21.172.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/172.21.65.233_udp@PTR;check="$$(dig +tcp +noall +answer +search 233.65.21.172.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/172.21.65.233_tcp@PTR;sleep 1; done + +STEP: Running these commands on jessie: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search dns-test-service A)" && test -n "$$check" && echo OK > /results/jessie_udp@dns-test-service;check="$$(dig +tcp +noall +answer +search dns-test-service A)" && test -n "$$check" && echo OK > /results/jessie_tcp@dns-test-service;check="$$(dig +notcp +noall +answer +search dns-test-service.e2e-tests-dns-s9vcd A)" && test -n "$$check" && echo OK > /results/jessie_udp@dns-test-service.e2e-tests-dns-s9vcd;check="$$(dig +tcp +noall +answer +search dns-test-service.e2e-tests-dns-s9vcd A)" && test -n "$$check" && echo OK > /results/jessie_tcp@dns-test-service.e2e-tests-dns-s9vcd;check="$$(dig +notcp +noall +answer +search dns-test-service.e2e-tests-dns-s9vcd.svc A)" && test -n "$$check" && echo OK > /results/jessie_udp@dns-test-service.e2e-tests-dns-s9vcd.svc;check="$$(dig +tcp +noall +answer +search dns-test-service.e2e-tests-dns-s9vcd.svc A)" && test -n "$$check" && echo OK > /results/jessie_tcp@dns-test-service.e2e-tests-dns-s9vcd.svc;check="$$(dig +notcp +noall +answer +search _http._tcp.dns-test-service.e2e-tests-dns-s9vcd.svc SRV)" && test -n "$$check" && echo OK > /results/jessie_udp@_http._tcp.dns-test-service.e2e-tests-dns-s9vcd.svc;check="$$(dig +tcp +noall +answer +search _http._tcp.dns-test-service.e2e-tests-dns-s9vcd.svc SRV)" && test -n "$$check" && echo OK > /results/jessie_tcp@_http._tcp.dns-test-service.e2e-tests-dns-s9vcd.svc;check="$$(dig +notcp +noall +answer +search _http._tcp.test-service-2.e2e-tests-dns-s9vcd.svc SRV)" && test -n "$$check" && echo OK > /results/jessie_udp@_http._tcp.test-service-2.e2e-tests-dns-s9vcd.svc;check="$$(dig +tcp +noall +answer +search _http._tcp.test-service-2.e2e-tests-dns-s9vcd.svc SRV)" && test -n "$$check" && echo OK > /results/jessie_tcp@_http._tcp.test-service-2.e2e-tests-dns-s9vcd.svc;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".e2e-tests-dns-s9vcd.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_tcp@PodARecord;check="$$(dig +notcp +noall +answer +search 233.65.21.172.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/172.21.65.233_udp@PTR;check="$$(dig +tcp +noall +answer +search 233.65.21.172.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/172.21.65.233_tcp@PTR;sleep 1; done + +STEP: creating a pod to probe DNS +STEP: submitting the pod to kubernetes +STEP: retrieving the pod +STEP: looking for the results for each expected name from probers +Jun 18 12:03:24.492: INFO: Unable to read wheezy_udp@dns-test-service from pod e2e-tests-dns-s9vcd/dns-test-0a7516a2-91c1-11e9-bce2-ae54e022189f: the server could not find the requested resource (get pods dns-test-0a7516a2-91c1-11e9-bce2-ae54e022189f) +Jun 18 12:03:24.910: INFO: Unable to read jessie_udp@dns-test-service from pod e2e-tests-dns-s9vcd/dns-test-0a7516a2-91c1-11e9-bce2-ae54e022189f: the server could not find the requested resource (get pods dns-test-0a7516a2-91c1-11e9-bce2-ae54e022189f) +Jun 18 12:03:24.935: INFO: Unable to read jessie_tcp@dns-test-service from pod e2e-tests-dns-s9vcd/dns-test-0a7516a2-91c1-11e9-bce2-ae54e022189f: the server could not find the requested resource (get pods dns-test-0a7516a2-91c1-11e9-bce2-ae54e022189f) +Jun 18 12:03:24.959: INFO: Unable to read jessie_udp@dns-test-service.e2e-tests-dns-s9vcd from pod e2e-tests-dns-s9vcd/dns-test-0a7516a2-91c1-11e9-bce2-ae54e022189f: the server could not find the requested resource (get pods dns-test-0a7516a2-91c1-11e9-bce2-ae54e022189f) +Jun 18 12:03:24.986: INFO: Unable to read jessie_tcp@dns-test-service.e2e-tests-dns-s9vcd from pod e2e-tests-dns-s9vcd/dns-test-0a7516a2-91c1-11e9-bce2-ae54e022189f: the server could not find the requested resource (get pods dns-test-0a7516a2-91c1-11e9-bce2-ae54e022189f) +Jun 18 12:03:25.014: INFO: Unable to read jessie_udp@dns-test-service.e2e-tests-dns-s9vcd.svc from pod e2e-tests-dns-s9vcd/dns-test-0a7516a2-91c1-11e9-bce2-ae54e022189f: the server could not find the requested resource (get pods dns-test-0a7516a2-91c1-11e9-bce2-ae54e022189f) +Jun 18 12:03:25.037: INFO: Unable to read jessie_tcp@dns-test-service.e2e-tests-dns-s9vcd.svc from pod e2e-tests-dns-s9vcd/dns-test-0a7516a2-91c1-11e9-bce2-ae54e022189f: the server could not find the requested resource (get pods dns-test-0a7516a2-91c1-11e9-bce2-ae54e022189f) +Jun 18 12:03:25.062: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.e2e-tests-dns-s9vcd.svc from pod e2e-tests-dns-s9vcd/dns-test-0a7516a2-91c1-11e9-bce2-ae54e022189f: the server could not find the requested resource (get pods dns-test-0a7516a2-91c1-11e9-bce2-ae54e022189f) +Jun 18 12:03:25.086: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.e2e-tests-dns-s9vcd.svc from pod e2e-tests-dns-s9vcd/dns-test-0a7516a2-91c1-11e9-bce2-ae54e022189f: the server could not find the requested resource (get pods dns-test-0a7516a2-91c1-11e9-bce2-ae54e022189f) +Jun 18 12:03:25.231: INFO: Lookups using e2e-tests-dns-s9vcd/dns-test-0a7516a2-91c1-11e9-bce2-ae54e022189f failed for: [wheezy_udp@dns-test-service jessie_udp@dns-test-service jessie_tcp@dns-test-service jessie_udp@dns-test-service.e2e-tests-dns-s9vcd jessie_tcp@dns-test-service.e2e-tests-dns-s9vcd jessie_udp@dns-test-service.e2e-tests-dns-s9vcd.svc jessie_tcp@dns-test-service.e2e-tests-dns-s9vcd.svc jessie_udp@_http._tcp.dns-test-service.e2e-tests-dns-s9vcd.svc jessie_tcp@_http._tcp.dns-test-service.e2e-tests-dns-s9vcd.svc] + +Jun 18 12:03:30.273: INFO: Unable to read wheezy_udp@dns-test-service from pod e2e-tests-dns-s9vcd/dns-test-0a7516a2-91c1-11e9-bce2-ae54e022189f: the server could not find the requested resource (get pods dns-test-0a7516a2-91c1-11e9-bce2-ae54e022189f) +Jun 18 12:03:30.605: INFO: Unable to read jessie_udp@dns-test-service from pod e2e-tests-dns-s9vcd/dns-test-0a7516a2-91c1-11e9-bce2-ae54e022189f: the server could not find the requested resource (get pods dns-test-0a7516a2-91c1-11e9-bce2-ae54e022189f) +Jun 18 12:03:30.634: INFO: Unable to read jessie_tcp@dns-test-service from pod e2e-tests-dns-s9vcd/dns-test-0a7516a2-91c1-11e9-bce2-ae54e022189f: the server could not find the requested resource (get pods dns-test-0a7516a2-91c1-11e9-bce2-ae54e022189f) +Jun 18 12:03:30.658: INFO: Unable to read jessie_udp@dns-test-service.e2e-tests-dns-s9vcd from pod e2e-tests-dns-s9vcd/dns-test-0a7516a2-91c1-11e9-bce2-ae54e022189f: the server could not find the requested resource (get pods dns-test-0a7516a2-91c1-11e9-bce2-ae54e022189f) +Jun 18 12:03:30.682: INFO: Unable to read jessie_tcp@dns-test-service.e2e-tests-dns-s9vcd from pod e2e-tests-dns-s9vcd/dns-test-0a7516a2-91c1-11e9-bce2-ae54e022189f: the server could not find the requested resource (get pods dns-test-0a7516a2-91c1-11e9-bce2-ae54e022189f) +Jun 18 12:03:30.705: INFO: Unable to read jessie_udp@dns-test-service.e2e-tests-dns-s9vcd.svc from pod e2e-tests-dns-s9vcd/dns-test-0a7516a2-91c1-11e9-bce2-ae54e022189f: the server could not find the requested resource (get pods dns-test-0a7516a2-91c1-11e9-bce2-ae54e022189f) +Jun 18 12:03:30.730: INFO: Unable to read jessie_tcp@dns-test-service.e2e-tests-dns-s9vcd.svc from pod e2e-tests-dns-s9vcd/dns-test-0a7516a2-91c1-11e9-bce2-ae54e022189f: the server could not find the requested resource (get pods dns-test-0a7516a2-91c1-11e9-bce2-ae54e022189f) +Jun 18 12:03:30.753: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.e2e-tests-dns-s9vcd.svc from pod e2e-tests-dns-s9vcd/dns-test-0a7516a2-91c1-11e9-bce2-ae54e022189f: the server could not find the requested resource (get pods dns-test-0a7516a2-91c1-11e9-bce2-ae54e022189f) +Jun 18 12:03:30.777: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.e2e-tests-dns-s9vcd.svc from pod e2e-tests-dns-s9vcd/dns-test-0a7516a2-91c1-11e9-bce2-ae54e022189f: the server could not find the requested resource (get pods dns-test-0a7516a2-91c1-11e9-bce2-ae54e022189f) +Jun 18 12:03:30.919: INFO: Lookups using e2e-tests-dns-s9vcd/dns-test-0a7516a2-91c1-11e9-bce2-ae54e022189f failed for: [wheezy_udp@dns-test-service jessie_udp@dns-test-service jessie_tcp@dns-test-service jessie_udp@dns-test-service.e2e-tests-dns-s9vcd jessie_tcp@dns-test-service.e2e-tests-dns-s9vcd jessie_udp@dns-test-service.e2e-tests-dns-s9vcd.svc jessie_tcp@dns-test-service.e2e-tests-dns-s9vcd.svc jessie_udp@_http._tcp.dns-test-service.e2e-tests-dns-s9vcd.svc jessie_tcp@_http._tcp.dns-test-service.e2e-tests-dns-s9vcd.svc] + +Jun 18 12:03:35.255: INFO: Unable to read wheezy_udp@dns-test-service from pod e2e-tests-dns-s9vcd/dns-test-0a7516a2-91c1-11e9-bce2-ae54e022189f: the server could not find the requested resource (get pods dns-test-0a7516a2-91c1-11e9-bce2-ae54e022189f) +Jun 18 12:03:35.630: INFO: Unable to read jessie_udp@dns-test-service from pod e2e-tests-dns-s9vcd/dns-test-0a7516a2-91c1-11e9-bce2-ae54e022189f: the server could not find the requested resource (get pods dns-test-0a7516a2-91c1-11e9-bce2-ae54e022189f) +Jun 18 12:03:35.653: INFO: Unable to read jessie_tcp@dns-test-service from pod e2e-tests-dns-s9vcd/dns-test-0a7516a2-91c1-11e9-bce2-ae54e022189f: the server could not find the requested resource (get pods dns-test-0a7516a2-91c1-11e9-bce2-ae54e022189f) +Jun 18 12:03:35.676: INFO: Unable to read jessie_udp@dns-test-service.e2e-tests-dns-s9vcd from pod e2e-tests-dns-s9vcd/dns-test-0a7516a2-91c1-11e9-bce2-ae54e022189f: the server could not find the requested resource (get pods dns-test-0a7516a2-91c1-11e9-bce2-ae54e022189f) +Jun 18 12:03:35.705: INFO: Unable to read jessie_tcp@dns-test-service.e2e-tests-dns-s9vcd from pod e2e-tests-dns-s9vcd/dns-test-0a7516a2-91c1-11e9-bce2-ae54e022189f: the server could not find the requested resource (get pods dns-test-0a7516a2-91c1-11e9-bce2-ae54e022189f) +Jun 18 12:03:35.731: INFO: Unable to read jessie_udp@dns-test-service.e2e-tests-dns-s9vcd.svc from pod e2e-tests-dns-s9vcd/dns-test-0a7516a2-91c1-11e9-bce2-ae54e022189f: the server could not find the requested resource (get pods dns-test-0a7516a2-91c1-11e9-bce2-ae54e022189f) +Jun 18 12:03:35.757: INFO: Unable to read jessie_tcp@dns-test-service.e2e-tests-dns-s9vcd.svc from pod e2e-tests-dns-s9vcd/dns-test-0a7516a2-91c1-11e9-bce2-ae54e022189f: the server could not find the requested resource (get pods dns-test-0a7516a2-91c1-11e9-bce2-ae54e022189f) +Jun 18 12:03:35.780: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.e2e-tests-dns-s9vcd.svc from pod e2e-tests-dns-s9vcd/dns-test-0a7516a2-91c1-11e9-bce2-ae54e022189f: the server could not find the requested resource (get pods dns-test-0a7516a2-91c1-11e9-bce2-ae54e022189f) +Jun 18 12:03:35.805: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.e2e-tests-dns-s9vcd.svc from pod e2e-tests-dns-s9vcd/dns-test-0a7516a2-91c1-11e9-bce2-ae54e022189f: the server could not find the requested resource (get pods dns-test-0a7516a2-91c1-11e9-bce2-ae54e022189f) +Jun 18 12:03:35.951: INFO: Lookups using e2e-tests-dns-s9vcd/dns-test-0a7516a2-91c1-11e9-bce2-ae54e022189f failed for: [wheezy_udp@dns-test-service jessie_udp@dns-test-service jessie_tcp@dns-test-service jessie_udp@dns-test-service.e2e-tests-dns-s9vcd jessie_tcp@dns-test-service.e2e-tests-dns-s9vcd jessie_udp@dns-test-service.e2e-tests-dns-s9vcd.svc jessie_tcp@dns-test-service.e2e-tests-dns-s9vcd.svc jessie_udp@_http._tcp.dns-test-service.e2e-tests-dns-s9vcd.svc jessie_tcp@_http._tcp.dns-test-service.e2e-tests-dns-s9vcd.svc] + +Jun 18 12:03:40.255: INFO: Unable to read wheezy_udp@dns-test-service from pod e2e-tests-dns-s9vcd/dns-test-0a7516a2-91c1-11e9-bce2-ae54e022189f: the server could not find the requested resource (get pods dns-test-0a7516a2-91c1-11e9-bce2-ae54e022189f) +Jun 18 12:03:40.658: INFO: Unable to read jessie_udp@dns-test-service from pod e2e-tests-dns-s9vcd/dns-test-0a7516a2-91c1-11e9-bce2-ae54e022189f: the server could not find the requested resource (get pods dns-test-0a7516a2-91c1-11e9-bce2-ae54e022189f) +Jun 18 12:03:40.682: INFO: Unable to read jessie_tcp@dns-test-service from pod e2e-tests-dns-s9vcd/dns-test-0a7516a2-91c1-11e9-bce2-ae54e022189f: the server could not find the requested resource (get pods dns-test-0a7516a2-91c1-11e9-bce2-ae54e022189f) +Jun 18 12:03:40.706: INFO: Unable to read jessie_udp@dns-test-service.e2e-tests-dns-s9vcd from pod e2e-tests-dns-s9vcd/dns-test-0a7516a2-91c1-11e9-bce2-ae54e022189f: the server could not find the requested resource (get pods dns-test-0a7516a2-91c1-11e9-bce2-ae54e022189f) +Jun 18 12:03:40.729: INFO: Unable to read jessie_tcp@dns-test-service.e2e-tests-dns-s9vcd from pod e2e-tests-dns-s9vcd/dns-test-0a7516a2-91c1-11e9-bce2-ae54e022189f: the server could not find the requested resource (get pods dns-test-0a7516a2-91c1-11e9-bce2-ae54e022189f) +Jun 18 12:03:40.784: INFO: Unable to read jessie_udp@dns-test-service.e2e-tests-dns-s9vcd.svc from pod e2e-tests-dns-s9vcd/dns-test-0a7516a2-91c1-11e9-bce2-ae54e022189f: the server could not find the requested resource (get pods dns-test-0a7516a2-91c1-11e9-bce2-ae54e022189f) +Jun 18 12:03:40.807: INFO: Unable to read jessie_tcp@dns-test-service.e2e-tests-dns-s9vcd.svc from pod e2e-tests-dns-s9vcd/dns-test-0a7516a2-91c1-11e9-bce2-ae54e022189f: the server could not find the requested resource (get pods dns-test-0a7516a2-91c1-11e9-bce2-ae54e022189f) +Jun 18 12:03:40.831: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.e2e-tests-dns-s9vcd.svc from pod e2e-tests-dns-s9vcd/dns-test-0a7516a2-91c1-11e9-bce2-ae54e022189f: the server could not find the requested resource (get pods dns-test-0a7516a2-91c1-11e9-bce2-ae54e022189f) +Jun 18 12:03:40.855: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.e2e-tests-dns-s9vcd.svc from pod e2e-tests-dns-s9vcd/dns-test-0a7516a2-91c1-11e9-bce2-ae54e022189f: the server could not find the requested resource (get pods dns-test-0a7516a2-91c1-11e9-bce2-ae54e022189f) +Jun 18 12:03:40.997: INFO: Lookups using e2e-tests-dns-s9vcd/dns-test-0a7516a2-91c1-11e9-bce2-ae54e022189f failed for: [wheezy_udp@dns-test-service jessie_udp@dns-test-service jessie_tcp@dns-test-service jessie_udp@dns-test-service.e2e-tests-dns-s9vcd jessie_tcp@dns-test-service.e2e-tests-dns-s9vcd jessie_udp@dns-test-service.e2e-tests-dns-s9vcd.svc jessie_tcp@dns-test-service.e2e-tests-dns-s9vcd.svc jessie_udp@_http._tcp.dns-test-service.e2e-tests-dns-s9vcd.svc jessie_tcp@_http._tcp.dns-test-service.e2e-tests-dns-s9vcd.svc] + +Jun 18 12:03:46.000: INFO: DNS probes using e2e-tests-dns-s9vcd/dns-test-0a7516a2-91c1-11e9-bce2-ae54e022189f succeeded + STEP: deleting the pod -[AfterEach] [k8s.io] Probing container +STEP: deleting the test service +STEP: deleting the test headless service +[AfterEach] [sig-network] DNS /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 19:20:10.453: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-container-probe-g862n" for this suite. -May 13 19:20:16.492: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 19:20:16.708: INFO: namespace: e2e-tests-container-probe-g862n, resource: bindings, ignored listing per whitelist -May 13 19:20:16.753: INFO: namespace e2e-tests-container-probe-g862n deletion completed in 6.289098111s +Jun 18 12:03:46.159: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "e2e-tests-dns-s9vcd" for this suite. +Jun 18 12:03:52.242: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:03:52.347: INFO: namespace: e2e-tests-dns-s9vcd, resource: bindings, ignored listing per whitelist +Jun 18 12:03:52.722: INFO: namespace e2e-tests-dns-s9vcd deletion completed in 6.549431869s -• [SLOW TEST:26.840 seconds] -[k8s.io] Probing container -/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 - should be restarted with a /healthz http liveness probe [NodeConformance] [Conformance] +• [SLOW TEST:43.100 seconds] +[sig-network] DNS +/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:22 + should provide DNS for services [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSS +SSS ------------------------------ -[k8s.io] Pods - should be updated [NodeConformance] [Conformance] +[sig-api-machinery] Garbage collector + should delete RS created by deployment when not orphaning [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -[BeforeEach] [k8s.io] Pods +[BeforeEach] [sig-api-machinery] Garbage collector /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client -May 13 19:20:16.754: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -STEP: Building a namespace api object, basename pods -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-pods-rkzg7 +Jun 18 12:03:52.723: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 +STEP: Building a namespace api object, basename gc +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-gc-bq2tp STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [k8s.io] Pods - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:132 -[It] should be updated [NodeConformance] [Conformance] +[It] should delete RS created by deployment when not orphaning [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -STEP: creating the pod -STEP: submitting the pod to kubernetes -STEP: verifying the pod is in kubernetes -STEP: updating the pod -May 13 19:20:19.692: INFO: Successfully updated pod "pod-update-2411f5f0-75b4-11e9-8f67-2632f168be36" -STEP: verifying the updated pod is in kubernetes -May 13 19:20:19.711: INFO: Pod update OK -[AfterEach] [k8s.io] Pods +STEP: create the deployment +STEP: Wait for the Deployment to create new ReplicaSet +STEP: delete the deployment +STEP: wait for all rs to be garbage collected +STEP: expected 0 rs, got 1 rs +STEP: expected 0 pods, got 2 pods +STEP: Gathering metrics +W0618 12:03:54.326628 17 metrics_grabber.go:81] Master node is not registered. Grabbing metrics from Scheduler, ControllerManager and ClusterAutoscaler is disabled. +Jun 18 12:03:54.326: INFO: For apiserver_request_count: +For apiserver_request_latencies_summary: +For etcd_helper_cache_entry_count: +For etcd_helper_cache_hit_count: +For etcd_helper_cache_miss_count: +For etcd_request_cache_add_latencies_summary: +For etcd_request_cache_get_latencies_summary: +For etcd_request_latencies_summary: +For garbage_collector_attempt_to_delete_queue_latency: +For garbage_collector_attempt_to_delete_work_duration: +For garbage_collector_attempt_to_orphan_queue_latency: +For garbage_collector_attempt_to_orphan_work_duration: +For garbage_collector_dirty_processing_latency_microseconds: +For garbage_collector_event_processing_latency_microseconds: +For garbage_collector_graph_changes_queue_latency: +For garbage_collector_graph_changes_work_duration: +For garbage_collector_orphan_processing_latency_microseconds: +For namespace_queue_latency: +For namespace_queue_latency_sum: +For namespace_queue_latency_count: +For namespace_retries: +For namespace_work_duration: +For namespace_work_duration_sum: +For namespace_work_duration_count: +For function_duration_seconds: +For errors_total: +For evicted_pods_total: + +[AfterEach] [sig-api-machinery] Garbage collector /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 19:20:19.711: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-pods-rkzg7" for this suite. -May 13 19:20:43.750: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 19:20:44.120: INFO: namespace: e2e-tests-pods-rkzg7, resource: bindings, ignored listing per whitelist -May 13 19:20:44.174: INFO: namespace e2e-tests-pods-rkzg7 deletion completed in 24.452672156s +Jun 18 12:03:54.326: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "e2e-tests-gc-bq2tp" for this suite. +Jun 18 12:04:02.415: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:04:02.530: INFO: namespace: e2e-tests-gc-bq2tp, resource: bindings, ignored listing per whitelist +Jun 18 12:04:02.920: INFO: namespace e2e-tests-gc-bq2tp deletion completed in 8.578276335s -• [SLOW TEST:27.420 seconds] -[k8s.io] Pods -/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 - should be updated [NodeConformance] [Conformance] +• [SLOW TEST:10.197 seconds] +[sig-api-machinery] Garbage collector +/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:22 + should delete RS created by deployment when not orphaning [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ -SSSSSSSSSSSSSSSSS +SSSSSSS ------------------------------ -[sig-storage] Projected downwardAPI - should update annotations on modification [NodeConformance] [Conformance] +[k8s.io] InitContainer [NodeConformance] + should invoke init containers on a RestartNever pod [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -[BeforeEach] [sig-storage] Projected downwardAPI +[BeforeEach] [k8s.io] InitContainer [NodeConformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client -May 13 19:20:44.175: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -STEP: Building a namespace api object, basename projected -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-projected-hv9gn +Jun 18 12:04:02.922: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 +STEP: Building a namespace api object, basename init-container +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-init-container-6jpjk STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [sig-storage] Projected downwardAPI - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:39 -[It] should update annotations on modification [NodeConformance] [Conformance] +[BeforeEach] [k8s.io] InitContainer [NodeConformance] + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/init_container.go:43 +[It] should invoke init containers on a RestartNever pod [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -STEP: Creating the pod -May 13 19:20:47.098: INFO: Successfully updated pod "annotationupdate345b6358-75b4-11e9-8f67-2632f168be36" -[AfterEach] [sig-storage] Projected downwardAPI +STEP: creating the pod +Jun 18 12:04:03.379: INFO: PodSpec: initContainers in spec.initContainers +[AfterEach] [k8s.io] InitContainer [NodeConformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 19:20:51.154: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-projected-hv9gn" for this suite. -May 13 19:21:13.250: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 19:21:13.489: INFO: namespace: e2e-tests-projected-hv9gn, resource: bindings, ignored listing per whitelist -May 13 19:21:13.577: INFO: namespace e2e-tests-projected-hv9gn deletion completed in 22.357577399s +Jun 18 12:04:07.428: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "e2e-tests-init-container-6jpjk" for this suite. +Jun 18 12:04:15.507: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:04:15.572: INFO: namespace: e2e-tests-init-container-6jpjk, resource: bindings, ignored listing per whitelist +Jun 18 12:04:16.007: INFO: namespace e2e-tests-init-container-6jpjk deletion completed in 8.554142094s -• [SLOW TEST:29.403 seconds] -[sig-storage] Projected downwardAPI -/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:33 - should update annotations on modification [NodeConformance] [Conformance] +• [SLOW TEST:13.086 seconds] +[k8s.io] InitContainer [NodeConformance] +/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 + should invoke init containers on a RestartNever pod [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ -SSSSSSSSS +SSSSSSSSSSS ------------------------------ -[sig-storage] Projected configMap - should be consumable from pods in volume with mappings and Item mode set [NodeConformance] [Conformance] +[sig-cli] Kubectl client [k8s.io] Kubectl cluster-info + should check if Kubernetes master services is included in cluster-info [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -[BeforeEach] [sig-storage] Projected configMap +[BeforeEach] [sig-cli] Kubectl client /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client -May 13 19:21:13.578: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -STEP: Building a namespace api object, basename projected -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-projected-wg2mx +Jun 18 12:04:16.008: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 +STEP: Building a namespace api object, basename kubectl +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-kubectl-lp5tr STEP: Waiting for a default service account to be provisioned in namespace -[It] should be consumable from pods in volume with mappings and Item mode set [NodeConformance] [Conformance] +[BeforeEach] [sig-cli] Kubectl client + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:243 +[It] should check if Kubernetes master services is included in cluster-info [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -STEP: Creating configMap with name projected-configmap-test-volume-map-45e38c23-75b4-11e9-8f67-2632f168be36 -STEP: Creating a pod to test consume configMaps -May 13 19:21:13.888: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-45e4ecdd-75b4-11e9-8f67-2632f168be36" in namespace "e2e-tests-projected-wg2mx" to be "success or failure" -May 13 19:21:13.896: INFO: Pod "pod-projected-configmaps-45e4ecdd-75b4-11e9-8f67-2632f168be36": Phase="Pending", Reason="", readiness=false. Elapsed: 7.82457ms -May 13 19:21:15.904: INFO: Pod "pod-projected-configmaps-45e4ecdd-75b4-11e9-8f67-2632f168be36": Phase="Running", Reason="", readiness=true. Elapsed: 2.015814573s -May 13 19:21:17.913: INFO: Pod "pod-projected-configmaps-45e4ecdd-75b4-11e9-8f67-2632f168be36": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.024391058s -STEP: Saw pod success -May 13 19:21:17.913: INFO: Pod "pod-projected-configmaps-45e4ecdd-75b4-11e9-8f67-2632f168be36" satisfied condition "success or failure" -May 13 19:21:17.922: INFO: Trying to get logs from node 10.170.219.177 pod pod-projected-configmaps-45e4ecdd-75b4-11e9-8f67-2632f168be36 container projected-configmap-volume-test: -STEP: delete the pod -May 13 19:21:17.965: INFO: Waiting for pod pod-projected-configmaps-45e4ecdd-75b4-11e9-8f67-2632f168be36 to disappear -May 13 19:21:17.973: INFO: Pod pod-projected-configmaps-45e4ecdd-75b4-11e9-8f67-2632f168be36 no longer exists -[AfterEach] [sig-storage] Projected configMap +STEP: validating cluster-info +Jun 18 12:04:16.457: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 cluster-info' +Jun 18 12:04:16.695: INFO: stderr: "" +Jun 18 12:04:16.695: INFO: stdout: "\x1b[0;32mKubernetes master\x1b[0m is running at \x1b[0;33mhttps://172.21.0.1:443\x1b[0m\n\x1b[0;32mCoreDNS\x1b[0m is running at \x1b[0;33mhttps://172.21.0.1:443/api/v1/namespaces/kube-system/services/kube-dns:dns/proxy\x1b[0m\n\x1b[0;32mkubernetes-dashboard\x1b[0m is running at \x1b[0;33mhttps://172.21.0.1:443/api/v1/namespaces/kube-system/services/https:kubernetes-dashboard:/proxy\x1b[0m\n\x1b[0;32mMetrics-server\x1b[0m is running at \x1b[0;33mhttps://172.21.0.1:443/api/v1/namespaces/kube-system/services/https:metrics-server:/proxy\x1b[0m\n\nTo further debug and diagnose cluster problems, use 'kubectl cluster-info dump'.\n" +[AfterEach] [sig-cli] Kubectl client /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 19:21:17.973: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-projected-wg2mx" for this suite. -May 13 19:21:24.012: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 19:21:24.162: INFO: namespace: e2e-tests-projected-wg2mx, resource: bindings, ignored listing per whitelist -May 13 19:21:24.340: INFO: namespace e2e-tests-projected-wg2mx deletion completed in 6.357017524s +Jun 18 12:04:16.696: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "e2e-tests-kubectl-lp5tr" for this suite. +Jun 18 12:04:22.798: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:04:23.311: INFO: namespace: e2e-tests-kubectl-lp5tr, resource: bindings, ignored listing per whitelist +Jun 18 12:04:23.461: INFO: namespace e2e-tests-kubectl-lp5tr deletion completed in 6.746004561s -• [SLOW TEST:10.762 seconds] -[sig-storage] Projected configMap -/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_configmap.go:34 - should be consumable from pods in volume with mappings and Item mode set [NodeConformance] [Conformance] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------- -SSSSSSSSS +• [SLOW TEST:7.453 seconds] +[sig-cli] Kubectl client +/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:22 + [k8s.io] Kubectl cluster-info + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 + should check if Kubernetes master services is included in cluster-info [Conformance] + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ -[sig-storage] EmptyDir volumes - volume on default medium should have the correct mode [NodeConformance] [Conformance] +[sig-storage] Secrets + should be consumable from pods in volume as non-root with defaultMode and fsGroup set [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -[BeforeEach] [sig-storage] EmptyDir volumes +[BeforeEach] [sig-storage] Secrets /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client -May 13 19:21:24.340: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -STEP: Building a namespace api object, basename emptydir -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-emptydir-vs5zx +Jun 18 12:04:23.462: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 +STEP: Building a namespace api object, basename secrets +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-secrets-dlss8 STEP: Waiting for a default service account to be provisioned in namespace -[It] volume on default medium should have the correct mode [NodeConformance] [Conformance] +[It] should be consumable from pods in volume as non-root with defaultMode and fsGroup set [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -STEP: Creating a pod to test emptydir volume type on node default medium -May 13 19:21:24.633: INFO: Waiting up to 5m0s for pod "pod-4c4cbc91-75b4-11e9-8f67-2632f168be36" in namespace "e2e-tests-emptydir-vs5zx" to be "success or failure" -May 13 19:21:24.642: INFO: Pod "pod-4c4cbc91-75b4-11e9-8f67-2632f168be36": Phase="Pending", Reason="", readiness=false. Elapsed: 8.572466ms -May 13 19:21:26.650: INFO: Pod "pod-4c4cbc91-75b4-11e9-8f67-2632f168be36": Phase="Pending", Reason="", readiness=false. Elapsed: 2.017402596s -May 13 19:21:28.659: INFO: Pod "pod-4c4cbc91-75b4-11e9-8f67-2632f168be36": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.026167806s +STEP: Creating secret with name secret-test-366c56eb-91c1-11e9-bce2-ae54e022189f +STEP: Creating a pod to test consume secrets +Jun 18 12:04:23.966: INFO: Waiting up to 5m0s for pod "pod-secrets-366e480a-91c1-11e9-bce2-ae54e022189f" in namespace "e2e-tests-secrets-dlss8" to be "success or failure" +Jun 18 12:04:23.980: INFO: Pod "pod-secrets-366e480a-91c1-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 13.804563ms +Jun 18 12:04:25.997: INFO: Pod "pod-secrets-366e480a-91c1-11e9-bce2-ae54e022189f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.03017606s STEP: Saw pod success -May 13 19:21:28.659: INFO: Pod "pod-4c4cbc91-75b4-11e9-8f67-2632f168be36" satisfied condition "success or failure" -May 13 19:21:28.666: INFO: Trying to get logs from node 10.170.219.153 pod pod-4c4cbc91-75b4-11e9-8f67-2632f168be36 container test-container: +Jun 18 12:04:25.997: INFO: Pod "pod-secrets-366e480a-91c1-11e9-bce2-ae54e022189f" satisfied condition "success or failure" +Jun 18 12:04:26.012: INFO: Trying to get logs from node 10.72.74.149 pod pod-secrets-366e480a-91c1-11e9-bce2-ae54e022189f container secret-volume-test: STEP: delete the pod -May 13 19:21:28.707: INFO: Waiting for pod pod-4c4cbc91-75b4-11e9-8f67-2632f168be36 to disappear -May 13 19:21:28.714: INFO: Pod pod-4c4cbc91-75b4-11e9-8f67-2632f168be36 no longer exists -[AfterEach] [sig-storage] EmptyDir volumes +Jun 18 12:04:26.999: INFO: Waiting for pod pod-secrets-366e480a-91c1-11e9-bce2-ae54e022189f to disappear +Jun 18 12:04:27.012: INFO: Pod pod-secrets-366e480a-91c1-11e9-bce2-ae54e022189f no longer exists +[AfterEach] [sig-storage] Secrets /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 19:21:28.714: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-emptydir-vs5zx" for this suite. -May 13 19:21:35.021: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 19:21:35.089: INFO: namespace: e2e-tests-emptydir-vs5zx, resource: bindings, ignored listing per whitelist -May 13 19:21:35.314: INFO: namespace e2e-tests-emptydir-vs5zx deletion completed in 6.593212954s +Jun 18 12:04:27.012: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "e2e-tests-secrets-dlss8" for this suite. +Jun 18 12:04:33.084: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:04:33.330: INFO: namespace: e2e-tests-secrets-dlss8, resource: bindings, ignored listing per whitelist +Jun 18 12:04:33.612: INFO: namespace e2e-tests-secrets-dlss8 deletion completed in 6.581174165s -• [SLOW TEST:10.974 seconds] -[sig-storage] EmptyDir volumes -/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:40 - volume on default medium should have the correct mode [NodeConformance] [Conformance] +• [SLOW TEST:10.151 seconds] +[sig-storage] Secrets +/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/secrets_volume.go:34 + should be consumable from pods in volume as non-root with defaultMode and fsGroup set [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ -SSSSSS +SSSSSSSSSSS ------------------------------ -[sig-storage] Subpath Atomic writer volumes - should support subpaths with downward pod [Conformance] +[sig-storage] Projected configMap + should be consumable from pods in volume with mappings [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -[BeforeEach] [sig-storage] Subpath +[BeforeEach] [sig-storage] Projected configMap /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client -May 13 19:21:35.315: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -STEP: Building a namespace api object, basename subpath -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-subpath-kd994 +Jun 18 12:04:33.613: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 +STEP: Building a namespace api object, basename projected +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-projected-z2ppl STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] Atomic writer volumes - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:38 -STEP: Setting up data -[It] should support subpaths with downward pod [Conformance] +[It] should be consumable from pods in volume with mappings [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -STEP: Creating pod pod-subpath-test-downwardapi-tp5n -STEP: Creating a pod to test atomic-volume-subpath -May 13 19:21:35.657: INFO: Waiting up to 5m0s for pod "pod-subpath-test-downwardapi-tp5n" in namespace "e2e-tests-subpath-kd994" to be "success or failure" -May 13 19:21:35.664: INFO: Pod "pod-subpath-test-downwardapi-tp5n": Phase="Pending", Reason="", readiness=false. Elapsed: 7.099253ms -May 13 19:21:37.673: INFO: Pod "pod-subpath-test-downwardapi-tp5n": Phase="Pending", Reason="", readiness=false. Elapsed: 2.015886019s -May 13 19:21:39.682: INFO: Pod "pod-subpath-test-downwardapi-tp5n": Phase="Running", Reason="", readiness=false. Elapsed: 4.024979946s -May 13 19:21:41.690: INFO: Pod "pod-subpath-test-downwardapi-tp5n": Phase="Running", Reason="", readiness=false. Elapsed: 6.032963089s -May 13 19:21:43.713: INFO: Pod "pod-subpath-test-downwardapi-tp5n": Phase="Running", Reason="", readiness=false. Elapsed: 8.055535393s -May 13 19:21:45.721: INFO: Pod "pod-subpath-test-downwardapi-tp5n": Phase="Running", Reason="", readiness=false. Elapsed: 10.063635456s -May 13 19:21:47.729: INFO: Pod "pod-subpath-test-downwardapi-tp5n": Phase="Running", Reason="", readiness=false. Elapsed: 12.07210163s -May 13 19:21:49.738: INFO: Pod "pod-subpath-test-downwardapi-tp5n": Phase="Running", Reason="", readiness=false. Elapsed: 14.080570452s -May 13 19:21:51.747: INFO: Pod "pod-subpath-test-downwardapi-tp5n": Phase="Running", Reason="", readiness=false. Elapsed: 16.090059962s -May 13 19:21:53.768: INFO: Pod "pod-subpath-test-downwardapi-tp5n": Phase="Running", Reason="", readiness=false. Elapsed: 18.111152538s -May 13 19:21:55.777: INFO: Pod "pod-subpath-test-downwardapi-tp5n": Phase="Running", Reason="", readiness=false. Elapsed: 20.119769774s -May 13 19:21:57.785: INFO: Pod "pod-subpath-test-downwardapi-tp5n": Phase="Running", Reason="", readiness=false. Elapsed: 22.128134697s -May 13 19:21:59.794: INFO: Pod "pod-subpath-test-downwardapi-tp5n": Phase="Succeeded", Reason="", readiness=false. Elapsed: 24.136487579s +STEP: Creating configMap with name projected-configmap-test-volume-map-3c78d8f5-91c1-11e9-bce2-ae54e022189f +STEP: Creating a pod to test consume configMaps +Jun 18 12:04:34.116: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-3c7ad519-91c1-11e9-bce2-ae54e022189f" in namespace "e2e-tests-projected-z2ppl" to be "success or failure" +Jun 18 12:04:34.131: INFO: Pod "pod-projected-configmaps-3c7ad519-91c1-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 14.571957ms +Jun 18 12:04:36.146: INFO: Pod "pod-projected-configmaps-3c7ad519-91c1-11e9-bce2-ae54e022189f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.029527828s STEP: Saw pod success -May 13 19:21:59.794: INFO: Pod "pod-subpath-test-downwardapi-tp5n" satisfied condition "success or failure" -May 13 19:21:59.801: INFO: Trying to get logs from node 10.170.219.177 pod pod-subpath-test-downwardapi-tp5n container test-container-subpath-downwardapi-tp5n: +Jun 18 12:04:36.146: INFO: Pod "pod-projected-configmaps-3c7ad519-91c1-11e9-bce2-ae54e022189f" satisfied condition "success or failure" +Jun 18 12:04:36.161: INFO: Trying to get logs from node 10.72.74.149 pod pod-projected-configmaps-3c7ad519-91c1-11e9-bce2-ae54e022189f container projected-configmap-volume-test: STEP: delete the pod -May 13 19:21:59.842: INFO: Waiting for pod pod-subpath-test-downwardapi-tp5n to disappear -May 13 19:21:59.849: INFO: Pod pod-subpath-test-downwardapi-tp5n no longer exists -STEP: Deleting pod pod-subpath-test-downwardapi-tp5n -May 13 19:21:59.849: INFO: Deleting pod "pod-subpath-test-downwardapi-tp5n" in namespace "e2e-tests-subpath-kd994" -[AfterEach] [sig-storage] Subpath +Jun 18 12:04:36.285: INFO: Waiting for pod pod-projected-configmaps-3c7ad519-91c1-11e9-bce2-ae54e022189f to disappear +Jun 18 12:04:36.300: INFO: Pod pod-projected-configmaps-3c7ad519-91c1-11e9-bce2-ae54e022189f no longer exists +[AfterEach] [sig-storage] Projected configMap /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 19:21:59.858: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-subpath-kd994" for this suite. -May 13 19:22:05.894: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 19:22:06.193: INFO: namespace: e2e-tests-subpath-kd994, resource: bindings, ignored listing per whitelist -May 13 19:22:06.229: INFO: namespace e2e-tests-subpath-kd994 deletion completed in 6.362944079s +Jun 18 12:04:36.300: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "e2e-tests-projected-z2ppl" for this suite. +Jun 18 12:04:42.369: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:04:42.520: INFO: namespace: e2e-tests-projected-z2ppl, resource: bindings, ignored listing per whitelist +Jun 18 12:04:42.903: INFO: namespace e2e-tests-projected-z2ppl deletion completed in 6.584518054s -• [SLOW TEST:30.915 seconds] -[sig-storage] Subpath -/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:22 - Atomic writer volumes - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:34 - should support subpaths with downward pod [Conformance] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 +• [SLOW TEST:9.290 seconds] +[sig-storage] Projected configMap +/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_configmap.go:34 + should be consumable from pods in volume with mappings [NodeConformance] [Conformance] + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ -SSS +SSSSSSSS ------------------------------ [k8s.io] Pods should support retrieving logs from the container over websockets [NodeConformance] [Conformance] @@ -4332,75 +3133,160 @@ SSS [BeforeEach] [k8s.io] Pods /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client -May 13 19:22:06.229: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 +Jun 18 12:04:42.904: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 STEP: Building a namespace api object, basename pods -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-pods-r7wt6 +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-pods-6pnq4 STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Pods /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:132 [It] should support retrieving logs from the container over websockets [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -May 13 19:22:06.511: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 +Jun 18 12:04:43.334: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 STEP: creating the pod STEP: submitting the pod to kubernetes [AfterEach] [k8s.io] Pods /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 19:22:08.620: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-pods-r7wt6" for this suite. -May 13 19:22:58.657: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 19:22:58.821: INFO: namespace: e2e-tests-pods-r7wt6, resource: bindings, ignored listing per whitelist -May 13 19:22:58.938: INFO: namespace e2e-tests-pods-r7wt6 deletion completed in 50.309356307s +Jun 18 12:04:47.506: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "e2e-tests-pods-6pnq4" for this suite. +Jun 18 12:05:37.606: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:05:38.198: INFO: namespace: e2e-tests-pods-6pnq4, resource: bindings, ignored listing per whitelist +Jun 18 12:05:38.326: INFO: namespace e2e-tests-pods-6pnq4 deletion completed in 50.799791403s -• [SLOW TEST:52.709 seconds] +• [SLOW TEST:55.422 seconds] [k8s.io] Pods /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 should support retrieving logs from the container over websockets [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ -SSSSSSSSSSSSSS +SSSSSSSSSS ------------------------------ -[k8s.io] InitContainer [NodeConformance] - should not start app containers if init containers fail on a RestartAlways pod [Conformance] +[sig-storage] Downward API volume + should provide container's cpu limit [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -[BeforeEach] [k8s.io] InitContainer [NodeConformance] +[BeforeEach] [sig-storage] Downward API volume /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client -May 13 19:22:58.939: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -STEP: Building a namespace api object, basename init-container -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-init-container-6t4v4 +Jun 18 12:05:38.326: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 +STEP: Building a namespace api object, basename downward-api +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-downward-api-nz5t2 STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [k8s.io] InitContainer [NodeConformance] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/init_container.go:43 -[It] should not start app containers if init containers fail on a RestartAlways pod [Conformance] +[BeforeEach] [sig-storage] Downward API volume + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:39 +[It] should provide container's cpu limit [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -STEP: creating the pod -May 13 19:22:59.224: INFO: PodSpec: initContainers in spec.initContainers -May 13 19:23:48.646: INFO: init container has failed twice: &v1.Pod{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"pod-init-84b0a898-75b4-11e9-8f67-2632f168be36", GenerateName:"", Namespace:"e2e-tests-init-container-6t4v4", SelfLink:"/api/v1/namespaces/e2e-tests-init-container-6t4v4/pods/pod-init-84b0a898-75b4-11e9-8f67-2632f168be36", UID:"84b207db-75b4-11e9-b786-da20024d205c", ResourceVersion:"41995", Generation:0, CreationTimestamp:v1.Time{Time:time.Time{wall:0x0, ext:63693372179, loc:(*time.Location)(0x7b33b80)}}, DeletionTimestamp:(*v1.Time)(nil), DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"name":"foo", "time":"224417083"}, Annotations:map[string]string{"kubernetes.io/psp":"e2e-test-privileged-psp"}, OwnerReferences:[]v1.OwnerReference(nil), Initializers:(*v1.Initializers)(nil), Finalizers:[]string(nil), ClusterName:""}, Spec:v1.PodSpec{Volumes:[]v1.Volume{v1.Volume{Name:"default-token-bphlg", VolumeSource:v1.VolumeSource{HostPath:(*v1.HostPathVolumeSource)(nil), EmptyDir:(*v1.EmptyDirVolumeSource)(nil), GCEPersistentDisk:(*v1.GCEPersistentDiskVolumeSource)(nil), AWSElasticBlockStore:(*v1.AWSElasticBlockStoreVolumeSource)(nil), GitRepo:(*v1.GitRepoVolumeSource)(nil), Secret:(*v1.SecretVolumeSource)(0xc0000c6f80), NFS:(*v1.NFSVolumeSource)(nil), ISCSI:(*v1.ISCSIVolumeSource)(nil), Glusterfs:(*v1.GlusterfsVolumeSource)(nil), PersistentVolumeClaim:(*v1.PersistentVolumeClaimVolumeSource)(nil), RBD:(*v1.RBDVolumeSource)(nil), FlexVolume:(*v1.FlexVolumeSource)(nil), Cinder:(*v1.CinderVolumeSource)(nil), CephFS:(*v1.CephFSVolumeSource)(nil), Flocker:(*v1.FlockerVolumeSource)(nil), DownwardAPI:(*v1.DownwardAPIVolumeSource)(nil), FC:(*v1.FCVolumeSource)(nil), AzureFile:(*v1.AzureFileVolumeSource)(nil), ConfigMap:(*v1.ConfigMapVolumeSource)(nil), VsphereVolume:(*v1.VsphereVirtualDiskVolumeSource)(nil), Quobyte:(*v1.QuobyteVolumeSource)(nil), AzureDisk:(*v1.AzureDiskVolumeSource)(nil), PhotonPersistentDisk:(*v1.PhotonPersistentDiskVolumeSource)(nil), Projected:(*v1.ProjectedVolumeSource)(nil), PortworxVolume:(*v1.PortworxVolumeSource)(nil), ScaleIO:(*v1.ScaleIOVolumeSource)(nil), StorageOS:(*v1.StorageOSVolumeSource)(nil)}}}, InitContainers:[]v1.Container{v1.Container{Name:"init1", Image:"docker.io/library/busybox:1.29", Command:[]string{"/bin/false"}, Args:[]string(nil), WorkingDir:"", Ports:[]v1.ContainerPort(nil), EnvFrom:[]v1.EnvFromSource(nil), Env:[]v1.EnvVar(nil), Resources:v1.ResourceRequirements{Limits:v1.ResourceList(nil), Requests:v1.ResourceList(nil)}, VolumeMounts:[]v1.VolumeMount{v1.VolumeMount{Name:"default-token-bphlg", ReadOnly:true, MountPath:"/var/run/secrets/kubernetes.io/serviceaccount", SubPath:"", MountPropagation:(*v1.MountPropagationMode)(nil)}}, VolumeDevices:[]v1.VolumeDevice(nil), LivenessProbe:(*v1.Probe)(nil), ReadinessProbe:(*v1.Probe)(nil), Lifecycle:(*v1.Lifecycle)(nil), TerminationMessagePath:"/dev/termination-log", TerminationMessagePolicy:"File", ImagePullPolicy:"IfNotPresent", SecurityContext:(*v1.SecurityContext)(nil), Stdin:false, StdinOnce:false, TTY:false}, v1.Container{Name:"init2", Image:"docker.io/library/busybox:1.29", Command:[]string{"/bin/true"}, Args:[]string(nil), WorkingDir:"", Ports:[]v1.ContainerPort(nil), EnvFrom:[]v1.EnvFromSource(nil), Env:[]v1.EnvVar(nil), Resources:v1.ResourceRequirements{Limits:v1.ResourceList(nil), Requests:v1.ResourceList(nil)}, VolumeMounts:[]v1.VolumeMount{v1.VolumeMount{Name:"default-token-bphlg", ReadOnly:true, MountPath:"/var/run/secrets/kubernetes.io/serviceaccount", SubPath:"", MountPropagation:(*v1.MountPropagationMode)(nil)}}, VolumeDevices:[]v1.VolumeDevice(nil), LivenessProbe:(*v1.Probe)(nil), ReadinessProbe:(*v1.Probe)(nil), Lifecycle:(*v1.Lifecycle)(nil), TerminationMessagePath:"/dev/termination-log", TerminationMessagePolicy:"File", ImagePullPolicy:"IfNotPresent", SecurityContext:(*v1.SecurityContext)(nil), Stdin:false, StdinOnce:false, TTY:false}}, Containers:[]v1.Container{v1.Container{Name:"run1", Image:"k8s.gcr.io/pause:3.1", Command:[]string(nil), Args:[]string(nil), WorkingDir:"", Ports:[]v1.ContainerPort(nil), EnvFrom:[]v1.EnvFromSource(nil), Env:[]v1.EnvVar(nil), Resources:v1.ResourceRequirements{Limits:v1.ResourceList{"cpu":resource.Quantity{i:resource.int64Amount{value:100, scale:-3}, d:resource.infDecAmount{Dec:(*inf.Dec)(nil)}, s:"100m", Format:"DecimalSI"}, "memory":resource.Quantity{i:resource.int64Amount{value:52428800, scale:0}, d:resource.infDecAmount{Dec:(*inf.Dec)(nil)}, s:"52428800", Format:"DecimalSI"}}, Requests:v1.ResourceList{"cpu":resource.Quantity{i:resource.int64Amount{value:100, scale:-3}, d:resource.infDecAmount{Dec:(*inf.Dec)(nil)}, s:"100m", Format:"DecimalSI"}, "memory":resource.Quantity{i:resource.int64Amount{value:52428800, scale:0}, d:resource.infDecAmount{Dec:(*inf.Dec)(nil)}, s:"52428800", Format:"DecimalSI"}}}, VolumeMounts:[]v1.VolumeMount{v1.VolumeMount{Name:"default-token-bphlg", ReadOnly:true, MountPath:"/var/run/secrets/kubernetes.io/serviceaccount", SubPath:"", MountPropagation:(*v1.MountPropagationMode)(nil)}}, VolumeDevices:[]v1.VolumeDevice(nil), LivenessProbe:(*v1.Probe)(nil), ReadinessProbe:(*v1.Probe)(nil), Lifecycle:(*v1.Lifecycle)(nil), TerminationMessagePath:"/dev/termination-log", TerminationMessagePolicy:"File", ImagePullPolicy:"IfNotPresent", SecurityContext:(*v1.SecurityContext)(nil), Stdin:false, StdinOnce:false, TTY:false}}, RestartPolicy:"Always", TerminationGracePeriodSeconds:(*int64)(0xc000374018), ActiveDeadlineSeconds:(*int64)(nil), DNSPolicy:"ClusterFirst", NodeSelector:map[string]string(nil), ServiceAccountName:"default", DeprecatedServiceAccount:"default", AutomountServiceAccountToken:(*bool)(nil), NodeName:"10.170.219.177", HostNetwork:false, HostPID:false, HostIPC:false, ShareProcessNamespace:(*bool)(nil), SecurityContext:(*v1.PodSecurityContext)(0xc00205e180), ImagePullSecrets:[]v1.LocalObjectReference(nil), Hostname:"", Subdomain:"", Affinity:(*v1.Affinity)(nil), SchedulerName:"default-scheduler", Tolerations:[]v1.Toleration{v1.Toleration{Key:"node.kubernetes.io/not-ready", Operator:"Exists", Value:"", Effect:"NoExecute", TolerationSeconds:(*int64)(0xc0003740a0)}, v1.Toleration{Key:"node.kubernetes.io/unreachable", Operator:"Exists", Value:"", Effect:"NoExecute", TolerationSeconds:(*int64)(0xc0003740c0)}}, HostAliases:[]v1.HostAlias(nil), PriorityClassName:"", Priority:(*int32)(0xc0003740c8), DNSConfig:(*v1.PodDNSConfig)(nil), ReadinessGates:[]v1.PodReadinessGate(nil), RuntimeClassName:(*string)(nil), EnableServiceLinks:(*bool)(0xc0003740cc)}, Status:v1.PodStatus{Phase:"Pending", Conditions:[]v1.PodCondition{v1.PodCondition{Type:"Initialized", Status:"False", LastProbeTime:v1.Time{Time:time.Time{wall:0x0, ext:0, loc:(*time.Location)(nil)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63693372179, loc:(*time.Location)(0x7b33b80)}}, Reason:"ContainersNotInitialized", Message:"containers with incomplete status: [init1 init2]"}, v1.PodCondition{Type:"Ready", Status:"False", LastProbeTime:v1.Time{Time:time.Time{wall:0x0, ext:0, loc:(*time.Location)(nil)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63693372179, loc:(*time.Location)(0x7b33b80)}}, Reason:"ContainersNotReady", Message:"containers with unready status: [run1]"}, v1.PodCondition{Type:"ContainersReady", Status:"False", LastProbeTime:v1.Time{Time:time.Time{wall:0x0, ext:0, loc:(*time.Location)(nil)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63693372179, loc:(*time.Location)(0x7b33b80)}}, Reason:"ContainersNotReady", Message:"containers with unready status: [run1]"}, v1.PodCondition{Type:"PodScheduled", Status:"True", LastProbeTime:v1.Time{Time:time.Time{wall:0x0, ext:0, loc:(*time.Location)(nil)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63693372179, loc:(*time.Location)(0x7b33b80)}}, Reason:"", Message:""}}, Message:"", Reason:"", NominatedNodeName:"", HostIP:"10.170.219.177", PodIP:"172.30.227.122", StartTime:(*v1.Time)(0xc001ae00e0), InitContainerStatuses:[]v1.ContainerStatus{v1.ContainerStatus{Name:"init1", State:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(nil), Running:(*v1.ContainerStateRunning)(0xc001ae0140), Terminated:(*v1.ContainerStateTerminated)(nil)}, LastTerminationState:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(nil), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(0xc0002f1420)}, Ready:false, RestartCount:3, Image:"docker.io/library/busybox:1.29", ImageID:"docker.io/library/busybox@sha256:8ccbac733d19c0dd4d70b4f0c1e12245b5fa3ad24758a11035ee505c629c0796", ContainerID:"containerd://6bffd483ad22fd14260b315773dece31fc3f0e2da84406f6419471c057e2e361"}, v1.ContainerStatus{Name:"init2", State:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(0xc001ae0160), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(nil)}, LastTerminationState:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(nil), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(nil)}, Ready:false, RestartCount:0, Image:"docker.io/library/busybox:1.29", ImageID:"", ContainerID:""}}, ContainerStatuses:[]v1.ContainerStatus{v1.ContainerStatus{Name:"run1", State:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(0xc001ae0100), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(nil)}, LastTerminationState:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(nil), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(nil)}, Ready:false, RestartCount:0, Image:"k8s.gcr.io/pause:3.1", ImageID:"", ContainerID:""}}, QOSClass:"Guaranteed"}} -[AfterEach] [k8s.io] InitContainer [NodeConformance] +STEP: Creating a pod to test downward API volume plugin +Jun 18 12:05:38.886: INFO: Waiting up to 5m0s for pod "downwardapi-volume-631627be-91c1-11e9-bce2-ae54e022189f" in namespace "e2e-tests-downward-api-nz5t2" to be "success or failure" +Jun 18 12:05:38.900: INFO: Pod "downwardapi-volume-631627be-91c1-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 13.556114ms +Jun 18 12:05:40.916: INFO: Pod "downwardapi-volume-631627be-91c1-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 2.029304337s +Jun 18 12:05:42.947: INFO: Pod "downwardapi-volume-631627be-91c1-11e9-bce2-ae54e022189f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.060313385s +STEP: Saw pod success +Jun 18 12:05:42.947: INFO: Pod "downwardapi-volume-631627be-91c1-11e9-bce2-ae54e022189f" satisfied condition "success or failure" +Jun 18 12:05:42.961: INFO: Trying to get logs from node 10.72.74.149 pod downwardapi-volume-631627be-91c1-11e9-bce2-ae54e022189f container client-container: +STEP: delete the pod +Jun 18 12:05:43.084: INFO: Waiting for pod downwardapi-volume-631627be-91c1-11e9-bce2-ae54e022189f to disappear +Jun 18 12:05:43.099: INFO: Pod downwardapi-volume-631627be-91c1-11e9-bce2-ae54e022189f no longer exists +[AfterEach] [sig-storage] Downward API volume /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 19:23:48.646: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-init-container-6t4v4" for this suite. -May 13 19:24:12.772: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 19:24:12.937: INFO: namespace: e2e-tests-init-container-6t4v4, resource: bindings, ignored listing per whitelist -May 13 19:24:13.068: INFO: namespace e2e-tests-init-container-6t4v4 deletion completed in 24.324635794s +Jun 18 12:05:43.099: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "e2e-tests-downward-api-nz5t2" for this suite. +Jun 18 12:05:49.168: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:05:49.603: INFO: namespace: e2e-tests-downward-api-nz5t2, resource: bindings, ignored listing per whitelist +Jun 18 12:05:49.696: INFO: namespace e2e-tests-downward-api-nz5t2 deletion completed in 6.576959561s -• [SLOW TEST:74.129 seconds] -[k8s.io] InitContainer [NodeConformance] -/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 - should not start app containers if init containers fail on a RestartAlways pod [Conformance] +• [SLOW TEST:11.369 seconds] +[sig-storage] Downward API volume +/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:34 + should provide container's cpu limit [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSSSSS ------------------------------ +[sig-storage] ConfigMap + should be consumable from pods in volume with defaultMode set [NodeConformance] [Conformance] + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 +[BeforeEach] [sig-storage] ConfigMap + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 +STEP: Creating a kubernetes client +Jun 18 12:05:49.696: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 +STEP: Building a namespace api object, basename configmap +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-configmap-kwmfl +STEP: Waiting for a default service account to be provisioned in namespace +[It] should be consumable from pods in volume with defaultMode set [NodeConformance] [Conformance] + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 +STEP: Creating configMap with name configmap-test-volume-69d72420-91c1-11e9-bce2-ae54e022189f +STEP: Creating a pod to test consume configMaps +Jun 18 12:05:50.236: INFO: Waiting up to 5m0s for pod "pod-configmaps-69d94c21-91c1-11e9-bce2-ae54e022189f" in namespace "e2e-tests-configmap-kwmfl" to be "success or failure" +Jun 18 12:05:50.253: INFO: Pod "pod-configmaps-69d94c21-91c1-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 17.549947ms +Jun 18 12:05:52.268: INFO: Pod "pod-configmaps-69d94c21-91c1-11e9-bce2-ae54e022189f": Phase="Running", Reason="", readiness=true. Elapsed: 2.032237237s +Jun 18 12:05:54.309: INFO: Pod "pod-configmaps-69d94c21-91c1-11e9-bce2-ae54e022189f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.073315985s +STEP: Saw pod success +Jun 18 12:05:54.309: INFO: Pod "pod-configmaps-69d94c21-91c1-11e9-bce2-ae54e022189f" satisfied condition "success or failure" +Jun 18 12:05:54.323: INFO: Trying to get logs from node 10.72.74.149 pod pod-configmaps-69d94c21-91c1-11e9-bce2-ae54e022189f container configmap-volume-test: +STEP: delete the pod +Jun 18 12:05:54.392: INFO: Waiting for pod pod-configmaps-69d94c21-91c1-11e9-bce2-ae54e022189f to disappear +Jun 18 12:05:54.484: INFO: Pod pod-configmaps-69d94c21-91c1-11e9-bce2-ae54e022189f no longer exists +[AfterEach] [sig-storage] ConfigMap + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 +Jun 18 12:05:54.484: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "e2e-tests-configmap-kwmfl" for this suite. +Jun 18 12:06:00.551: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:06:00.706: INFO: namespace: e2e-tests-configmap-kwmfl, resource: bindings, ignored listing per whitelist +Jun 18 12:06:01.051: INFO: namespace e2e-tests-configmap-kwmfl deletion completed in 6.547863095s + +• [SLOW TEST:11.356 seconds] +[sig-storage] ConfigMap +/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:33 + should be consumable from pods in volume with defaultMode set [NodeConformance] [Conformance] + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 +------------------------------ +[sig-storage] EmptyDir volumes + should support (non-root,0666,tmpfs) [NodeConformance] [Conformance] + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 +[BeforeEach] [sig-storage] EmptyDir volumes + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 +STEP: Creating a kubernetes client +Jun 18 12:06:01.052: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 +STEP: Building a namespace api object, basename emptydir +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-emptydir-2xrmr +STEP: Waiting for a default service account to be provisioned in namespace +[It] should support (non-root,0666,tmpfs) [NodeConformance] [Conformance] + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 +STEP: Creating a pod to test emptydir 0666 on tmpfs +Jun 18 12:06:01.685: INFO: Waiting up to 5m0s for pod "pod-709e0ec0-91c1-11e9-bce2-ae54e022189f" in namespace "e2e-tests-emptydir-2xrmr" to be "success or failure" +Jun 18 12:06:01.699: INFO: Pod "pod-709e0ec0-91c1-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 14.514272ms +Jun 18 12:06:03.715: INFO: Pod "pod-709e0ec0-91c1-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 2.030440953s +Jun 18 12:06:06.351: INFO: Pod "pod-709e0ec0-91c1-11e9-bce2-ae54e022189f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.666792389s +STEP: Saw pod success +Jun 18 12:06:06.352: INFO: Pod "pod-709e0ec0-91c1-11e9-bce2-ae54e022189f" satisfied condition "success or failure" +Jun 18 12:06:06.369: INFO: Trying to get logs from node 10.72.74.143 pod pod-709e0ec0-91c1-11e9-bce2-ae54e022189f container test-container: +STEP: delete the pod +Jun 18 12:06:06.441: INFO: Waiting for pod pod-709e0ec0-91c1-11e9-bce2-ae54e022189f to disappear +Jun 18 12:06:06.456: INFO: Pod pod-709e0ec0-91c1-11e9-bce2-ae54e022189f no longer exists +[AfterEach] [sig-storage] EmptyDir volumes + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 +Jun 18 12:06:06.456: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "e2e-tests-emptydir-2xrmr" for this suite. +Jun 18 12:06:12.531: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:06:12.691: INFO: namespace: e2e-tests-emptydir-2xrmr, resource: bindings, ignored listing per whitelist +Jun 18 12:06:13.030: INFO: namespace e2e-tests-emptydir-2xrmr deletion completed in 6.554259993s + +• [SLOW TEST:11.977 seconds] +[sig-storage] EmptyDir volumes +/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:40 + should support (non-root,0666,tmpfs) [NodeConformance] [Conformance] + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 +------------------------------ +SSSSSSSSSSSS +------------------------------ [sig-cli] Kubectl client [k8s.io] Update Demo should create and stop a replication controller [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-cli] Kubectl client /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client -May 13 19:24:13.068: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 +Jun 18 12:06:13.031: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 STEP: Building a namespace api object, basename kubectl -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-kubectl-hpsw9 +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-kubectl-mhr68 STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:243 @@ -4409,65 +3295,65 @@ STEP: Waiting for a default service account to be provisioned in namespace [It] should create and stop a replication controller [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: creating a replication controller -May 13 19:24:13.355: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 create -f - --namespace=e2e-tests-kubectl-hpsw9' -May 13 19:24:13.705: INFO: stderr: "" -May 13 19:24:13.705: INFO: stdout: "replicationcontroller/update-demo-nautilus created\n" +Jun 18 12:06:13.568: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 create -f - --namespace=e2e-tests-kubectl-mhr68' +Jun 18 12:06:13.814: INFO: stderr: "" +Jun 18 12:06:13.814: INFO: stdout: "replicationcontroller/update-demo-nautilus created\n" STEP: waiting for all containers in name=update-demo pods to come up. -May 13 19:24:13.705: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=e2e-tests-kubectl-hpsw9' -May 13 19:24:13.821: INFO: stderr: "" -May 13 19:24:13.821: INFO: stdout: "update-demo-nautilus-wpr65 update-demo-nautilus-wzhx9 " -May 13 19:24:13.821: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 get pods update-demo-nautilus-wpr65 -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=e2e-tests-kubectl-hpsw9' -May 13 19:24:13.935: INFO: stderr: "" -May 13 19:24:13.935: INFO: stdout: "" -May 13 19:24:13.936: INFO: update-demo-nautilus-wpr65 is created but not running -May 13 19:24:18.936: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=e2e-tests-kubectl-hpsw9' -May 13 19:24:19.061: INFO: stderr: "" -May 13 19:24:19.061: INFO: stdout: "update-demo-nautilus-wpr65 update-demo-nautilus-wzhx9 " -May 13 19:24:19.061: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 get pods update-demo-nautilus-wpr65 -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=e2e-tests-kubectl-hpsw9' -May 13 19:24:19.156: INFO: stderr: "" -May 13 19:24:19.156: INFO: stdout: "true" -May 13 19:24:19.156: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 get pods update-demo-nautilus-wpr65 -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=e2e-tests-kubectl-hpsw9' -May 13 19:24:19.258: INFO: stderr: "" -May 13 19:24:19.258: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0" -May 13 19:24:19.258: INFO: validating pod update-demo-nautilus-wpr65 -May 13 19:24:19.280: INFO: got data: { +Jun 18 12:06:13.814: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=e2e-tests-kubectl-mhr68' +Jun 18 12:06:13.958: INFO: stderr: "" +Jun 18 12:06:13.958: INFO: stdout: "update-demo-nautilus-429rs update-demo-nautilus-ld8l8 " +Jun 18 12:06:13.958: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 get pods update-demo-nautilus-429rs -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=e2e-tests-kubectl-mhr68' +Jun 18 12:06:14.096: INFO: stderr: "" +Jun 18 12:06:14.096: INFO: stdout: "" +Jun 18 12:06:14.096: INFO: update-demo-nautilus-429rs is created but not running +Jun 18 12:06:19.096: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=e2e-tests-kubectl-mhr68' +Jun 18 12:06:19.258: INFO: stderr: "" +Jun 18 12:06:19.258: INFO: stdout: "update-demo-nautilus-429rs update-demo-nautilus-ld8l8 " +Jun 18 12:06:19.258: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 get pods update-demo-nautilus-429rs -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=e2e-tests-kubectl-mhr68' +Jun 18 12:06:19.388: INFO: stderr: "" +Jun 18 12:06:19.388: INFO: stdout: "true" +Jun 18 12:06:19.388: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 get pods update-demo-nautilus-429rs -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=e2e-tests-kubectl-mhr68' +Jun 18 12:06:19.535: INFO: stderr: "" +Jun 18 12:06:19.535: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0" +Jun 18 12:06:19.535: INFO: validating pod update-demo-nautilus-429rs +Jun 18 12:06:19.568: INFO: got data: { "image": "nautilus.jpg" } -May 13 19:24:19.280: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . -May 13 19:24:19.280: INFO: update-demo-nautilus-wpr65 is verified up and running -May 13 19:24:19.280: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 get pods update-demo-nautilus-wzhx9 -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=e2e-tests-kubectl-hpsw9' -May 13 19:24:19.384: INFO: stderr: "" -May 13 19:24:19.384: INFO: stdout: "true" -May 13 19:24:19.384: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 get pods update-demo-nautilus-wzhx9 -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=e2e-tests-kubectl-hpsw9' -May 13 19:24:19.476: INFO: stderr: "" -May 13 19:24:19.476: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0" -May 13 19:24:19.476: INFO: validating pod update-demo-nautilus-wzhx9 -May 13 19:24:19.490: INFO: got data: { +Jun 18 12:06:19.568: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . +Jun 18 12:06:19.568: INFO: update-demo-nautilus-429rs is verified up and running +Jun 18 12:06:19.568: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 get pods update-demo-nautilus-ld8l8 -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=e2e-tests-kubectl-mhr68' +Jun 18 12:06:19.698: INFO: stderr: "" +Jun 18 12:06:19.698: INFO: stdout: "true" +Jun 18 12:06:19.698: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 get pods update-demo-nautilus-ld8l8 -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=e2e-tests-kubectl-mhr68' +Jun 18 12:06:20.638: INFO: stderr: "" +Jun 18 12:06:20.638: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0" +Jun 18 12:06:20.638: INFO: validating pod update-demo-nautilus-ld8l8 +Jun 18 12:06:20.671: INFO: got data: { "image": "nautilus.jpg" } -May 13 19:24:19.490: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . -May 13 19:24:19.490: INFO: update-demo-nautilus-wzhx9 is verified up and running +Jun 18 12:06:20.671: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . +Jun 18 12:06:20.671: INFO: update-demo-nautilus-ld8l8 is verified up and running STEP: using delete to clean up resources -May 13 19:24:19.491: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 delete --grace-period=0 --force -f - --namespace=e2e-tests-kubectl-hpsw9' -May 13 19:24:19.622: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" -May 13 19:24:19.622: INFO: stdout: "replicationcontroller \"update-demo-nautilus\" force deleted\n" -May 13 19:24:19.622: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 get rc,svc -l name=update-demo --no-headers --namespace=e2e-tests-kubectl-hpsw9' -May 13 19:24:19.744: INFO: stderr: "No resources found.\n" -May 13 19:24:19.744: INFO: stdout: "" -May 13 19:24:19.744: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 get pods -l name=update-demo --namespace=e2e-tests-kubectl-hpsw9 -o go-template={{ range .items }}{{ if not .metadata.deletionTimestamp }}{{ .metadata.name }}{{ "\n" }}{{ end }}{{ end }}' -May 13 19:24:19.850: INFO: stderr: "" -May 13 19:24:19.850: INFO: stdout: "" +Jun 18 12:06:20.671: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 delete --grace-period=0 --force -f - --namespace=e2e-tests-kubectl-mhr68' +Jun 18 12:06:20.812: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" +Jun 18 12:06:20.812: INFO: stdout: "replicationcontroller \"update-demo-nautilus\" force deleted\n" +Jun 18 12:06:20.812: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 get rc,svc -l name=update-demo --no-headers --namespace=e2e-tests-kubectl-mhr68' +Jun 18 12:06:20.988: INFO: stderr: "No resources found.\n" +Jun 18 12:06:20.988: INFO: stdout: "" +Jun 18 12:06:20.988: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 get pods -l name=update-demo --namespace=e2e-tests-kubectl-mhr68 -o go-template={{ range .items }}{{ if not .metadata.deletionTimestamp }}{{ .metadata.name }}{{ "\n" }}{{ end }}{{ end }}' +Jun 18 12:06:21.124: INFO: stderr: "" +Jun 18 12:06:21.124: INFO: stdout: "" [AfterEach] [sig-cli] Kubectl client /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 19:24:19.850: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-kubectl-hpsw9" for this suite. -May 13 19:24:43.894: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 19:24:44.127: INFO: namespace: e2e-tests-kubectl-hpsw9, resource: bindings, ignored listing per whitelist -May 13 19:24:44.168: INFO: namespace e2e-tests-kubectl-hpsw9 deletion completed in 24.304907253s +Jun 18 12:06:21.124: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "e2e-tests-kubectl-mhr68" for this suite. +Jun 18 12:06:45.195: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:06:46.774: INFO: namespace: e2e-tests-kubectl-mhr68, resource: bindings, ignored listing per whitelist +Jun 18 12:06:46.801: INFO: namespace e2e-tests-kubectl-mhr68 deletion completed in 25.655984162s -• [SLOW TEST:31.100 seconds] +• [SLOW TEST:33.770 seconds] [sig-cli] Kubectl client /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:22 [k8s.io] Update Demo @@ -4475,834 +3361,1712 @@ May 13 19:24:44.168: INFO: namespace e2e-tests-kubectl-hpsw9 deletion completed should create and stop a replication controller [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ -SS +SSSSSSSSSSS ------------------------------ -[sig-storage] EmptyDir volumes - should support (non-root,0666,default) [NodeConformance] [Conformance] +[k8s.io] Container Lifecycle Hook when create a pod with lifecycle hook + should execute prestop exec hook properly [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -[BeforeEach] [sig-storage] EmptyDir volumes +[BeforeEach] [k8s.io] Container Lifecycle Hook /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client -May 13 19:24:44.168: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -STEP: Building a namespace api object, basename emptydir -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-emptydir-wzvrx +Jun 18 12:06:46.801: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 +STEP: Building a namespace api object, basename container-lifecycle-hook +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-container-lifecycle-hook-bx8lz STEP: Waiting for a default service account to be provisioned in namespace -[It] should support (non-root,0666,default) [NodeConformance] [Conformance] +[BeforeEach] when create a pod with lifecycle hook + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:61 +STEP: create the container to handle the HTTPGet hook request. +[It] should execute prestop exec hook properly [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -STEP: Creating a pod to test emptydir 0666 on node default medium -May 13 19:24:44.609: INFO: Waiting up to 5m0s for pod "pod-c36e5381-75b4-11e9-8f67-2632f168be36" in namespace "e2e-tests-emptydir-wzvrx" to be "success or failure" -May 13 19:24:44.620: INFO: Pod "pod-c36e5381-75b4-11e9-8f67-2632f168be36": Phase="Pending", Reason="", readiness=false. Elapsed: 11.309201ms -May 13 19:24:46.636: INFO: Pod "pod-c36e5381-75b4-11e9-8f67-2632f168be36": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.027308097s +STEP: create the pod with lifecycle hook +STEP: delete the pod with lifecycle hook +Jun 18 12:06:51.500: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear +Jun 18 12:06:51.515: INFO: Pod pod-with-prestop-exec-hook still exists +Jun 18 12:06:53.515: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear +Jun 18 12:06:53.549: INFO: Pod pod-with-prestop-exec-hook still exists +Jun 18 12:06:55.515: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear +Jun 18 12:06:55.532: INFO: Pod pod-with-prestop-exec-hook still exists +Jun 18 12:06:57.515: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear +Jun 18 12:06:57.530: INFO: Pod pod-with-prestop-exec-hook still exists +Jun 18 12:06:59.515: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear +Jun 18 12:06:59.530: INFO: Pod pod-with-prestop-exec-hook still exists +Jun 18 12:07:01.515: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear +Jun 18 12:07:01.531: INFO: Pod pod-with-prestop-exec-hook still exists +Jun 18 12:07:03.515: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear +Jun 18 12:07:03.530: INFO: Pod pod-with-prestop-exec-hook still exists +Jun 18 12:07:05.515: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear +Jun 18 12:07:05.553: INFO: Pod pod-with-prestop-exec-hook still exists +Jun 18 12:07:07.515: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear +Jun 18 12:07:07.530: INFO: Pod pod-with-prestop-exec-hook still exists +Jun 18 12:07:09.515: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear +Jun 18 12:07:09.532: INFO: Pod pod-with-prestop-exec-hook still exists +Jun 18 12:07:11.515: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear +Jun 18 12:07:11.531: INFO: Pod pod-with-prestop-exec-hook still exists +Jun 18 12:07:13.515: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear +Jun 18 12:07:13.533: INFO: Pod pod-with-prestop-exec-hook still exists +Jun 18 12:07:15.515: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear +Jun 18 12:07:15.531: INFO: Pod pod-with-prestop-exec-hook still exists +Jun 18 12:07:17.515: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear +Jun 18 12:07:17.552: INFO: Pod pod-with-prestop-exec-hook still exists +Jun 18 12:07:19.515: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear +Jun 18 12:07:19.530: INFO: Pod pod-with-prestop-exec-hook no longer exists +STEP: check prestop hook +[AfterEach] [k8s.io] Container Lifecycle Hook + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 +Jun 18 12:07:19.584: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "e2e-tests-container-lifecycle-hook-bx8lz" for this suite. +Jun 18 12:07:43.660: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:07:43.706: INFO: namespace: e2e-tests-container-lifecycle-hook-bx8lz, resource: bindings, ignored listing per whitelist +Jun 18 12:07:44.158: INFO: namespace e2e-tests-container-lifecycle-hook-bx8lz deletion completed in 24.55459548s + +• [SLOW TEST:57.357 seconds] +[k8s.io] Container Lifecycle Hook +/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 + when create a pod with lifecycle hook + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:40 + should execute prestop exec hook properly [NodeConformance] [Conformance] + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 +------------------------------ +SSSSSSSSSSSSSSSSSS +------------------------------ +[sig-apps] Daemon set [Serial] + should retry creating failed daemon pods [Conformance] + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 +[BeforeEach] [sig-apps] Daemon set [Serial] + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 +STEP: Creating a kubernetes client +Jun 18 12:07:44.159: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 +STEP: Building a namespace api object, basename daemonsets +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-daemonsets-sgtlc +STEP: Waiting for a default service account to be provisioned in namespace +[BeforeEach] [sig-apps] Daemon set [Serial] + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:102 +[It] should retry creating failed daemon pods [Conformance] + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 +STEP: Creating a simple DaemonSet "daemon-set" +STEP: Check that daemon pods launch on every node of the cluster. +Jun 18 12:07:44.823: INFO: Number of nodes with available pods: 0 +Jun 18 12:07:44.823: INFO: Node 10.72.74.143 is running more than one daemon pod +Jun 18 12:07:45.860: INFO: Number of nodes with available pods: 0 +Jun 18 12:07:45.860: INFO: Node 10.72.74.143 is running more than one daemon pod +Jun 18 12:07:46.885: INFO: Number of nodes with available pods: 1 +Jun 18 12:07:46.885: INFO: Node 10.72.74.143 is running more than one daemon pod +Jun 18 12:07:47.858: INFO: Number of nodes with available pods: 3 +Jun 18 12:07:47.859: INFO: Number of running nodes: 3, number of available pods: 3 +STEP: Set a daemon pod's phase to 'Failed', check that the daemon pod is revived. +Jun 18 12:07:47.930: INFO: Number of nodes with available pods: 2 +Jun 18 12:07:47.930: INFO: Node 10.72.74.143 is running more than one daemon pod +Jun 18 12:07:48.979: INFO: Number of nodes with available pods: 2 +Jun 18 12:07:48.979: INFO: Node 10.72.74.143 is running more than one daemon pod +Jun 18 12:07:50.373: INFO: Number of nodes with available pods: 3 +Jun 18 12:07:50.373: INFO: Number of running nodes: 3, number of available pods: 3 +STEP: Wait for the failed daemon pod to be completely deleted. +[AfterEach] [sig-apps] Daemon set [Serial] + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:68 +STEP: Deleting DaemonSet "daemon-set" +STEP: deleting DaemonSet.extensions daemon-set in namespace e2e-tests-daemonsets-sgtlc, will wait for the garbage collector to delete the pods +Jun 18 12:07:50.492: INFO: Deleting DaemonSet.extensions daemon-set took: 29.404003ms +Jun 18 12:07:50.593: INFO: Terminating DaemonSet.extensions daemon-set pods took: 100.354605ms +Jun 18 12:08:31.185: INFO: Number of nodes with available pods: 0 +Jun 18 12:08:31.185: INFO: Number of running nodes: 0, number of available pods: 0 +Jun 18 12:08:31.198: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"selfLink":"/apis/apps/v1/namespaces/e2e-tests-daemonsets-sgtlc/daemonsets","resourceVersion":"95589"},"items":null} + +Jun 18 12:08:31.212: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"selfLink":"/api/v1/namespaces/e2e-tests-daemonsets-sgtlc/pods","resourceVersion":"95589"},"items":null} + +[AfterEach] [sig-apps] Daemon set [Serial] + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 +Jun 18 12:08:31.273: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "e2e-tests-daemonsets-sgtlc" for this suite. +Jun 18 12:08:39.337: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:08:39.416: INFO: namespace: e2e-tests-daemonsets-sgtlc, resource: bindings, ignored listing per whitelist +Jun 18 12:08:39.834: INFO: namespace e2e-tests-daemonsets-sgtlc deletion completed in 8.545417349s + +• [SLOW TEST:55.675 seconds] +[sig-apps] Daemon set [Serial] +/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22 + should retry creating failed daemon pods [Conformance] + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 +------------------------------ +[k8s.io] Variable Expansion + should allow substituting values in a container's command [NodeConformance] [Conformance] + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 +[BeforeEach] [k8s.io] Variable Expansion + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 +STEP: Creating a kubernetes client +Jun 18 12:08:39.834: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 +STEP: Building a namespace api object, basename var-expansion +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-var-expansion-dpjsn +STEP: Waiting for a default service account to be provisioned in namespace +[It] should allow substituting values in a container's command [NodeConformance] [Conformance] + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 +STEP: Creating a pod to test substitution in container's command +Jun 18 12:08:40.355: INFO: Waiting up to 5m0s for pod "var-expansion-cf4036c1-91c1-11e9-bce2-ae54e022189f" in namespace "e2e-tests-var-expansion-dpjsn" to be "success or failure" +Jun 18 12:08:40.370: INFO: Pod "var-expansion-cf4036c1-91c1-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 14.561912ms +Jun 18 12:08:42.403: INFO: Pod "var-expansion-cf4036c1-91c1-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 2.047077524s +Jun 18 12:08:44.418: INFO: Pod "var-expansion-cf4036c1-91c1-11e9-bce2-ae54e022189f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.062284293s STEP: Saw pod success -May 13 19:24:46.636: INFO: Pod "pod-c36e5381-75b4-11e9-8f67-2632f168be36" satisfied condition "success or failure" -May 13 19:24:46.644: INFO: Trying to get logs from node 10.170.219.153 pod pod-c36e5381-75b4-11e9-8f67-2632f168be36 container test-container: +Jun 18 12:08:44.418: INFO: Pod "var-expansion-cf4036c1-91c1-11e9-bce2-ae54e022189f" satisfied condition "success or failure" +Jun 18 12:08:44.432: INFO: Trying to get logs from node 10.72.74.149 pod var-expansion-cf4036c1-91c1-11e9-bce2-ae54e022189f container dapi-container: STEP: delete the pod -May 13 19:24:46.689: INFO: Waiting for pod pod-c36e5381-75b4-11e9-8f67-2632f168be36 to disappear -May 13 19:24:46.697: INFO: Pod pod-c36e5381-75b4-11e9-8f67-2632f168be36 no longer exists -[AfterEach] [sig-storage] EmptyDir volumes +Jun 18 12:08:44.502: INFO: Waiting for pod var-expansion-cf4036c1-91c1-11e9-bce2-ae54e022189f to disappear +Jun 18 12:08:44.516: INFO: Pod var-expansion-cf4036c1-91c1-11e9-bce2-ae54e022189f no longer exists +[AfterEach] [k8s.io] Variable Expansion /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 19:24:46.697: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-emptydir-wzvrx" for this suite. -May 13 19:24:52.756: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 19:24:52.920: INFO: namespace: e2e-tests-emptydir-wzvrx, resource: bindings, ignored listing per whitelist -May 13 19:24:53.099: INFO: namespace e2e-tests-emptydir-wzvrx deletion completed in 6.371071965s +Jun 18 12:08:44.516: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "e2e-tests-var-expansion-dpjsn" for this suite. +Jun 18 12:08:50.583: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:08:50.972: INFO: namespace: e2e-tests-var-expansion-dpjsn, resource: bindings, ignored listing per whitelist +Jun 18 12:08:51.335: INFO: namespace e2e-tests-var-expansion-dpjsn deletion completed in 6.800770061s -• [SLOW TEST:8.930 seconds] -[sig-storage] EmptyDir volumes -/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:40 - should support (non-root,0666,default) [NodeConformance] [Conformance] +• [SLOW TEST:11.501 seconds] +[k8s.io] Variable Expansion +/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 + should allow substituting values in a container's command [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +SSSSSSSSSS ------------------------------ -[sig-storage] Downward API volume - should provide container's memory limit [NodeConformance] [Conformance] +[sig-storage] Projected configMap + should be consumable from pods in volume [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -[BeforeEach] [sig-storage] Downward API volume +[BeforeEach] [sig-storage] Projected configMap /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client -May 13 19:24:53.101: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -STEP: Building a namespace api object, basename downward-api -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-downward-api-pcn7f +Jun 18 12:08:51.335: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 +STEP: Building a namespace api object, basename projected +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-projected-8r9nx STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [sig-storage] Downward API volume - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:39 -[It] should provide container's memory limit [NodeConformance] [Conformance] +[It] should be consumable from pods in volume [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -STEP: Creating a pod to test downward API volume plugin -May 13 19:24:53.394: INFO: Waiting up to 5m0s for pod "downwardapi-volume-c8bad340-75b4-11e9-8f67-2632f168be36" in namespace "e2e-tests-downward-api-pcn7f" to be "success or failure" -May 13 19:24:53.406: INFO: Pod "downwardapi-volume-c8bad340-75b4-11e9-8f67-2632f168be36": Phase="Pending", Reason="", readiness=false. Elapsed: 12.538494ms -May 13 19:24:55.426: INFO: Pod "downwardapi-volume-c8bad340-75b4-11e9-8f67-2632f168be36": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.032878373s +STEP: Creating configMap with name projected-configmap-test-volume-d6150c58-91c1-11e9-bce2-ae54e022189f +STEP: Creating a pod to test consume configMaps +Jun 18 12:08:51.831: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-d6170c2f-91c1-11e9-bce2-ae54e022189f" in namespace "e2e-tests-projected-8r9nx" to be "success or failure" +Jun 18 12:08:51.846: INFO: Pod "pod-projected-configmaps-d6170c2f-91c1-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 14.785808ms +Jun 18 12:08:53.880: INFO: Pod "pod-projected-configmaps-d6170c2f-91c1-11e9-bce2-ae54e022189f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.04882872s STEP: Saw pod success -May 13 19:24:55.426: INFO: Pod "downwardapi-volume-c8bad340-75b4-11e9-8f67-2632f168be36" satisfied condition "success or failure" -May 13 19:24:55.528: INFO: Trying to get logs from node 10.170.219.153 pod downwardapi-volume-c8bad340-75b4-11e9-8f67-2632f168be36 container client-container: +Jun 18 12:08:53.880: INFO: Pod "pod-projected-configmaps-d6170c2f-91c1-11e9-bce2-ae54e022189f" satisfied condition "success or failure" +Jun 18 12:08:53.894: INFO: Trying to get logs from node 10.72.74.143 pod pod-projected-configmaps-d6170c2f-91c1-11e9-bce2-ae54e022189f container projected-configmap-volume-test: STEP: delete the pod -May 13 19:24:55.566: INFO: Waiting for pod downwardapi-volume-c8bad340-75b4-11e9-8f67-2632f168be36 to disappear -May 13 19:24:55.573: INFO: Pod downwardapi-volume-c8bad340-75b4-11e9-8f67-2632f168be36 no longer exists -[AfterEach] [sig-storage] Downward API volume +Jun 18 12:08:53.969: INFO: Waiting for pod pod-projected-configmaps-d6170c2f-91c1-11e9-bce2-ae54e022189f to disappear +Jun 18 12:08:53.983: INFO: Pod pod-projected-configmaps-d6170c2f-91c1-11e9-bce2-ae54e022189f no longer exists +[AfterEach] [sig-storage] Projected configMap /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 19:24:55.573: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-downward-api-pcn7f" for this suite. -May 13 19:25:01.611: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 19:25:01.821: INFO: namespace: e2e-tests-downward-api-pcn7f, resource: bindings, ignored listing per whitelist -May 13 19:25:01.879: INFO: namespace e2e-tests-downward-api-pcn7f deletion completed in 6.297084832s +Jun 18 12:08:53.983: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "e2e-tests-projected-8r9nx" for this suite. +Jun 18 12:09:00.052: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:09:00.387: INFO: namespace: e2e-tests-projected-8r9nx, resource: bindings, ignored listing per whitelist +Jun 18 12:09:00.572: INFO: namespace e2e-tests-projected-8r9nx deletion completed in 6.567598408s -• [SLOW TEST:8.779 seconds] -[sig-storage] Downward API volume -/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:34 - should provide container's memory limit [NodeConformance] [Conformance] +• [SLOW TEST:9.236 seconds] +[sig-storage] Projected configMap +/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_configmap.go:34 + should be consumable from pods in volume [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ -SSSSSSSS +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-api-machinery] Garbage collector - should delete RS created by deployment when not orphaning [Conformance] +[sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] + should perform rolling updates and roll backs of template modifications [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -[BeforeEach] [sig-api-machinery] Garbage collector +[BeforeEach] [sig-apps] StatefulSet /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client -May 13 19:25:01.880: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -STEP: Building a namespace api object, basename gc -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-gc-zmx24 +Jun 18 12:09:00.573: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 +STEP: Building a namespace api object, basename statefulset +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-statefulset-b2kqz STEP: Waiting for a default service account to be provisioned in namespace -[It] should delete RS created by deployment when not orphaning [Conformance] +[BeforeEach] [sig-apps] StatefulSet + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:59 +[BeforeEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:74 +STEP: Creating service test in namespace e2e-tests-statefulset-b2kqz +[It] should perform rolling updates and roll backs of template modifications [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -STEP: create the deployment -STEP: Wait for the Deployment to create new ReplicaSet -STEP: delete the deployment -STEP: wait for all rs to be garbage collected -STEP: expected 0 rs, got 1 rs -STEP: expected 0 pods, got 2 pods -STEP: Gathering metrics -W0513 19:25:03.264121 16 metrics_grabber.go:81] Master node is not registered. Grabbing metrics from Scheduler, ControllerManager and ClusterAutoscaler is disabled. -May 13 19:25:03.264: INFO: For apiserver_request_count: -For apiserver_request_latencies_summary: -For etcd_helper_cache_entry_count: -For etcd_helper_cache_hit_count: -For etcd_helper_cache_miss_count: -For etcd_request_cache_add_latencies_summary: -For etcd_request_cache_get_latencies_summary: -For etcd_request_latencies_summary: -For garbage_collector_attempt_to_delete_queue_latency: -For garbage_collector_attempt_to_delete_work_duration: -For garbage_collector_attempt_to_orphan_queue_latency: -For garbage_collector_attempt_to_orphan_work_duration: -For garbage_collector_dirty_processing_latency_microseconds: -For garbage_collector_event_processing_latency_microseconds: -For garbage_collector_graph_changes_queue_latency: -For garbage_collector_graph_changes_work_duration: -For garbage_collector_orphan_processing_latency_microseconds: -For namespace_queue_latency: -For namespace_queue_latency_sum: -For namespace_queue_latency_count: -For namespace_retries: -For namespace_work_duration: -For namespace_work_duration_sum: -For namespace_work_duration_count: -For function_duration_seconds: -For errors_total: -For evicted_pods_total: +STEP: Creating a new StatefulSet +Jun 18 12:09:01.136: INFO: Found 0 stateful pods, waiting for 3 +Jun 18 12:09:11.216: INFO: Waiting for pod ss2-0 to enter Running - Ready=true, currently Running - Ready=true +Jun 18 12:09:11.216: INFO: Waiting for pod ss2-1 to enter Running - Ready=true, currently Running - Ready=true +Jun 18 12:09:11.216: INFO: Waiting for pod ss2-2 to enter Running - Ready=true, currently Running - Ready=true +Jun 18 12:09:11.264: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 exec --namespace=e2e-tests-statefulset-b2kqz ss2-1 -- /bin/sh -c mv -v /usr/share/nginx/html/index.html /tmp/ || true' +Jun 18 12:09:11.718: INFO: stderr: "" +Jun 18 12:09:11.718: INFO: stdout: "'/usr/share/nginx/html/index.html' -> '/tmp/index.html'\n" +Jun 18 12:09:11.718: INFO: stdout of mv -v /usr/share/nginx/html/index.html /tmp/ || true on ss2-1: '/usr/share/nginx/html/index.html' -> '/tmp/index.html' -[AfterEach] [sig-api-machinery] Garbage collector +STEP: Updating StatefulSet template: update image from docker.io/library/nginx:1.14-alpine to docker.io/library/nginx:1.15-alpine +Jun 18 12:09:21.900: INFO: Updating stateful set ss2 +STEP: Creating a new revision +STEP: Updating Pods in reverse ordinal order +Jun 18 12:09:31.999: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 exec --namespace=e2e-tests-statefulset-b2kqz ss2-1 -- /bin/sh -c mv -v /tmp/index.html /usr/share/nginx/html/ || true' +Jun 18 12:09:32.379: INFO: stderr: "" +Jun 18 12:09:32.379: INFO: stdout: "'/tmp/index.html' -> '/usr/share/nginx/html/index.html'\n" +Jun 18 12:09:32.379: INFO: stdout of mv -v /tmp/index.html /usr/share/nginx/html/ || true on ss2-1: '/tmp/index.html' -> '/usr/share/nginx/html/index.html' + +Jun 18 12:09:43.759: INFO: Waiting for StatefulSet e2e-tests-statefulset-b2kqz/ss2 to complete update +Jun 18 12:09:43.759: INFO: Waiting for Pod e2e-tests-statefulset-b2kqz/ss2-0 to have revision ss2-c79899b9 update revision ss2-787997d666 +Jun 18 12:09:43.759: INFO: Waiting for Pod e2e-tests-statefulset-b2kqz/ss2-1 to have revision ss2-c79899b9 update revision ss2-787997d666 +Jun 18 12:09:53.807: INFO: Waiting for StatefulSet e2e-tests-statefulset-b2kqz/ss2 to complete update +STEP: Rolling back to a previous revision +Jun 18 12:10:03.791: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 exec --namespace=e2e-tests-statefulset-b2kqz ss2-1 -- /bin/sh -c mv -v /usr/share/nginx/html/index.html /tmp/ || true' +Jun 18 12:10:04.208: INFO: stderr: "" +Jun 18 12:10:04.208: INFO: stdout: "'/usr/share/nginx/html/index.html' -> '/tmp/index.html'\n" +Jun 18 12:10:04.208: INFO: stdout of mv -v /usr/share/nginx/html/index.html /tmp/ || true on ss2-1: '/usr/share/nginx/html/index.html' -> '/tmp/index.html' + +Jun 18 12:10:14.356: INFO: Updating stateful set ss2 +STEP: Rolling back update in reverse ordinal order +Jun 18 12:10:24.449: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 exec --namespace=e2e-tests-statefulset-b2kqz ss2-1 -- /bin/sh -c mv -v /tmp/index.html /usr/share/nginx/html/ || true' +Jun 18 12:10:24.884: INFO: stderr: "" +Jun 18 12:10:24.884: INFO: stdout: "'/tmp/index.html' -> '/usr/share/nginx/html/index.html'\n" +Jun 18 12:10:24.884: INFO: stdout of mv -v /tmp/index.html /usr/share/nginx/html/ || true on ss2-1: '/tmp/index.html' -> '/usr/share/nginx/html/index.html' + +Jun 18 12:10:35.084: INFO: Waiting for StatefulSet e2e-tests-statefulset-b2kqz/ss2 to complete update +Jun 18 12:10:35.084: INFO: Waiting for Pod e2e-tests-statefulset-b2kqz/ss2-0 to have revision ss2-787997d666 update revision ss2-c79899b9 +Jun 18 12:10:35.084: INFO: Waiting for Pod e2e-tests-statefulset-b2kqz/ss2-1 to have revision ss2-787997d666 update revision ss2-c79899b9 +Jun 18 12:10:35.084: INFO: Waiting for Pod e2e-tests-statefulset-b2kqz/ss2-2 to have revision ss2-787997d666 update revision ss2-c79899b9 +Jun 18 12:10:45.134: INFO: Waiting for StatefulSet e2e-tests-statefulset-b2kqz/ss2 to complete update +Jun 18 12:10:45.134: INFO: Waiting for Pod e2e-tests-statefulset-b2kqz/ss2-0 to have revision ss2-787997d666 update revision ss2-c79899b9 +Jun 18 12:10:45.134: INFO: Waiting for Pod e2e-tests-statefulset-b2kqz/ss2-1 to have revision ss2-787997d666 update revision ss2-c79899b9 +Jun 18 12:10:55.115: INFO: Waiting for StatefulSet e2e-tests-statefulset-b2kqz/ss2 to complete update +Jun 18 12:10:55.116: INFO: Waiting for Pod e2e-tests-statefulset-b2kqz/ss2-0 to have revision ss2-787997d666 update revision ss2-c79899b9 +[AfterEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:85 +Jun 18 12:11:05.133: INFO: Deleting all statefulset in ns e2e-tests-statefulset-b2kqz +Jun 18 12:11:05.147: INFO: Scaling statefulset ss2 to 0 +Jun 18 12:11:35.227: INFO: Waiting for statefulset status.replicas updated to 0 +Jun 18 12:11:35.242: INFO: Deleting statefulset ss2 +[AfterEach] [sig-apps] StatefulSet /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 19:25:03.264: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-gc-zmx24" for this suite. -May 13 19:25:11.302: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 19:25:11.327: INFO: namespace: e2e-tests-gc-zmx24, resource: bindings, ignored listing per whitelist -May 13 19:25:11.623: INFO: namespace e2e-tests-gc-zmx24 deletion completed in 8.35220014s +Jun 18 12:11:35.310: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "e2e-tests-statefulset-b2kqz" for this suite. +Jun 18 12:11:43.383: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:11:43.626: INFO: namespace: e2e-tests-statefulset-b2kqz, resource: bindings, ignored listing per whitelist +Jun 18 12:11:43.900: INFO: namespace e2e-tests-statefulset-b2kqz deletion completed in 8.571896658s -• [SLOW TEST:9.743 seconds] -[sig-api-machinery] Garbage collector -/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:22 - should delete RS created by deployment when not orphaning [Conformance] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 +• [SLOW TEST:163.327 seconds] +[sig-apps] StatefulSet +/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22 + [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 + should perform rolling updates and roll backs of template modifications [Conformance] + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ -SSSSS +SSSSSSSSSSSSSS ------------------------------ -[k8s.io] [sig-node] Events - should be sent by kubelets and the scheduler about pods scheduling and running [Conformance] +[sig-apps] ReplicationController + should adopt matching pods on creation [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -[BeforeEach] [k8s.io] [sig-node] Events +[BeforeEach] [sig-apps] ReplicationController /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client -May 13 19:25:11.623: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -STEP: Building a namespace api object, basename events -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-events-6m5bj +Jun 18 12:11:43.903: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 +STEP: Building a namespace api object, basename replication-controller +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-replication-controller-trhgx STEP: Waiting for a default service account to be provisioned in namespace -[It] should be sent by kubelets and the scheduler about pods scheduling and running [Conformance] +[It] should adopt matching pods on creation [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -STEP: creating the pod -STEP: submitting the pod to kubernetes -STEP: verifying the pod is in kubernetes -STEP: retrieving the pod -May 13 19:25:16.035: INFO: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:send-events-d3c6b603-75b4-11e9-8f67-2632f168be36,GenerateName:,Namespace:e2e-tests-events-6m5bj,SelfLink:/api/v1/namespaces/e2e-tests-events-6m5bj/pods/send-events-d3c6b603-75b4-11e9-8f67-2632f168be36,UID:d3c84ad6-75b4-11e9-b786-da20024d205c,ResourceVersion:42367,Generation:0,CreationTimestamp:2019-05-13 19:25:11 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: foo,time: 908922510,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-dlbb9 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-dlbb9,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{p gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1 [] [] [{ 0 80 TCP }] [] [] {map[] map[]} [{default-token-dlbb9 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*30,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.170.219.153,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc001959d80} {node.kubernetes.io/unreachable Exists NoExecute 0xc001959da0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Running,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:25:11 +0000 UTC } {Ready True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:25:14 +0000 UTC } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:25:14 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:25:11 +0000 UTC }],Message:,Reason:,HostIP:10.170.219.153,PodIP:172.30.63.32,StartTime:2019-05-13 19:25:11 +0000 UTC,ContainerStatuses:[{p {nil ContainerStateRunning{StartedAt:2019-05-13 19:25:13 +0000 UTC,} nil} {nil nil nil} true 0 gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1 gcr.io/kubernetes-e2e-test-images/serve-hostname@sha256:bab70473a6d8ef65a22625dc9a1b0f0452e811530fdbe77e4408523460177ff1 containerd://6aa1104d809be5bdef34bf1f3d031c1498735350697db8fed51a5967447ab2c0}],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} - -STEP: checking for scheduler event about the pod -May 13 19:25:18.056: INFO: Saw scheduler event for our pod. -STEP: checking for kubelet event about the pod -May 13 19:25:20.063: INFO: Saw kubelet event for our pod. -STEP: deleting the pod -[AfterEach] [k8s.io] [sig-node] Events +STEP: Given a Pod with a 'name' label pod-adoption is created +STEP: When a replication controller with a matching selector is created +STEP: Then the orphan pod is adopted +[AfterEach] [sig-apps] ReplicationController /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 19:25:20.077: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-events-6m5bj" for this suite. -May 13 19:26:00.119: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 19:26:00.283: INFO: namespace: e2e-tests-events-6m5bj, resource: bindings, ignored listing per whitelist -May 13 19:26:00.455: INFO: namespace e2e-tests-events-6m5bj deletion completed in 40.368618704s +Jun 18 12:11:52.125: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "e2e-tests-replication-controller-trhgx" for this suite. +Jun 18 12:12:18.197: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:12:18.320: INFO: namespace: e2e-tests-replication-controller-trhgx, resource: bindings, ignored listing per whitelist +Jun 18 12:12:18.762: INFO: namespace e2e-tests-replication-controller-trhgx deletion completed in 26.618058139s -• [SLOW TEST:48.832 seconds] -[k8s.io] [sig-node] Events -/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 - should be sent by kubelets and the scheduler about pods scheduling and running [Conformance] +• [SLOW TEST:34.860 seconds] +[sig-apps] ReplicationController +/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22 + should adopt matching pods on creation [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SS ------------------------------ -[sig-storage] Projected secret - optional updates should be reflected in volume [NodeConformance] [Conformance] +[sig-network] Service endpoints latency + should not be very high [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -[BeforeEach] [sig-storage] Projected secret +[BeforeEach] [sig-network] Service endpoints latency /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client -May 13 19:26:00.455: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -STEP: Building a namespace api object, basename projected -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-projected-ltvkg +Jun 18 12:12:18.763: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 +STEP: Building a namespace api object, basename svc-latency +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-svc-latency-2s59d STEP: Waiting for a default service account to be provisioned in namespace -[It] optional updates should be reflected in volume [NodeConformance] [Conformance] +[It] should not be very high [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -STEP: Creating secret with name s-test-opt-del-f0e124f8-75b4-11e9-8f67-2632f168be36 -STEP: Creating secret with name s-test-opt-upd-f0e1259d-75b4-11e9-8f67-2632f168be36 -STEP: Creating the pod -STEP: Deleting secret s-test-opt-del-f0e124f8-75b4-11e9-8f67-2632f168be36 -STEP: Updating secret s-test-opt-upd-f0e1259d-75b4-11e9-8f67-2632f168be36 -STEP: Creating secret with name s-test-opt-create-f0e125bb-75b4-11e9-8f67-2632f168be36 -STEP: waiting to observe update in volume -[AfterEach] [sig-storage] Projected secret +STEP: creating replication controller svc-latency-rc in namespace e2e-tests-svc-latency-2s59d +I0618 12:12:19.218107 17 runners.go:184] Created replication controller with name: svc-latency-rc, namespace: e2e-tests-svc-latency-2s59d, replica count: 1 +I0618 12:12:20.268511 17 runners.go:184] svc-latency-rc Pods: 1 out of 1 created, 0 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady +I0618 12:12:21.268804 17 runners.go:184] svc-latency-rc Pods: 1 out of 1 created, 1 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady +Jun 18 12:12:21.400: INFO: Created: latency-svc-zvd9c +Jun 18 12:12:21.423: INFO: Got endpoints: latency-svc-zvd9c [54.631568ms] +Jun 18 12:12:21.517: INFO: Created: latency-svc-x2tf4 +Jun 18 12:12:21.526: INFO: Got endpoints: latency-svc-x2tf4 [102.180344ms] +Jun 18 12:12:21.536: INFO: Created: latency-svc-x7dmn +Jun 18 12:12:21.545: INFO: Got endpoints: latency-svc-x7dmn [120.248753ms] +Jun 18 12:12:21.551: INFO: Created: latency-svc-n6frb +Jun 18 12:12:21.560: INFO: Got endpoints: latency-svc-n6frb [135.813347ms] +Jun 18 12:12:21.569: INFO: Created: latency-svc-8bzd5 +Jun 18 12:12:21.578: INFO: Got endpoints: latency-svc-8bzd5 [153.228533ms] +Jun 18 12:12:21.587: INFO: Created: latency-svc-k5z5s +Jun 18 12:12:21.596: INFO: Got endpoints: latency-svc-k5z5s [170.926692ms] +Jun 18 12:12:21.608: INFO: Created: latency-svc-42z2z +Jun 18 12:12:21.621: INFO: Got endpoints: latency-svc-42z2z [196.4617ms] +Jun 18 12:12:21.628: INFO: Created: latency-svc-2pn4s +Jun 18 12:12:21.637: INFO: Got endpoints: latency-svc-2pn4s [211.843627ms] +Jun 18 12:12:21.644: INFO: Created: latency-svc-w2jmx +Jun 18 12:12:21.653: INFO: Got endpoints: latency-svc-w2jmx [228.146356ms] +Jun 18 12:12:21.663: INFO: Created: latency-svc-jq9d8 +Jun 18 12:12:21.673: INFO: Got endpoints: latency-svc-jq9d8 [248.48844ms] +Jun 18 12:12:21.683: INFO: Created: latency-svc-n2x47 +Jun 18 12:12:21.692: INFO: Got endpoints: latency-svc-n2x47 [267.272925ms] +Jun 18 12:12:21.701: INFO: Created: latency-svc-fsv9r +Jun 18 12:12:21.710: INFO: Got endpoints: latency-svc-fsv9r [284.725913ms] +Jun 18 12:12:21.718: INFO: Created: latency-svc-psc9p +Jun 18 12:12:21.728: INFO: Got endpoints: latency-svc-psc9p [302.997155ms] +Jun 18 12:12:21.803: INFO: Created: latency-svc-qtpxr +Jun 18 12:12:21.815: INFO: Got endpoints: latency-svc-qtpxr [389.670609ms] +Jun 18 12:12:21.823: INFO: Created: latency-svc-kv2wg +Jun 18 12:12:21.831: INFO: Got endpoints: latency-svc-kv2wg [405.426607ms] +Jun 18 12:12:21.845: INFO: Created: latency-svc-l95fd +Jun 18 12:12:21.854: INFO: Got endpoints: latency-svc-l95fd [428.527881ms] +Jun 18 12:12:21.861: INFO: Created: latency-svc-h68sr +Jun 18 12:12:21.868: INFO: Got endpoints: latency-svc-h68sr [341.839025ms] +Jun 18 12:12:21.879: INFO: Created: latency-svc-l585t +Jun 18 12:12:21.890: INFO: Got endpoints: latency-svc-l585t [344.938025ms] +Jun 18 12:12:21.984: INFO: Created: latency-svc-snst8 +Jun 18 12:12:21.993: INFO: Got endpoints: latency-svc-snst8 [432.485424ms] +Jun 18 12:12:22.000: INFO: Created: latency-svc-c4qk7 +Jun 18 12:12:22.012: INFO: Got endpoints: latency-svc-c4qk7 [434.290341ms] +Jun 18 12:12:22.026: INFO: Created: latency-svc-jll8z +Jun 18 12:12:22.036: INFO: Got endpoints: latency-svc-jll8z [440.261183ms] +Jun 18 12:12:22.047: INFO: Created: latency-svc-6t5bl +Jun 18 12:12:22.056: INFO: Got endpoints: latency-svc-6t5bl [434.702512ms] +Jun 18 12:12:22.067: INFO: Created: latency-svc-zwjsb +Jun 18 12:12:22.075: INFO: Got endpoints: latency-svc-zwjsb [438.540147ms] +Jun 18 12:12:22.083: INFO: Created: latency-svc-56wsq +Jun 18 12:12:22.092: INFO: Got endpoints: latency-svc-56wsq [439.011195ms] +Jun 18 12:12:22.101: INFO: Created: latency-svc-5j8l6 +Jun 18 12:12:22.111: INFO: Got endpoints: latency-svc-5j8l6 [437.191241ms] +Jun 18 12:12:22.125: INFO: Created: latency-svc-rd5p8 +Jun 18 12:12:22.135: INFO: Got endpoints: latency-svc-rd5p8 [442.107251ms] +Jun 18 12:12:22.143: INFO: Created: latency-svc-xm7ht +Jun 18 12:12:22.152: INFO: Got endpoints: latency-svc-xm7ht [441.875561ms] +Jun 18 12:12:22.164: INFO: Created: latency-svc-w865v +Jun 18 12:12:22.172: INFO: Got endpoints: latency-svc-w865v [443.521062ms] +Jun 18 12:12:22.179: INFO: Created: latency-svc-l9fll +Jun 18 12:12:22.188: INFO: Got endpoints: latency-svc-l9fll [373.35196ms] +Jun 18 12:12:22.196: INFO: Created: latency-svc-nzmf2 +Jun 18 12:12:22.206: INFO: Got endpoints: latency-svc-nzmf2 [375.398314ms] +Jun 18 12:12:22.218: INFO: Created: latency-svc-wc595 +Jun 18 12:12:22.229: INFO: Got endpoints: latency-svc-wc595 [374.987012ms] +Jun 18 12:12:22.239: INFO: Created: latency-svc-d5j62 +Jun 18 12:12:22.251: INFO: Got endpoints: latency-svc-d5j62 [383.391309ms] +Jun 18 12:12:22.263: INFO: Created: latency-svc-249pf +Jun 18 12:12:22.273: INFO: Got endpoints: latency-svc-249pf [382.877396ms] +Jun 18 12:12:22.279: INFO: Created: latency-svc-v9jvn +Jun 18 12:12:22.290: INFO: Got endpoints: latency-svc-v9jvn [297.268496ms] +Jun 18 12:12:22.294: INFO: Created: latency-svc-9wfd5 +Jun 18 12:12:22.303: INFO: Got endpoints: latency-svc-9wfd5 [290.935501ms] +Jun 18 12:12:22.311: INFO: Created: latency-svc-2r2nw +Jun 18 12:12:22.320: INFO: Got endpoints: latency-svc-2r2nw [284.238801ms] +Jun 18 12:12:22.333: INFO: Created: latency-svc-m5m92 +Jun 18 12:12:22.342: INFO: Got endpoints: latency-svc-m5m92 [286.358506ms] +Jun 18 12:12:22.352: INFO: Created: latency-svc-bk8j8 +Jun 18 12:12:22.363: INFO: Got endpoints: latency-svc-bk8j8 [287.497253ms] +Jun 18 12:12:22.371: INFO: Created: latency-svc-t42z5 +Jun 18 12:12:22.380: INFO: Got endpoints: latency-svc-t42z5 [287.198588ms] +Jun 18 12:12:22.390: INFO: Created: latency-svc-hjm4j +Jun 18 12:12:22.397: INFO: Got endpoints: latency-svc-hjm4j [286.57377ms] +Jun 18 12:12:22.405: INFO: Created: latency-svc-9tpk5 +Jun 18 12:12:22.416: INFO: Got endpoints: latency-svc-9tpk5 [281.112973ms] +Jun 18 12:12:22.425: INFO: Created: latency-svc-ldrdk +Jun 18 12:12:22.434: INFO: Got endpoints: latency-svc-ldrdk [281.590495ms] +Jun 18 12:12:22.442: INFO: Created: latency-svc-69rvn +Jun 18 12:12:22.453: INFO: Got endpoints: latency-svc-69rvn [280.518968ms] +Jun 18 12:12:22.461: INFO: Created: latency-svc-mkd7g +Jun 18 12:12:22.472: INFO: Got endpoints: latency-svc-mkd7g [283.805804ms] +Jun 18 12:12:22.479: INFO: Created: latency-svc-7hdjv +Jun 18 12:12:22.488: INFO: Got endpoints: latency-svc-7hdjv [281.886503ms] +Jun 18 12:12:22.501: INFO: Created: latency-svc-nx7n9 +Jun 18 12:12:22.514: INFO: Got endpoints: latency-svc-nx7n9 [285.012838ms] +Jun 18 12:12:22.544: INFO: Created: latency-svc-h6wrd +Jun 18 12:12:22.556: INFO: Got endpoints: latency-svc-h6wrd [304.58996ms] +Jun 18 12:12:22.631: INFO: Created: latency-svc-ng6v4 +Jun 18 12:12:22.642: INFO: Got endpoints: latency-svc-ng6v4 [369.624646ms] +Jun 18 12:12:22.656: INFO: Created: latency-svc-g74lw +Jun 18 12:12:22.664: INFO: Got endpoints: latency-svc-g74lw [373.289216ms] +Jun 18 12:12:22.673: INFO: Created: latency-svc-7xs6f +Jun 18 12:12:22.682: INFO: Got endpoints: latency-svc-7xs6f [378.573447ms] +Jun 18 12:12:22.692: INFO: Created: latency-svc-6hqhj +Jun 18 12:12:22.704: INFO: Got endpoints: latency-svc-6hqhj [383.253552ms] +Jun 18 12:12:22.719: INFO: Created: latency-svc-8fkjf +Jun 18 12:12:22.728: INFO: Created: latency-svc-fsv8x +Jun 18 12:12:22.729: INFO: Got endpoints: latency-svc-8fkjf [386.851459ms] +Jun 18 12:12:22.739: INFO: Got endpoints: latency-svc-fsv8x [375.200433ms] +Jun 18 12:12:22.747: INFO: Created: latency-svc-r5swg +Jun 18 12:12:22.760: INFO: Got endpoints: latency-svc-r5swg [380.438672ms] +Jun 18 12:12:22.766: INFO: Created: latency-svc-vmddc +Jun 18 12:12:22.777: INFO: Got endpoints: latency-svc-vmddc [379.656037ms] +Jun 18 12:12:22.785: INFO: Created: latency-svc-25pq4 +Jun 18 12:12:22.794: INFO: Got endpoints: latency-svc-25pq4 [377.626535ms] +Jun 18 12:12:22.803: INFO: Created: latency-svc-xn2sh +Jun 18 12:12:22.819: INFO: Created: latency-svc-94qgl +Jun 18 12:12:22.828: INFO: Got endpoints: latency-svc-xn2sh [394.031427ms] +Jun 18 12:12:22.840: INFO: Created: latency-svc-qlxq5 +Jun 18 12:12:22.864: INFO: Created: latency-svc-t8dtq +Jun 18 12:12:22.876: INFO: Got endpoints: latency-svc-94qgl [423.582557ms] +Jun 18 12:12:22.884: INFO: Created: latency-svc-djsgj +Jun 18 12:12:22.898: INFO: Created: latency-svc-bj7vx +Jun 18 12:12:22.920: INFO: Created: latency-svc-lvnfs +Jun 18 12:12:22.926: INFO: Got endpoints: latency-svc-qlxq5 [453.549171ms] +Jun 18 12:12:22.934: INFO: Created: latency-svc-lwpp9 +Jun 18 12:12:22.951: INFO: Created: latency-svc-pnh9s +Jun 18 12:12:22.967: INFO: Created: latency-svc-hzlbw +Jun 18 12:12:22.975: INFO: Got endpoints: latency-svc-t8dtq [486.431252ms] +Jun 18 12:12:22.983: INFO: Created: latency-svc-f6fhp +Jun 18 12:12:23.001: INFO: Created: latency-svc-6c7lc +Jun 18 12:12:23.020: INFO: Created: latency-svc-ctpkh +Jun 18 12:12:23.025: INFO: Got endpoints: latency-svc-djsgj [510.893392ms] +Jun 18 12:12:23.036: INFO: Created: latency-svc-jlrp2 +Jun 18 12:12:23.050: INFO: Created: latency-svc-d7hnr +Jun 18 12:12:23.068: INFO: Created: latency-svc-tz8bs +Jun 18 12:12:23.077: INFO: Got endpoints: latency-svc-bj7vx [520.951558ms] +Jun 18 12:12:23.085: INFO: Created: latency-svc-tq9zk +Jun 18 12:12:23.101: INFO: Created: latency-svc-rp9jc +Jun 18 12:12:23.119: INFO: Created: latency-svc-jspjt +Jun 18 12:12:23.126: INFO: Got endpoints: latency-svc-lvnfs [483.589495ms] +Jun 18 12:12:23.135: INFO: Created: latency-svc-x9xkh +Jun 18 12:12:23.155: INFO: Created: latency-svc-lvxfx +Jun 18 12:12:23.171: INFO: Created: latency-svc-vxjbx +Jun 18 12:12:23.176: INFO: Got endpoints: latency-svc-lwpp9 [512.443396ms] +Jun 18 12:12:23.216: INFO: Created: latency-svc-fvj7k +Jun 18 12:12:23.226: INFO: Got endpoints: latency-svc-pnh9s [544.048989ms] +Jun 18 12:12:23.264: INFO: Created: latency-svc-qzbkk +Jun 18 12:12:23.276: INFO: Got endpoints: latency-svc-hzlbw [571.793858ms] +Jun 18 12:12:23.306: INFO: Created: latency-svc-nm8ks +Jun 18 12:12:23.326: INFO: Got endpoints: latency-svc-f6fhp [596.991912ms] +Jun 18 12:12:23.356: INFO: Created: latency-svc-tzjf9 +Jun 18 12:12:23.376: INFO: Got endpoints: latency-svc-6c7lc [637.256025ms] +Jun 18 12:12:23.406: INFO: Created: latency-svc-dtl62 +Jun 18 12:12:23.426: INFO: Got endpoints: latency-svc-ctpkh [665.823313ms] +Jun 18 12:12:23.456: INFO: Created: latency-svc-4hlg4 +Jun 18 12:12:23.476: INFO: Got endpoints: latency-svc-jlrp2 [699.051929ms] +Jun 18 12:12:23.507: INFO: Created: latency-svc-4bp8s +Jun 18 12:12:23.525: INFO: Got endpoints: latency-svc-d7hnr [731.611388ms] +Jun 18 12:12:23.558: INFO: Created: latency-svc-lv5tp +Jun 18 12:12:23.576: INFO: Got endpoints: latency-svc-tz8bs [748.586102ms] +Jun 18 12:12:23.608: INFO: Created: latency-svc-8vksv +Jun 18 12:12:23.625: INFO: Got endpoints: latency-svc-tq9zk [748.793237ms] +Jun 18 12:12:23.657: INFO: Created: latency-svc-vd4bh +Jun 18 12:12:23.676: INFO: Got endpoints: latency-svc-rp9jc [750.322886ms] +Jun 18 12:12:23.714: INFO: Created: latency-svc-gsfv7 +Jun 18 12:12:23.725: INFO: Got endpoints: latency-svc-jspjt [749.91922ms] +Jun 18 12:12:23.757: INFO: Created: latency-svc-xx82k +Jun 18 12:12:23.778: INFO: Got endpoints: latency-svc-x9xkh [752.656617ms] +Jun 18 12:12:23.816: INFO: Created: latency-svc-fmx5t +Jun 18 12:12:23.826: INFO: Got endpoints: latency-svc-lvxfx [748.626796ms] +Jun 18 12:12:23.860: INFO: Created: latency-svc-44952 +Jun 18 12:12:23.875: INFO: Got endpoints: latency-svc-vxjbx [749.226253ms] +Jun 18 12:12:23.908: INFO: Created: latency-svc-g7vw6 +Jun 18 12:12:23.927: INFO: Got endpoints: latency-svc-fvj7k [750.384058ms] +Jun 18 12:12:23.959: INFO: Created: latency-svc-6p58p +Jun 18 12:12:23.975: INFO: Got endpoints: latency-svc-qzbkk [749.287416ms] +Jun 18 12:12:24.007: INFO: Created: latency-svc-n9k9d +Jun 18 12:12:24.025: INFO: Got endpoints: latency-svc-nm8ks [749.741971ms] +Jun 18 12:12:24.057: INFO: Created: latency-svc-7b2dh +Jun 18 12:12:24.076: INFO: Got endpoints: latency-svc-tzjf9 [750.22507ms] +Jun 18 12:12:24.107: INFO: Created: latency-svc-69l2t +Jun 18 12:12:24.125: INFO: Got endpoints: latency-svc-dtl62 [747.721798ms] +Jun 18 12:12:24.157: INFO: Created: latency-svc-s45ck +Jun 18 12:12:24.176: INFO: Got endpoints: latency-svc-4hlg4 [750.161137ms] +Jun 18 12:12:24.208: INFO: Created: latency-svc-vtnpl +Jun 18 12:12:24.226: INFO: Got endpoints: latency-svc-4bp8s [750.360782ms] +Jun 18 12:12:24.265: INFO: Created: latency-svc-447lx +Jun 18 12:12:24.277: INFO: Got endpoints: latency-svc-lv5tp [751.553736ms] +Jun 18 12:12:24.323: INFO: Created: latency-svc-9ct7t +Jun 18 12:12:24.326: INFO: Got endpoints: latency-svc-8vksv [749.710213ms] +Jun 18 12:12:24.359: INFO: Created: latency-svc-lll4x +Jun 18 12:12:24.378: INFO: Got endpoints: latency-svc-vd4bh [752.696787ms] +Jun 18 12:12:24.408: INFO: Created: latency-svc-4vfs7 +Jun 18 12:12:24.433: INFO: Got endpoints: latency-svc-gsfv7 [756.498871ms] +Jun 18 12:12:24.464: INFO: Created: latency-svc-9dh5g +Jun 18 12:12:24.475: INFO: Got endpoints: latency-svc-xx82k [749.566906ms] +Jun 18 12:12:24.504: INFO: Created: latency-svc-mn4dp +Jun 18 12:12:24.526: INFO: Got endpoints: latency-svc-fmx5t [748.091555ms] +Jun 18 12:12:24.558: INFO: Created: latency-svc-xkpt4 +Jun 18 12:12:24.576: INFO: Got endpoints: latency-svc-44952 [749.71911ms] +Jun 18 12:12:24.605: INFO: Created: latency-svc-4ftnc +Jun 18 12:12:24.625: INFO: Got endpoints: latency-svc-g7vw6 [749.775561ms] +Jun 18 12:12:24.656: INFO: Created: latency-svc-zncqc +Jun 18 12:12:24.676: INFO: Got endpoints: latency-svc-6p58p [748.836015ms] +Jun 18 12:12:24.706: INFO: Created: latency-svc-qt6gx +Jun 18 12:12:24.725: INFO: Got endpoints: latency-svc-n9k9d [750.077609ms] +Jun 18 12:12:24.763: INFO: Created: latency-svc-nqqzn +Jun 18 12:12:24.775: INFO: Got endpoints: latency-svc-7b2dh [749.862677ms] +Jun 18 12:12:24.805: INFO: Created: latency-svc-sm794 +Jun 18 12:12:24.826: INFO: Got endpoints: latency-svc-69l2t [749.959385ms] +Jun 18 12:12:24.857: INFO: Created: latency-svc-k7d7c +Jun 18 12:12:24.875: INFO: Got endpoints: latency-svc-s45ck [750.330996ms] +Jun 18 12:12:24.907: INFO: Created: latency-svc-9cd4j +Jun 18 12:12:24.925: INFO: Got endpoints: latency-svc-vtnpl [749.295845ms] +Jun 18 12:12:24.955: INFO: Created: latency-svc-q69vb +Jun 18 12:12:24.980: INFO: Got endpoints: latency-svc-447lx [754.117246ms] +Jun 18 12:12:25.014: INFO: Created: latency-svc-wcqpc +Jun 18 12:12:25.026: INFO: Got endpoints: latency-svc-9ct7t [749.103554ms] +Jun 18 12:12:25.057: INFO: Created: latency-svc-h8bsm +Jun 18 12:12:25.075: INFO: Got endpoints: latency-svc-lll4x [748.176977ms] +Jun 18 12:12:25.107: INFO: Created: latency-svc-fvmcd +Jun 18 12:12:25.125: INFO: Got endpoints: latency-svc-4vfs7 [747.535821ms] +Jun 18 12:12:25.156: INFO: Created: latency-svc-t6scc +Jun 18 12:12:25.176: INFO: Got endpoints: latency-svc-9dh5g [743.039834ms] +Jun 18 12:12:25.214: INFO: Created: latency-svc-cthd5 +Jun 18 12:12:25.229: INFO: Got endpoints: latency-svc-mn4dp [754.331939ms] +Jun 18 12:12:25.261: INFO: Created: latency-svc-lcdcg +Jun 18 12:12:25.275: INFO: Got endpoints: latency-svc-xkpt4 [749.346325ms] +Jun 18 12:12:25.308: INFO: Created: latency-svc-rjdh4 +Jun 18 12:12:25.326: INFO: Got endpoints: latency-svc-4ftnc [750.561246ms] +Jun 18 12:12:25.375: INFO: Got endpoints: latency-svc-zncqc [749.717394ms] +Jun 18 12:12:25.377: INFO: Created: latency-svc-rqg4l +Jun 18 12:12:25.404: INFO: Created: latency-svc-pdjkv +Jun 18 12:12:25.426: INFO: Got endpoints: latency-svc-qt6gx [750.411356ms] +Jun 18 12:12:25.460: INFO: Created: latency-svc-vhrvj +Jun 18 12:12:25.476: INFO: Got endpoints: latency-svc-nqqzn [750.576879ms] +Jun 18 12:12:25.506: INFO: Created: latency-svc-9ln45 +Jun 18 12:12:25.528: INFO: Got endpoints: latency-svc-sm794 [752.321294ms] +Jun 18 12:12:25.578: INFO: Got endpoints: latency-svc-k7d7c [751.608325ms] +Jun 18 12:12:25.627: INFO: Got endpoints: latency-svc-9cd4j [751.247572ms] +Jun 18 12:12:25.679: INFO: Got endpoints: latency-svc-q69vb [752.914833ms] +Jun 18 12:12:25.729: INFO: Got endpoints: latency-svc-wcqpc [748.15402ms] +Jun 18 12:12:25.760: INFO: Created: latency-svc-s8nww +Jun 18 12:12:25.776: INFO: Got endpoints: latency-svc-h8bsm [749.707081ms] +Jun 18 12:12:25.776: INFO: Created: latency-svc-ngpb9 +Jun 18 12:12:25.795: INFO: Created: latency-svc-zlw8c +Jun 18 12:12:25.812: INFO: Created: latency-svc-kn8k8 +Jun 18 12:12:25.826: INFO: Got endpoints: latency-svc-fvmcd [750.92495ms] +Jun 18 12:12:25.829: INFO: Created: latency-svc-pzfcd +Jun 18 12:12:25.847: INFO: Created: latency-svc-cbj4p +Jun 18 12:12:25.865: INFO: Created: latency-svc-nvcvx +Jun 18 12:12:25.876: INFO: Got endpoints: latency-svc-t6scc [751.020321ms] +Jun 18 12:12:25.910: INFO: Created: latency-svc-cqp55 +Jun 18 12:12:25.926: INFO: Got endpoints: latency-svc-cthd5 [749.909019ms] +Jun 18 12:12:25.957: INFO: Created: latency-svc-mj4vk +Jun 18 12:12:25.977: INFO: Got endpoints: latency-svc-lcdcg [747.610365ms] +Jun 18 12:12:26.010: INFO: Created: latency-svc-4zfrf +Jun 18 12:12:26.030: INFO: Got endpoints: latency-svc-rjdh4 [754.429987ms] +Jun 18 12:12:26.064: INFO: Created: latency-svc-z5mnz +Jun 18 12:12:26.079: INFO: Got endpoints: latency-svc-rqg4l [752.219013ms] +Jun 18 12:12:26.116: INFO: Created: latency-svc-zxvfz +Jun 18 12:12:26.127: INFO: Got endpoints: latency-svc-pdjkv [752.406732ms] +Jun 18 12:12:26.158: INFO: Created: latency-svc-plcrx +Jun 18 12:12:26.176: INFO: Got endpoints: latency-svc-vhrvj [749.63327ms] +Jun 18 12:12:26.207: INFO: Created: latency-svc-z2ddm +Jun 18 12:12:26.226: INFO: Got endpoints: latency-svc-9ln45 [750.117632ms] +Jun 18 12:12:26.257: INFO: Created: latency-svc-8cckf +Jun 18 12:12:26.276: INFO: Got endpoints: latency-svc-s8nww [748.858357ms] +Jun 18 12:12:26.310: INFO: Created: latency-svc-qtp6n +Jun 18 12:12:26.327: INFO: Got endpoints: latency-svc-ngpb9 [748.614553ms] +Jun 18 12:12:26.361: INFO: Created: latency-svc-82v2d +Jun 18 12:12:26.379: INFO: Got endpoints: latency-svc-zlw8c [751.878489ms] +Jun 18 12:12:26.416: INFO: Created: latency-svc-pvctk +Jun 18 12:12:26.427: INFO: Got endpoints: latency-svc-kn8k8 [748.838258ms] +Jun 18 12:12:26.458: INFO: Created: latency-svc-vnmh9 +Jun 18 12:12:26.477: INFO: Got endpoints: latency-svc-pzfcd [747.869033ms] +Jun 18 12:12:26.511: INFO: Created: latency-svc-z5d8r +Jun 18 12:12:26.525: INFO: Got endpoints: latency-svc-cbj4p [749.547835ms] +Jun 18 12:12:26.561: INFO: Created: latency-svc-2fjzk +Jun 18 12:12:26.576: INFO: Got endpoints: latency-svc-nvcvx [749.31538ms] +Jun 18 12:12:26.606: INFO: Created: latency-svc-xn4l2 +Jun 18 12:12:26.630: INFO: Got endpoints: latency-svc-cqp55 [753.646306ms] +Jun 18 12:12:26.660: INFO: Created: latency-svc-bv86n +Jun 18 12:12:26.677: INFO: Got endpoints: latency-svc-mj4vk [750.908599ms] +Jun 18 12:12:26.707: INFO: Created: latency-svc-xgfsj +Jun 18 12:12:26.726: INFO: Got endpoints: latency-svc-4zfrf [749.120544ms] +Jun 18 12:12:26.757: INFO: Created: latency-svc-qs66c +Jun 18 12:12:26.776: INFO: Got endpoints: latency-svc-z5mnz [746.378785ms] +Jun 18 12:12:26.810: INFO: Created: latency-svc-xtjvn +Jun 18 12:12:26.825: INFO: Got endpoints: latency-svc-zxvfz [746.329454ms] +Jun 18 12:12:26.858: INFO: Created: latency-svc-49lm2 +Jun 18 12:12:26.876: INFO: Got endpoints: latency-svc-plcrx [749.145907ms] +Jun 18 12:12:26.909: INFO: Created: latency-svc-hsmzx +Jun 18 12:12:26.926: INFO: Got endpoints: latency-svc-z2ddm [750.184676ms] +Jun 18 12:12:26.969: INFO: Created: latency-svc-j9gc2 +Jun 18 12:12:26.975: INFO: Got endpoints: latency-svc-8cckf [748.941277ms] +Jun 18 12:12:27.025: INFO: Got endpoints: latency-svc-qtp6n [748.95904ms] +Jun 18 12:12:27.077: INFO: Got endpoints: latency-svc-82v2d [750.49935ms] +Jun 18 12:12:27.126: INFO: Got endpoints: latency-svc-pvctk [747.116792ms] +Jun 18 12:12:27.176: INFO: Got endpoints: latency-svc-vnmh9 [748.448081ms] +Jun 18 12:12:27.228: INFO: Got endpoints: latency-svc-z5d8r [751.711252ms] +Jun 18 12:12:27.277: INFO: Got endpoints: latency-svc-2fjzk [750.846441ms] +Jun 18 12:12:27.326: INFO: Got endpoints: latency-svc-xn4l2 [750.578321ms] +Jun 18 12:12:27.376: INFO: Got endpoints: latency-svc-bv86n [746.172119ms] +Jun 18 12:12:27.425: INFO: Got endpoints: latency-svc-xgfsj [748.666435ms] +Jun 18 12:12:27.476: INFO: Got endpoints: latency-svc-qs66c [749.814621ms] +Jun 18 12:12:27.526: INFO: Got endpoints: latency-svc-xtjvn [750.083959ms] +Jun 18 12:12:27.575: INFO: Got endpoints: latency-svc-49lm2 [750.233063ms] +Jun 18 12:12:27.627: INFO: Got endpoints: latency-svc-hsmzx [750.745316ms] +Jun 18 12:12:27.675: INFO: Got endpoints: latency-svc-j9gc2 [748.488411ms] +Jun 18 12:12:28.131: INFO: Created: latency-svc-8tcs9 +Jun 18 12:12:28.136: INFO: Created: latency-svc-bcbft +Jun 18 12:12:28.136: INFO: Created: latency-svc-hmdqr +Jun 18 12:12:28.138: INFO: Created: latency-svc-t8vpp +Jun 18 12:12:28.139: INFO: Created: latency-svc-zv4hn +Jun 18 12:12:28.141: INFO: Created: latency-svc-zjm4w +Jun 18 12:12:28.141: INFO: Created: latency-svc-dvwdp +Jun 18 12:12:28.142: INFO: Got endpoints: latency-svc-8tcs9 [1.166667247s] +Jun 18 12:12:28.145: INFO: Created: latency-svc-zpd95 +Jun 18 12:12:28.146: INFO: Created: latency-svc-nczdw +Jun 18 12:12:28.148: INFO: Created: latency-svc-hpm7d +Jun 18 12:12:28.150: INFO: Created: latency-svc-5r6bp +Jun 18 12:12:28.155: INFO: Got endpoints: latency-svc-bcbft [1.129471701s] +Jun 18 12:12:28.156: INFO: Created: latency-svc-vzkt7 +Jun 18 12:12:28.157: INFO: Created: latency-svc-22gxv +Jun 18 12:12:28.157: INFO: Created: latency-svc-84qhf +Jun 18 12:12:28.158: INFO: Got endpoints: latency-svc-hmdqr [1.08090827s] +Jun 18 12:12:28.160: INFO: Created: latency-svc-t64k2 +Jun 18 12:12:28.160: INFO: Got endpoints: latency-svc-zv4hn [1.034064557s] +Jun 18 12:12:28.161: INFO: Got endpoints: latency-svc-t8vpp [985.272271ms] +Jun 18 12:12:28.162: INFO: Got endpoints: latency-svc-dvwdp [933.908882ms] +Jun 18 12:12:28.169: INFO: Got endpoints: latency-svc-zjm4w [892.781499ms] +Jun 18 12:12:28.171: INFO: Got endpoints: latency-svc-zpd95 [844.081324ms] +Jun 18 12:12:28.176: INFO: Got endpoints: latency-svc-hpm7d [750.602847ms] +Jun 18 12:12:28.176: INFO: Got endpoints: latency-svc-nczdw [799.958753ms] +Jun 18 12:12:28.186: INFO: Created: latency-svc-52wt6 +Jun 18 12:12:28.200: INFO: Created: latency-svc-d7bvm +Jun 18 12:12:28.228: INFO: Got endpoints: latency-svc-84qhf [752.135028ms] +Jun 18 12:12:28.229: INFO: Created: latency-svc-2jl2k +Jun 18 12:12:28.250: INFO: Created: latency-svc-cqghj +Jun 18 12:12:28.270: INFO: Created: latency-svc-ncrdb +Jun 18 12:12:28.279: INFO: Got endpoints: latency-svc-5r6bp [752.667581ms] +Jun 18 12:12:28.295: INFO: Created: latency-svc-4fwzs +Jun 18 12:12:28.322: INFO: Created: latency-svc-b95cq +Jun 18 12:12:28.330: INFO: Got endpoints: latency-svc-t64k2 [754.285679ms] +Jun 18 12:12:28.344: INFO: Created: latency-svc-j99nc +Jun 18 12:12:28.360: INFO: Created: latency-svc-wkr9v +Jun 18 12:12:28.377: INFO: Got endpoints: latency-svc-22gxv [749.941382ms] +Jun 18 12:12:28.382: INFO: Created: latency-svc-swcx9 +Jun 18 12:12:28.398: INFO: Created: latency-svc-l2fwd +Jun 18 12:12:28.418: INFO: Created: latency-svc-vz84w +Jun 18 12:12:28.428: INFO: Got endpoints: latency-svc-vzkt7 [753.486004ms] +Jun 18 12:12:28.434: INFO: Created: latency-svc-89vp4 +Jun 18 12:12:28.450: INFO: Created: latency-svc-rw98k +Jun 18 12:12:28.471: INFO: Created: latency-svc-42wt7 +Jun 18 12:12:28.475: INFO: Got endpoints: latency-svc-52wt6 [333.13616ms] +Jun 18 12:12:28.508: INFO: Created: latency-svc-7qbr9 +Jun 18 12:12:28.527: INFO: Got endpoints: latency-svc-d7bvm [371.468125ms] +Jun 18 12:12:28.561: INFO: Created: latency-svc-kd2nn +Jun 18 12:12:28.577: INFO: Got endpoints: latency-svc-2jl2k [418.916476ms] +Jun 18 12:12:28.609: INFO: Created: latency-svc-vtc8k +Jun 18 12:12:28.627: INFO: Got endpoints: latency-svc-cqghj [466.690955ms] +Jun 18 12:12:28.662: INFO: Created: latency-svc-spmmj +Jun 18 12:12:28.675: INFO: Got endpoints: latency-svc-ncrdb [514.045884ms] +Jun 18 12:12:28.708: INFO: Created: latency-svc-j2zv7 +Jun 18 12:12:28.725: INFO: Got endpoints: latency-svc-4fwzs [562.686061ms] +Jun 18 12:12:28.759: INFO: Created: latency-svc-8plwf +Jun 18 12:12:28.777: INFO: Got endpoints: latency-svc-b95cq [607.072685ms] +Jun 18 12:12:28.808: INFO: Created: latency-svc-k4wvc +Jun 18 12:12:28.828: INFO: Got endpoints: latency-svc-j99nc [657.540814ms] +Jun 18 12:12:28.860: INFO: Created: latency-svc-8mlkj +Jun 18 12:12:28.876: INFO: Got endpoints: latency-svc-wkr9v [700.298727ms] +Jun 18 12:12:28.915: INFO: Created: latency-svc-gctpm +Jun 18 12:12:28.926: INFO: Got endpoints: latency-svc-swcx9 [749.146795ms] +Jun 18 12:12:28.959: INFO: Created: latency-svc-cr8jc +Jun 18 12:12:28.977: INFO: Got endpoints: latency-svc-l2fwd [748.729938ms] +Jun 18 12:12:29.009: INFO: Created: latency-svc-jpbx9 +Jun 18 12:12:29.027: INFO: Got endpoints: latency-svc-vz84w [747.296959ms] +Jun 18 12:12:29.057: INFO: Created: latency-svc-7qq42 +Jun 18 12:12:29.076: INFO: Got endpoints: latency-svc-89vp4 [746.108379ms] +Jun 18 12:12:29.108: INFO: Created: latency-svc-jq49v +Jun 18 12:12:29.125: INFO: Got endpoints: latency-svc-rw98k [747.66056ms] +Jun 18 12:12:29.156: INFO: Created: latency-svc-qvm9t +Jun 18 12:12:29.184: INFO: Got endpoints: latency-svc-42wt7 [755.843493ms] +Jun 18 12:12:29.216: INFO: Created: latency-svc-7xc8f +Jun 18 12:12:29.225: INFO: Got endpoints: latency-svc-7qbr9 [750.027218ms] +Jun 18 12:12:29.255: INFO: Created: latency-svc-zpzhv +Jun 18 12:12:29.276: INFO: Got endpoints: latency-svc-kd2nn [748.916509ms] +Jun 18 12:12:29.312: INFO: Created: latency-svc-zzf7m +Jun 18 12:12:29.330: INFO: Got endpoints: latency-svc-vtc8k [752.874824ms] +Jun 18 12:12:29.380: INFO: Got endpoints: latency-svc-spmmj [753.179238ms] +Jun 18 12:12:29.429: INFO: Got endpoints: latency-svc-j2zv7 [753.564559ms] +Jun 18 12:12:29.476: INFO: Got endpoints: latency-svc-8plwf [750.420558ms] +Jun 18 12:12:29.528: INFO: Got endpoints: latency-svc-k4wvc [751.389777ms] +Jun 18 12:12:29.576: INFO: Got endpoints: latency-svc-8mlkj [747.821518ms] +Jun 18 12:12:29.626: INFO: Got endpoints: latency-svc-gctpm [749.194014ms] +Jun 18 12:12:29.677: INFO: Got endpoints: latency-svc-cr8jc [751.327994ms] +Jun 18 12:12:29.726: INFO: Got endpoints: latency-svc-jpbx9 [749.138402ms] +Jun 18 12:12:29.789: INFO: Got endpoints: latency-svc-7qq42 [762.672184ms] +Jun 18 12:12:29.825: INFO: Got endpoints: latency-svc-jq49v [749.469013ms] +Jun 18 12:12:29.877: INFO: Got endpoints: latency-svc-qvm9t [751.535542ms] +Jun 18 12:12:29.926: INFO: Got endpoints: latency-svc-7xc8f [741.586066ms] +Jun 18 12:12:29.975: INFO: Got endpoints: latency-svc-zpzhv [750.127033ms] +Jun 18 12:12:30.026: INFO: Got endpoints: latency-svc-zzf7m [749.972907ms] +Jun 18 12:12:30.026: INFO: Latencies: [102.180344ms 120.248753ms 135.813347ms 153.228533ms 170.926692ms 196.4617ms 211.843627ms 228.146356ms 248.48844ms 267.272925ms 280.518968ms 281.112973ms 281.590495ms 281.886503ms 283.805804ms 284.238801ms 284.725913ms 285.012838ms 286.358506ms 286.57377ms 287.198588ms 287.497253ms 290.935501ms 297.268496ms 302.997155ms 304.58996ms 333.13616ms 341.839025ms 344.938025ms 369.624646ms 371.468125ms 373.289216ms 373.35196ms 374.987012ms 375.200433ms 375.398314ms 377.626535ms 378.573447ms 379.656037ms 380.438672ms 382.877396ms 383.253552ms 383.391309ms 386.851459ms 389.670609ms 394.031427ms 405.426607ms 418.916476ms 423.582557ms 428.527881ms 432.485424ms 434.290341ms 434.702512ms 437.191241ms 438.540147ms 439.011195ms 440.261183ms 441.875561ms 442.107251ms 443.521062ms 453.549171ms 466.690955ms 483.589495ms 486.431252ms 510.893392ms 512.443396ms 514.045884ms 520.951558ms 544.048989ms 562.686061ms 571.793858ms 596.991912ms 607.072685ms 637.256025ms 657.540814ms 665.823313ms 699.051929ms 700.298727ms 731.611388ms 741.586066ms 743.039834ms 746.108379ms 746.172119ms 746.329454ms 746.378785ms 747.116792ms 747.296959ms 747.535821ms 747.610365ms 747.66056ms 747.721798ms 747.821518ms 747.869033ms 748.091555ms 748.15402ms 748.176977ms 748.448081ms 748.488411ms 748.586102ms 748.614553ms 748.626796ms 748.666435ms 748.729938ms 748.793237ms 748.836015ms 748.838258ms 748.858357ms 748.916509ms 748.941277ms 748.95904ms 749.103554ms 749.120544ms 749.138402ms 749.145907ms 749.146795ms 749.194014ms 749.226253ms 749.287416ms 749.295845ms 749.31538ms 749.346325ms 749.469013ms 749.547835ms 749.566906ms 749.63327ms 749.707081ms 749.710213ms 749.717394ms 749.71911ms 749.741971ms 749.775561ms 749.814621ms 749.862677ms 749.909019ms 749.91922ms 749.941382ms 749.959385ms 749.972907ms 750.027218ms 750.077609ms 750.083959ms 750.117632ms 750.127033ms 750.161137ms 750.184676ms 750.22507ms 750.233063ms 750.322886ms 750.330996ms 750.360782ms 750.384058ms 750.411356ms 750.420558ms 750.49935ms 750.561246ms 750.576879ms 750.578321ms 750.602847ms 750.745316ms 750.846441ms 750.908599ms 750.92495ms 751.020321ms 751.247572ms 751.327994ms 751.389777ms 751.535542ms 751.553736ms 751.608325ms 751.711252ms 751.878489ms 752.135028ms 752.219013ms 752.321294ms 752.406732ms 752.656617ms 752.667581ms 752.696787ms 752.874824ms 752.914833ms 753.179238ms 753.486004ms 753.564559ms 753.646306ms 754.117246ms 754.285679ms 754.331939ms 754.429987ms 755.843493ms 756.498871ms 762.672184ms 799.958753ms 844.081324ms 892.781499ms 933.908882ms 985.272271ms 1.034064557s 1.08090827s 1.129471701s 1.166667247s] +Jun 18 12:12:30.026: INFO: 50 %ile: 748.626796ms +Jun 18 12:12:30.026: INFO: 90 %ile: 753.179238ms +Jun 18 12:12:30.026: INFO: 99 %ile: 1.129471701s +Jun 18 12:12:30.026: INFO: Total sample count: 200 +[AfterEach] [sig-network] Service endpoints latency /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 19:26:06.989: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-projected-ltvkg" for this suite. -May 13 19:26:31.055: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 19:26:31.148: INFO: namespace: e2e-tests-projected-ltvkg, resource: bindings, ignored listing per whitelist -May 13 19:26:31.416: INFO: namespace e2e-tests-projected-ltvkg deletion completed in 24.395774311s +Jun 18 12:12:30.027: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "e2e-tests-svc-latency-2s59d" for this suite. +Jun 18 12:12:56.118: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:12:56.460: INFO: namespace: e2e-tests-svc-latency-2s59d, resource: bindings, ignored listing per whitelist +Jun 18 12:12:56.661: INFO: namespace e2e-tests-svc-latency-2s59d deletion completed in 26.620577396s -• [SLOW TEST:30.961 seconds] -[sig-storage] Projected secret -/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_secret.go:34 - optional updates should be reflected in volume [NodeConformance] [Conformance] +• [SLOW TEST:37.899 seconds] +[sig-network] Service endpoints latency +/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:22 + should not be very high [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ -[sig-storage] Projected secret - should be consumable from pods in volume as non-root with defaultMode and fsGroup set [NodeConformance] [Conformance] +SSSSSSSSSSS +------------------------------ +[sig-network] Services + should provide secure master service [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -[BeforeEach] [sig-storage] Projected secret +[BeforeEach] [sig-network] Services /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client -May 13 19:26:31.416: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -STEP: Building a namespace api object, basename projected -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-projected-pbwqd +Jun 18 12:12:56.664: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 +STEP: Building a namespace api object, basename services +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-services-kr6vw STEP: Waiting for a default service account to be provisioned in namespace -[It] should be consumable from pods in volume as non-root with defaultMode and fsGroup set [NodeConformance] [Conformance] +[BeforeEach] [sig-network] Services + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:85 +[It] should provide secure master service [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -STEP: Creating projection with secret that has name projected-secret-test-0358efbb-75b5-11e9-8f67-2632f168be36 -STEP: Creating a pod to test consume secrets -May 13 19:26:31.746: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-035a6934-75b5-11e9-8f67-2632f168be36" in namespace "e2e-tests-projected-pbwqd" to be "success or failure" -May 13 19:26:31.757: INFO: Pod "pod-projected-secrets-035a6934-75b5-11e9-8f67-2632f168be36": Phase="Pending", Reason="", readiness=false. Elapsed: 10.657002ms -May 13 19:26:33.766: INFO: Pod "pod-projected-secrets-035a6934-75b5-11e9-8f67-2632f168be36": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.018988086s -STEP: Saw pod success -May 13 19:26:33.766: INFO: Pod "pod-projected-secrets-035a6934-75b5-11e9-8f67-2632f168be36" satisfied condition "success or failure" -May 13 19:26:33.773: INFO: Trying to get logs from node 10.170.219.153 pod pod-projected-secrets-035a6934-75b5-11e9-8f67-2632f168be36 container projected-secret-volume-test: -STEP: delete the pod -May 13 19:26:33.919: INFO: Waiting for pod pod-projected-secrets-035a6934-75b5-11e9-8f67-2632f168be36 to disappear -May 13 19:26:33.930: INFO: Pod pod-projected-secrets-035a6934-75b5-11e9-8f67-2632f168be36 no longer exists -[AfterEach] [sig-storage] Projected secret +[AfterEach] [sig-network] Services /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 19:26:33.930: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-projected-pbwqd" for this suite. -May 13 19:26:39.969: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 19:26:40.108: INFO: namespace: e2e-tests-projected-pbwqd, resource: bindings, ignored listing per whitelist -May 13 19:26:40.398: INFO: namespace e2e-tests-projected-pbwqd deletion completed in 6.459401889s +Jun 18 12:12:57.149: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "e2e-tests-services-kr6vw" for this suite. +Jun 18 12:13:03.213: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:13:04.945: INFO: namespace: e2e-tests-services-kr6vw, resource: bindings, ignored listing per whitelist +Jun 18 12:13:05.083: INFO: namespace e2e-tests-services-kr6vw deletion completed in 7.919125189s +[AfterEach] [sig-network] Services + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:90 -• [SLOW TEST:8.982 seconds] -[sig-storage] Projected secret -/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_secret.go:34 - should be consumable from pods in volume as non-root with defaultMode and fsGroup set [NodeConformance] [Conformance] +• [SLOW TEST:8.420 seconds] +[sig-network] Services +/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:22 + should provide secure master service [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ -SSS +SSSS ------------------------------ -[k8s.io] Kubelet when scheduling a busybox Pod with hostAliases - should write entries to /etc/hosts [NodeConformance] [Conformance] +[sig-network] Proxy version v1 + should proxy through a service and a pod [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -[BeforeEach] [k8s.io] Kubelet +[BeforeEach] version v1 /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client -May 13 19:26:40.398: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -STEP: Building a namespace api object, basename kubelet-test -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-kubelet-test-k454t +Jun 18 12:13:05.085: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 +STEP: Building a namespace api object, basename proxy +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-proxy-k27sq STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [k8s.io] Kubelet - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:37 -[It] should write entries to /etc/hosts [NodeConformance] [Conformance] +[It] should proxy through a service and a pod [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -[AfterEach] [k8s.io] Kubelet +STEP: starting an echo server on multiple ports +STEP: creating replication controller proxy-service-plvhh in namespace e2e-tests-proxy-k27sq +I0618 12:13:05.603606 17 runners.go:184] Created replication controller with name: proxy-service-plvhh, namespace: e2e-tests-proxy-k27sq, replica count: 1 +I0618 12:13:06.654052 17 runners.go:184] proxy-service-plvhh Pods: 1 out of 1 created, 0 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady +I0618 12:13:07.654344 17 runners.go:184] proxy-service-plvhh Pods: 1 out of 1 created, 0 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady +I0618 12:13:08.654567 17 runners.go:184] proxy-service-plvhh Pods: 1 out of 1 created, 0 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady +I0618 12:13:09.654805 17 runners.go:184] proxy-service-plvhh Pods: 1 out of 1 created, 0 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady +I0618 12:13:10.655108 17 runners.go:184] proxy-service-plvhh Pods: 1 out of 1 created, 0 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 1 runningButNotReady +I0618 12:13:11.655471 17 runners.go:184] proxy-service-plvhh Pods: 1 out of 1 created, 0 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 1 runningButNotReady +I0618 12:13:12.655742 17 runners.go:184] proxy-service-plvhh Pods: 1 out of 1 created, 0 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 1 runningButNotReady +I0618 12:13:13.656041 17 runners.go:184] proxy-service-plvhh Pods: 1 out of 1 created, 0 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 1 runningButNotReady +I0618 12:13:14.656289 17 runners.go:184] proxy-service-plvhh Pods: 1 out of 1 created, 0 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 1 runningButNotReady +I0618 12:13:15.656602 17 runners.go:184] proxy-service-plvhh Pods: 1 out of 1 created, 0 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 1 runningButNotReady +I0618 12:13:16.656832 17 runners.go:184] proxy-service-plvhh Pods: 1 out of 1 created, 0 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 1 runningButNotReady +I0618 12:13:17.657047 17 runners.go:184] proxy-service-plvhh Pods: 1 out of 1 created, 0 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 1 runningButNotReady +I0618 12:13:18.657259 17 runners.go:184] proxy-service-plvhh Pods: 1 out of 1 created, 0 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 1 runningButNotReady +I0618 12:13:19.657453 17 runners.go:184] proxy-service-plvhh Pods: 1 out of 1 created, 1 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady +Jun 18 12:13:19.698: INFO: setup took 14.143300995s, starting test cases +STEP: running 16 cases, 20 attempts per case, 320 total attempts +Jun 18 12:13:19.736: INFO: (0) /api/v1/namespaces/e2e-tests-proxy-k27sq/pods/proxy-service-plvhh-h9x46:1080/proxy/: >> kubeConfig: /tmp/kubeconfig-555138423 -STEP: Building a namespace api object, basename secrets -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-secrets-mlsk9 +Jun 18 12:13:34.683: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 +STEP: Building a namespace api object, basename projected +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-projected-2gfx4 STEP: Waiting for a default service account to be provisioned in namespace -[It] should be consumable from pods in volume as non-root with defaultMode and fsGroup set [NodeConformance] [Conformance] +[BeforeEach] [sig-storage] Projected downwardAPI + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:39 +[It] should provide podname only [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -STEP: Creating secret with name secret-test-2230bb2b-75b5-11e9-8f67-2632f168be36 -STEP: Creating a pod to test consume secrets -May 13 19:27:23.496: INFO: Waiting up to 5m0s for pod "pod-secrets-2233238d-75b5-11e9-8f67-2632f168be36" in namespace "e2e-tests-secrets-mlsk9" to be "success or failure" -May 13 19:27:23.504: INFO: Pod "pod-secrets-2233238d-75b5-11e9-8f67-2632f168be36": Phase="Pending", Reason="", readiness=false. Elapsed: 7.176266ms -May 13 19:27:25.513: INFO: Pod "pod-secrets-2233238d-75b5-11e9-8f67-2632f168be36": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.016238406s +STEP: Creating a pod to test downward API volume plugin +Jun 18 12:13:35.210: INFO: Waiting up to 5m0s for pod "downwardapi-volume-7ef69536-91c2-11e9-bce2-ae54e022189f" in namespace "e2e-tests-projected-2gfx4" to be "success or failure" +Jun 18 12:13:35.225: INFO: Pod "downwardapi-volume-7ef69536-91c2-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 15.145091ms +Jun 18 12:13:37.241: INFO: Pod "downwardapi-volume-7ef69536-91c2-11e9-bce2-ae54e022189f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.030458474s STEP: Saw pod success -May 13 19:27:25.513: INFO: Pod "pod-secrets-2233238d-75b5-11e9-8f67-2632f168be36" satisfied condition "success or failure" -May 13 19:27:25.520: INFO: Trying to get logs from node 10.170.219.153 pod pod-secrets-2233238d-75b5-11e9-8f67-2632f168be36 container secret-volume-test: +Jun 18 12:13:37.241: INFO: Pod "downwardapi-volume-7ef69536-91c2-11e9-bce2-ae54e022189f" satisfied condition "success or failure" +Jun 18 12:13:37.255: INFO: Trying to get logs from node 10.72.74.144 pod downwardapi-volume-7ef69536-91c2-11e9-bce2-ae54e022189f container client-container: STEP: delete the pod -May 13 19:27:25.820: INFO: Waiting for pod pod-secrets-2233238d-75b5-11e9-8f67-2632f168be36 to disappear -May 13 19:27:25.827: INFO: Pod pod-secrets-2233238d-75b5-11e9-8f67-2632f168be36 no longer exists -[AfterEach] [sig-storage] Secrets +Jun 18 12:13:37.346: INFO: Waiting for pod downwardapi-volume-7ef69536-91c2-11e9-bce2-ae54e022189f to disappear +Jun 18 12:13:37.361: INFO: Pod downwardapi-volume-7ef69536-91c2-11e9-bce2-ae54e022189f no longer exists +[AfterEach] [sig-storage] Projected downwardAPI /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 19:27:25.827: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-secrets-mlsk9" for this suite. -May 13 19:27:31.866: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 19:27:31.921: INFO: namespace: e2e-tests-secrets-mlsk9, resource: bindings, ignored listing per whitelist -May 13 19:27:32.152: INFO: namespace e2e-tests-secrets-mlsk9 deletion completed in 6.316469901s +Jun 18 12:13:37.361: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "e2e-tests-projected-2gfx4" for this suite. +Jun 18 12:13:43.518: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:13:44.605: INFO: namespace: e2e-tests-projected-2gfx4, resource: bindings, ignored listing per whitelist +Jun 18 12:13:44.942: INFO: namespace e2e-tests-projected-2gfx4 deletion completed in 7.561471216s -• [SLOW TEST:9.032 seconds] -[sig-storage] Secrets -/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/secrets_volume.go:34 - should be consumable from pods in volume as non-root with defaultMode and fsGroup set [NodeConformance] [Conformance] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------- -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------- -[sig-cli] Kubectl client [k8s.io] Kubectl version - should check is all data is printed [Conformance] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -[BeforeEach] [sig-cli] Kubectl client - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 -STEP: Creating a kubernetes client -May 13 19:27:32.154: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -STEP: Building a namespace api object, basename kubectl -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-kubectl-r9jsz -STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [sig-cli] Kubectl client - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:243 -[It] should check is all data is printed [Conformance] +• [SLOW TEST:10.259 seconds] +[sig-storage] Projected downwardAPI +/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:33 + should provide podname only [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -May 13 19:27:32.418: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 version' -May 13 19:27:32.508: INFO: stderr: "" -May 13 19:27:32.508: INFO: stdout: "Client Version: version.Info{Major:\"1\", Minor:\"13\", GitVersion:\"v1.13.0\", GitCommit:\"ddf47ac13c1a9483ea035a79cd7c10005ff21a6d\", GitTreeState:\"clean\", BuildDate:\"2018-12-03T21:04:45Z\", GoVersion:\"go1.11.2\", Compiler:\"gc\", Platform:\"linux/amd64\"}\nServer Version: version.Info{Major:\"1\", Minor:\"13\", GitVersion:\"v1.13.6+IKS\", GitCommit:\"ac5f7341d5d0ce8ea8f206ba5b030dc9e9d4cc97\", GitTreeState:\"clean\", BuildDate:\"2019-05-09T13:26:51Z\", GoVersion:\"go1.11.5\", Compiler:\"gc\", Platform:\"linux/amd64\"}\n" -[AfterEach] [sig-cli] Kubectl client - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 19:27:32.508: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-kubectl-r9jsz" for this suite. -May 13 19:27:38.548: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 19:27:38.640: INFO: namespace: e2e-tests-kubectl-r9jsz, resource: bindings, ignored listing per whitelist -May 13 19:27:38.871: INFO: namespace e2e-tests-kubectl-r9jsz deletion completed in 6.355745814s - -• [SLOW TEST:6.717 seconds] -[sig-cli] Kubectl client -/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:22 - [k8s.io] Kubectl version - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 - should check is all data is printed [Conformance] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ -SSSSS +SS ------------------------------ -[k8s.io] Probing container - should have monotonically increasing restart count [Slow][NodeConformance] [Conformance] +[sig-storage] Projected downwardAPI + should set DefaultMode on files [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -[BeforeEach] [k8s.io] Probing container +[BeforeEach] [sig-storage] Projected downwardAPI /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client -May 13 19:27:38.871: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -STEP: Building a namespace api object, basename container-probe -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-container-probe-8s6zb +Jun 18 12:13:44.942: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 +STEP: Building a namespace api object, basename projected +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-projected-xxmfj STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [k8s.io] Probing container - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/container_probe.go:48 -[It] should have monotonically increasing restart count [Slow][NodeConformance] [Conformance] +[BeforeEach] [sig-storage] Projected downwardAPI + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:39 +[It] should set DefaultMode on files [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -STEP: Creating pod liveness-http in namespace e2e-tests-container-probe-8s6zb -May 13 19:27:41.171: INFO: Started pod liveness-http in namespace e2e-tests-container-probe-8s6zb -STEP: checking the pod's current state and verifying that restartCount is present -May 13 19:27:41.179: INFO: Initial restart count of pod liveness-http is 0 -May 13 19:27:55.273: INFO: Restart count of pod e2e-tests-container-probe-8s6zb/liveness-http is now 1 (14.093476926s elapsed) -May 13 19:28:15.386: INFO: Restart count of pod e2e-tests-container-probe-8s6zb/liveness-http is now 2 (34.207046824s elapsed) -May 13 19:28:35.962: INFO: Restart count of pod e2e-tests-container-probe-8s6zb/liveness-http is now 3 (54.782986412s elapsed) -May 13 19:28:56.073: INFO: Restart count of pod e2e-tests-container-probe-8s6zb/liveness-http is now 4 (1m14.893546281s elapsed) -May 13 19:29:54.982: INFO: Restart count of pod e2e-tests-container-probe-8s6zb/liveness-http is now 5 (2m13.80255479s elapsed) -STEP: deleting the pod -[AfterEach] [k8s.io] Probing container +STEP: Creating a pod to test downward API volume plugin +Jun 18 12:13:45.511: INFO: Waiting up to 5m0s for pod "downwardapi-volume-8522dd8b-91c2-11e9-bce2-ae54e022189f" in namespace "e2e-tests-projected-xxmfj" to be "success or failure" +Jun 18 12:13:45.528: INFO: Pod "downwardapi-volume-8522dd8b-91c2-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 16.788664ms +Jun 18 12:13:47.543: INFO: Pod "downwardapi-volume-8522dd8b-91c2-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 2.031190614s +Jun 18 12:13:49.560: INFO: Pod "downwardapi-volume-8522dd8b-91c2-11e9-bce2-ae54e022189f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.048455776s +STEP: Saw pod success +Jun 18 12:13:49.560: INFO: Pod "downwardapi-volume-8522dd8b-91c2-11e9-bce2-ae54e022189f" satisfied condition "success or failure" +Jun 18 12:13:49.574: INFO: Trying to get logs from node 10.72.74.143 pod downwardapi-volume-8522dd8b-91c2-11e9-bce2-ae54e022189f container client-container: +STEP: delete the pod +Jun 18 12:13:49.665: INFO: Waiting for pod downwardapi-volume-8522dd8b-91c2-11e9-bce2-ae54e022189f to disappear +Jun 18 12:13:49.679: INFO: Pod downwardapi-volume-8522dd8b-91c2-11e9-bce2-ae54e022189f no longer exists +[AfterEach] [sig-storage] Projected downwardAPI /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 19:29:55.006: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-container-probe-8s6zb" for this suite. -May 13 19:30:01.045: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 19:30:01.294: INFO: namespace: e2e-tests-container-probe-8s6zb, resource: bindings, ignored listing per whitelist -May 13 19:30:01.342: INFO: namespace e2e-tests-container-probe-8s6zb deletion completed in 6.324745553s +Jun 18 12:13:49.679: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "e2e-tests-projected-xxmfj" for this suite. +Jun 18 12:13:55.747: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:13:57.341: INFO: namespace: e2e-tests-projected-xxmfj, resource: bindings, ignored listing per whitelist +Jun 18 12:13:57.423: INFO: namespace e2e-tests-projected-xxmfj deletion completed in 7.72432526s -• [SLOW TEST:142.471 seconds] -[k8s.io] Probing container -/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 - should have monotonically increasing restart count [Slow][NodeConformance] [Conformance] +• [SLOW TEST:12.481 seconds] +[sig-storage] Projected downwardAPI +/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:33 + should set DefaultMode on files [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------- [sig-storage] EmptyDir volumes - should support (root,0666,tmpfs) [NodeConformance] [Conformance] + should support (root,0777,default) [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-storage] EmptyDir volumes /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client -May 13 19:30:01.344: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 +Jun 18 12:13:57.423: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 STEP: Building a namespace api object, basename emptydir -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-emptydir-trswb +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-emptydir-8tqhz STEP: Waiting for a default service account to be provisioned in namespace -[It] should support (root,0666,tmpfs) [NodeConformance] [Conformance] +[It] should support (root,0777,default) [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -STEP: Creating a pod to test emptydir 0666 on tmpfs -May 13 19:30:01.737: INFO: Waiting up to 5m0s for pod "pod-80846756-75b5-11e9-8f67-2632f168be36" in namespace "e2e-tests-emptydir-trswb" to be "success or failure" -May 13 19:30:01.744: INFO: Pod "pod-80846756-75b5-11e9-8f67-2632f168be36": Phase="Pending", Reason="", readiness=false. Elapsed: 7.796949ms -May 13 19:30:03.753: INFO: Pod "pod-80846756-75b5-11e9-8f67-2632f168be36": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.016090026s +STEP: Creating a pod to test emptydir 0777 on node default medium +Jun 18 12:13:57.906: INFO: Waiting up to 5m0s for pod "pod-8c86065a-91c2-11e9-bce2-ae54e022189f" in namespace "e2e-tests-emptydir-8tqhz" to be "success or failure" +Jun 18 12:13:57.919: INFO: Pod "pod-8c86065a-91c2-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 13.546728ms +Jun 18 12:13:59.935: INFO: Pod "pod-8c86065a-91c2-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 2.029018846s +Jun 18 12:14:01.968: INFO: Pod "pod-8c86065a-91c2-11e9-bce2-ae54e022189f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.062227719s STEP: Saw pod success -May 13 19:30:03.753: INFO: Pod "pod-80846756-75b5-11e9-8f67-2632f168be36" satisfied condition "success or failure" -May 13 19:30:03.760: INFO: Trying to get logs from node 10.170.219.153 pod pod-80846756-75b5-11e9-8f67-2632f168be36 container test-container: +Jun 18 12:14:01.968: INFO: Pod "pod-8c86065a-91c2-11e9-bce2-ae54e022189f" satisfied condition "success or failure" +Jun 18 12:14:01.985: INFO: Trying to get logs from node 10.72.74.143 pod pod-8c86065a-91c2-11e9-bce2-ae54e022189f container test-container: STEP: delete the pod -May 13 19:30:03.880: INFO: Waiting for pod pod-80846756-75b5-11e9-8f67-2632f168be36 to disappear -May 13 19:30:03.887: INFO: Pod pod-80846756-75b5-11e9-8f67-2632f168be36 no longer exists +Jun 18 12:14:02.058: INFO: Waiting for pod pod-8c86065a-91c2-11e9-bce2-ae54e022189f to disappear +Jun 18 12:14:02.072: INFO: Pod pod-8c86065a-91c2-11e9-bce2-ae54e022189f no longer exists [AfterEach] [sig-storage] EmptyDir volumes /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 19:30:03.887: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-emptydir-trswb" for this suite. -May 13 19:30:09.923: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 19:30:10.189: INFO: namespace: e2e-tests-emptydir-trswb, resource: bindings, ignored listing per whitelist -May 13 19:30:10.206: INFO: namespace e2e-tests-emptydir-trswb deletion completed in 6.311474267s +Jun 18 12:14:02.073: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "e2e-tests-emptydir-8tqhz" for this suite. +Jun 18 12:14:10.145: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:14:10.326: INFO: namespace: e2e-tests-emptydir-8tqhz, resource: bindings, ignored listing per whitelist +Jun 18 12:14:10.657: INFO: namespace e2e-tests-emptydir-8tqhz deletion completed in 8.560759941s -• [SLOW TEST:8.862 seconds] +• [SLOW TEST:13.234 seconds] [sig-storage] EmptyDir volumes /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:40 - should support (root,0666,tmpfs) [NodeConformance] [Conformance] + should support (root,0777,default) [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ -SS +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-scheduling] SchedulerPredicates [Serial] - validates that NodeSelector is respected if matching [Conformance] +[sig-cli] Kubectl client [k8s.io] Kubectl replace + should update a single-container pod's image [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -[BeforeEach] [sig-scheduling] SchedulerPredicates [Serial] +[BeforeEach] [sig-cli] Kubectl client /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client -May 13 19:30:10.206: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -STEP: Building a namespace api object, basename sched-pred -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-sched-pred-5d84j +Jun 18 12:14:10.660: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 +STEP: Building a namespace api object, basename kubectl +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-kubectl-xv9w4 STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [sig-scheduling] SchedulerPredicates [Serial] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:79 -May 13 19:30:10.464: INFO: Waiting up to 1m0s for all (but 0) nodes to be ready -May 13 19:30:10.478: INFO: Waiting for terminating namespaces to be deleted... -May 13 19:30:10.485: INFO: -Logging pods the kubelet thinks is on node 10.170.219.151 before test -May 13 19:30:10.537: INFO: public-cr0ef1a5f1e33c44fe9c21650356d908ee-alb1-8f59cf484-9v5pm from kube-system started at 2019-05-13 16:35:03 +0000 UTC (4 container statuses recorded) -May 13 19:30:10.537: INFO: Container ingress-auth-1 ready: true, restart count 0 -May 13 19:30:10.537: INFO: Container ingress-auth-2 ready: true, restart count 0 -May 13 19:30:10.537: INFO: Container ingress-auth-3 ready: true, restart count 0 -May 13 19:30:10.537: INFO: Container nginx-ingress ready: true, restart count 0 -May 13 19:30:10.537: INFO: ibm-master-proxy-static-10.170.219.151 from kube-system started at (0 container statuses recorded) -May 13 19:30:10.537: INFO: calico-kube-controllers-7dd978d898-r8w74 from kube-system started at 2019-05-13 16:27:35 +0000 UTC (1 container statuses recorded) -May 13 19:30:10.537: INFO: Container calico-kube-controllers ready: true, restart count 0 -May 13 19:30:10.537: INFO: ibm-keepalived-watcher-8wshf from kube-system started at 2019-05-13 16:27:05 +0000 UTC (1 container statuses recorded) -May 13 19:30:10.537: INFO: Container keepalived-watcher ready: true, restart count 0 -May 13 19:30:10.537: INFO: coredns-58d696879-rrkth from kube-system started at 2019-05-13 16:27:35 +0000 UTC (1 container statuses recorded) -May 13 19:30:10.537: INFO: Container coredns ready: true, restart count 0 -May 13 19:30:10.537: INFO: ibm-file-plugin-7f6d8979bd-p5smn from kube-system started at 2019-05-13 16:27:35 +0000 UTC (1 container statuses recorded) -May 13 19:30:10.537: INFO: Container ibm-file-plugin-container ready: true, restart count 0 -May 13 19:30:10.537: INFO: coredns-58d696879-zmj2g from kube-system started at 2019-05-13 16:27:49 +0000 UTC (1 container statuses recorded) -May 13 19:30:10.537: INFO: Container coredns ready: true, restart count 0 -May 13 19:30:10.537: INFO: calico-node-d9d6s from kube-system started at 2019-05-13 16:27:05 +0000 UTC (1 container statuses recorded) -May 13 19:30:10.537: INFO: Container calico-node ready: true, restart count 0 -May 13 19:30:10.537: INFO: vpn-774cf5c6d4-6p6cr from kube-system started at 2019-05-13 16:27:35 +0000 UTC (1 container statuses recorded) -May 13 19:30:10.537: INFO: Container vpn ready: true, restart count 0 -May 13 19:30:10.537: INFO: kubernetes-dashboard-7996b848f4-mzbwj from kube-system started at 2019-05-13 16:27:35 +0000 UTC (1 container statuses recorded) -May 13 19:30:10.537: INFO: Container kubernetes-dashboard ready: true, restart count 0 -May 13 19:30:10.537: INFO: ibm-kube-fluentd-6fw9c from kube-system started at 2019-05-13 16:32:50 +0000 UTC (1 container statuses recorded) -May 13 19:30:10.537: INFO: Container fluentd ready: true, restart count 0 -May 13 19:30:10.537: INFO: sonobuoy-systemd-logs-daemon-set-8e5ce9bdf4a94ece-nlv42 from heptio-sonobuoy started at 2019-05-13 18:40:58 +0000 UTC (2 container statuses recorded) -May 13 19:30:10.537: INFO: Container sonobuoy-worker ready: true, restart count 0 -May 13 19:30:10.537: INFO: Container systemd-logs ready: true, restart count 0 -May 13 19:30:10.537: INFO: ibm-storage-watcher-845946d5b5-knxcm from kube-system started at 2019-05-13 16:27:35 +0000 UTC (1 container statuses recorded) -May 13 19:30:10.537: INFO: Container ibm-storage-watcher-container ready: true, restart count 0 -May 13 19:30:10.537: INFO: coredns-autoscaler-64f9c5b4df-hzq57 from kube-system started at 2019-05-13 16:27:35 +0000 UTC (1 container statuses recorded) -May 13 19:30:10.537: INFO: Container autoscaler ready: true, restart count 0 -May 13 19:30:10.537: INFO: -Logging pods the kubelet thinks is on node 10.170.219.153 before test -May 13 19:30:10.591: INFO: calico-node-phw8q from kube-system started at 2019-05-13 16:27:29 +0000 UTC (1 container statuses recorded) -May 13 19:30:10.591: INFO: Container calico-node ready: true, restart count 0 -May 13 19:30:10.591: INFO: ibm-cloud-provider-ip-169-45-218-130-66c489dcb-z58z4 from ibm-system started at 2019-05-13 16:30:08 +0000 UTC (1 container statuses recorded) -May 13 19:30:10.591: INFO: Container ibm-cloud-provider-ip-169-45-218-130 ready: true, restart count 0 -May 13 19:30:10.591: INFO: ibm-kube-fluentd-cdmr5 from kube-system started at 2019-05-13 16:32:50 +0000 UTC (1 container statuses recorded) -May 13 19:30:10.591: INFO: Container fluentd ready: true, restart count 0 -May 13 19:30:10.591: INFO: sonobuoy from heptio-sonobuoy started at 2019-05-13 18:40:57 +0000 UTC (1 container statuses recorded) -May 13 19:30:10.591: INFO: Container kube-sonobuoy ready: true, restart count 0 -May 13 19:30:10.591: INFO: metrics-server-58dd96c6bd-sggps from kube-system started at 2019-05-13 16:28:04 +0000 UTC (2 container statuses recorded) -May 13 19:30:10.591: INFO: Container metrics-server ready: true, restart count 0 -May 13 19:30:10.591: INFO: Container metrics-server-nanny ready: true, restart count 0 -May 13 19:30:10.591: INFO: test-k8s-e2e-pvg-master-verification from default started at 2019-05-13 18:40:52 +0000 UTC (1 container statuses recorded) -May 13 19:30:10.591: INFO: Container test-k8s-e2e-pvg-master-verification ready: true, restart count 0 -May 13 19:30:10.591: INFO: sonobuoy-systemd-logs-daemon-set-8e5ce9bdf4a94ece-tgp2l from heptio-sonobuoy started at 2019-05-13 18:40:58 +0000 UTC (2 container statuses recorded) -May 13 19:30:10.591: INFO: Container sonobuoy-worker ready: true, restart count 0 -May 13 19:30:10.591: INFO: Container systemd-logs ready: true, restart count 0 -May 13 19:30:10.591: INFO: ibm-master-proxy-static-10.170.219.153 from kube-system started at (0 container statuses recorded) -May 13 19:30:10.591: INFO: ibm-keepalived-watcher-cwtfg from kube-system started at 2019-05-13 16:27:29 +0000 UTC (1 container statuses recorded) -May 13 19:30:10.591: INFO: Container keepalived-watcher ready: true, restart count 0 -May 13 19:30:10.591: INFO: -Logging pods the kubelet thinks is on node 10.170.219.177 before test -May 13 19:30:10.613: INFO: ibm-keepalived-watcher-qszvz from kube-system started at 2019-05-13 16:27:34 +0000 UTC (1 container statuses recorded) -May 13 19:30:10.613: INFO: Container keepalived-watcher ready: true, restart count 0 -May 13 19:30:10.613: INFO: calico-node-twx7l from kube-system started at 2019-05-13 16:27:34 +0000 UTC (1 container statuses recorded) -May 13 19:30:10.613: INFO: Container calico-node ready: true, restart count 0 -May 13 19:30:10.613: INFO: ibm-master-proxy-static-10.170.219.177 from kube-system started at (0 container statuses recorded) -May 13 19:30:10.613: INFO: public-cr0ef1a5f1e33c44fe9c21650356d908ee-alb1-8f59cf484-7r5vq from kube-system started at 2019-05-13 16:35:03 +0000 UTC (4 container statuses recorded) -May 13 19:30:10.613: INFO: Container ingress-auth-1 ready: true, restart count 0 -May 13 19:30:10.613: INFO: Container ingress-auth-2 ready: true, restart count 0 -May 13 19:30:10.613: INFO: Container ingress-auth-3 ready: true, restart count 0 -May 13 19:30:10.613: INFO: Container nginx-ingress ready: true, restart count 0 -May 13 19:30:10.613: INFO: ibm-cloud-provider-ip-169-45-218-130-66c489dcb-kd74s from ibm-system started at 2019-05-13 16:30:08 +0000 UTC (1 container statuses recorded) -May 13 19:30:10.613: INFO: Container ibm-cloud-provider-ip-169-45-218-130 ready: true, restart count 0 -May 13 19:30:10.613: INFO: sonobuoy-e2e-job-afa7e83446444f4e from heptio-sonobuoy started at 2019-05-13 18:40:58 +0000 UTC (2 container statuses recorded) -May 13 19:30:10.613: INFO: Container e2e ready: true, restart count 0 -May 13 19:30:10.613: INFO: Container sonobuoy-worker ready: true, restart count 0 -May 13 19:30:10.613: INFO: ibm-kube-fluentd-zzl5d from kube-system started at 2019-05-13 16:32:50 +0000 UTC (1 container statuses recorded) -May 13 19:30:10.613: INFO: Container fluentd ready: true, restart count 0 -May 13 19:30:10.613: INFO: sonobuoy-systemd-logs-daemon-set-8e5ce9bdf4a94ece-kjc69 from heptio-sonobuoy started at 2019-05-13 18:40:58 +0000 UTC (2 container statuses recorded) -May 13 19:30:10.613: INFO: Container sonobuoy-worker ready: true, restart count 0 -May 13 19:30:10.613: INFO: Container systemd-logs ready: true, restart count 0 -[It] validates that NodeSelector is respected if matching [Conformance] +[BeforeEach] [sig-cli] Kubectl client + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:243 +[BeforeEach] [k8s.io] Kubectl replace + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1563 +[It] should update a single-container pod's image [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -STEP: Trying to launch a pod without a label to get a node which can launch it. -STEP: Explicitly delete pod here to free the resource it takes. -STEP: Trying to apply a random label on the found node. -STEP: verifying the node has the label kubernetes.io/e2e-870c4b7d-75b5-11e9-8f67-2632f168be36 42 -STEP: Trying to relaunch the pod, now with labels. -STEP: removing the label kubernetes.io/e2e-870c4b7d-75b5-11e9-8f67-2632f168be36 off the node 10.170.219.177 -STEP: verifying the node doesn't have the label kubernetes.io/e2e-870c4b7d-75b5-11e9-8f67-2632f168be36 -[AfterEach] [sig-scheduling] SchedulerPredicates [Serial] +STEP: running the image docker.io/library/nginx:1.14-alpine +Jun 18 12:14:11.119: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 run e2e-test-nginx-pod --generator=run-pod/v1 --image=docker.io/library/nginx:1.14-alpine --labels=run=e2e-test-nginx-pod --namespace=e2e-tests-kubectl-xv9w4' +Jun 18 12:14:11.275: INFO: stderr: "" +Jun 18 12:14:11.275: INFO: stdout: "pod/e2e-test-nginx-pod created\n" +STEP: verifying the pod e2e-test-nginx-pod is running +STEP: verifying the pod e2e-test-nginx-pod was created +Jun 18 12:14:16.325: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 get pod e2e-test-nginx-pod --namespace=e2e-tests-kubectl-xv9w4 -o json' +Jun 18 12:14:16.453: INFO: stderr: "" +Jun 18 12:14:16.453: INFO: stdout: "{\n \"apiVersion\": \"v1\",\n \"kind\": \"Pod\",\n \"metadata\": {\n \"annotations\": {\n \"kubernetes.io/psp\": \"e2e-test-privileged-psp\"\n },\n \"creationTimestamp\": \"2019-06-18T12:14:11Z\",\n \"labels\": {\n \"run\": \"e2e-test-nginx-pod\"\n },\n \"name\": \"e2e-test-nginx-pod\",\n \"namespace\": \"e2e-tests-kubectl-xv9w4\",\n \"resourceVersion\": \"98337\",\n \"selfLink\": \"/api/v1/namespaces/e2e-tests-kubectl-xv9w4/pods/e2e-test-nginx-pod\",\n \"uid\": \"947f4be7-91c2-11e9-bf44-fa6f350b29f0\"\n },\n \"spec\": {\n \"containers\": [\n {\n \"image\": \"docker.io/library/nginx:1.14-alpine\",\n \"imagePullPolicy\": \"IfNotPresent\",\n \"name\": \"e2e-test-nginx-pod\",\n \"resources\": {},\n \"terminationMessagePath\": \"/dev/termination-log\",\n \"terminationMessagePolicy\": \"File\",\n \"volumeMounts\": [\n {\n \"mountPath\": \"/var/run/secrets/kubernetes.io/serviceaccount\",\n \"name\": \"default-token-q2jrc\",\n \"readOnly\": true\n }\n ]\n }\n ],\n \"dnsPolicy\": \"ClusterFirst\",\n \"enableServiceLinks\": true,\n \"nodeName\": \"10.72.74.144\",\n \"priority\": 0,\n \"restartPolicy\": \"Always\",\n \"schedulerName\": \"default-scheduler\",\n \"securityContext\": {},\n \"serviceAccount\": \"default\",\n \"serviceAccountName\": \"default\",\n \"terminationGracePeriodSeconds\": 30,\n \"tolerations\": [\n {\n \"effect\": \"NoExecute\",\n \"key\": \"node.kubernetes.io/not-ready\",\n \"operator\": \"Exists\",\n \"tolerationSeconds\": 300\n },\n {\n \"effect\": \"NoExecute\",\n \"key\": \"node.kubernetes.io/unreachable\",\n \"operator\": \"Exists\",\n \"tolerationSeconds\": 300\n }\n ],\n \"volumes\": [\n {\n \"name\": \"default-token-q2jrc\",\n \"secret\": {\n \"defaultMode\": 420,\n \"secretName\": \"default-token-q2jrc\"\n }\n }\n ]\n },\n \"status\": {\n \"conditions\": [\n {\n \"lastProbeTime\": null,\n \"lastTransitionTime\": \"2019-06-18T12:14:11Z\",\n \"status\": \"True\",\n \"type\": \"Initialized\"\n },\n {\n \"lastProbeTime\": null,\n \"lastTransitionTime\": \"2019-06-18T12:14:12Z\",\n \"status\": \"True\",\n \"type\": \"Ready\"\n },\n {\n \"lastProbeTime\": null,\n \"lastTransitionTime\": \"2019-06-18T12:14:12Z\",\n \"status\": \"True\",\n \"type\": \"ContainersReady\"\n },\n {\n \"lastProbeTime\": null,\n \"lastTransitionTime\": \"2019-06-18T12:14:11Z\",\n \"status\": \"True\",\n \"type\": \"PodScheduled\"\n }\n ],\n \"containerStatuses\": [\n {\n \"containerID\": \"containerd://b332dd0683b1ed72440faea31ebea2f8fe19a233b3272929f92d4b672d56473d\",\n \"image\": \"docker.io/library/nginx:1.14-alpine\",\n \"imageID\": \"docker.io/library/nginx@sha256:485b610fefec7ff6c463ced9623314a04ed67e3945b9c08d7e53a47f6d108dc7\",\n \"lastState\": {},\n \"name\": \"e2e-test-nginx-pod\",\n \"ready\": true,\n \"restartCount\": 0,\n \"state\": {\n \"running\": {\n \"startedAt\": \"2019-06-18T12:14:12Z\"\n }\n }\n }\n ],\n \"hostIP\": \"10.72.74.144\",\n \"phase\": \"Running\",\n \"podIP\": \"172.30.114.33\",\n \"qosClass\": \"BestEffort\",\n \"startTime\": \"2019-06-18T12:14:11Z\"\n }\n}\n" +STEP: replace the image in the pod +Jun 18 12:14:16.453: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 replace -f - --namespace=e2e-tests-kubectl-xv9w4' +Jun 18 12:14:17.049: INFO: stderr: "" +Jun 18 12:14:17.049: INFO: stdout: "pod/e2e-test-nginx-pod replaced\n" +STEP: verifying the pod e2e-test-nginx-pod has the right image docker.io/library/busybox:1.29 +[AfterEach] [k8s.io] Kubectl replace + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1568 +Jun 18 12:14:17.082: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 delete pods e2e-test-nginx-pod --namespace=e2e-tests-kubectl-xv9w4' +Jun 18 12:14:19.076: INFO: stderr: "" +Jun 18 12:14:19.076: INFO: stdout: "pod \"e2e-test-nginx-pod\" deleted\n" +[AfterEach] [sig-cli] Kubectl client /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 19:30:14.825: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-sched-pred-5d84j" for this suite. -May 13 19:30:36.860: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 19:30:36.933: INFO: namespace: e2e-tests-sched-pred-5d84j, resource: bindings, ignored listing per whitelist -May 13 19:30:37.153: INFO: namespace e2e-tests-sched-pred-5d84j deletion completed in 22.320040485s -[AfterEach] [sig-scheduling] SchedulerPredicates [Serial] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:70 +Jun 18 12:14:19.076: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "e2e-tests-kubectl-xv9w4" for this suite. +Jun 18 12:14:25.148: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:14:25.608: INFO: namespace: e2e-tests-kubectl-xv9w4, resource: bindings, ignored listing per whitelist +Jun 18 12:14:25.647: INFO: namespace e2e-tests-kubectl-xv9w4 deletion completed in 6.5493873s -• [SLOW TEST:26.947 seconds] -[sig-scheduling] SchedulerPredicates [Serial] -/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/framework.go:22 - validates that NodeSelector is respected if matching [Conformance] +• [SLOW TEST:14.988 seconds] +[sig-cli] Kubectl client +/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:22 + [k8s.io] Kubectl replace + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 + should update a single-container pod's image [Conformance] + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 +------------------------------ +SSSSSSS +------------------------------ +[sig-cli] Kubectl client [k8s.io] Kubectl run default + should create an rc or deployment from an image [Conformance] + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 +[BeforeEach] [sig-cli] Kubectl client + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 +STEP: Creating a kubernetes client +Jun 18 12:14:25.648: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 +STEP: Building a namespace api object, basename kubectl +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-kubectl-6zvks +STEP: Waiting for a default service account to be provisioned in namespace +[BeforeEach] [sig-cli] Kubectl client + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:243 +[BeforeEach] [k8s.io] Kubectl run default + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1262 +[It] should create an rc or deployment from an image [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 +STEP: running the image docker.io/library/nginx:1.14-alpine +Jun 18 12:14:26.191: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 run e2e-test-nginx-deployment --image=docker.io/library/nginx:1.14-alpine --namespace=e2e-tests-kubectl-6zvks' +Jun 18 12:14:26.342: INFO: stderr: "kubectl run --generator=deployment/apps.v1 is DEPRECATED and will be removed in a future version. Use kubectl run --generator=run-pod/v1 or kubectl create instead.\n" +Jun 18 12:14:26.342: INFO: stdout: "deployment.apps/e2e-test-nginx-deployment created\n" +STEP: verifying the pod controlled by e2e-test-nginx-deployment gets created +[AfterEach] [k8s.io] Kubectl run default + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1268 +Jun 18 12:14:28.390: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 delete deployment e2e-test-nginx-deployment --namespace=e2e-tests-kubectl-6zvks' +Jun 18 12:14:28.545: INFO: stderr: "" +Jun 18 12:14:28.546: INFO: stdout: "deployment.extensions \"e2e-test-nginx-deployment\" deleted\n" +[AfterEach] [sig-cli] Kubectl client + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 +Jun 18 12:14:28.546: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "e2e-tests-kubectl-6zvks" for this suite. +Jun 18 12:14:36.661: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:14:36.726: INFO: namespace: e2e-tests-kubectl-6zvks, resource: bindings, ignored listing per whitelist +Jun 18 12:14:37.166: INFO: namespace e2e-tests-kubectl-6zvks deletion completed in 8.560805443s + +• [SLOW TEST:11.518 seconds] +[sig-cli] Kubectl client +/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:22 + [k8s.io] Kubectl run default + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 + should create an rc or deployment from an image [Conformance] + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ -SSSSSSSSS +SSSSSSSS ------------------------------ -[sig-storage] Projected downwardAPI - should update labels on modification [NodeConformance] [Conformance] +[sig-cli] Kubectl client [k8s.io] Kubectl label + should update the label on a resource [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -[BeforeEach] [sig-storage] Projected downwardAPI +[BeforeEach] [sig-cli] Kubectl client /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client -May 13 19:30:37.153: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -STEP: Building a namespace api object, basename projected -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-projected-27hkk +Jun 18 12:14:37.166: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 +STEP: Building a namespace api object, basename kubectl +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-kubectl-fx2mp STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [sig-storage] Projected downwardAPI - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:39 -[It] should update labels on modification [NodeConformance] [Conformance] +[BeforeEach] [sig-cli] Kubectl client + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:243 +[BeforeEach] [k8s.io] Kubectl label + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1052 +STEP: creating the pod +Jun 18 12:14:37.621: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 create -f - --namespace=e2e-tests-kubectl-fx2mp' +Jun 18 12:14:37.903: INFO: stderr: "" +Jun 18 12:14:37.903: INFO: stdout: "pod/pause created\n" +Jun 18 12:14:37.903: INFO: Waiting up to 5m0s for 1 pods to be running and ready: [pause] +Jun 18 12:14:37.903: INFO: Waiting up to 5m0s for pod "pause" in namespace "e2e-tests-kubectl-fx2mp" to be "running and ready" +Jun 18 12:14:37.918: INFO: Pod "pause": Phase="Pending", Reason="", readiness=false. Elapsed: 14.593961ms +Jun 18 12:14:39.950: INFO: Pod "pause": Phase="Pending", Reason="", readiness=false. Elapsed: 2.046181934s +Jun 18 12:14:41.964: INFO: Pod "pause": Phase="Running", Reason="", readiness=true. Elapsed: 4.060668848s +Jun 18 12:14:41.964: INFO: Pod "pause" satisfied condition "running and ready" +Jun 18 12:14:41.964: INFO: Wanted all 1 pods to be running and ready. Result: true. Pods: [pause] +[It] should update the label on a resource [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -STEP: Creating the pod -May 13 19:30:40.021: INFO: Successfully updated pod "labelsupdate95d01c4f-75b5-11e9-8f67-2632f168be36" -[AfterEach] [sig-storage] Projected downwardAPI +STEP: adding the label testing-label with value testing-label-value to a pod +Jun 18 12:14:41.964: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 label pods pause testing-label=testing-label-value --namespace=e2e-tests-kubectl-fx2mp' +Jun 18 12:14:42.204: INFO: stderr: "" +Jun 18 12:14:42.204: INFO: stdout: "pod/pause labeled\n" +STEP: verifying the pod has the label testing-label with the value testing-label-value +Jun 18 12:14:42.204: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 get pod pause -L testing-label --namespace=e2e-tests-kubectl-fx2mp' +Jun 18 12:14:42.339: INFO: stderr: "" +Jun 18 12:14:42.339: INFO: stdout: "NAME READY STATUS RESTARTS AGE TESTING-LABEL\npause 1/1 Running 0 5s testing-label-value\n" +STEP: removing the label testing-label of a pod +Jun 18 12:14:42.339: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 label pods pause testing-label- --namespace=e2e-tests-kubectl-fx2mp' +Jun 18 12:14:42.495: INFO: stderr: "" +Jun 18 12:14:42.495: INFO: stdout: "pod/pause labeled\n" +STEP: verifying the pod doesn't have the label testing-label +Jun 18 12:14:42.495: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 get pod pause -L testing-label --namespace=e2e-tests-kubectl-fx2mp' +Jun 18 12:14:42.624: INFO: stderr: "" +Jun 18 12:14:42.624: INFO: stdout: "NAME READY STATUS RESTARTS AGE TESTING-LABEL\npause 1/1 Running 0 5s \n" +[AfterEach] [k8s.io] Kubectl label + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1059 +STEP: using delete to clean up resources +Jun 18 12:14:42.624: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 delete --grace-period=0 --force -f - --namespace=e2e-tests-kubectl-fx2mp' +Jun 18 12:14:42.811: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" +Jun 18 12:14:42.811: INFO: stdout: "pod \"pause\" force deleted\n" +Jun 18 12:14:42.811: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 get rc,svc -l name=pause --no-headers --namespace=e2e-tests-kubectl-fx2mp' +Jun 18 12:14:42.960: INFO: stderr: "No resources found.\n" +Jun 18 12:14:42.960: INFO: stdout: "" +Jun 18 12:14:42.960: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 get pods -l name=pause --namespace=e2e-tests-kubectl-fx2mp -o go-template={{ range .items }}{{ if not .metadata.deletionTimestamp }}{{ .metadata.name }}{{ "\n" }}{{ end }}{{ end }}' +Jun 18 12:14:43.112: INFO: stderr: "" +Jun 18 12:14:43.112: INFO: stdout: "" +[AfterEach] [sig-cli] Kubectl client /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 19:30:42.070: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-projected-27hkk" for this suite. -May 13 19:31:06.108: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 19:31:06.197: INFO: namespace: e2e-tests-projected-27hkk, resource: bindings, ignored listing per whitelist -May 13 19:31:06.358: INFO: namespace e2e-tests-projected-27hkk deletion completed in 24.277945764s +Jun 18 12:14:43.112: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "e2e-tests-kubectl-fx2mp" for this suite. +Jun 18 12:14:51.201: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:14:51.347: INFO: namespace: e2e-tests-kubectl-fx2mp, resource: bindings, ignored listing per whitelist +Jun 18 12:14:51.838: INFO: namespace e2e-tests-kubectl-fx2mp deletion completed in 8.707340522s -• [SLOW TEST:29.205 seconds] -[sig-storage] Projected downwardAPI -/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:33 - should update labels on modification [NodeConformance] [Conformance] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------- -SSSSSSSSSSSSS +• [SLOW TEST:14.672 seconds] +[sig-cli] Kubectl client +/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:22 + [k8s.io] Kubectl label + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 + should update the label on a resource [Conformance] + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ -[k8s.io] Container Lifecycle Hook when create a pod with lifecycle hook - should execute poststart http hook properly [NodeConformance] [Conformance] +[sig-storage] ConfigMap + updates should be reflected in volume [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -[BeforeEach] [k8s.io] Container Lifecycle Hook +[BeforeEach] [sig-storage] ConfigMap /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client -May 13 19:31:06.358: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -STEP: Building a namespace api object, basename container-lifecycle-hook -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-container-lifecycle-hook-qt79p +Jun 18 12:14:51.839: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 +STEP: Building a namespace api object, basename configmap +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-configmap-kx8vz STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] when create a pod with lifecycle hook - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:61 -STEP: create the container to handle the HTTPGet hook request. -[It] should execute poststart http hook properly [NodeConformance] [Conformance] +[It] updates should be reflected in volume [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -STEP: create the pod with lifecycle hook -STEP: check poststart hook -STEP: delete the pod with lifecycle hook -May 13 19:31:13.296: INFO: Waiting for pod pod-with-poststart-http-hook to disappear -May 13 19:31:13.307: INFO: Pod pod-with-poststart-http-hook still exists -May 13 19:31:15.307: INFO: Waiting for pod pod-with-poststart-http-hook to disappear -May 13 19:31:15.327: INFO: Pod pod-with-poststart-http-hook no longer exists -[AfterEach] [k8s.io] Container Lifecycle Hook +STEP: Creating configMap with name configmap-test-upd-ad023d3d-91c2-11e9-bce2-ae54e022189f +STEP: Creating the pod +STEP: Updating configmap configmap-test-upd-ad023d3d-91c2-11e9-bce2-ae54e022189f +STEP: waiting to observe update in volume +[AfterEach] [sig-storage] ConfigMap /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 19:31:15.327: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-container-lifecycle-hook-qt79p" for this suite. -May 13 19:31:39.449: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 19:31:39.750: INFO: namespace: e2e-tests-container-lifecycle-hook-qt79p, resource: bindings, ignored listing per whitelist -May 13 19:31:39.758: INFO: namespace e2e-tests-container-lifecycle-hook-qt79p deletion completed in 24.422294059s +Jun 18 12:14:56.590: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "e2e-tests-configmap-kx8vz" for this suite. +Jun 18 12:15:20.733: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:15:20.869: INFO: namespace: e2e-tests-configmap-kx8vz, resource: bindings, ignored listing per whitelist +Jun 18 12:15:21.238: INFO: namespace e2e-tests-configmap-kx8vz deletion completed in 24.552791258s -• [SLOW TEST:33.400 seconds] -[k8s.io] Container Lifecycle Hook -/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 - when create a pod with lifecycle hook - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:40 - should execute poststart http hook properly [NodeConformance] [Conformance] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 +• [SLOW TEST:29.399 seconds] +[sig-storage] ConfigMap +/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:33 + updates should be reflected in volume [NodeConformance] [Conformance] + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ -SSSSSSSSSSSSSSSSSS +SSSSSSSSSSS ------------------------------ -[sig-apps] Deployment - deployment should support proportional scaling [Conformance] +[k8s.io] [sig-node] Events + should be sent by kubelets and the scheduler about pods scheduling and running [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -[BeforeEach] [sig-apps] Deployment +[BeforeEach] [k8s.io] [sig-node] Events /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client -May 13 19:31:39.759: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -STEP: Building a namespace api object, basename deployment -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-deployment-27brj +Jun 18 12:15:21.238: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 +STEP: Building a namespace api object, basename events +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-events-bz2kw STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [sig-apps] Deployment - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:65 -[It] deployment should support proportional scaling [Conformance] +[It] should be sent by kubelets and the scheduler about pods scheduling and running [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -May 13 19:31:40.122: INFO: Creating deployment "nginx-deployment" -May 13 19:31:40.129: INFO: Waiting for observed generation 1 -May 13 19:31:42.140: INFO: Waiting for all required pods to come up -May 13 19:31:42.150: INFO: Pod name nginx: Found 10 pods out of 10 -STEP: ensuring each pod is running -May 13 19:31:46.181: INFO: Waiting for deployment "nginx-deployment" to complete -May 13 19:31:46.194: INFO: Updating deployment "nginx-deployment" with a non-existent image -May 13 19:31:46.219: INFO: Updating deployment nginx-deployment -May 13 19:31:46.219: INFO: Waiting for observed generation 2 -May 13 19:31:48.459: INFO: Waiting for the first rollout's replicaset to have .status.availableReplicas = 8 -May 13 19:31:48.466: INFO: Waiting for the first rollout's replicaset to have .spec.replicas = 8 -May 13 19:31:48.474: INFO: Waiting for the first rollout's replicaset of deployment "nginx-deployment" to have desired number of replicas -May 13 19:31:48.499: INFO: Verifying that the second rollout's replicaset has .status.availableReplicas = 0 -May 13 19:31:48.499: INFO: Waiting for the second rollout's replicaset to have .spec.replicas = 5 -May 13 19:31:48.508: INFO: Waiting for the second rollout's replicaset of deployment "nginx-deployment" to have desired number of replicas -May 13 19:31:48.524: INFO: Verifying that deployment "nginx-deployment" has minimum required number of available replicas -May 13 19:31:48.524: INFO: Scaling up the deployment "nginx-deployment" from 10 to 30 -May 13 19:31:48.538: INFO: Updating deployment nginx-deployment -May 13 19:31:48.538: INFO: Waiting for the replicasets of deployment "nginx-deployment" to have desired number of replicas -May 13 19:31:48.553: INFO: Verifying that first rollout's replicaset has .spec.replicas = 20 -May 13 19:31:48.561: INFO: Verifying that second rollout's replicaset has .spec.replicas = 13 -[AfterEach] [sig-apps] Deployment - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:59 -May 13 19:31:48.576: INFO: Deployment "nginx-deployment": -&Deployment{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment,GenerateName:,Namespace:e2e-tests-deployment-27brj,SelfLink:/apis/apps/v1/namespaces/e2e-tests-deployment-27brj/deployments/nginx-deployment,UID:bb2bea6a-75b5-11e9-b786-da20024d205c,ResourceVersion:43766,Generation:3,CreationTimestamp:2019-05-13 19:31:40 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,},Annotations:map[string]string{deployment.kubernetes.io/revision: 2,},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Spec:DeploymentSpec{Replicas:*30,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: nginx,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[],Containers:[{nginx nginx:404 [] [] [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,},},Strategy:DeploymentStrategy{Type:RollingUpdate,RollingUpdate:&RollingUpdateDeployment{MaxUnavailable:2,MaxSurge:3,},},MinReadySeconds:0,RevisionHistoryLimit:*10,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:3,Replicas:13,UpdatedReplicas:5,AvailableReplicas:8,UnavailableReplicas:5,Conditions:[{Progressing True 2019-05-13 19:31:46 +0000 UTC 2019-05-13 19:31:40 +0000 UTC ReplicaSetUpdated ReplicaSet "nginx-deployment-65bbdb5f8" is progressing.} {Available False 2019-05-13 19:31:48 +0000 UTC 2019-05-13 19:31:48 +0000 UTC MinimumReplicasUnavailable Deployment does not have minimum availability.}],ReadyReplicas:8,CollisionCount:nil,},} - -May 13 19:31:48.595: INFO: New ReplicaSet "nginx-deployment-65bbdb5f8" of Deployment "nginx-deployment": -&ReplicaSet{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-65bbdb5f8,GenerateName:,Namespace:e2e-tests-deployment-27brj,SelfLink:/apis/apps/v1/namespaces/e2e-tests-deployment-27brj/replicasets/nginx-deployment-65bbdb5f8,UID:becc99fe-75b5-11e9-83a8-7e0242378207,ResourceVersion:43759,Generation:3,CreationTimestamp:2019-05-13 19:31:46 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 65bbdb5f8,},Annotations:map[string]string{deployment.kubernetes.io/desired-replicas: 30,deployment.kubernetes.io/max-replicas: 33,deployment.kubernetes.io/revision: 2,},OwnerReferences:[{apps/v1 Deployment nginx-deployment bb2bea6a-75b5-11e9-b786-da20024d205c 0xc0024bcbf7 0xc0024bcbf8}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:ReplicaSetSpec{Replicas:*13,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: nginx,pod-template-hash: 65bbdb5f8,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 65bbdb5f8,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[],Containers:[{nginx nginx:404 [] [] [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,},},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:5,FullyLabeledReplicas:5,ObservedGeneration:2,ReadyReplicas:0,AvailableReplicas:0,Conditions:[],},} -May 13 19:31:48.595: INFO: All old ReplicaSets of Deployment "nginx-deployment": -May 13 19:31:48.595: INFO: &ReplicaSet{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-555b55d965,GenerateName:,Namespace:e2e-tests-deployment-27brj,SelfLink:/apis/apps/v1/namespaces/e2e-tests-deployment-27brj/replicasets/nginx-deployment-555b55d965,UID:bb2fe888-75b5-11e9-83a8-7e0242378207,ResourceVersion:43757,Generation:3,CreationTimestamp:2019-05-13 19:31:40 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 555b55d965,},Annotations:map[string]string{deployment.kubernetes.io/desired-replicas: 30,deployment.kubernetes.io/max-replicas: 33,deployment.kubernetes.io/revision: 1,},OwnerReferences:[{apps/v1 Deployment nginx-deployment bb2bea6a-75b5-11e9-b786-da20024d205c 0xc0024bcb37 0xc0024bcb38}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:ReplicaSetSpec{Replicas:*20,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: nginx,pod-template-hash: 555b55d965,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 555b55d965,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,},},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:8,FullyLabeledReplicas:8,ObservedGeneration:2,ReadyReplicas:8,AvailableReplicas:8,Conditions:[],},} -May 13 19:31:48.612: INFO: Pod "nginx-deployment-555b55d965-2998d" is available: -&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-555b55d965-2998d,GenerateName:nginx-deployment-555b55d965-,Namespace:e2e-tests-deployment-27brj,SelfLink:/api/v1/namespaces/e2e-tests-deployment-27brj/pods/nginx-deployment-555b55d965-2998d,UID:bb36a0f5-75b5-11e9-83a8-7e0242378207,ResourceVersion:43603,Generation:0,CreationTimestamp:2019-05-13 19:31:40 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 555b55d965,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-555b55d965 bb2fe888-75b5-11e9-83a8-7e0242378207 0xc0024bd6f7 0xc0024bd6f8}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-8bx99 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-8bx99,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-8bx99 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.170.219.177,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc0024bd770} {node.kubernetes.io/unreachable Exists NoExecute 0xc0024bd790}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Running,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:31:40 +0000 UTC } {Ready True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:31:41 +0000 UTC } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:31:41 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:31:40 +0000 UTC }],Message:,Reason:,HostIP:10.170.219.177,PodIP:172.30.227.76,StartTime:2019-05-13 19:31:40 +0000 UTC,ContainerStatuses:[{nginx {nil ContainerStateRunning{StartedAt:2019-05-13 19:31:41 +0000 UTC,} nil} {nil nil nil} true 0 docker.io/library/nginx:1.14-alpine docker.io/library/nginx@sha256:485b610fefec7ff6c463ced9623314a04ed67e3945b9c08d7e53a47f6d108dc7 containerd://7eede40d812d0deb8826a0e9e1bac88ecea438993e290cead47e5d038c83d6cd}],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} -May 13 19:31:48.612: INFO: Pod "nginx-deployment-555b55d965-554km" is available: -&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-555b55d965-554km,GenerateName:nginx-deployment-555b55d965-,Namespace:e2e-tests-deployment-27brj,SelfLink:/api/v1/namespaces/e2e-tests-deployment-27brj/pods/nginx-deployment-555b55d965-554km,UID:bb3510f5-75b5-11e9-83a8-7e0242378207,ResourceVersion:43629,Generation:0,CreationTimestamp:2019-05-13 19:31:40 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 555b55d965,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-555b55d965 bb2fe888-75b5-11e9-83a8-7e0242378207 0xc0024bd917 0xc0024bd918}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-8bx99 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-8bx99,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-8bx99 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.170.219.153,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc0024bd990} {node.kubernetes.io/unreachable Exists NoExecute 0xc0024bd9b0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Running,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:31:40 +0000 UTC } {Ready True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:31:42 +0000 UTC } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:31:42 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:31:40 +0000 UTC }],Message:,Reason:,HostIP:10.170.219.153,PodIP:172.30.63.2,StartTime:2019-05-13 19:31:40 +0000 UTC,ContainerStatuses:[{nginx {nil ContainerStateRunning{StartedAt:2019-05-13 19:31:41 +0000 UTC,} nil} {nil nil nil} true 0 docker.io/library/nginx:1.14-alpine docker.io/library/nginx@sha256:485b610fefec7ff6c463ced9623314a04ed67e3945b9c08d7e53a47f6d108dc7 containerd://e496a86f07680649c7c9b7bce3edf1537dd7d74f4f6e36b43a7a28db1d35d6f0}],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} -May 13 19:31:48.612: INFO: Pod "nginx-deployment-555b55d965-77phz" is available: -&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-555b55d965-77phz,GenerateName:nginx-deployment-555b55d965-,Namespace:e2e-tests-deployment-27brj,SelfLink:/api/v1/namespaces/e2e-tests-deployment-27brj/pods/nginx-deployment-555b55d965-77phz,UID:bb37b3cc-75b5-11e9-83a8-7e0242378207,ResourceVersion:43618,Generation:0,CreationTimestamp:2019-05-13 19:31:40 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 555b55d965,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-555b55d965 bb2fe888-75b5-11e9-83a8-7e0242378207 0xc0024bdad0 0xc0024bdad1}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-8bx99 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-8bx99,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-8bx99 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.170.219.151,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc0024bdb40} {node.kubernetes.io/unreachable Exists NoExecute 0xc0024bdb60}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Running,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:31:40 +0000 UTC } {Ready True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:31:41 +0000 UTC } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:31:41 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:31:40 +0000 UTC }],Message:,Reason:,HostIP:10.170.219.151,PodIP:172.30.19.103,StartTime:2019-05-13 19:31:40 +0000 UTC,ContainerStatuses:[{nginx {nil ContainerStateRunning{StartedAt:2019-05-13 19:31:41 +0000 UTC,} nil} {nil nil nil} true 0 docker.io/library/nginx:1.14-alpine docker.io/library/nginx@sha256:485b610fefec7ff6c463ced9623314a04ed67e3945b9c08d7e53a47f6d108dc7 containerd://d52e1267c9f87b0ce7ce5c0397d2b9df2617018b6fe1b09b719a238095ce30a7}],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} -May 13 19:31:48.612: INFO: Pod "nginx-deployment-555b55d965-86wd9" is available: -&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-555b55d965-86wd9,GenerateName:nginx-deployment-555b55d965-,Namespace:e2e-tests-deployment-27brj,SelfLink:/api/v1/namespaces/e2e-tests-deployment-27brj/pods/nginx-deployment-555b55d965-86wd9,UID:bb365f9f-75b5-11e9-83a8-7e0242378207,ResourceVersion:43641,Generation:0,CreationTimestamp:2019-05-13 19:31:40 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 555b55d965,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-555b55d965 bb2fe888-75b5-11e9-83a8-7e0242378207 0xc0024bdc77 0xc0024bdc78}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-8bx99 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-8bx99,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-8bx99 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.170.219.153,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc0024bdcf0} {node.kubernetes.io/unreachable Exists NoExecute 0xc0024bdd10}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Running,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:31:40 +0000 UTC } {Ready True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:31:42 +0000 UTC } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:31:42 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:31:40 +0000 UTC }],Message:,Reason:,HostIP:10.170.219.153,PodIP:172.30.63.3,StartTime:2019-05-13 19:31:40 +0000 UTC,ContainerStatuses:[{nginx {nil ContainerStateRunning{StartedAt:2019-05-13 19:31:41 +0000 UTC,} nil} {nil nil nil} true 0 docker.io/library/nginx:1.14-alpine docker.io/library/nginx@sha256:485b610fefec7ff6c463ced9623314a04ed67e3945b9c08d7e53a47f6d108dc7 containerd://24ecc598cf97f6ab2180a5008e310f335710ca8d55185c3566b55ea24f227d71}],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} -May 13 19:31:48.612: INFO: Pod "nginx-deployment-555b55d965-92zkd" is not available: -&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-555b55d965-92zkd,GenerateName:nginx-deployment-555b55d965-,Namespace:e2e-tests-deployment-27brj,SelfLink:/api/v1/namespaces/e2e-tests-deployment-27brj/pods/nginx-deployment-555b55d965-92zkd,UID:c035f44d-75b5-11e9-83a8-7e0242378207,ResourceVersion:43799,Generation:0,CreationTimestamp:2019-05-13 19:31:48 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 555b55d965,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-555b55d965 bb2fe888-75b5-11e9-83a8-7e0242378207 0xc0024bddd0 0xc0024bddd1}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-8bx99 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-8bx99,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-8bx99 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.170.219.177,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc0024bde40} {node.kubernetes.io/unreachable Exists NoExecute 0xc0024bde60}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:31:48 +0000 UTC }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} -May 13 19:31:48.612: INFO: Pod "nginx-deployment-555b55d965-cpftm" is not available: -&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-555b55d965-cpftm,GenerateName:nginx-deployment-555b55d965-,Namespace:e2e-tests-deployment-27brj,SelfLink:/api/v1/namespaces/e2e-tests-deployment-27brj/pods/nginx-deployment-555b55d965-cpftm,UID:c03478e3-75b5-11e9-83a8-7e0242378207,ResourceVersion:43779,Generation:0,CreationTimestamp:2019-05-13 19:31:48 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 555b55d965,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-555b55d965 bb2fe888-75b5-11e9-83a8-7e0242378207 0xc0024bdf50 0xc0024bdf51}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-8bx99 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-8bx99,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-8bx99 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.170.219.151,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc0024bdfc0} {node.kubernetes.io/unreachable Exists NoExecute 0xc0024bdfe0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:31:48 +0000 UTC }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} -May 13 19:31:48.612: INFO: Pod "nginx-deployment-555b55d965-h26p8" is not available: -&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-555b55d965-h26p8,GenerateName:nginx-deployment-555b55d965-,Namespace:e2e-tests-deployment-27brj,SelfLink:/api/v1/namespaces/e2e-tests-deployment-27brj/pods/nginx-deployment-555b55d965-h26p8,UID:c035fc70-75b5-11e9-83a8-7e0242378207,ResourceVersion:43802,Generation:0,CreationTimestamp:2019-05-13 19:31:48 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 555b55d965,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-555b55d965 bb2fe888-75b5-11e9-83a8-7e0242378207 0xc002378050 0xc002378051}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-8bx99 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-8bx99,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-8bx99 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.170.219.177,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc002378130} {node.kubernetes.io/unreachable Exists NoExecute 0xc002378150}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:31:48 +0000 UTC }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} -May 13 19:31:48.612: INFO: Pod "nginx-deployment-555b55d965-hf268" is not available: -&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-555b55d965-hf268,GenerateName:nginx-deployment-555b55d965-,Namespace:e2e-tests-deployment-27brj,SelfLink:/api/v1/namespaces/e2e-tests-deployment-27brj/pods/nginx-deployment-555b55d965-hf268,UID:c0348611-75b5-11e9-83a8-7e0242378207,ResourceVersion:43781,Generation:0,CreationTimestamp:2019-05-13 19:31:48 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 555b55d965,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-555b55d965 bb2fe888-75b5-11e9-83a8-7e0242378207 0xc0023781c0 0xc0023781c1}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-8bx99 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-8bx99,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-8bx99 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.170.219.177,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc002378230} {node.kubernetes.io/unreachable Exists NoExecute 0xc002378250}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:31:48 +0000 UTC }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} -May 13 19:31:48.613: INFO: Pod "nginx-deployment-555b55d965-hqh7q" is available: -&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-555b55d965-hqh7q,GenerateName:nginx-deployment-555b55d965-,Namespace:e2e-tests-deployment-27brj,SelfLink:/api/v1/namespaces/e2e-tests-deployment-27brj/pods/nginx-deployment-555b55d965-hqh7q,UID:bb367dac-75b5-11e9-83a8-7e0242378207,ResourceVersion:43638,Generation:0,CreationTimestamp:2019-05-13 19:31:40 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 555b55d965,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-555b55d965 bb2fe888-75b5-11e9-83a8-7e0242378207 0xc0023782d0 0xc0023782d1}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-8bx99 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-8bx99,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-8bx99 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.170.219.153,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc002378350} {node.kubernetes.io/unreachable Exists NoExecute 0xc002378370}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Running,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:31:40 +0000 UTC } {Ready True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:31:42 +0000 UTC } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:31:42 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:31:40 +0000 UTC }],Message:,Reason:,HostIP:10.170.219.153,PodIP:172.30.63.41,StartTime:2019-05-13 19:31:40 +0000 UTC,ContainerStatuses:[{nginx {nil ContainerStateRunning{StartedAt:2019-05-13 19:31:41 +0000 UTC,} nil} {nil nil nil} true 0 docker.io/library/nginx:1.14-alpine docker.io/library/nginx@sha256:485b610fefec7ff6c463ced9623314a04ed67e3945b9c08d7e53a47f6d108dc7 containerd://f6f22defecf54564328f141fffe2a8451f15abb9444065d3bf16e4719e64a47e}],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} -May 13 19:31:48.613: INFO: Pod "nginx-deployment-555b55d965-k2fgw" is not available: -&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-555b55d965-k2fgw,GenerateName:nginx-deployment-555b55d965-,Namespace:e2e-tests-deployment-27brj,SelfLink:/api/v1/namespaces/e2e-tests-deployment-27brj/pods/nginx-deployment-555b55d965-k2fgw,UID:c0319682-75b5-11e9-83a8-7e0242378207,ResourceVersion:43787,Generation:0,CreationTimestamp:2019-05-13 19:31:48 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 555b55d965,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-555b55d965 bb2fe888-75b5-11e9-83a8-7e0242378207 0xc0023784a7 0xc0023784a8}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-8bx99 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-8bx99,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-8bx99 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.170.219.151,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc002378520} {node.kubernetes.io/unreachable Exists NoExecute 0xc002378540}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:31:48 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:31:48 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:31:48 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:31:48 +0000 UTC }],Message:,Reason:,HostIP:10.170.219.151,PodIP:,StartTime:2019-05-13 19:31:48 +0000 UTC,ContainerStatuses:[{nginx {ContainerStateWaiting{Reason:ContainerCreating,Message:,} nil nil} {nil nil nil} false 0 docker.io/library/nginx:1.14-alpine }],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} -May 13 19:31:48.613: INFO: Pod "nginx-deployment-555b55d965-kbxp8" is not available: -&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-555b55d965-kbxp8,GenerateName:nginx-deployment-555b55d965-,Namespace:e2e-tests-deployment-27brj,SelfLink:/api/v1/namespaces/e2e-tests-deployment-27brj/pods/nginx-deployment-555b55d965-kbxp8,UID:c0330026-75b5-11e9-83a8-7e0242378207,ResourceVersion:43810,Generation:0,CreationTimestamp:2019-05-13 19:31:48 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 555b55d965,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-555b55d965 bb2fe888-75b5-11e9-83a8-7e0242378207 0xc0023785f7 0xc0023785f8}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-8bx99 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-8bx99,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-8bx99 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.170.219.153,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc002378720} {node.kubernetes.io/unreachable Exists NoExecute 0xc002378740}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:31:48 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:31:48 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:31:48 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:31:48 +0000 UTC }],Message:,Reason:,HostIP:10.170.219.153,PodIP:,StartTime:2019-05-13 19:31:48 +0000 UTC,ContainerStatuses:[{nginx {ContainerStateWaiting{Reason:ContainerCreating,Message:,} nil nil} {nil nil nil} false 0 docker.io/library/nginx:1.14-alpine }],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} -May 13 19:31:48.613: INFO: Pod "nginx-deployment-555b55d965-mkm2v" is available: -&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-555b55d965-mkm2v,GenerateName:nginx-deployment-555b55d965-,Namespace:e2e-tests-deployment-27brj,SelfLink:/api/v1/namespaces/e2e-tests-deployment-27brj/pods/nginx-deployment-555b55d965-mkm2v,UID:bb366f28-75b5-11e9-83a8-7e0242378207,ResourceVersion:43600,Generation:0,CreationTimestamp:2019-05-13 19:31:40 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 555b55d965,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-555b55d965 bb2fe888-75b5-11e9-83a8-7e0242378207 0xc0023787f7 0xc0023787f8}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-8bx99 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-8bx99,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-8bx99 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.170.219.177,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc002378870} {node.kubernetes.io/unreachable Exists NoExecute 0xc002378890}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Running,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:31:40 +0000 UTC } {Ready True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:31:41 +0000 UTC } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:31:41 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:31:40 +0000 UTC }],Message:,Reason:,HostIP:10.170.219.177,PodIP:172.30.227.74,StartTime:2019-05-13 19:31:40 +0000 UTC,ContainerStatuses:[{nginx {nil ContainerStateRunning{StartedAt:2019-05-13 19:31:41 +0000 UTC,} nil} {nil nil nil} true 0 docker.io/library/nginx:1.14-alpine docker.io/library/nginx@sha256:485b610fefec7ff6c463ced9623314a04ed67e3945b9c08d7e53a47f6d108dc7 containerd://040875ef68a06abcb48a80991443b1312ff3853ce715e7426947ff70bae2a9d7}],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} -May 13 19:31:48.613: INFO: Pod "nginx-deployment-555b55d965-mwqd7" is not available: -&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-555b55d965-mwqd7,GenerateName:nginx-deployment-555b55d965-,Namespace:e2e-tests-deployment-27brj,SelfLink:/api/v1/namespaces/e2e-tests-deployment-27brj/pods/nginx-deployment-555b55d965-mwqd7,UID:c0347ba3-75b5-11e9-83a8-7e0242378207,ResourceVersion:43788,Generation:0,CreationTimestamp:2019-05-13 19:31:48 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 555b55d965,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-555b55d965 bb2fe888-75b5-11e9-83a8-7e0242378207 0xc002378957 0xc002378958}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-8bx99 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-8bx99,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-8bx99 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.170.219.153,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc0023789d0} {node.kubernetes.io/unreachable Exists NoExecute 0xc0023789f0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:31:48 +0000 UTC }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} -May 13 19:31:48.613: INFO: Pod "nginx-deployment-555b55d965-nqkw2" is not available: -&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-555b55d965-nqkw2,GenerateName:nginx-deployment-555b55d965-,Namespace:e2e-tests-deployment-27brj,SelfLink:/api/v1/namespaces/e2e-tests-deployment-27brj/pods/nginx-deployment-555b55d965-nqkw2,UID:c035eaac-75b5-11e9-83a8-7e0242378207,ResourceVersion:43796,Generation:0,CreationTimestamp:2019-05-13 19:31:48 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 555b55d965,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-555b55d965 bb2fe888-75b5-11e9-83a8-7e0242378207 0xc002378a60 0xc002378a61}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-8bx99 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-8bx99,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-8bx99 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.170.219.151,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc002378ad0} {node.kubernetes.io/unreachable Exists NoExecute 0xc002378ba0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:31:48 +0000 UTC }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} -May 13 19:31:48.613: INFO: Pod "nginx-deployment-555b55d965-q6f9l" is not available: -&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-555b55d965-q6f9l,GenerateName:nginx-deployment-555b55d965-,Namespace:e2e-tests-deployment-27brj,SelfLink:/api/v1/namespaces/e2e-tests-deployment-27brj/pods/nginx-deployment-555b55d965-q6f9l,UID:c03323cc-75b5-11e9-83a8-7e0242378207,ResourceVersion:43770,Generation:0,CreationTimestamp:2019-05-13 19:31:48 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 555b55d965,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-555b55d965 bb2fe888-75b5-11e9-83a8-7e0242378207 0xc002378c10 0xc002378c11}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-8bx99 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-8bx99,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-8bx99 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.170.219.153,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc002378c80} {node.kubernetes.io/unreachable Exists NoExecute 0xc002378ca0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:31:48 +0000 UTC }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} -May 13 19:31:48.614: INFO: Pod "nginx-deployment-555b55d965-w5fzl" is not available: -&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-555b55d965-w5fzl,GenerateName:nginx-deployment-555b55d965-,Namespace:e2e-tests-deployment-27brj,SelfLink:/api/v1/namespaces/e2e-tests-deployment-27brj/pods/nginx-deployment-555b55d965-w5fzl,UID:c035fbd2-75b5-11e9-83a8-7e0242378207,ResourceVersion:43798,Generation:0,CreationTimestamp:2019-05-13 19:31:48 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 555b55d965,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-555b55d965 bb2fe888-75b5-11e9-83a8-7e0242378207 0xc002378d10 0xc002378d11}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-8bx99 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-8bx99,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-8bx99 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.170.219.153,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc002378d80} {node.kubernetes.io/unreachable Exists NoExecute 0xc002378da0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:31:48 +0000 UTC }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} -May 13 19:31:48.614: INFO: Pod "nginx-deployment-555b55d965-wttt7" is not available: -&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-555b55d965-wttt7,GenerateName:nginx-deployment-555b55d965-,Namespace:e2e-tests-deployment-27brj,SelfLink:/api/v1/namespaces/e2e-tests-deployment-27brj/pods/nginx-deployment-555b55d965-wttt7,UID:c0360ab2-75b5-11e9-83a8-7e0242378207,ResourceVersion:43801,Generation:0,CreationTimestamp:2019-05-13 19:31:48 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 555b55d965,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-555b55d965 bb2fe888-75b5-11e9-83a8-7e0242378207 0xc002378f50 0xc002378f51}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-8bx99 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-8bx99,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-8bx99 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.170.219.151,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc002378fc0} {node.kubernetes.io/unreachable Exists NoExecute 0xc002378fe0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:31:48 +0000 UTC }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} -May 13 19:31:48.614: INFO: Pod "nginx-deployment-555b55d965-wx87v" is available: -&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-555b55d965-wx87v,GenerateName:nginx-deployment-555b55d965-,Namespace:e2e-tests-deployment-27brj,SelfLink:/api/v1/namespaces/e2e-tests-deployment-27brj/pods/nginx-deployment-555b55d965-wx87v,UID:bb351315-75b5-11e9-83a8-7e0242378207,ResourceVersion:43622,Generation:0,CreationTimestamp:2019-05-13 19:31:40 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 555b55d965,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-555b55d965 bb2fe888-75b5-11e9-83a8-7e0242378207 0xc002379050 0xc002379051}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-8bx99 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-8bx99,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-8bx99 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.170.219.151,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc0023790c0} {node.kubernetes.io/unreachable Exists NoExecute 0xc0023790e0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Running,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:31:40 +0000 UTC } {Ready True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:31:41 +0000 UTC } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:31:41 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:31:40 +0000 UTC }],Message:,Reason:,HostIP:10.170.219.151,PodIP:172.30.19.101,StartTime:2019-05-13 19:31:40 +0000 UTC,ContainerStatuses:[{nginx {nil ContainerStateRunning{StartedAt:2019-05-13 19:31:41 +0000 UTC,} nil} {nil nil nil} true 0 docker.io/library/nginx:1.14-alpine docker.io/library/nginx@sha256:485b610fefec7ff6c463ced9623314a04ed67e3945b9c08d7e53a47f6d108dc7 containerd://b69cffd9aaf4c232a491edc3c82dba290b8d4e7052105728422a6a300694cd72}],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} -May 13 19:31:48.614: INFO: Pod "nginx-deployment-555b55d965-z8nk6" is available: -&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-555b55d965-z8nk6,GenerateName:nginx-deployment-555b55d965-,Namespace:e2e-tests-deployment-27brj,SelfLink:/api/v1/namespaces/e2e-tests-deployment-27brj/pods/nginx-deployment-555b55d965-z8nk6,UID:bb33ed79-75b5-11e9-83a8-7e0242378207,ResourceVersion:43607,Generation:0,CreationTimestamp:2019-05-13 19:31:40 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 555b55d965,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-555b55d965 bb2fe888-75b5-11e9-83a8-7e0242378207 0xc002379287 0xc002379288}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-8bx99 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-8bx99,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-8bx99 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.170.219.177,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc002379300} {node.kubernetes.io/unreachable Exists NoExecute 0xc002379320}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Running,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:31:40 +0000 UTC } {Ready True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:31:41 +0000 UTC } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:31:41 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:31:40 +0000 UTC }],Message:,Reason:,HostIP:10.170.219.177,PodIP:172.30.227.72,StartTime:2019-05-13 19:31:40 +0000 UTC,ContainerStatuses:[{nginx {nil ContainerStateRunning{StartedAt:2019-05-13 19:31:41 +0000 UTC,} nil} {nil nil nil} true 0 docker.io/library/nginx:1.14-alpine docker.io/library/nginx@sha256:485b610fefec7ff6c463ced9623314a04ed67e3945b9c08d7e53a47f6d108dc7 containerd://4b1fdb1470f6529e8c0ae942ba88238295c45254ac43817136ce6a428cc4180c}],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} -May 13 19:31:48.614: INFO: Pod "nginx-deployment-555b55d965-zpzld" is not available: -&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-555b55d965-zpzld,GenerateName:nginx-deployment-555b55d965-,Namespace:e2e-tests-deployment-27brj,SelfLink:/api/v1/namespaces/e2e-tests-deployment-27brj/pods/nginx-deployment-555b55d965-zpzld,UID:c03488a0-75b5-11e9-83a8-7e0242378207,ResourceVersion:43786,Generation:0,CreationTimestamp:2019-05-13 19:31:48 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 555b55d965,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-555b55d965 bb2fe888-75b5-11e9-83a8-7e0242378207 0xc0023793e7 0xc0023793e8}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-8bx99 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-8bx99,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-8bx99 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.170.219.177,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc0023794c0} {node.kubernetes.io/unreachable Exists NoExecute 0xc0023794e0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:31:48 +0000 UTC }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} -May 13 19:31:48.614: INFO: Pod "nginx-deployment-65bbdb5f8-4xbfd" is not available: -&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-65bbdb5f8-4xbfd,GenerateName:nginx-deployment-65bbdb5f8-,Namespace:e2e-tests-deployment-27brj,SelfLink:/api/v1/namespaces/e2e-tests-deployment-27brj/pods/nginx-deployment-65bbdb5f8-4xbfd,UID:c0367776-75b5-11e9-83a8-7e0242378207,ResourceVersion:43805,Generation:0,CreationTimestamp:2019-05-13 19:31:48 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 65bbdb5f8,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-65bbdb5f8 becc99fe-75b5-11e9-83a8-7e0242378207 0xc002379580 0xc002379581}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-8bx99 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-8bx99,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] [] [] [] [] {map[] map[]} [{default-token-8bx99 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.170.219.151,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc002379600} {node.kubernetes.io/unreachable Exists NoExecute 0xc002379620}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:31:48 +0000 UTC }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} -May 13 19:31:48.614: INFO: Pod "nginx-deployment-65bbdb5f8-6zxs8" is not available: -&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-65bbdb5f8-6zxs8,GenerateName:nginx-deployment-65bbdb5f8-,Namespace:e2e-tests-deployment-27brj,SelfLink:/api/v1/namespaces/e2e-tests-deployment-27brj/pods/nginx-deployment-65bbdb5f8-6zxs8,UID:c0373023-75b5-11e9-83a8-7e0242378207,ResourceVersion:43808,Generation:0,CreationTimestamp:2019-05-13 19:31:48 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 65bbdb5f8,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-65bbdb5f8 becc99fe-75b5-11e9-83a8-7e0242378207 0xc002379690 0xc002379691}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-8bx99 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-8bx99,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] [] [] [] [] {map[] map[]} [{default-token-8bx99 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.170.219.153,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc002379710} {node.kubernetes.io/unreachable Exists NoExecute 0xc002379730}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:31:48 +0000 UTC }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} -May 13 19:31:48.614: INFO: Pod "nginx-deployment-65bbdb5f8-7mc6b" is not available: -&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-65bbdb5f8-7mc6b,GenerateName:nginx-deployment-65bbdb5f8-,Namespace:e2e-tests-deployment-27brj,SelfLink:/api/v1/namespaces/e2e-tests-deployment-27brj/pods/nginx-deployment-65bbdb5f8-7mc6b,UID:c036b96a-75b5-11e9-83a8-7e0242378207,ResourceVersion:43807,Generation:0,CreationTimestamp:2019-05-13 19:31:48 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 65bbdb5f8,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-65bbdb5f8 becc99fe-75b5-11e9-83a8-7e0242378207 0xc0023797a0 0xc0023797a1}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-8bx99 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-8bx99,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] [] [] [] [] {map[] map[]} [{default-token-8bx99 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.170.219.177,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc0020da030} {node.kubernetes.io/unreachable Exists NoExecute 0xc0020da050}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:31:48 +0000 UTC }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} -May 13 19:31:48.615: INFO: Pod "nginx-deployment-65bbdb5f8-9z5ms" is not available: -&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-65bbdb5f8-9z5ms,GenerateName:nginx-deployment-65bbdb5f8-,Namespace:e2e-tests-deployment-27brj,SelfLink:/api/v1/namespaces/e2e-tests-deployment-27brj/pods/nginx-deployment-65bbdb5f8-9z5ms,UID:bed766b7-75b5-11e9-83a8-7e0242378207,ResourceVersion:43722,Generation:0,CreationTimestamp:2019-05-13 19:31:46 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 65bbdb5f8,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-65bbdb5f8 becc99fe-75b5-11e9-83a8-7e0242378207 0xc0020da0c0 0xc0020da0c1}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-8bx99 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-8bx99,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] [] [] [] [] {map[] map[]} [{default-token-8bx99 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.170.219.153,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc0020da140} {node.kubernetes.io/unreachable Exists NoExecute 0xc0020da160}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:31:46 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:31:46 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:31:46 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:31:46 +0000 UTC }],Message:,Reason:,HostIP:10.170.219.153,PodIP:,StartTime:2019-05-13 19:31:46 +0000 UTC,ContainerStatuses:[{nginx {ContainerStateWaiting{Reason:ContainerCreating,Message:,} nil nil} {nil nil nil} false 0 nginx:404 }],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} -May 13 19:31:48.615: INFO: Pod "nginx-deployment-65bbdb5f8-bkfvq" is not available: -&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-65bbdb5f8-bkfvq,GenerateName:nginx-deployment-65bbdb5f8-,Namespace:e2e-tests-deployment-27brj,SelfLink:/api/v1/namespaces/e2e-tests-deployment-27brj/pods/nginx-deployment-65bbdb5f8-bkfvq,UID:c0367654-75b5-11e9-83a8-7e0242378207,ResourceVersion:43804,Generation:0,CreationTimestamp:2019-05-13 19:31:48 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 65bbdb5f8,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-65bbdb5f8 becc99fe-75b5-11e9-83a8-7e0242378207 0xc0020da220 0xc0020da221}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-8bx99 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-8bx99,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] [] [] [] [] {map[] map[]} [{default-token-8bx99 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.170.219.177,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc0020da2a0} {node.kubernetes.io/unreachable Exists NoExecute 0xc0020da2c0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:31:48 +0000 UTC }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} -May 13 19:31:48.615: INFO: Pod "nginx-deployment-65bbdb5f8-ft2jk" is not available: -&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-65bbdb5f8-ft2jk,GenerateName:nginx-deployment-65bbdb5f8-,Namespace:e2e-tests-deployment-27brj,SelfLink:/api/v1/namespaces/e2e-tests-deployment-27brj/pods/nginx-deployment-65bbdb5f8-ft2jk,UID:c03507b9-75b5-11e9-83a8-7e0242378207,ResourceVersion:43789,Generation:0,CreationTimestamp:2019-05-13 19:31:48 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 65bbdb5f8,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-65bbdb5f8 becc99fe-75b5-11e9-83a8-7e0242378207 0xc0020da330 0xc0020da331}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-8bx99 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-8bx99,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] [] [] [] [] {map[] map[]} [{default-token-8bx99 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.170.219.151,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc0020da3b0} {node.kubernetes.io/unreachable Exists NoExecute 0xc0020da3d0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:31:48 +0000 UTC }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} -May 13 19:31:48.615: INFO: Pod "nginx-deployment-65bbdb5f8-j82v8" is not available: -&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-65bbdb5f8-j82v8,GenerateName:nginx-deployment-65bbdb5f8-,Namespace:e2e-tests-deployment-27brj,SelfLink:/api/v1/namespaces/e2e-tests-deployment-27brj/pods/nginx-deployment-65bbdb5f8-j82v8,UID:c0387663-75b5-11e9-83a8-7e0242378207,ResourceVersion:43806,Generation:0,CreationTimestamp:2019-05-13 19:31:48 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 65bbdb5f8,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-65bbdb5f8 becc99fe-75b5-11e9-83a8-7e0242378207 0xc0020da440 0xc0020da441}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-8bx99 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-8bx99,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] [] [] [] [] {map[] map[]} [{default-token-8bx99 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc0020da4b0} {node.kubernetes.io/unreachable Exists NoExecute 0xc0020da4d0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} -May 13 19:31:48.615: INFO: Pod "nginx-deployment-65bbdb5f8-mmsfj" is not available: -&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-65bbdb5f8-mmsfj,GenerateName:nginx-deployment-65bbdb5f8-,Namespace:e2e-tests-deployment-27brj,SelfLink:/api/v1/namespaces/e2e-tests-deployment-27brj/pods/nginx-deployment-65bbdb5f8-mmsfj,UID:c0330ebc-75b5-11e9-83a8-7e0242378207,ResourceVersion:43809,Generation:0,CreationTimestamp:2019-05-13 19:31:48 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 65bbdb5f8,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-65bbdb5f8 becc99fe-75b5-11e9-83a8-7e0242378207 0xc0020da527 0xc0020da528}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-8bx99 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-8bx99,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] [] [] [] [] {map[] map[]} [{default-token-8bx99 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.170.219.177,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc0020da5a0} {node.kubernetes.io/unreachable Exists NoExecute 0xc0020da5c0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:31:48 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:31:48 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:31:48 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:31:48 +0000 UTC }],Message:,Reason:,HostIP:10.170.219.177,PodIP:,StartTime:2019-05-13 19:31:48 +0000 UTC,ContainerStatuses:[{nginx {ContainerStateWaiting{Reason:ContainerCreating,Message:,} nil nil} {nil nil nil} false 0 nginx:404 }],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} -May 13 19:31:48.615: INFO: Pod "nginx-deployment-65bbdb5f8-nrzsj" is not available: -&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-65bbdb5f8-nrzsj,GenerateName:nginx-deployment-65bbdb5f8-,Namespace:e2e-tests-deployment-27brj,SelfLink:/api/v1/namespaces/e2e-tests-deployment-27brj/pods/nginx-deployment-65bbdb5f8-nrzsj,UID:becf3e74-75b5-11e9-83a8-7e0242378207,ResourceVersion:43755,Generation:0,CreationTimestamp:2019-05-13 19:31:46 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 65bbdb5f8,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-65bbdb5f8 becc99fe-75b5-11e9-83a8-7e0242378207 0xc0020da680 0xc0020da681}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-8bx99 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-8bx99,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] [] [] [] [] {map[] map[]} [{default-token-8bx99 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.170.219.177,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc0020da710} {node.kubernetes.io/unreachable Exists NoExecute 0xc0020da730}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:31:46 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:31:46 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:31:46 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:31:46 +0000 UTC }],Message:,Reason:,HostIP:10.170.219.177,PodIP:172.30.227.73,StartTime:2019-05-13 19:31:46 +0000 UTC,ContainerStatuses:[{nginx {ContainerStateWaiting{Reason:ErrImagePull,Message:rpc error: code = Unknown desc = failed to resolve image "docker.io/library/nginx:404": no available registry endpoint: docker.io/library/nginx:404 not found,} nil nil} {nil nil nil} false 0 nginx:404 }],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} -May 13 19:31:48.615: INFO: Pod "nginx-deployment-65bbdb5f8-ppx74" is not available: -&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-65bbdb5f8-ppx74,GenerateName:nginx-deployment-65bbdb5f8-,Namespace:e2e-tests-deployment-27brj,SelfLink:/api/v1/namespaces/e2e-tests-deployment-27brj/pods/nginx-deployment-65bbdb5f8-ppx74,UID:becdf59e-75b5-11e9-83a8-7e0242378207,ResourceVersion:43745,Generation:0,CreationTimestamp:2019-05-13 19:31:46 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 65bbdb5f8,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-65bbdb5f8 becc99fe-75b5-11e9-83a8-7e0242378207 0xc0020da820 0xc0020da821}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-8bx99 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-8bx99,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] [] [] [] [] {map[] map[]} [{default-token-8bx99 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.170.219.153,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc0020da8a0} {node.kubernetes.io/unreachable Exists NoExecute 0xc0020da8c0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:31:46 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:31:46 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:31:46 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:31:46 +0000 UTC }],Message:,Reason:,HostIP:10.170.219.153,PodIP:172.30.63.44,StartTime:2019-05-13 19:31:46 +0000 UTC,ContainerStatuses:[{nginx {ContainerStateWaiting{Reason:ErrImagePull,Message:rpc error: code = Unknown desc = failed to resolve image "docker.io/library/nginx:404": no available registry endpoint: docker.io/library/nginx:404 not found,} nil nil} {nil nil nil} false 0 nginx:404 }],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} -May 13 19:31:48.617: INFO: Pod "nginx-deployment-65bbdb5f8-xnj7x" is not available: -&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-65bbdb5f8-xnj7x,GenerateName:nginx-deployment-65bbdb5f8-,Namespace:e2e-tests-deployment-27brj,SelfLink:/api/v1/namespaces/e2e-tests-deployment-27brj/pods/nginx-deployment-65bbdb5f8-xnj7x,UID:becf3a2a-75b5-11e9-83a8-7e0242378207,ResourceVersion:43736,Generation:0,CreationTimestamp:2019-05-13 19:31:46 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 65bbdb5f8,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-65bbdb5f8 becc99fe-75b5-11e9-83a8-7e0242378207 0xc0020dac00 0xc0020dac01}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-8bx99 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-8bx99,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] [] [] [] [] {map[] map[]} [{default-token-8bx99 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.170.219.151,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc0020dac80} {node.kubernetes.io/unreachable Exists NoExecute 0xc0020daca0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:31:46 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:31:46 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:31:46 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:31:46 +0000 UTC }],Message:,Reason:,HostIP:10.170.219.151,PodIP:172.30.19.104,StartTime:2019-05-13 19:31:46 +0000 UTC,ContainerStatuses:[{nginx {ContainerStateWaiting{Reason:ErrImagePull,Message:rpc error: code = Unknown desc = failed to resolve image "docker.io/library/nginx:404": no available registry endpoint: docker.io/library/nginx:404 not found,} nil nil} {nil nil nil} false 0 nginx:404 }],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} -May 13 19:31:48.617: INFO: Pod "nginx-deployment-65bbdb5f8-xrbrd" is not available: -&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-65bbdb5f8-xrbrd,GenerateName:nginx-deployment-65bbdb5f8-,Namespace:e2e-tests-deployment-27brj,SelfLink:/api/v1/namespaces/e2e-tests-deployment-27brj/pods/nginx-deployment-65bbdb5f8-xrbrd,UID:bed61018-75b5-11e9-83a8-7e0242378207,ResourceVersion:43703,Generation:0,CreationTimestamp:2019-05-13 19:31:46 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 65bbdb5f8,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-65bbdb5f8 becc99fe-75b5-11e9-83a8-7e0242378207 0xc0020db550 0xc0020db551}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-8bx99 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-8bx99,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] [] [] [] [] {map[] map[]} [{default-token-8bx99 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.170.219.153,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc0020db5e0} {node.kubernetes.io/unreachable Exists NoExecute 0xc0020db600}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:31:46 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:31:46 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:31:46 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:31:46 +0000 UTC }],Message:,Reason:,HostIP:10.170.219.153,PodIP:,StartTime:2019-05-13 19:31:46 +0000 UTC,ContainerStatuses:[{nginx {ContainerStateWaiting{Reason:ContainerCreating,Message:,} nil nil} {nil nil nil} false 0 nginx:404 }],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} -May 13 19:31:48.617: INFO: Pod "nginx-deployment-65bbdb5f8-zdsgw" is not available: -&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-65bbdb5f8-zdsgw,GenerateName:nginx-deployment-65bbdb5f8-,Namespace:e2e-tests-deployment-27brj,SelfLink:/api/v1/namespaces/e2e-tests-deployment-27brj/pods/nginx-deployment-65bbdb5f8-zdsgw,UID:c0356dd0-75b5-11e9-83a8-7e0242378207,ResourceVersion:43793,Generation:0,CreationTimestamp:2019-05-13 19:31:48 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 65bbdb5f8,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-65bbdb5f8 becc99fe-75b5-11e9-83a8-7e0242378207 0xc0020dba50 0xc0020dba51}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-8bx99 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-8bx99,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] [] [] [] [] {map[] map[]} [{default-token-8bx99 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.170.219.153,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc0020dbad0} {node.kubernetes.io/unreachable Exists NoExecute 0xc0020dbaf0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:31:48 +0000 UTC }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} -[AfterEach] [sig-apps] Deployment +STEP: creating the pod +STEP: submitting the pod to kubernetes +STEP: verifying the pod is in kubernetes +STEP: retrieving the pod +Jun 18 12:15:23.761: INFO: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:send-events-be790474-91c2-11e9-bce2-ae54e022189f,GenerateName:,Namespace:e2e-tests-events-bz2kw,SelfLink:/api/v1/namespaces/e2e-tests-events-bz2kw/pods/send-events-be790474-91c2-11e9-bce2-ae54e022189f,UID:be7bb817-91c2-11e9-bf44-fa6f350b29f0,ResourceVersion:98634,Generation:0,CreationTimestamp:2019-06-18 12:15:21 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: foo,time: 679583437,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-zlxb5 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-zlxb5,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{p gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1 [] [] [{ 0 80 TCP }] [] [] {map[] map[]} [{default-token-zlxb5 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*30,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.72.74.149,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc0002f4f30} {node.kubernetes.io/unreachable Exists NoExecute 0xc0002f59f0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Running,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:15:21 +0000 UTC } {Ready True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:15:23 +0000 UTC } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:15:23 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:15:21 +0000 UTC }],Message:,Reason:,HostIP:10.72.74.149,PodIP:172.30.39.49,StartTime:2019-06-18 12:15:21 +0000 UTC,ContainerStatuses:[{p {nil ContainerStateRunning{StartedAt:2019-06-18 12:15:23 +0000 UTC,} nil} {nil nil nil} true 0 gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1 gcr.io/kubernetes-e2e-test-images/serve-hostname@sha256:bab70473a6d8ef65a22625dc9a1b0f0452e811530fdbe77e4408523460177ff1 containerd://1a091482f806d7420678be509ce49e38fabecc3b22f6117e46a5b544be78e3cc}],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} + +STEP: checking for scheduler event about the pod +Jun 18 12:15:26.023: INFO: Saw scheduler event for our pod. +STEP: checking for kubelet event about the pod +Jun 18 12:15:28.036: INFO: Saw kubelet event for our pod. +STEP: deleting the pod +[AfterEach] [k8s.io] [sig-node] Events /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 19:31:48.617: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-deployment-27brj" for this suite. -May 13 19:31:56.669: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 19:31:56.776: INFO: namespace: e2e-tests-deployment-27brj, resource: bindings, ignored listing per whitelist -May 13 19:31:56.981: INFO: namespace e2e-tests-deployment-27brj deletion completed in 8.357901321s +Jun 18 12:15:28.061: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "e2e-tests-events-bz2kw" for this suite. +Jun 18 12:16:10.862: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:16:11.309: INFO: namespace: e2e-tests-events-bz2kw, resource: bindings, ignored listing per whitelist +Jun 18 12:16:11.620: INFO: namespace e2e-tests-events-bz2kw deletion completed in 43.539530001s -• [SLOW TEST:17.222 seconds] -[sig-apps] Deployment -/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22 - deployment should support proportional scaling [Conformance] +• [SLOW TEST:50.382 seconds] +[k8s.io] [sig-node] Events +/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 + should be sent by kubelets and the scheduler about pods scheduling and running [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSSS ------------------------------ -[k8s.io] InitContainer [NodeConformance] - should invoke init containers on a RestartAlways pod [Conformance] +[sig-storage] Projected combined + should project all components that make up the projection API [Projection][NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -[BeforeEach] [k8s.io] InitContainer [NodeConformance] +[BeforeEach] [sig-storage] Projected combined /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client -May 13 19:31:56.981: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -STEP: Building a namespace api object, basename init-container -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-init-container-7wbmf +Jun 18 12:16:11.621: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 +STEP: Building a namespace api object, basename projected +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-projected-4pwmq STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [k8s.io] InitContainer [NodeConformance] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/init_container.go:43 -[It] should invoke init containers on a RestartAlways pod [Conformance] +[It] should project all components that make up the projection API [Projection][NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -STEP: creating the pod -May 13 19:31:57.349: INFO: PodSpec: initContainers in spec.initContainers -[AfterEach] [k8s.io] InitContainer [NodeConformance] +STEP: Creating configMap with name configmap-projected-all-test-volume-dc811107-91c2-11e9-bce2-ae54e022189f +STEP: Creating secret with name secret-projected-all-test-volume-dc8110de-91c2-11e9-bce2-ae54e022189f +STEP: Creating a pod to test Check all projections for projected volume plugin +Jun 18 12:16:12.119: INFO: Waiting up to 5m0s for pod "projected-volume-dc811086-91c2-11e9-bce2-ae54e022189f" in namespace "e2e-tests-projected-4pwmq" to be "success or failure" +Jun 18 12:16:12.133: INFO: Pod "projected-volume-dc811086-91c2-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 13.896431ms +Jun 18 12:16:14.147: INFO: Pod "projected-volume-dc811086-91c2-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 2.028153616s +Jun 18 12:16:16.163: INFO: Pod "projected-volume-dc811086-91c2-11e9-bce2-ae54e022189f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.043891054s +STEP: Saw pod success +Jun 18 12:16:16.163: INFO: Pod "projected-volume-dc811086-91c2-11e9-bce2-ae54e022189f" satisfied condition "success or failure" +Jun 18 12:16:16.181: INFO: Trying to get logs from node 10.72.74.143 pod projected-volume-dc811086-91c2-11e9-bce2-ae54e022189f container projected-all-volume-test: +STEP: delete the pod +Jun 18 12:16:16.253: INFO: Waiting for pod projected-volume-dc811086-91c2-11e9-bce2-ae54e022189f to disappear +Jun 18 12:16:16.268: INFO: Pod projected-volume-dc811086-91c2-11e9-bce2-ae54e022189f no longer exists +[AfterEach] [sig-storage] Projected combined /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 19:32:02.461: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-init-container-7wbmf" for this suite. -May 13 19:32:26.528: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 19:32:26.735: INFO: namespace: e2e-tests-init-container-7wbmf, resource: bindings, ignored listing per whitelist -May 13 19:32:26.775: INFO: namespace e2e-tests-init-container-7wbmf deletion completed in 24.306678168s +Jun 18 12:16:16.268: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "e2e-tests-projected-4pwmq" for this suite. +Jun 18 12:16:22.542: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:16:22.902: INFO: namespace: e2e-tests-projected-4pwmq, resource: bindings, ignored listing per whitelist +Jun 18 12:16:23.035: INFO: namespace e2e-tests-projected-4pwmq deletion completed in 6.744241471s -• [SLOW TEST:29.794 seconds] -[k8s.io] InitContainer [NodeConformance] -/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 - should invoke init containers on a RestartAlways pod [Conformance] +• [SLOW TEST:11.415 seconds] +[sig-storage] Projected combined +/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_combined.go:31 + should project all components that make up the projection API [Projection][NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ -SSSSSSSSSSSSS +SS ------------------------------ -[sig-cli] Kubectl client [k8s.io] Kubectl run deployment - should create a deployment from an image [Conformance] +[sig-cli] Kubectl client [k8s.io] Update Demo + should do a rolling update of a replication controller [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-cli] Kubectl client /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client -May 13 19:32:26.775: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 +Jun 18 12:16:23.036: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 STEP: Building a namespace api object, basename kubectl -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-kubectl-rfb5b +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-kubectl-ltw4d STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:243 -[BeforeEach] [k8s.io] Kubectl run deployment - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1399 -[It] should create a deployment from an image [Conformance] +[BeforeEach] [k8s.io] Update Demo + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:295 +[It] should do a rolling update of a replication controller [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -STEP: running the image docker.io/library/nginx:1.14-alpine -May 13 19:32:27.065: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 run e2e-test-nginx-deployment --image=docker.io/library/nginx:1.14-alpine --generator=deployment/v1beta1 --namespace=e2e-tests-kubectl-rfb5b' -May 13 19:32:27.381: INFO: stderr: "kubectl run --generator=deployment/v1beta1 is DEPRECATED and will be removed in a future version. Use kubectl run --generator=run-pod/v1 or kubectl create instead.\n" -May 13 19:32:27.381: INFO: stdout: "deployment.extensions/e2e-test-nginx-deployment created\n" -STEP: verifying the deployment e2e-test-nginx-deployment was created -STEP: verifying the pod controlled by deployment e2e-test-nginx-deployment was created -[AfterEach] [k8s.io] Kubectl run deployment - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1404 -May 13 19:32:31.411: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 delete deployment e2e-test-nginx-deployment --namespace=e2e-tests-kubectl-rfb5b' -May 13 19:32:31.547: INFO: stderr: "" -May 13 19:32:31.547: INFO: stdout: "deployment.extensions \"e2e-test-nginx-deployment\" deleted\n" +STEP: creating the initial replication controller +Jun 18 12:16:23.488: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 create -f - --namespace=e2e-tests-kubectl-ltw4d' +Jun 18 12:16:23.789: INFO: stderr: "" +Jun 18 12:16:23.789: INFO: stdout: "replicationcontroller/update-demo-nautilus created\n" +STEP: waiting for all containers in name=update-demo pods to come up. +Jun 18 12:16:23.789: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=e2e-tests-kubectl-ltw4d' +Jun 18 12:16:23.932: INFO: stderr: "" +Jun 18 12:16:23.932: INFO: stdout: "update-demo-nautilus-7lslx update-demo-nautilus-fmm4k " +Jun 18 12:16:23.932: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 get pods update-demo-nautilus-7lslx -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=e2e-tests-kubectl-ltw4d' +Jun 18 12:16:24.146: INFO: stderr: "" +Jun 18 12:16:24.146: INFO: stdout: "" +Jun 18 12:16:24.146: INFO: update-demo-nautilus-7lslx is created but not running +Jun 18 12:16:29.146: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=e2e-tests-kubectl-ltw4d' +Jun 18 12:16:29.978: INFO: stderr: "" +Jun 18 12:16:29.978: INFO: stdout: "update-demo-nautilus-7lslx update-demo-nautilus-fmm4k " +Jun 18 12:16:29.978: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 get pods update-demo-nautilus-7lslx -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=e2e-tests-kubectl-ltw4d' +Jun 18 12:16:30.124: INFO: stderr: "" +Jun 18 12:16:30.124: INFO: stdout: "true" +Jun 18 12:16:30.124: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 get pods update-demo-nautilus-7lslx -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=e2e-tests-kubectl-ltw4d' +Jun 18 12:16:37.520: INFO: stderr: "" +Jun 18 12:16:37.520: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0" +Jun 18 12:16:37.520: INFO: validating pod update-demo-nautilus-7lslx +Jun 18 12:16:37.574: INFO: got data: { + "image": "nautilus.jpg" +} + +Jun 18 12:16:37.574: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . +Jun 18 12:16:37.574: INFO: update-demo-nautilus-7lslx is verified up and running +Jun 18 12:16:37.574: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 get pods update-demo-nautilus-fmm4k -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=e2e-tests-kubectl-ltw4d' +Jun 18 12:16:37.787: INFO: stderr: "" +Jun 18 12:16:37.787: INFO: stdout: "true" +Jun 18 12:16:37.787: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 get pods update-demo-nautilus-fmm4k -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=e2e-tests-kubectl-ltw4d' +Jun 18 12:16:37.946: INFO: stderr: "" +Jun 18 12:16:37.946: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0" +Jun 18 12:16:37.946: INFO: validating pod update-demo-nautilus-fmm4k +Jun 18 12:16:37.986: INFO: got data: { + "image": "nautilus.jpg" +} + +Jun 18 12:16:37.986: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . +Jun 18 12:16:37.986: INFO: update-demo-nautilus-fmm4k is verified up and running +STEP: rolling-update to new replication controller +Jun 18 12:16:37.988: INFO: scanned /root for discovery docs: +Jun 18 12:16:37.988: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 rolling-update update-demo-nautilus --update-period=1s -f - --namespace=e2e-tests-kubectl-ltw4d' +Jun 18 12:17:05.676: INFO: stderr: "Command \"rolling-update\" is deprecated, use \"rollout\" instead\n" +Jun 18 12:17:05.676: INFO: stdout: "Created update-demo-kitten\nScaling up update-demo-kitten from 0 to 2, scaling down update-demo-nautilus from 2 to 0 (keep 2 pods available, don't exceed 3 pods)\nScaling update-demo-kitten up to 1\nScaling update-demo-nautilus down to 1\nScaling update-demo-kitten up to 2\nScaling update-demo-nautilus down to 0\nUpdate succeeded. Deleting old controller: update-demo-nautilus\nRenaming update-demo-kitten to update-demo-nautilus\nreplicationcontroller/update-demo-nautilus rolling updated\n" +STEP: waiting for all containers in name=update-demo pods to come up. +Jun 18 12:17:05.676: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=e2e-tests-kubectl-ltw4d' +Jun 18 12:17:05.890: INFO: stderr: "" +Jun 18 12:17:05.890: INFO: stdout: "update-demo-kitten-mx6df update-demo-kitten-vgzp6 " +Jun 18 12:17:05.890: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 get pods update-demo-kitten-mx6df -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=e2e-tests-kubectl-ltw4d' +Jun 18 12:17:06.256: INFO: stderr: "" +Jun 18 12:17:06.257: INFO: stdout: "true" +Jun 18 12:17:06.257: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 get pods update-demo-kitten-mx6df -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=e2e-tests-kubectl-ltw4d' +Jun 18 12:17:06.400: INFO: stderr: "" +Jun 18 12:17:06.400: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/kitten:1.0" +Jun 18 12:17:06.400: INFO: validating pod update-demo-kitten-mx6df +Jun 18 12:17:06.433: INFO: got data: { + "image": "kitten.jpg" +} + +Jun 18 12:17:06.433: INFO: Unmarshalled json jpg/img => {kitten.jpg} , expecting kitten.jpg . +Jun 18 12:17:06.433: INFO: update-demo-kitten-mx6df is verified up and running +Jun 18 12:17:06.433: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 get pods update-demo-kitten-vgzp6 -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=e2e-tests-kubectl-ltw4d' +Jun 18 12:17:06.566: INFO: stderr: "" +Jun 18 12:17:06.566: INFO: stdout: "true" +Jun 18 12:17:06.566: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 get pods update-demo-kitten-vgzp6 -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=e2e-tests-kubectl-ltw4d' +Jun 18 12:17:06.711: INFO: stderr: "" +Jun 18 12:17:06.711: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/kitten:1.0" +Jun 18 12:17:06.711: INFO: validating pod update-demo-kitten-vgzp6 +Jun 18 12:17:06.745: INFO: got data: { + "image": "kitten.jpg" +} + +Jun 18 12:17:06.745: INFO: Unmarshalled json jpg/img => {kitten.jpg} , expecting kitten.jpg . +Jun 18 12:17:06.745: INFO: update-demo-kitten-vgzp6 is verified up and running [AfterEach] [sig-cli] Kubectl client /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 19:32:31.547: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-kubectl-rfb5b" for this suite. -May 13 19:32:37.596: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 19:32:37.708: INFO: namespace: e2e-tests-kubectl-rfb5b, resource: bindings, ignored listing per whitelist -May 13 19:32:37.849: INFO: namespace e2e-tests-kubectl-rfb5b deletion completed in 6.282383058s +Jun 18 12:17:06.745: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "e2e-tests-kubectl-ltw4d" for this suite. +Jun 18 12:17:32.812: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:17:33.180: INFO: namespace: e2e-tests-kubectl-ltw4d, resource: bindings, ignored listing per whitelist +Jun 18 12:17:33.751: INFO: namespace e2e-tests-kubectl-ltw4d deletion completed in 26.98677573s -• [SLOW TEST:11.074 seconds] +• [SLOW TEST:70.715 seconds] [sig-cli] Kubectl client /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:22 - [k8s.io] Kubectl run deployment + [k8s.io] Update Demo /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 - should create a deployment from an image [Conformance] + should do a rolling update of a replication controller [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ -S +SSSSSSSSSSSSSSSSSSSSSS +------------------------------ +[k8s.io] Probing container + should *not* be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance] + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 +[BeforeEach] [k8s.io] Probing container + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 +STEP: Creating a kubernetes client +Jun 18 12:17:33.752: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 +STEP: Building a namespace api object, basename container-probe +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-container-probe-t5xbb +STEP: Waiting for a default service account to be provisioned in namespace +[BeforeEach] [k8s.io] Probing container + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/container_probe.go:48 +[It] should *not* be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance] + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 +STEP: Creating pod liveness-exec in namespace e2e-tests-container-probe-t5xbb +Jun 18 12:17:36.295: INFO: Started pod liveness-exec in namespace e2e-tests-container-probe-t5xbb +STEP: checking the pod's current state and verifying that restartCount is present +Jun 18 12:17:36.309: INFO: Initial restart count of pod liveness-exec is 0 +STEP: deleting the pod +[AfterEach] [k8s.io] Probing container + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 +Jun 18 12:21:37.268: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "e2e-tests-container-probe-t5xbb" for this suite. +Jun 18 12:21:43.340: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:21:43.571: INFO: namespace: e2e-tests-container-probe-t5xbb, resource: bindings, ignored listing per whitelist +Jun 18 12:21:43.822: INFO: namespace e2e-tests-container-probe-t5xbb deletion completed in 6.534344079s + +• [SLOW TEST:250.070 seconds] +[k8s.io] Probing container +/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 + should *not* be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance] + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 +------------------------------ +SSSSS ------------------------------ [sig-storage] Secrets optional updates should be reflected in volume [NodeConformance] [Conformance] @@ -5310,28 +5074,28 @@ S [BeforeEach] [sig-storage] Secrets /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client -May 13 19:32:37.850: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 +Jun 18 12:21:43.824: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 STEP: Building a namespace api object, basename secrets -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-secrets-c9mkh +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-secrets-hsv6v STEP: Waiting for a default service account to be provisioned in namespace [It] optional updates should be reflected in volume [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -STEP: Creating secret with name s-test-opt-del-ddcc6929-75b5-11e9-8f67-2632f168be36 -STEP: Creating secret with name s-test-opt-upd-ddcc6975-75b5-11e9-8f67-2632f168be36 +STEP: Creating secret with name s-test-opt-del-a2a15f79-91c3-11e9-bce2-ae54e022189f +STEP: Creating secret with name s-test-opt-upd-a2a15fd9-91c3-11e9-bce2-ae54e022189f STEP: Creating the pod -STEP: Deleting secret s-test-opt-del-ddcc6929-75b5-11e9-8f67-2632f168be36 -STEP: Updating secret s-test-opt-upd-ddcc6975-75b5-11e9-8f67-2632f168be36 -STEP: Creating secret with name s-test-opt-create-ddcc6993-75b5-11e9-8f67-2632f168be36 +STEP: Deleting secret s-test-opt-del-a2a15f79-91c3-11e9-bce2-ae54e022189f +STEP: Updating secret s-test-opt-upd-a2a15fd9-91c3-11e9-bce2-ae54e022189f +STEP: Creating secret with name s-test-opt-create-a2a16005-91c3-11e9-bce2-ae54e022189f STEP: waiting to observe update in volume [AfterEach] [sig-storage] Secrets /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 19:34:06.082: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-secrets-c9mkh" for this suite. -May 13 19:34:30.117: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 19:34:30.385: INFO: namespace: e2e-tests-secrets-c9mkh, resource: bindings, ignored listing per whitelist -May 13 19:34:30.434: INFO: namespace e2e-tests-secrets-c9mkh deletion completed in 24.344089387s +Jun 18 12:23:13.850: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "e2e-tests-secrets-hsv6v" for this suite. +Jun 18 12:23:39.917: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:23:40.357: INFO: namespace: e2e-tests-secrets-hsv6v, resource: bindings, ignored listing per whitelist +Jun 18 12:23:40.430: INFO: namespace e2e-tests-secrets-hsv6v deletion completed in 26.561063679s -• [SLOW TEST:112.584 seconds] +• [SLOW TEST:116.606 seconds] [sig-storage] Secrets /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/secrets_volume.go:34 optional updates should be reflected in volume [NodeConformance] [Conformance] @@ -5339,2116 +5103,2157 @@ May 13 19:34:30.434: INFO: namespace e2e-tests-secrets-c9mkh deletion completed ------------------------------ S ------------------------------ -[sig-storage] Projected downwardAPI - should provide container's cpu request [NodeConformance] [Conformance] +[sig-network] Networking Granular Checks: Pods + should function for node-pod communication: http [NodeConformance] [Conformance] + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 +[BeforeEach] [sig-network] Networking + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 +STEP: Creating a kubernetes client +Jun 18 12:23:40.430: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 +STEP: Building a namespace api object, basename pod-network-test +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-pod-network-test-sxqbn +STEP: Waiting for a default service account to be provisioned in namespace +[It] should function for node-pod communication: http [NodeConformance] [Conformance] + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 +STEP: Performing setup for networking test in namespace e2e-tests-pod-network-test-sxqbn +STEP: creating a selector +STEP: Creating the service pods in kubernetes +Jun 18 12:23:40.883: INFO: Waiting up to 10m0s for all (but 0) nodes to be schedulable +STEP: Creating test pods +Jun 18 12:23:59.201: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s --max-time 15 --connect-timeout 1 http://172.30.39.53:8080/hostName | grep -v '^\s*$'] Namespace:e2e-tests-pod-network-test-sxqbn PodName:host-test-container-pod ContainerName:hostexec Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} +Jun 18 12:23:59.201: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 +Jun 18 12:23:59.449: INFO: Found all expected endpoints: [netserver-0] +Jun 18 12:23:59.464: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s --max-time 15 --connect-timeout 1 http://172.30.58.183:8080/hostName | grep -v '^\s*$'] Namespace:e2e-tests-pod-network-test-sxqbn PodName:host-test-container-pod ContainerName:hostexec Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} +Jun 18 12:23:59.464: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 +Jun 18 12:23:59.786: INFO: Found all expected endpoints: [netserver-1] +Jun 18 12:23:59.800: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s --max-time 15 --connect-timeout 1 http://172.30.114.35:8080/hostName | grep -v '^\s*$'] Namespace:e2e-tests-pod-network-test-sxqbn PodName:host-test-container-pod ContainerName:hostexec Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} +Jun 18 12:23:59.800: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 +Jun 18 12:24:00.049: INFO: Found all expected endpoints: [netserver-2] +[AfterEach] [sig-network] Networking + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 +Jun 18 12:24:00.049: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "e2e-tests-pod-network-test-sxqbn" for this suite. +Jun 18 12:24:24.118: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:24:24.549: INFO: namespace: e2e-tests-pod-network-test-sxqbn, resource: bindings, ignored listing per whitelist +Jun 18 12:24:24.613: INFO: namespace e2e-tests-pod-network-test-sxqbn deletion completed in 24.544492874s + +• [SLOW TEST:44.183 seconds] +[sig-network] Networking +/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:25 + Granular Checks: Pods + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:28 + should function for node-pod communication: http [NodeConformance] [Conformance] + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 +------------------------------ +SS +------------------------------ +[sig-scheduling] SchedulerPredicates [Serial] + validates resource limits of pods that are allowed to run [Conformance] + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 +[BeforeEach] [sig-scheduling] SchedulerPredicates [Serial] + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 +STEP: Creating a kubernetes client +Jun 18 12:24:24.614: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 +STEP: Building a namespace api object, basename sched-pred +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-sched-pred-hpfc8 +STEP: Waiting for a default service account to be provisioned in namespace +[BeforeEach] [sig-scheduling] SchedulerPredicates [Serial] + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:79 +Jun 18 12:24:25.047: INFO: Waiting up to 1m0s for all (but 0) nodes to be ready +Jun 18 12:24:25.076: INFO: Waiting for terminating namespaces to be deleted... +Jun 18 12:24:25.089: INFO: +Logging pods the kubelet thinks is on node 10.72.74.143 before test +Jun 18 12:24:25.140: INFO: sonobuoy-e2e-job-4f826760f7504668 from heptio-sonobuoy started at 2019-06-18 11:33:59 +0000 UTC (2 container statuses recorded) +Jun 18 12:24:25.140: INFO: Container e2e ready: true, restart count 0 +Jun 18 12:24:25.140: INFO: Container sonobuoy-worker ready: true, restart count 0 +Jun 18 12:24:25.140: INFO: ibm-kube-fluentd-7spm2 from kube-system started at 2019-06-17 21:43:21 +0000 UTC (1 container statuses recorded) +Jun 18 12:24:25.140: INFO: Container fluentd ready: true, restart count 0 +Jun 18 12:24:25.140: INFO: ibm-file-plugin-bf4cc7987-jwdjh from kube-system started at 2019-06-17 21:36:47 +0000 UTC (1 container statuses recorded) +Jun 18 12:24:25.140: INFO: Container ibm-file-plugin-container ready: true, restart count 0 +Jun 18 12:24:25.141: INFO: ibm-storage-watcher-64989c44d-tp68k from kube-system started at 2019-06-17 21:36:47 +0000 UTC (1 container statuses recorded) +Jun 18 12:24:25.141: INFO: Container ibm-storage-watcher-container ready: true, restart count 0 +Jun 18 12:24:25.141: INFO: sonobuoy from heptio-sonobuoy started at 2019-06-18 11:33:50 +0000 UTC (1 container statuses recorded) +Jun 18 12:24:25.141: INFO: Container kube-sonobuoy ready: true, restart count 0 +Jun 18 12:24:25.141: INFO: calico-node-fw2l9 from kube-system started at 2019-06-17 21:36:37 +0000 UTC (1 container statuses recorded) +Jun 18 12:24:25.141: INFO: Container calico-node ready: true, restart count 0 +Jun 18 12:24:25.141: INFO: coredns-autoscaler-5c7646547d-dshx6 from kube-system started at 2019-06-17 21:36:47 +0000 UTC (1 container statuses recorded) +Jun 18 12:24:25.141: INFO: Container autoscaler ready: true, restart count 0 +Jun 18 12:24:25.141: INFO: calico-kube-controllers-54d47c87f-kwkh9 from kube-system started at 2019-06-17 21:36:47 +0000 UTC (1 container statuses recorded) +Jun 18 12:24:25.141: INFO: Container calico-kube-controllers ready: true, restart count 0 +Jun 18 12:24:25.141: INFO: sonobuoy-systemd-logs-daemon-set-562f76bc52c447d0-zt85r from heptio-sonobuoy started at 2019-06-18 11:33:59 +0000 UTC (2 container statuses recorded) +Jun 18 12:24:25.141: INFO: Container sonobuoy-worker ready: true, restart count 0 +Jun 18 12:24:25.141: INFO: Container systemd-logs ready: true, restart count 0 +Jun 18 12:24:25.141: INFO: ibm-master-proxy-static-10.72.74.143 from kube-system started at (0 container statuses recorded) +Jun 18 12:24:25.141: INFO: ibm-keepalived-watcher-5z7h2 from kube-system started at 2019-06-17 21:36:37 +0000 UTC (1 container statuses recorded) +Jun 18 12:24:25.141: INFO: Container keepalived-watcher ready: true, restart count 0 +Jun 18 12:24:25.141: INFO: kubernetes-dashboard-6cf8b975c-prz8l from kube-system started at 2019-06-17 21:36:47 +0000 UTC (1 container statuses recorded) +Jun 18 12:24:25.141: INFO: Container kubernetes-dashboard ready: true, restart count 0 +Jun 18 12:24:25.141: INFO: coredns-5545c6ddc4-dxkvs from kube-system started at 2019-06-17 21:36:47 +0000 UTC (1 container statuses recorded) +Jun 18 12:24:25.141: INFO: Container coredns ready: true, restart count 0 +Jun 18 12:24:25.141: INFO: vpn-7f677b8cb5-29tf9 from kube-system started at 2019-06-17 21:36:47 +0000 UTC (1 container statuses recorded) +Jun 18 12:24:25.141: INFO: Container vpn ready: true, restart count 0 +Jun 18 12:24:25.141: INFO: test-k8s-e2e-pvg-master-verification from default started at 2019-06-18 11:33:43 +0000 UTC (1 container statuses recorded) +Jun 18 12:24:25.141: INFO: Container test-k8s-e2e-pvg-master-verification ready: true, restart count 0 +Jun 18 12:24:25.141: INFO: +Logging pods the kubelet thinks is on node 10.72.74.144 before test +Jun 18 12:24:25.195: INFO: ibm-cloud-provider-ip-158-176-120-130-699ff5cfd-z4hhb from ibm-system started at 2019-06-17 21:40:39 +0000 UTC (1 container statuses recorded) +Jun 18 12:24:25.195: INFO: Container ibm-cloud-provider-ip-158-176-120-130 ready: true, restart count 0 +Jun 18 12:24:25.195: INFO: public-cr49a3e8d7011b436d9b4596ba0f279008-alb1-778b7ff477-tpktg from kube-system started at 2019-06-17 21:41:04 +0000 UTC (4 container statuses recorded) +Jun 18 12:24:25.195: INFO: Container ingress-auth-1 ready: true, restart count 0 +Jun 18 12:24:25.195: INFO: Container ingress-auth-2 ready: true, restart count 0 +Jun 18 12:24:25.195: INFO: Container ingress-auth-3 ready: true, restart count 0 +Jun 18 12:24:25.195: INFO: Container nginx-ingress ready: true, restart count 0 +Jun 18 12:24:25.195: INFO: sonobuoy-systemd-logs-daemon-set-562f76bc52c447d0-ct76c from heptio-sonobuoy started at 2019-06-18 11:33:59 +0000 UTC (2 container statuses recorded) +Jun 18 12:24:25.195: INFO: Container sonobuoy-worker ready: true, restart count 0 +Jun 18 12:24:25.195: INFO: Container systemd-logs ready: true, restart count 0 +Jun 18 12:24:25.195: INFO: ibm-master-proxy-static-10.72.74.144 from kube-system started at (0 container statuses recorded) +Jun 18 12:24:25.195: INFO: calico-node-rptvs from kube-system started at 2019-06-17 21:36:43 +0000 UTC (1 container statuses recorded) +Jun 18 12:24:25.195: INFO: Container calico-node ready: true, restart count 0 +Jun 18 12:24:25.195: INFO: ibm-keepalived-watcher-drbmt from kube-system started at 2019-06-17 21:36:43 +0000 UTC (1 container statuses recorded) +Jun 18 12:24:25.195: INFO: Container keepalived-watcher ready: true, restart count 0 +Jun 18 12:24:25.195: INFO: coredns-5545c6ddc4-4s87g from kube-system started at 2019-06-17 21:37:04 +0000 UTC (1 container statuses recorded) +Jun 18 12:24:25.195: INFO: Container coredns ready: true, restart count 0 +Jun 18 12:24:25.195: INFO: ibm-kube-fluentd-g5hgb from kube-system started at 2019-06-17 21:43:21 +0000 UTC (1 container statuses recorded) +Jun 18 12:24:25.195: INFO: Container fluentd ready: true, restart count 0 +Jun 18 12:24:25.195: INFO: +Logging pods the kubelet thinks is on node 10.72.74.149 before test +Jun 18 12:24:25.254: INFO: ibm-master-proxy-static-10.72.74.149 from kube-system started at (0 container statuses recorded) +Jun 18 12:24:25.255: INFO: calico-node-4pqtj from kube-system started at 2019-06-17 21:36:50 +0000 UTC (1 container statuses recorded) +Jun 18 12:24:25.255: INFO: Container calico-node ready: true, restart count 0 +Jun 18 12:24:25.255: INFO: ibm-cloud-provider-ip-158-176-120-130-699ff5cfd-td8hg from ibm-system started at 2019-06-17 21:40:39 +0000 UTC (1 container statuses recorded) +Jun 18 12:24:25.255: INFO: Container ibm-cloud-provider-ip-158-176-120-130 ready: true, restart count 0 +Jun 18 12:24:25.255: INFO: public-cr49a3e8d7011b436d9b4596ba0f279008-alb1-778b7ff477-sxttq from kube-system started at 2019-06-17 21:41:04 +0000 UTC (4 container statuses recorded) +Jun 18 12:24:25.255: INFO: Container ingress-auth-1 ready: true, restart count 0 +Jun 18 12:24:25.255: INFO: Container ingress-auth-2 ready: true, restart count 0 +Jun 18 12:24:25.255: INFO: Container ingress-auth-3 ready: true, restart count 0 +Jun 18 12:24:25.255: INFO: Container nginx-ingress ready: true, restart count 0 +Jun 18 12:24:25.255: INFO: ibm-kube-fluentd-c6kth from kube-system started at 2019-06-17 21:43:21 +0000 UTC (1 container statuses recorded) +Jun 18 12:24:25.255: INFO: Container fluentd ready: true, restart count 0 +Jun 18 12:24:25.255: INFO: metrics-server-6ccf788d5b-6gwxm from kube-system started at 2019-06-17 21:37:11 +0000 UTC (2 container statuses recorded) +Jun 18 12:24:25.255: INFO: Container metrics-server ready: true, restart count 0 +Jun 18 12:24:25.255: INFO: Container metrics-server-nanny ready: true, restart count 0 +Jun 18 12:24:25.255: INFO: ibm-keepalived-watcher-6846v from kube-system started at 2019-06-17 21:36:50 +0000 UTC (1 container statuses recorded) +Jun 18 12:24:25.255: INFO: Container keepalived-watcher ready: true, restart count 0 +Jun 18 12:24:25.255: INFO: sonobuoy-systemd-logs-daemon-set-562f76bc52c447d0-btfpp from heptio-sonobuoy started at 2019-06-18 11:33:59 +0000 UTC (2 container statuses recorded) +Jun 18 12:24:25.255: INFO: Container sonobuoy-worker ready: true, restart count 0 +Jun 18 12:24:25.255: INFO: Container systemd-logs ready: true, restart count 0 +[It] validates resource limits of pods that are allowed to run [Conformance] + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 +STEP: verifying the node has the label node 10.72.74.143 +STEP: verifying the node has the label node 10.72.74.144 +STEP: verifying the node has the label node 10.72.74.149 +Jun 18 12:24:25.420: INFO: Pod test-k8s-e2e-pvg-master-verification requesting resource cpu=0m on Node 10.72.74.143 +Jun 18 12:24:25.420: INFO: Pod sonobuoy requesting resource cpu=0m on Node 10.72.74.143 +Jun 18 12:24:25.420: INFO: Pod sonobuoy-e2e-job-4f826760f7504668 requesting resource cpu=0m on Node 10.72.74.143 +Jun 18 12:24:25.420: INFO: Pod sonobuoy-systemd-logs-daemon-set-562f76bc52c447d0-btfpp requesting resource cpu=0m on Node 10.72.74.149 +Jun 18 12:24:25.420: INFO: Pod sonobuoy-systemd-logs-daemon-set-562f76bc52c447d0-ct76c requesting resource cpu=0m on Node 10.72.74.144 +Jun 18 12:24:25.420: INFO: Pod sonobuoy-systemd-logs-daemon-set-562f76bc52c447d0-zt85r requesting resource cpu=0m on Node 10.72.74.143 +Jun 18 12:24:25.420: INFO: Pod ibm-cloud-provider-ip-158-176-120-130-699ff5cfd-td8hg requesting resource cpu=5m on Node 10.72.74.149 +Jun 18 12:24:25.420: INFO: Pod ibm-cloud-provider-ip-158-176-120-130-699ff5cfd-z4hhb requesting resource cpu=5m on Node 10.72.74.144 +Jun 18 12:24:25.420: INFO: Pod calico-kube-controllers-54d47c87f-kwkh9 requesting resource cpu=10m on Node 10.72.74.143 +Jun 18 12:24:25.420: INFO: Pod calico-node-4pqtj requesting resource cpu=250m on Node 10.72.74.149 +Jun 18 12:24:25.420: INFO: Pod calico-node-fw2l9 requesting resource cpu=250m on Node 10.72.74.143 +Jun 18 12:24:25.420: INFO: Pod calico-node-rptvs requesting resource cpu=250m on Node 10.72.74.144 +Jun 18 12:24:25.420: INFO: Pod coredns-5545c6ddc4-4s87g requesting resource cpu=100m on Node 10.72.74.144 +Jun 18 12:24:25.420: INFO: Pod coredns-5545c6ddc4-dxkvs requesting resource cpu=100m on Node 10.72.74.143 +Jun 18 12:24:25.420: INFO: Pod coredns-autoscaler-5c7646547d-dshx6 requesting resource cpu=20m on Node 10.72.74.143 +Jun 18 12:24:25.420: INFO: Pod ibm-file-plugin-bf4cc7987-jwdjh requesting resource cpu=50m on Node 10.72.74.143 +Jun 18 12:24:25.420: INFO: Pod ibm-keepalived-watcher-5z7h2 requesting resource cpu=5m on Node 10.72.74.143 +Jun 18 12:24:25.420: INFO: Pod ibm-keepalived-watcher-6846v requesting resource cpu=5m on Node 10.72.74.149 +Jun 18 12:24:25.420: INFO: Pod ibm-keepalived-watcher-drbmt requesting resource cpu=5m on Node 10.72.74.144 +Jun 18 12:24:25.420: INFO: Pod ibm-kube-fluentd-7spm2 requesting resource cpu=25m on Node 10.72.74.143 +Jun 18 12:24:25.420: INFO: Pod ibm-kube-fluentd-c6kth requesting resource cpu=25m on Node 10.72.74.149 +Jun 18 12:24:25.420: INFO: Pod ibm-kube-fluentd-g5hgb requesting resource cpu=25m on Node 10.72.74.144 +Jun 18 12:24:25.420: INFO: Pod ibm-master-proxy-static-10.72.74.143 requesting resource cpu=25m on Node 10.72.74.143 +Jun 18 12:24:25.420: INFO: Pod ibm-master-proxy-static-10.72.74.144 requesting resource cpu=25m on Node 10.72.74.144 +Jun 18 12:24:25.420: INFO: Pod ibm-master-proxy-static-10.72.74.149 requesting resource cpu=25m on Node 10.72.74.149 +Jun 18 12:24:25.420: INFO: Pod ibm-storage-watcher-64989c44d-tp68k requesting resource cpu=50m on Node 10.72.74.143 +Jun 18 12:24:25.420: INFO: Pod kubernetes-dashboard-6cf8b975c-prz8l requesting resource cpu=50m on Node 10.72.74.143 +Jun 18 12:24:25.420: INFO: Pod metrics-server-6ccf788d5b-6gwxm requesting resource cpu=53m on Node 10.72.74.149 +Jun 18 12:24:25.420: INFO: Pod public-cr49a3e8d7011b436d9b4596ba0f279008-alb1-778b7ff477-sxttq requesting resource cpu=0m on Node 10.72.74.149 +Jun 18 12:24:25.420: INFO: Pod public-cr49a3e8d7011b436d9b4596ba0f279008-alb1-778b7ff477-tpktg requesting resource cpu=0m on Node 10.72.74.144 +Jun 18 12:24:25.420: INFO: Pod vpn-7f677b8cb5-29tf9 requesting resource cpu=5m on Node 10.72.74.143 +STEP: Starting Pods to consume most of the cluster CPU. +STEP: Creating another pod that requires unavailable amount of CPU. +STEP: Considering event: +Type = [Normal], Name = [filler-pod-0291545e-91c4-11e9-bce2-ae54e022189f.15a94a81651122ee], Reason = [Scheduled], Message = [Successfully assigned e2e-tests-sched-pred-hpfc8/filler-pod-0291545e-91c4-11e9-bce2-ae54e022189f to 10.72.74.144] +STEP: Considering event: +Type = [Normal], Name = [filler-pod-0291545e-91c4-11e9-bce2-ae54e022189f.15a94a81a5523b26], Reason = [Pulled], Message = [Container image "k8s.gcr.io/pause:3.1" already present on machine] +STEP: Considering event: +Type = [Normal], Name = [filler-pod-0291545e-91c4-11e9-bce2-ae54e022189f.15a94a81a8776faa], Reason = [Created], Message = [Created container] +STEP: Considering event: +Type = [Normal], Name = [filler-pod-0291545e-91c4-11e9-bce2-ae54e022189f.15a94a81b7c08230], Reason = [Started], Message = [Started container] +STEP: Considering event: +Type = [Normal], Name = [filler-pod-02961fdc-91c4-11e9-bce2-ae54e022189f.15a94a8166a0776e], Reason = [Scheduled], Message = [Successfully assigned e2e-tests-sched-pred-hpfc8/filler-pod-02961fdc-91c4-11e9-bce2-ae54e022189f to 10.72.74.149] +STEP: Considering event: +Type = [Normal], Name = [filler-pod-02961fdc-91c4-11e9-bce2-ae54e022189f.15a94a81a97c7ff9], Reason = [Pulled], Message = [Container image "k8s.gcr.io/pause:3.1" already present on machine] +STEP: Considering event: +Type = [Normal], Name = [filler-pod-02961fdc-91c4-11e9-bce2-ae54e022189f.15a94a81acfd0f37], Reason = [Created], Message = [Created container] +STEP: Considering event: +Type = [Normal], Name = [filler-pod-02961fdc-91c4-11e9-bce2-ae54e022189f.15a94a81bb7491ed], Reason = [Started], Message = [Started container] +STEP: Considering event: +Type = [Normal], Name = [filler-pod-0298b9a6-91c4-11e9-bce2-ae54e022189f.15a94a816792590a], Reason = [Scheduled], Message = [Successfully assigned e2e-tests-sched-pred-hpfc8/filler-pod-0298b9a6-91c4-11e9-bce2-ae54e022189f to 10.72.74.143] +STEP: Considering event: +Type = [Normal], Name = [filler-pod-0298b9a6-91c4-11e9-bce2-ae54e022189f.15a94a81a8ca7762], Reason = [Pulled], Message = [Container image "k8s.gcr.io/pause:3.1" already present on machine] +STEP: Considering event: +Type = [Normal], Name = [filler-pod-0298b9a6-91c4-11e9-bce2-ae54e022189f.15a94a81ace91807], Reason = [Created], Message = [Created container] +STEP: Considering event: +Type = [Normal], Name = [filler-pod-0298b9a6-91c4-11e9-bce2-ae54e022189f.15a94a81ba433cde], Reason = [Started], Message = [Started container] +STEP: Considering event: +Type = [Warning], Name = [additional-pod.15a94a81e3dcfc48], Reason = [FailedScheduling], Message = [0/3 nodes are available: 3 Insufficient cpu.] +STEP: removing the label node off the node 10.72.74.143 +STEP: verifying the node doesn't have the label node +STEP: removing the label node off the node 10.72.74.144 +STEP: verifying the node doesn't have the label node +STEP: removing the label node off the node 10.72.74.149 +STEP: verifying the node doesn't have the label node +[AfterEach] [sig-scheduling] SchedulerPredicates [Serial] + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 +Jun 18 12:24:28.743: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "e2e-tests-sched-pred-hpfc8" for this suite. +Jun 18 12:24:36.812: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:24:37.076: INFO: namespace: e2e-tests-sched-pred-hpfc8, resource: bindings, ignored listing per whitelist +Jun 18 12:24:37.422: INFO: namespace e2e-tests-sched-pred-hpfc8 deletion completed in 8.659467646s +[AfterEach] [sig-scheduling] SchedulerPredicates [Serial] + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:70 + +• [SLOW TEST:12.808 seconds] +[sig-scheduling] SchedulerPredicates [Serial] +/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/framework.go:22 + validates resource limits of pods that are allowed to run [Conformance] + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 +------------------------------ +SSSSSSSSS +------------------------------ +[sig-network] Networking Granular Checks: Pods + should function for node-pod communication: udp [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -[BeforeEach] [sig-storage] Projected downwardAPI +[BeforeEach] [sig-network] Networking /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client -May 13 19:34:30.434: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -STEP: Building a namespace api object, basename projected -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-projected-flv2d +Jun 18 12:24:37.423: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 +STEP: Building a namespace api object, basename pod-network-test +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-pod-network-test-j2k4q STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [sig-storage] Projected downwardAPI - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:39 -[It] should provide container's cpu request [NodeConformance] [Conformance] +[It] should function for node-pod communication: udp [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -STEP: Creating a pod to test downward API volume plugin -May 13 19:34:30.721: INFO: Waiting up to 5m0s for pod "downwardapi-volume-20d84759-75b6-11e9-8f67-2632f168be36" in namespace "e2e-tests-projected-flv2d" to be "success or failure" -May 13 19:34:30.731: INFO: Pod "downwardapi-volume-20d84759-75b6-11e9-8f67-2632f168be36": Phase="Pending", Reason="", readiness=false. Elapsed: 9.147675ms -May 13 19:34:32.739: INFO: Pod "downwardapi-volume-20d84759-75b6-11e9-8f67-2632f168be36": Phase="Running", Reason="", readiness=true. Elapsed: 2.017380564s -May 13 19:34:34.747: INFO: Pod "downwardapi-volume-20d84759-75b6-11e9-8f67-2632f168be36": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.025751229s -STEP: Saw pod success -May 13 19:34:34.747: INFO: Pod "downwardapi-volume-20d84759-75b6-11e9-8f67-2632f168be36" satisfied condition "success or failure" -May 13 19:34:34.754: INFO: Trying to get logs from node 10.170.219.153 pod downwardapi-volume-20d84759-75b6-11e9-8f67-2632f168be36 container client-container: -STEP: delete the pod -May 13 19:34:34.794: INFO: Waiting for pod downwardapi-volume-20d84759-75b6-11e9-8f67-2632f168be36 to disappear -May 13 19:34:34.801: INFO: Pod downwardapi-volume-20d84759-75b6-11e9-8f67-2632f168be36 no longer exists -[AfterEach] [sig-storage] Projected downwardAPI +STEP: Performing setup for networking test in namespace e2e-tests-pod-network-test-j2k4q +STEP: creating a selector +STEP: Creating the service pods in kubernetes +Jun 18 12:24:37.895: INFO: Waiting up to 10m0s for all (but 0) nodes to be schedulable +STEP: Creating test pods +Jun 18 12:25:04.267: INFO: ExecWithOptions {Command:[/bin/sh -c echo 'hostName' | nc -w 1 -u 172.30.58.185 8081 | grep -v '^\s*$'] Namespace:e2e-tests-pod-network-test-j2k4q PodName:host-test-container-pod ContainerName:hostexec Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} +Jun 18 12:25:04.267: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 +Jun 18 12:25:05.541: INFO: Found all expected endpoints: [netserver-0] +Jun 18 12:25:05.555: INFO: ExecWithOptions {Command:[/bin/sh -c echo 'hostName' | nc -w 1 -u 172.30.114.38 8081 | grep -v '^\s*$'] Namespace:e2e-tests-pod-network-test-j2k4q PodName:host-test-container-pod ContainerName:hostexec Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} +Jun 18 12:25:05.555: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 +Jun 18 12:25:06.802: INFO: Found all expected endpoints: [netserver-1] +Jun 18 12:25:06.818: INFO: ExecWithOptions {Command:[/bin/sh -c echo 'hostName' | nc -w 1 -u 172.30.39.55 8081 | grep -v '^\s*$'] Namespace:e2e-tests-pod-network-test-j2k4q PodName:host-test-container-pod ContainerName:hostexec Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} +Jun 18 12:25:06.818: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 +Jun 18 12:25:08.069: INFO: Found all expected endpoints: [netserver-2] +[AfterEach] [sig-network] Networking /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 19:34:34.801: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-projected-flv2d" for this suite. -May 13 19:34:40.839: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 19:34:40.989: INFO: namespace: e2e-tests-projected-flv2d, resource: bindings, ignored listing per whitelist -May 13 19:34:41.141: INFO: namespace e2e-tests-projected-flv2d deletion completed in 6.332352622s +Jun 18 12:25:08.069: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "e2e-tests-pod-network-test-j2k4q" for this suite. +Jun 18 12:25:32.143: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:25:32.547: INFO: namespace: e2e-tests-pod-network-test-j2k4q, resource: bindings, ignored listing per whitelist +Jun 18 12:25:32.679: INFO: namespace e2e-tests-pod-network-test-j2k4q deletion completed in 24.5896822s -• [SLOW TEST:10.707 seconds] -[sig-storage] Projected downwardAPI -/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:33 - should provide container's cpu request [NodeConformance] [Conformance] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------- -SSSSSSSSSSSS +• [SLOW TEST:55.256 seconds] +[sig-network] Networking +/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:25 + Granular Checks: Pods + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:28 + should function for node-pod communication: udp [NodeConformance] [Conformance] + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ [sig-storage] EmptyDir volumes - should support (root,0644,default) [NodeConformance] [Conformance] + should support (non-root,0777,default) [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-storage] EmptyDir volumes /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client -May 13 19:34:41.143: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 +Jun 18 12:25:32.679: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 STEP: Building a namespace api object, basename emptydir -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-emptydir-xwpmh +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-emptydir-rcb72 STEP: Waiting for a default service account to be provisioned in namespace -[It] should support (root,0644,default) [NodeConformance] [Conformance] +[It] should support (non-root,0777,default) [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -STEP: Creating a pod to test emptydir 0644 on node default medium -May 13 19:34:41.535: INFO: Waiting up to 5m0s for pod "pod-273c0f8d-75b6-11e9-8f67-2632f168be36" in namespace "e2e-tests-emptydir-xwpmh" to be "success or failure" -May 13 19:34:41.542: INFO: Pod "pod-273c0f8d-75b6-11e9-8f67-2632f168be36": Phase="Pending", Reason="", readiness=false. Elapsed: 7.503589ms -May 13 19:34:43.551: INFO: Pod "pod-273c0f8d-75b6-11e9-8f67-2632f168be36": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.016079158s +STEP: Creating a pod to test emptydir 0777 on node default medium +Jun 18 12:25:33.172: INFO: Waiting up to 5m0s for pod "pod-2aef12ce-91c4-11e9-bce2-ae54e022189f" in namespace "e2e-tests-emptydir-rcb72" to be "success or failure" +Jun 18 12:25:33.190: INFO: Pod "pod-2aef12ce-91c4-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 17.73837ms +Jun 18 12:25:35.210: INFO: Pod "pod-2aef12ce-91c4-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 2.037301392s +Jun 18 12:25:37.227: INFO: Pod "pod-2aef12ce-91c4-11e9-bce2-ae54e022189f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.054498212s STEP: Saw pod success -May 13 19:34:43.551: INFO: Pod "pod-273c0f8d-75b6-11e9-8f67-2632f168be36" satisfied condition "success or failure" -May 13 19:34:43.620: INFO: Trying to get logs from node 10.170.219.153 pod pod-273c0f8d-75b6-11e9-8f67-2632f168be36 container test-container: +Jun 18 12:25:37.227: INFO: Pod "pod-2aef12ce-91c4-11e9-bce2-ae54e022189f" satisfied condition "success or failure" +Jun 18 12:25:37.301: INFO: Trying to get logs from node 10.72.74.149 pod pod-2aef12ce-91c4-11e9-bce2-ae54e022189f container test-container: STEP: delete the pod -May 13 19:34:43.661: INFO: Waiting for pod pod-273c0f8d-75b6-11e9-8f67-2632f168be36 to disappear -May 13 19:34:43.669: INFO: Pod pod-273c0f8d-75b6-11e9-8f67-2632f168be36 no longer exists +Jun 18 12:25:37.377: INFO: Waiting for pod pod-2aef12ce-91c4-11e9-bce2-ae54e022189f to disappear +Jun 18 12:25:37.391: INFO: Pod pod-2aef12ce-91c4-11e9-bce2-ae54e022189f no longer exists [AfterEach] [sig-storage] EmptyDir volumes /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 19:34:43.669: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-emptydir-xwpmh" for this suite. -May 13 19:34:49.707: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 19:34:49.737: INFO: namespace: e2e-tests-emptydir-xwpmh, resource: bindings, ignored listing per whitelist -May 13 19:34:49.967: INFO: namespace e2e-tests-emptydir-xwpmh deletion completed in 6.289827803s +Jun 18 12:25:37.391: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "e2e-tests-emptydir-rcb72" for this suite. +Jun 18 12:25:43.472: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:25:43.638: INFO: namespace: e2e-tests-emptydir-rcb72, resource: bindings, ignored listing per whitelist +Jun 18 12:25:44.058: INFO: namespace e2e-tests-emptydir-rcb72 deletion completed in 6.643437675s -• [SLOW TEST:8.825 seconds] +• [SLOW TEST:11.380 seconds] [sig-storage] EmptyDir volumes /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:40 - should support (root,0644,default) [NodeConformance] [Conformance] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------- -SSSSSS ------------------------------- -[sig-api-machinery] Namespaces [Serial] - should ensure that all pods are removed when a namespace is deleted [Conformance] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -[BeforeEach] [sig-api-machinery] Namespaces [Serial] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 -STEP: Creating a kubernetes client -May 13 19:34:49.968: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -STEP: Building a namespace api object, basename namespaces -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-namespaces-crlgb -STEP: Waiting for a default service account to be provisioned in namespace -[It] should ensure that all pods are removed when a namespace is deleted [Conformance] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -STEP: Creating a test namespace -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-nsdeletetest-czlgx -STEP: Waiting for a default service account to be provisioned in namespace -STEP: Creating a pod in the namespace -STEP: Waiting for the pod to have running status -STEP: Creating an uninitialized pod in the namespace -STEP: Deleting the namespace -STEP: Waiting for the namespace to be removed. -May 13 19:34:59.720: INFO: error from create uninitialized namespace: Internal error occurred: object deleted while waiting for creation -STEP: Recreating the namespace -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-nsdeletetest-57j8x -STEP: Verifying there are no pods in the namespace -[AfterEach] [sig-api-machinery] Namespaces [Serial] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 19:35:17.756: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-namespaces-crlgb" for this suite. -May 13 19:35:23.792: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 19:35:24.219: INFO: namespace: e2e-tests-namespaces-crlgb, resource: bindings, ignored listing per whitelist -May 13 19:35:24.225: INFO: namespace e2e-tests-namespaces-crlgb deletion completed in 6.460832776s -STEP: Destroying namespace "e2e-tests-nsdeletetest-czlgx" for this suite. -May 13 19:35:24.232: INFO: Namespace e2e-tests-nsdeletetest-czlgx was already deleted -STEP: Destroying namespace "e2e-tests-nsdeletetest-57j8x" for this suite. -May 13 19:35:30.281: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 19:35:30.437: INFO: namespace: e2e-tests-nsdeletetest-57j8x, resource: bindings, ignored listing per whitelist -May 13 19:35:30.538: INFO: namespace e2e-tests-nsdeletetest-57j8x deletion completed in 6.305588102s - -• [SLOW TEST:40.570 seconds] -[sig-api-machinery] Namespaces [Serial] -/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:22 - should ensure that all pods are removed when a namespace is deleted [Conformance] + should support (non-root,0777,default) [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SS ------------------------------ -[sig-api-machinery] Garbage collector - should orphan RS created by deployment when deleteOptions.PropagationPolicy is Orphan [Conformance] +[k8s.io] [sig-node] Pods Extended [k8s.io] Pods Set QOS Class + should be submitted and removed [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -[BeforeEach] [sig-api-machinery] Garbage collector +[BeforeEach] [k8s.io] [sig-node] Pods Extended /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client -May 13 19:35:30.539: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -STEP: Building a namespace api object, basename gc -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-gc-kjz4j +Jun 18 12:25:44.058: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 +STEP: Building a namespace api object, basename pods +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-pods-4kl8k STEP: Waiting for a default service account to be provisioned in namespace -[It] should orphan RS created by deployment when deleteOptions.PropagationPolicy is Orphan [Conformance] +[BeforeEach] [k8s.io] Pods Set QOS Class + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/node/pods.go:204 +[It] should be submitted and removed [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -STEP: create the deployment -STEP: Wait for the Deployment to create new ReplicaSet -STEP: delete the deployment -STEP: wait for 30 seconds to see if the garbage collector mistakenly deletes the rs -STEP: Gathering metrics -W0513 19:36:01.421834 16 metrics_grabber.go:81] Master node is not registered. Grabbing metrics from Scheduler, ControllerManager and ClusterAutoscaler is disabled. -May 13 19:36:01.421: INFO: For apiserver_request_count: -For apiserver_request_latencies_summary: -For etcd_helper_cache_entry_count: -For etcd_helper_cache_hit_count: -For etcd_helper_cache_miss_count: -For etcd_request_cache_add_latencies_summary: -For etcd_request_cache_get_latencies_summary: -For etcd_request_latencies_summary: -For garbage_collector_attempt_to_delete_queue_latency: -For garbage_collector_attempt_to_delete_work_duration: -For garbage_collector_attempt_to_orphan_queue_latency: -For garbage_collector_attempt_to_orphan_work_duration: -For garbage_collector_dirty_processing_latency_microseconds: -For garbage_collector_event_processing_latency_microseconds: -For garbage_collector_graph_changes_queue_latency: -For garbage_collector_graph_changes_work_duration: -For garbage_collector_orphan_processing_latency_microseconds: -For namespace_queue_latency: -For namespace_queue_latency_sum: -For namespace_queue_latency_count: -For namespace_retries: -For namespace_work_duration: -For namespace_work_duration_sum: -For namespace_work_duration_count: -For function_duration_seconds: -For errors_total: -For evicted_pods_total: - -[AfterEach] [sig-api-machinery] Garbage collector +STEP: creating the pod +STEP: submitting the pod to kubernetes +STEP: verifying QOS class is set on the pod +[AfterEach] [k8s.io] [sig-node] Pods Extended /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 19:36:01.421: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-gc-kjz4j" for this suite. -May 13 19:36:07.457: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 19:36:07.515: INFO: namespace: e2e-tests-gc-kjz4j, resource: bindings, ignored listing per whitelist -May 13 19:36:09.391: INFO: namespace e2e-tests-gc-kjz4j deletion completed in 7.964354227s +Jun 18 12:25:44.554: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "e2e-tests-pods-4kl8k" for this suite. +Jun 18 12:26:08.697: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:26:08.886: INFO: namespace: e2e-tests-pods-4kl8k, resource: bindings, ignored listing per whitelist +Jun 18 12:26:09.319: INFO: namespace e2e-tests-pods-4kl8k deletion completed in 24.746909285s -• [SLOW TEST:38.853 seconds] -[sig-api-machinery] Garbage collector -/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:22 - should orphan RS created by deployment when deleteOptions.PropagationPolicy is Orphan [Conformance] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 +• [SLOW TEST:25.260 seconds] +[k8s.io] [sig-node] Pods Extended +/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 + [k8s.io] Pods Set QOS Class + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 + should be submitted and removed [Conformance] + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ -SSSS +S ------------------------------ -[k8s.io] Container Lifecycle Hook when create a pod with lifecycle hook - should execute prestop exec hook properly [NodeConformance] [Conformance] +[k8s.io] Docker Containers + should be able to override the image's default command and arguments [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -[BeforeEach] [k8s.io] Container Lifecycle Hook +[BeforeEach] [k8s.io] Docker Containers /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client -May 13 19:36:09.392: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -STEP: Building a namespace api object, basename container-lifecycle-hook -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-container-lifecycle-hook-sc86z +Jun 18 12:26:09.319: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 +STEP: Building a namespace api object, basename containers +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-containers-r9btl STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] when create a pod with lifecycle hook - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:61 -STEP: create the container to handle the HTTPGet hook request. -[It] should execute prestop exec hook properly [NodeConformance] [Conformance] +[It] should be able to override the image's default command and arguments [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -STEP: create the pod with lifecycle hook -STEP: delete the pod with lifecycle hook -May 13 19:36:13.772: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear -May 13 19:36:13.783: INFO: Pod pod-with-prestop-exec-hook still exists -May 13 19:36:15.783: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear -May 13 19:36:15.791: INFO: Pod pod-with-prestop-exec-hook still exists -May 13 19:36:17.783: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear -May 13 19:36:17.796: INFO: Pod pod-with-prestop-exec-hook still exists -May 13 19:36:19.783: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear -May 13 19:36:19.792: INFO: Pod pod-with-prestop-exec-hook still exists -May 13 19:36:21.783: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear -May 13 19:36:21.805: INFO: Pod pod-with-prestop-exec-hook still exists -May 13 19:36:23.783: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear -May 13 19:36:23.791: INFO: Pod pod-with-prestop-exec-hook still exists -May 13 19:36:25.783: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear -May 13 19:36:25.792: INFO: Pod pod-with-prestop-exec-hook still exists -May 13 19:36:27.783: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear -May 13 19:36:27.792: INFO: Pod pod-with-prestop-exec-hook still exists -May 13 19:36:29.783: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear -May 13 19:36:29.792: INFO: Pod pod-with-prestop-exec-hook still exists -May 13 19:36:31.783: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear -May 13 19:36:31.792: INFO: Pod pod-with-prestop-exec-hook still exists -May 13 19:36:33.783: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear -May 13 19:36:33.803: INFO: Pod pod-with-prestop-exec-hook still exists -May 13 19:36:35.783: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear -May 13 19:36:35.792: INFO: Pod pod-with-prestop-exec-hook still exists -May 13 19:36:37.783: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear -May 13 19:36:37.792: INFO: Pod pod-with-prestop-exec-hook still exists -May 13 19:36:39.783: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear -May 13 19:36:39.791: INFO: Pod pod-with-prestop-exec-hook still exists -May 13 19:36:41.783: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear -May 13 19:36:41.791: INFO: Pod pod-with-prestop-exec-hook no longer exists -STEP: check prestop hook -[AfterEach] [k8s.io] Container Lifecycle Hook +STEP: Creating a pod to test override all +Jun 18 12:26:09.937: INFO: Waiting up to 5m0s for pod "client-containers-40d94c72-91c4-11e9-bce2-ae54e022189f" in namespace "e2e-tests-containers-r9btl" to be "success or failure" +Jun 18 12:26:09.984: INFO: Pod "client-containers-40d94c72-91c4-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 47.075252ms +Jun 18 12:26:12.000: INFO: Pod "client-containers-40d94c72-91c4-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 2.063179342s +Jun 18 12:26:14.035: INFO: Pod "client-containers-40d94c72-91c4-11e9-bce2-ae54e022189f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.097338025s +STEP: Saw pod success +Jun 18 12:26:14.036: INFO: Pod "client-containers-40d94c72-91c4-11e9-bce2-ae54e022189f" satisfied condition "success or failure" +Jun 18 12:26:14.052: INFO: Trying to get logs from node 10.72.74.144 pod client-containers-40d94c72-91c4-11e9-bce2-ae54e022189f container test-container: +STEP: delete the pod +Jun 18 12:26:14.126: INFO: Waiting for pod client-containers-40d94c72-91c4-11e9-bce2-ae54e022189f to disappear +Jun 18 12:26:14.143: INFO: Pod client-containers-40d94c72-91c4-11e9-bce2-ae54e022189f no longer exists +[AfterEach] [k8s.io] Docker Containers /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 19:36:41.825: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-container-lifecycle-hook-sc86z" for this suite. -May 13 19:37:05.866: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 19:37:05.911: INFO: namespace: e2e-tests-container-lifecycle-hook-sc86z, resource: bindings, ignored listing per whitelist -May 13 19:37:06.115: INFO: namespace e2e-tests-container-lifecycle-hook-sc86z deletion completed in 24.278478072s +Jun 18 12:26:14.143: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "e2e-tests-containers-r9btl" for this suite. +Jun 18 12:26:20.209: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:26:20.293: INFO: namespace: e2e-tests-containers-r9btl, resource: bindings, ignored listing per whitelist +Jun 18 12:26:20.747: INFO: namespace e2e-tests-containers-r9btl deletion completed in 6.585084949s -• [SLOW TEST:56.724 seconds] -[k8s.io] Container Lifecycle Hook +• [SLOW TEST:11.428 seconds] +[k8s.io] Docker Containers /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 - when create a pod with lifecycle hook - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:40 - should execute prestop exec hook properly [NodeConformance] [Conformance] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 + should be able to override the image's default command and arguments [NodeConformance] [Conformance] + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ -[sig-storage] Projected secret - should be consumable from pods in volume with defaultMode set [NodeConformance] [Conformance] +SS +------------------------------ +[sig-apps] Deployment + deployment should support proportional scaling [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -[BeforeEach] [sig-storage] Projected secret +[BeforeEach] [sig-apps] Deployment /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client -May 13 19:37:06.116: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -STEP: Building a namespace api object, basename projected -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-projected-7fgn8 +Jun 18 12:26:20.748: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 +STEP: Building a namespace api object, basename deployment +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-deployment-jzshx STEP: Waiting for a default service account to be provisioned in namespace -[It] should be consumable from pods in volume with defaultMode set [NodeConformance] [Conformance] +[BeforeEach] [sig-apps] Deployment + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:65 +[It] deployment should support proportional scaling [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -STEP: Creating projection with secret that has name projected-secret-test-7da50e3a-75b6-11e9-8f67-2632f168be36 -STEP: Creating a pod to test consume secrets -May 13 19:37:06.424: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-7da6874a-75b6-11e9-8f67-2632f168be36" in namespace "e2e-tests-projected-7fgn8" to be "success or failure" -May 13 19:37:06.433: INFO: Pod "pod-projected-secrets-7da6874a-75b6-11e9-8f67-2632f168be36": Phase="Pending", Reason="", readiness=false. Elapsed: 9.101074ms -May 13 19:37:08.441: INFO: Pod "pod-projected-secrets-7da6874a-75b6-11e9-8f67-2632f168be36": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.017144216s -STEP: Saw pod success -May 13 19:37:08.441: INFO: Pod "pod-projected-secrets-7da6874a-75b6-11e9-8f67-2632f168be36" satisfied condition "success or failure" -May 13 19:37:08.448: INFO: Trying to get logs from node 10.170.219.153 pod pod-projected-secrets-7da6874a-75b6-11e9-8f67-2632f168be36 container projected-secret-volume-test: -STEP: delete the pod -May 13 19:37:08.493: INFO: Waiting for pod pod-projected-secrets-7da6874a-75b6-11e9-8f67-2632f168be36 to disappear -May 13 19:37:08.530: INFO: Pod pod-projected-secrets-7da6874a-75b6-11e9-8f67-2632f168be36 no longer exists -[AfterEach] [sig-storage] Projected secret +Jun 18 12:26:21.290: INFO: Creating deployment "nginx-deployment" +Jun 18 12:26:21.305: INFO: Waiting for observed generation 1 +Jun 18 12:26:23.332: INFO: Waiting for all required pods to come up +Jun 18 12:26:23.358: INFO: Pod name nginx: Found 10 pods out of 10 +STEP: ensuring each pod is running +Jun 18 12:26:25.485: INFO: Waiting for deployment "nginx-deployment" to complete +Jun 18 12:26:25.512: INFO: Updating deployment "nginx-deployment" with a non-existent image +Jun 18 12:26:25.537: INFO: Updating deployment nginx-deployment +Jun 18 12:26:25.537: INFO: Waiting for observed generation 2 +Jun 18 12:26:27.564: INFO: Waiting for the first rollout's replicaset to have .status.availableReplicas = 8 +Jun 18 12:26:27.577: INFO: Waiting for the first rollout's replicaset to have .spec.replicas = 8 +Jun 18 12:26:27.590: INFO: Waiting for the first rollout's replicaset of deployment "nginx-deployment" to have desired number of replicas +Jun 18 12:26:27.636: INFO: Verifying that the second rollout's replicaset has .status.availableReplicas = 0 +Jun 18 12:26:27.636: INFO: Waiting for the second rollout's replicaset to have .spec.replicas = 5 +Jun 18 12:26:27.648: INFO: Waiting for the second rollout's replicaset of deployment "nginx-deployment" to have desired number of replicas +Jun 18 12:26:27.675: INFO: Verifying that deployment "nginx-deployment" has minimum required number of available replicas +Jun 18 12:26:27.675: INFO: Scaling up the deployment "nginx-deployment" from 10 to 30 +Jun 18 12:26:27.717: INFO: Updating deployment nginx-deployment +Jun 18 12:26:27.717: INFO: Waiting for the replicasets of deployment "nginx-deployment" to have desired number of replicas +Jun 18 12:26:27.742: INFO: Verifying that first rollout's replicaset has .spec.replicas = 20 +Jun 18 12:26:29.770: INFO: Verifying that second rollout's replicaset has .spec.replicas = 13 +[AfterEach] [sig-apps] Deployment + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:59 +Jun 18 12:26:29.797: INFO: Deployment "nginx-deployment": +&Deployment{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment,GenerateName:,Namespace:e2e-tests-deployment-jzshx,SelfLink:/apis/apps/v1/namespaces/e2e-tests-deployment-jzshx/deployments/nginx-deployment,UID:47a2e90c-91c4-11e9-bf44-fa6f350b29f0,ResourceVersion:100836,Generation:3,CreationTimestamp:2019-06-18 12:26:21 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,},Annotations:map[string]string{deployment.kubernetes.io/revision: 2,},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Spec:DeploymentSpec{Replicas:*30,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: nginx,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[],Containers:[{nginx nginx:404 [] [] [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,},},Strategy:DeploymentStrategy{Type:RollingUpdate,RollingUpdate:&RollingUpdateDeployment{MaxUnavailable:2,MaxSurge:3,},},MinReadySeconds:0,RevisionHistoryLimit:*10,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:3,Replicas:33,UpdatedReplicas:13,AvailableReplicas:8,UnavailableReplicas:25,Conditions:[{Available False 2019-06-18 12:26:27 +0000 UTC 2019-06-18 12:26:27 +0000 UTC MinimumReplicasUnavailable Deployment does not have minimum availability.} {Progressing True 2019-06-18 12:26:27 +0000 UTC 2019-06-18 12:26:21 +0000 UTC ReplicaSetUpdated ReplicaSet "nginx-deployment-65bbdb5f8" is progressing.}],ReadyReplicas:8,CollisionCount:nil,},} + +Jun 18 12:26:29.813: INFO: New ReplicaSet "nginx-deployment-65bbdb5f8" of Deployment "nginx-deployment": +&ReplicaSet{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-65bbdb5f8,GenerateName:,Namespace:e2e-tests-deployment-jzshx,SelfLink:/apis/apps/v1/namespaces/e2e-tests-deployment-jzshx/replicasets/nginx-deployment-65bbdb5f8,UID:4a2a98be-91c4-11e9-bf44-fa6f350b29f0,ResourceVersion:100819,Generation:3,CreationTimestamp:2019-06-18 12:26:25 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 65bbdb5f8,},Annotations:map[string]string{deployment.kubernetes.io/desired-replicas: 30,deployment.kubernetes.io/max-replicas: 33,deployment.kubernetes.io/revision: 2,},OwnerReferences:[{apps/v1 Deployment nginx-deployment 47a2e90c-91c4-11e9-bf44-fa6f350b29f0 0xc001ed3ef7 0xc001ed3ef8}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:ReplicaSetSpec{Replicas:*13,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: nginx,pod-template-hash: 65bbdb5f8,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 65bbdb5f8,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[],Containers:[{nginx nginx:404 [] [] [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,},},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:13,FullyLabeledReplicas:13,ObservedGeneration:3,ReadyReplicas:0,AvailableReplicas:0,Conditions:[],},} +Jun 18 12:26:29.813: INFO: All old ReplicaSets of Deployment "nginx-deployment": +Jun 18 12:26:29.813: INFO: &ReplicaSet{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-555b55d965,GenerateName:,Namespace:e2e-tests-deployment-jzshx,SelfLink:/apis/apps/v1/namespaces/e2e-tests-deployment-jzshx/replicasets/nginx-deployment-555b55d965,UID:47a72cb1-91c4-11e9-bf44-fa6f350b29f0,ResourceVersion:100833,Generation:3,CreationTimestamp:2019-06-18 12:26:21 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 555b55d965,},Annotations:map[string]string{deployment.kubernetes.io/desired-replicas: 30,deployment.kubernetes.io/max-replicas: 33,deployment.kubernetes.io/revision: 1,},OwnerReferences:[{apps/v1 Deployment nginx-deployment 47a2e90c-91c4-11e9-bf44-fa6f350b29f0 0xc001ed3e37 0xc001ed3e38}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:ReplicaSetSpec{Replicas:*20,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: nginx,pod-template-hash: 555b55d965,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 555b55d965,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,},},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:20,FullyLabeledReplicas:20,ObservedGeneration:3,ReadyReplicas:8,AvailableReplicas:8,Conditions:[],},} +Jun 18 12:26:29.842: INFO: Pod "nginx-deployment-555b55d965-5bj56" is available: +&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-555b55d965-5bj56,GenerateName:nginx-deployment-555b55d965-,Namespace:e2e-tests-deployment-jzshx,SelfLink:/api/v1/namespaces/e2e-tests-deployment-jzshx/pods/nginx-deployment-555b55d965-5bj56,UID:47b0f76a-91c4-11e9-bf44-fa6f350b29f0,ResourceVersion:100642,Generation:0,CreationTimestamp:2019-06-18 12:26:21 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 555b55d965,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-555b55d965 47a72cb1-91c4-11e9-bf44-fa6f350b29f0 0xc002278c27 0xc002278c28}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-g4nvn {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-g4nvn,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-g4nvn true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.72.74.149,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc002279d20} {node.kubernetes.io/unreachable Exists NoExecute 0xc002279d40}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Running,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:21 +0000 UTC } {Ready True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:24 +0000 UTC } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:24 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:21 +0000 UTC }],Message:,Reason:,HostIP:10.72.74.149,PodIP:172.30.39.57,StartTime:2019-06-18 12:26:21 +0000 UTC,ContainerStatuses:[{nginx {nil ContainerStateRunning{StartedAt:2019-06-18 12:26:23 +0000 UTC,} nil} {nil nil nil} true 0 docker.io/library/nginx:1.14-alpine docker.io/library/nginx@sha256:485b610fefec7ff6c463ced9623314a04ed67e3945b9c08d7e53a47f6d108dc7 containerd://01b4097e1fbf7970ccee4f476cc9c6c91971f93c1c422b8b1179ddedb4f1b915}],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} +Jun 18 12:26:29.842: INFO: Pod "nginx-deployment-555b55d965-6b9lc" is not available: +&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-555b55d965-6b9lc,GenerateName:nginx-deployment-555b55d965-,Namespace:e2e-tests-deployment-jzshx,SelfLink:/api/v1/namespaces/e2e-tests-deployment-jzshx/pods/nginx-deployment-555b55d965-6b9lc,UID:4b7a70f5-91c4-11e9-bf44-fa6f350b29f0,ResourceVersion:100797,Generation:0,CreationTimestamp:2019-06-18 12:26:27 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 555b55d965,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-555b55d965 47a72cb1-91c4-11e9-bf44-fa6f350b29f0 0xc0012da037 0xc0012da038}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-g4nvn {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-g4nvn,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-g4nvn true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.72.74.144,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc0012da0b0} {node.kubernetes.io/unreachable Exists NoExecute 0xc0012da0d0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:27 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:27 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:27 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:27 +0000 UTC }],Message:,Reason:,HostIP:10.72.74.144,PodIP:,StartTime:2019-06-18 12:26:27 +0000 UTC,ContainerStatuses:[{nginx {ContainerStateWaiting{Reason:ContainerCreating,Message:,} nil nil} {nil nil nil} false 0 docker.io/library/nginx:1.14-alpine }],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} +Jun 18 12:26:29.843: INFO: Pod "nginx-deployment-555b55d965-6fd5f" is available: +&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-555b55d965-6fd5f,GenerateName:nginx-deployment-555b55d965-,Namespace:e2e-tests-deployment-jzshx,SelfLink:/api/v1/namespaces/e2e-tests-deployment-jzshx/pods/nginx-deployment-555b55d965-6fd5f,UID:47ae7d83-91c4-11e9-bf44-fa6f350b29f0,ResourceVersion:100631,Generation:0,CreationTimestamp:2019-06-18 12:26:21 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 555b55d965,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-555b55d965 47a72cb1-91c4-11e9-bf44-fa6f350b29f0 0xc0012da187 0xc0012da188}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-g4nvn {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-g4nvn,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-g4nvn true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.72.74.143,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc0012da200} {node.kubernetes.io/unreachable Exists NoExecute 0xc0012da220}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Running,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:21 +0000 UTC } {Ready True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:23 +0000 UTC } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:23 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:21 +0000 UTC }],Message:,Reason:,HostIP:10.72.74.143,PodIP:172.30.58.188,StartTime:2019-06-18 12:26:21 +0000 UTC,ContainerStatuses:[{nginx {nil ContainerStateRunning{StartedAt:2019-06-18 12:26:23 +0000 UTC,} nil} {nil nil nil} true 0 docker.io/library/nginx:1.14-alpine docker.io/library/nginx@sha256:485b610fefec7ff6c463ced9623314a04ed67e3945b9c08d7e53a47f6d108dc7 containerd://33fa3eb4726d71d8e3075e986f7e8eb1483205a2a6b99fcdffdff6267ced9209}],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} +Jun 18 12:26:29.843: INFO: Pod "nginx-deployment-555b55d965-74q7h" is not available: +&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-555b55d965-74q7h,GenerateName:nginx-deployment-555b55d965-,Namespace:e2e-tests-deployment-jzshx,SelfLink:/api/v1/namespaces/e2e-tests-deployment-jzshx/pods/nginx-deployment-555b55d965-74q7h,UID:4b7fcdec-91c4-11e9-bf44-fa6f350b29f0,ResourceVersion:100889,Generation:0,CreationTimestamp:2019-06-18 12:26:27 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 555b55d965,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-555b55d965 47a72cb1-91c4-11e9-bf44-fa6f350b29f0 0xc0012da2e7 0xc0012da2e8}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-g4nvn {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-g4nvn,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-g4nvn true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.72.74.143,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc0012da360} {node.kubernetes.io/unreachable Exists NoExecute 0xc0012da380}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:27 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:27 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:27 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:27 +0000 UTC }],Message:,Reason:,HostIP:10.72.74.143,PodIP:,StartTime:2019-06-18 12:26:27 +0000 UTC,ContainerStatuses:[{nginx {ContainerStateWaiting{Reason:ContainerCreating,Message:,} nil nil} {nil nil nil} false 0 docker.io/library/nginx:1.14-alpine }],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} +Jun 18 12:26:29.843: INFO: Pod "nginx-deployment-555b55d965-7tgt2" is not available: +&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-555b55d965-7tgt2,GenerateName:nginx-deployment-555b55d965-,Namespace:e2e-tests-deployment-jzshx,SelfLink:/api/v1/namespaces/e2e-tests-deployment-jzshx/pods/nginx-deployment-555b55d965-7tgt2,UID:4b8043bf-91c4-11e9-bf44-fa6f350b29f0,ResourceVersion:100870,Generation:0,CreationTimestamp:2019-06-18 12:26:27 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 555b55d965,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-555b55d965 47a72cb1-91c4-11e9-bf44-fa6f350b29f0 0xc0012da437 0xc0012da438}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-g4nvn {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-g4nvn,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-g4nvn true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.72.74.149,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc0012da4b0} {node.kubernetes.io/unreachable Exists NoExecute 0xc0012da4d0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:27 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:27 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:27 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:27 +0000 UTC }],Message:,Reason:,HostIP:10.72.74.149,PodIP:,StartTime:2019-06-18 12:26:27 +0000 UTC,ContainerStatuses:[{nginx {ContainerStateWaiting{Reason:ContainerCreating,Message:,} nil nil} {nil nil nil} false 0 docker.io/library/nginx:1.14-alpine }],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} +Jun 18 12:26:29.843: INFO: Pod "nginx-deployment-555b55d965-848zb" is not available: +&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-555b55d965-848zb,GenerateName:nginx-deployment-555b55d965-,Namespace:e2e-tests-deployment-jzshx,SelfLink:/api/v1/namespaces/e2e-tests-deployment-jzshx/pods/nginx-deployment-555b55d965-848zb,UID:4b800bce-91c4-11e9-bf44-fa6f350b29f0,ResourceVersion:100834,Generation:0,CreationTimestamp:2019-06-18 12:26:27 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 555b55d965,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-555b55d965 47a72cb1-91c4-11e9-bf44-fa6f350b29f0 0xc0012da587 0xc0012da588}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-g4nvn {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-g4nvn,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-g4nvn true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.72.74.143,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc0012da600} {node.kubernetes.io/unreachable Exists NoExecute 0xc0012da620}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:27 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:27 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:27 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:27 +0000 UTC }],Message:,Reason:,HostIP:10.72.74.143,PodIP:,StartTime:2019-06-18 12:26:27 +0000 UTC,ContainerStatuses:[{nginx {ContainerStateWaiting{Reason:ContainerCreating,Message:,} nil nil} {nil nil nil} false 0 docker.io/library/nginx:1.14-alpine }],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} +Jun 18 12:26:29.843: INFO: Pod "nginx-deployment-555b55d965-8wkq4" is not available: +&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-555b55d965-8wkq4,GenerateName:nginx-deployment-555b55d965-,Namespace:e2e-tests-deployment-jzshx,SelfLink:/api/v1/namespaces/e2e-tests-deployment-jzshx/pods/nginx-deployment-555b55d965-8wkq4,UID:4b7a727e-91c4-11e9-bf44-fa6f350b29f0,ResourceVersion:100822,Generation:0,CreationTimestamp:2019-06-18 12:26:27 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 555b55d965,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-555b55d965 47a72cb1-91c4-11e9-bf44-fa6f350b29f0 0xc0012da6d7 0xc0012da6d8}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-g4nvn {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-g4nvn,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-g4nvn true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.72.74.149,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc0012da760} {node.kubernetes.io/unreachable Exists NoExecute 0xc0012da780}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:27 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:27 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:27 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:27 +0000 UTC }],Message:,Reason:,HostIP:10.72.74.149,PodIP:,StartTime:2019-06-18 12:26:27 +0000 UTC,ContainerStatuses:[{nginx {ContainerStateWaiting{Reason:ContainerCreating,Message:,} nil nil} {nil nil nil} false 0 docker.io/library/nginx:1.14-alpine }],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} +Jun 18 12:26:29.843: INFO: Pod "nginx-deployment-555b55d965-cppb2" is not available: +&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-555b55d965-cppb2,GenerateName:nginx-deployment-555b55d965-,Namespace:e2e-tests-deployment-jzshx,SelfLink:/api/v1/namespaces/e2e-tests-deployment-jzshx/pods/nginx-deployment-555b55d965-cppb2,UID:4b789bd1-91c4-11e9-bf44-fa6f350b29f0,ResourceVersion:100780,Generation:0,CreationTimestamp:2019-06-18 12:26:27 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 555b55d965,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-555b55d965 47a72cb1-91c4-11e9-bf44-fa6f350b29f0 0xc0012da837 0xc0012da838}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-g4nvn {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-g4nvn,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-g4nvn true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.72.74.143,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc0012da8b0} {node.kubernetes.io/unreachable Exists NoExecute 0xc0012da8d0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:27 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:27 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:27 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:27 +0000 UTC }],Message:,Reason:,HostIP:10.72.74.143,PodIP:,StartTime:2019-06-18 12:26:27 +0000 UTC,ContainerStatuses:[{nginx {ContainerStateWaiting{Reason:ContainerCreating,Message:,} nil nil} {nil nil nil} false 0 docker.io/library/nginx:1.14-alpine }],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} +Jun 18 12:26:29.843: INFO: Pod "nginx-deployment-555b55d965-hkjgv" is available: +&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-555b55d965-hkjgv,GenerateName:nginx-deployment-555b55d965-,Namespace:e2e-tests-deployment-jzshx,SelfLink:/api/v1/namespaces/e2e-tests-deployment-jzshx/pods/nginx-deployment-555b55d965-hkjgv,UID:47b340de-91c4-11e9-bf44-fa6f350b29f0,ResourceVersion:100648,Generation:0,CreationTimestamp:2019-06-18 12:26:21 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 555b55d965,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-555b55d965 47a72cb1-91c4-11e9-bf44-fa6f350b29f0 0xc0012da987 0xc0012da988}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-g4nvn {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-g4nvn,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-g4nvn true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.72.74.149,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc0012daa00} {node.kubernetes.io/unreachable Exists NoExecute 0xc0012daa20}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Running,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:21 +0000 UTC } {Ready True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:24 +0000 UTC } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:24 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:21 +0000 UTC }],Message:,Reason:,HostIP:10.72.74.149,PodIP:172.30.39.58,StartTime:2019-06-18 12:26:21 +0000 UTC,ContainerStatuses:[{nginx {nil ContainerStateRunning{StartedAt:2019-06-18 12:26:23 +0000 UTC,} nil} {nil nil nil} true 0 docker.io/library/nginx:1.14-alpine docker.io/library/nginx@sha256:485b610fefec7ff6c463ced9623314a04ed67e3945b9c08d7e53a47f6d108dc7 containerd://4adad8f0224e913cf784f81655b4d207efa8c1a6de6b3b813110cd9a411dfcb4}],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} +Jun 18 12:26:29.844: INFO: Pod "nginx-deployment-555b55d965-hxlbp" is not available: +&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-555b55d965-hxlbp,GenerateName:nginx-deployment-555b55d965-,Namespace:e2e-tests-deployment-jzshx,SelfLink:/api/v1/namespaces/e2e-tests-deployment-jzshx/pods/nginx-deployment-555b55d965-hxlbp,UID:4b7d35cd-91c4-11e9-bf44-fa6f350b29f0,ResourceVersion:100854,Generation:0,CreationTimestamp:2019-06-18 12:26:27 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 555b55d965,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-555b55d965 47a72cb1-91c4-11e9-bf44-fa6f350b29f0 0xc0012daae7 0xc0012daae8}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-g4nvn {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-g4nvn,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-g4nvn true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.72.74.144,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc0012dab60} {node.kubernetes.io/unreachable Exists NoExecute 0xc0012dab80}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:27 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:27 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:27 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:27 +0000 UTC }],Message:,Reason:,HostIP:10.72.74.144,PodIP:,StartTime:2019-06-18 12:26:27 +0000 UTC,ContainerStatuses:[{nginx {ContainerStateWaiting{Reason:ContainerCreating,Message:,} nil nil} {nil nil nil} false 0 docker.io/library/nginx:1.14-alpine }],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} +Jun 18 12:26:29.844: INFO: Pod "nginx-deployment-555b55d965-jr74f" is not available: +&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-555b55d965-jr74f,GenerateName:nginx-deployment-555b55d965-,Namespace:e2e-tests-deployment-jzshx,SelfLink:/api/v1/namespaces/e2e-tests-deployment-jzshx/pods/nginx-deployment-555b55d965-jr74f,UID:4b807ddf-91c4-11e9-bf44-fa6f350b29f0,ResourceVersion:100808,Generation:0,CreationTimestamp:2019-06-18 12:26:27 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 555b55d965,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-555b55d965 47a72cb1-91c4-11e9-bf44-fa6f350b29f0 0xc0012dac37 0xc0012dac38}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-g4nvn {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-g4nvn,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-g4nvn true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.72.74.144,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc0012dacb0} {node.kubernetes.io/unreachable Exists NoExecute 0xc0012dacd0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:27 +0000 UTC }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} +Jun 18 12:26:29.844: INFO: Pod "nginx-deployment-555b55d965-kvlp7" is not available: +&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-555b55d965-kvlp7,GenerateName:nginx-deployment-555b55d965-,Namespace:e2e-tests-deployment-jzshx,SelfLink:/api/v1/namespaces/e2e-tests-deployment-jzshx/pods/nginx-deployment-555b55d965-kvlp7,UID:4b7d339b-91c4-11e9-bf44-fa6f350b29f0,ResourceVersion:100827,Generation:0,CreationTimestamp:2019-06-18 12:26:27 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 555b55d965,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-555b55d965 47a72cb1-91c4-11e9-bf44-fa6f350b29f0 0xc0012dad40 0xc0012dad41}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-g4nvn {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-g4nvn,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-g4nvn true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.72.74.144,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc0012dadb0} {node.kubernetes.io/unreachable Exists NoExecute 0xc0012dadd0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:27 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:27 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:27 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:27 +0000 UTC }],Message:,Reason:,HostIP:10.72.74.144,PodIP:,StartTime:2019-06-18 12:26:27 +0000 UTC,ContainerStatuses:[{nginx {ContainerStateWaiting{Reason:ContainerCreating,Message:,} nil nil} {nil nil nil} false 0 docker.io/library/nginx:1.14-alpine }],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} +Jun 18 12:26:29.844: INFO: Pod "nginx-deployment-555b55d965-lc57v" is available: +&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-555b55d965-lc57v,GenerateName:nginx-deployment-555b55d965-,Namespace:e2e-tests-deployment-jzshx,SelfLink:/api/v1/namespaces/e2e-tests-deployment-jzshx/pods/nginx-deployment-555b55d965-lc57v,UID:47b3246f-91c4-11e9-bf44-fa6f350b29f0,ResourceVersion:100643,Generation:0,CreationTimestamp:2019-06-18 12:26:21 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 555b55d965,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-555b55d965 47a72cb1-91c4-11e9-bf44-fa6f350b29f0 0xc0012dae87 0xc0012dae88}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-g4nvn {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-g4nvn,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-g4nvn true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.72.74.144,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc0012daf00} {node.kubernetes.io/unreachable Exists NoExecute 0xc0012daf20}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Running,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:21 +0000 UTC } {Ready True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:24 +0000 UTC } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:24 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:21 +0000 UTC }],Message:,Reason:,HostIP:10.72.74.144,PodIP:172.30.114.40,StartTime:2019-06-18 12:26:21 +0000 UTC,ContainerStatuses:[{nginx {nil ContainerStateRunning{StartedAt:2019-06-18 12:26:23 +0000 UTC,} nil} {nil nil nil} true 0 docker.io/library/nginx:1.14-alpine docker.io/library/nginx@sha256:485b610fefec7ff6c463ced9623314a04ed67e3945b9c08d7e53a47f6d108dc7 containerd://395d7e2b2a551c4ecaa72c76fa93bfa940f8a39c3906a344795569ab9d9a7d56}],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} +Jun 18 12:26:29.844: INFO: Pod "nginx-deployment-555b55d965-nq2gt" is not available: +&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-555b55d965-nq2gt,GenerateName:nginx-deployment-555b55d965-,Namespace:e2e-tests-deployment-jzshx,SelfLink:/api/v1/namespaces/e2e-tests-deployment-jzshx/pods/nginx-deployment-555b55d965-nq2gt,UID:4b7cecb9-91c4-11e9-bf44-fa6f350b29f0,ResourceVersion:100830,Generation:0,CreationTimestamp:2019-06-18 12:26:27 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 555b55d965,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-555b55d965 47a72cb1-91c4-11e9-bf44-fa6f350b29f0 0xc0012dafe7 0xc0012dafe8}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-g4nvn {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-g4nvn,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-g4nvn true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.72.74.149,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc0012db070} {node.kubernetes.io/unreachable Exists NoExecute 0xc0012db090}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:27 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:27 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:27 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:27 +0000 UTC }],Message:,Reason:,HostIP:10.72.74.149,PodIP:,StartTime:2019-06-18 12:26:27 +0000 UTC,ContainerStatuses:[{nginx {ContainerStateWaiting{Reason:ContainerCreating,Message:,} nil nil} {nil nil nil} false 0 docker.io/library/nginx:1.14-alpine }],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} +Jun 18 12:26:29.844: INFO: Pod "nginx-deployment-555b55d965-p9wr9" is available: +&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-555b55d965-p9wr9,GenerateName:nginx-deployment-555b55d965-,Namespace:e2e-tests-deployment-jzshx,SelfLink:/api/v1/namespaces/e2e-tests-deployment-jzshx/pods/nginx-deployment-555b55d965-p9wr9,UID:47b0db8a-91c4-11e9-bf44-fa6f350b29f0,ResourceVersion:100633,Generation:0,CreationTimestamp:2019-06-18 12:26:21 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 555b55d965,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-555b55d965 47a72cb1-91c4-11e9-bf44-fa6f350b29f0 0xc0012db147 0xc0012db148}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-g4nvn {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-g4nvn,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-g4nvn true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.72.74.143,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc0012db1c0} {node.kubernetes.io/unreachable Exists NoExecute 0xc0012db1e0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Running,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:21 +0000 UTC } {Ready True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:23 +0000 UTC } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:23 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:21 +0000 UTC }],Message:,Reason:,HostIP:10.72.74.143,PodIP:172.30.58.189,StartTime:2019-06-18 12:26:21 +0000 UTC,ContainerStatuses:[{nginx {nil ContainerStateRunning{StartedAt:2019-06-18 12:26:23 +0000 UTC,} nil} {nil nil nil} true 0 docker.io/library/nginx:1.14-alpine docker.io/library/nginx@sha256:485b610fefec7ff6c463ced9623314a04ed67e3945b9c08d7e53a47f6d108dc7 containerd://717f7979042dac3f0c280a3b344980260a99d3cb4f53c82d3906b2b6f560294e}],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} +Jun 18 12:26:29.844: INFO: Pod "nginx-deployment-555b55d965-qltqg" is available: +&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-555b55d965-qltqg,GenerateName:nginx-deployment-555b55d965-,Namespace:e2e-tests-deployment-jzshx,SelfLink:/api/v1/namespaces/e2e-tests-deployment-jzshx/pods/nginx-deployment-555b55d965-qltqg,UID:47b0c684-91c4-11e9-bf44-fa6f350b29f0,ResourceVersion:100640,Generation:0,CreationTimestamp:2019-06-18 12:26:21 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 555b55d965,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-555b55d965 47a72cb1-91c4-11e9-bf44-fa6f350b29f0 0xc0012db2a7 0xc0012db2a8}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-g4nvn {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-g4nvn,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-g4nvn true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.72.74.144,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc0012db320} {node.kubernetes.io/unreachable Exists NoExecute 0xc0012db340}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Running,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:21 +0000 UTC } {Ready True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:24 +0000 UTC } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:24 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:21 +0000 UTC }],Message:,Reason:,HostIP:10.72.74.144,PodIP:172.30.114.41,StartTime:2019-06-18 12:26:21 +0000 UTC,ContainerStatuses:[{nginx {nil ContainerStateRunning{StartedAt:2019-06-18 12:26:23 +0000 UTC,} nil} {nil nil nil} true 0 docker.io/library/nginx:1.14-alpine docker.io/library/nginx@sha256:485b610fefec7ff6c463ced9623314a04ed67e3945b9c08d7e53a47f6d108dc7 containerd://cb7d9f910d3de18a607384e08c982d116fcb849811477017dd117f0c827014e6}],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} +Jun 18 12:26:29.845: INFO: Pod "nginx-deployment-555b55d965-slxr9" is available: +&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-555b55d965-slxr9,GenerateName:nginx-deployment-555b55d965-,Namespace:e2e-tests-deployment-jzshx,SelfLink:/api/v1/namespaces/e2e-tests-deployment-jzshx/pods/nginx-deployment-555b55d965-slxr9,UID:47acc168-91c4-11e9-bf44-fa6f350b29f0,ResourceVersion:100653,Generation:0,CreationTimestamp:2019-06-18 12:26:21 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 555b55d965,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-555b55d965 47a72cb1-91c4-11e9-bf44-fa6f350b29f0 0xc0012db407 0xc0012db408}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-g4nvn {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-g4nvn,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-g4nvn true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.72.74.149,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc0012db480} {node.kubernetes.io/unreachable Exists NoExecute 0xc0012db4a0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Running,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:21 +0000 UTC } {Ready True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:24 +0000 UTC } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:24 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:21 +0000 UTC }],Message:,Reason:,HostIP:10.72.74.149,PodIP:172.30.39.60,StartTime:2019-06-18 12:26:21 +0000 UTC,ContainerStatuses:[{nginx {nil ContainerStateRunning{StartedAt:2019-06-18 12:26:24 +0000 UTC,} nil} {nil nil nil} true 0 docker.io/library/nginx:1.14-alpine docker.io/library/nginx@sha256:485b610fefec7ff6c463ced9623314a04ed67e3945b9c08d7e53a47f6d108dc7 containerd://f60b6e26dfc4f0de54d7dddec6ca3a97d49002f3c94cb679d66508eb2284a01c}],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} +Jun 18 12:26:29.845: INFO: Pod "nginx-deployment-555b55d965-w6mz4" is available: +&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-555b55d965-w6mz4,GenerateName:nginx-deployment-555b55d965-,Namespace:e2e-tests-deployment-jzshx,SelfLink:/api/v1/namespaces/e2e-tests-deployment-jzshx/pods/nginx-deployment-555b55d965-w6mz4,UID:47aeb70d-91c4-11e9-bf44-fa6f350b29f0,ResourceVersion:100647,Generation:0,CreationTimestamp:2019-06-18 12:26:21 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 555b55d965,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-555b55d965 47a72cb1-91c4-11e9-bf44-fa6f350b29f0 0xc0012db567 0xc0012db568}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-g4nvn {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-g4nvn,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-g4nvn true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.72.74.144,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc0012db5e0} {node.kubernetes.io/unreachable Exists NoExecute 0xc0012db600}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Running,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:21 +0000 UTC } {Ready True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:24 +0000 UTC } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:24 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:21 +0000 UTC }],Message:,Reason:,HostIP:10.72.74.144,PodIP:172.30.114.42,StartTime:2019-06-18 12:26:21 +0000 UTC,ContainerStatuses:[{nginx {nil ContainerStateRunning{StartedAt:2019-06-18 12:26:23 +0000 UTC,} nil} {nil nil nil} true 0 docker.io/library/nginx:1.14-alpine docker.io/library/nginx@sha256:485b610fefec7ff6c463ced9623314a04ed67e3945b9c08d7e53a47f6d108dc7 containerd://f94cab596649fefe951931b63560d02c5cd5c8ed32af9ab0fc63ce3724260b9b}],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} +Jun 18 12:26:29.845: INFO: Pod "nginx-deployment-555b55d965-xwfpx" is not available: +&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-555b55d965-xwfpx,GenerateName:nginx-deployment-555b55d965-,Namespace:e2e-tests-deployment-jzshx,SelfLink:/api/v1/namespaces/e2e-tests-deployment-jzshx/pods/nginx-deployment-555b55d965-xwfpx,UID:4b7fdb9c-91c4-11e9-bf44-fa6f350b29f0,ResourceVersion:100899,Generation:0,CreationTimestamp:2019-06-18 12:26:27 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 555b55d965,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-555b55d965 47a72cb1-91c4-11e9-bf44-fa6f350b29f0 0xc0012db6c7 0xc0012db6c8}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-g4nvn {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-g4nvn,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-g4nvn true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.72.74.149,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc0012db740} {node.kubernetes.io/unreachable Exists NoExecute 0xc0012db760}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:27 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:27 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:27 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:27 +0000 UTC }],Message:,Reason:,HostIP:10.72.74.149,PodIP:,StartTime:2019-06-18 12:26:27 +0000 UTC,ContainerStatuses:[{nginx {ContainerStateWaiting{Reason:ContainerCreating,Message:,} nil nil} {nil nil nil} false 0 docker.io/library/nginx:1.14-alpine }],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} +Jun 18 12:26:29.845: INFO: Pod "nginx-deployment-555b55d965-zhw7h" is not available: +&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-555b55d965-zhw7h,GenerateName:nginx-deployment-555b55d965-,Namespace:e2e-tests-deployment-jzshx,SelfLink:/api/v1/namespaces/e2e-tests-deployment-jzshx/pods/nginx-deployment-555b55d965-zhw7h,UID:4b7d05fd-91c4-11e9-bf44-fa6f350b29f0,ResourceVersion:100821,Generation:0,CreationTimestamp:2019-06-18 12:26:27 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 555b55d965,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-555b55d965 47a72cb1-91c4-11e9-bf44-fa6f350b29f0 0xc0012db817 0xc0012db818}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-g4nvn {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-g4nvn,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-g4nvn true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.72.74.143,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc0012db8a0} {node.kubernetes.io/unreachable Exists NoExecute 0xc0012db8c0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:27 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:27 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:27 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:27 +0000 UTC }],Message:,Reason:,HostIP:10.72.74.143,PodIP:,StartTime:2019-06-18 12:26:27 +0000 UTC,ContainerStatuses:[{nginx {ContainerStateWaiting{Reason:ContainerCreating,Message:,} nil nil} {nil nil nil} false 0 docker.io/library/nginx:1.14-alpine }],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} +Jun 18 12:26:29.845: INFO: Pod "nginx-deployment-65bbdb5f8-457mh" is not available: +&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-65bbdb5f8-457mh,GenerateName:nginx-deployment-65bbdb5f8-,Namespace:e2e-tests-deployment-jzshx,SelfLink:/api/v1/namespaces/e2e-tests-deployment-jzshx/pods/nginx-deployment-65bbdb5f8-457mh,UID:4a2e83f7-91c4-11e9-bf44-fa6f350b29f0,ResourceVersion:100690,Generation:0,CreationTimestamp:2019-06-18 12:26:25 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 65bbdb5f8,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-65bbdb5f8 4a2a98be-91c4-11e9-bf44-fa6f350b29f0 0xc0012db977 0xc0012db978}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-g4nvn {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-g4nvn,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] [] [] [] [] {map[] map[]} [{default-token-g4nvn true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.72.74.143,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc0012db9f0} {node.kubernetes.io/unreachable Exists NoExecute 0xc0012dba20}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:25 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:25 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:25 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:25 +0000 UTC }],Message:,Reason:,HostIP:10.72.74.143,PodIP:,StartTime:2019-06-18 12:26:25 +0000 UTC,ContainerStatuses:[{nginx {ContainerStateWaiting{Reason:ContainerCreating,Message:,} nil nil} {nil nil nil} false 0 nginx:404 }],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} +Jun 18 12:26:29.845: INFO: Pod "nginx-deployment-65bbdb5f8-5vl8m" is not available: +&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-65bbdb5f8-5vl8m,GenerateName:nginx-deployment-65bbdb5f8-,Namespace:e2e-tests-deployment-jzshx,SelfLink:/api/v1/namespaces/e2e-tests-deployment-jzshx/pods/nginx-deployment-65bbdb5f8-5vl8m,UID:4a2e95c8-91c4-11e9-bf44-fa6f350b29f0,ResourceVersion:100691,Generation:0,CreationTimestamp:2019-06-18 12:26:25 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 65bbdb5f8,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-65bbdb5f8 4a2a98be-91c4-11e9-bf44-fa6f350b29f0 0xc0012dbaf0 0xc0012dbaf1}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-g4nvn {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-g4nvn,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] [] [] [] [] {map[] map[]} [{default-token-g4nvn true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.72.74.144,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc0012dbb70} {node.kubernetes.io/unreachable Exists NoExecute 0xc0012dbb90}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:25 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:25 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:25 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:25 +0000 UTC }],Message:,Reason:,HostIP:10.72.74.144,PodIP:,StartTime:2019-06-18 12:26:25 +0000 UTC,ContainerStatuses:[{nginx {ContainerStateWaiting{Reason:ContainerCreating,Message:,} nil nil} {nil nil nil} false 0 nginx:404 }],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} +Jun 18 12:26:29.845: INFO: Pod "nginx-deployment-65bbdb5f8-95khm" is not available: +&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-65bbdb5f8-95khm,GenerateName:nginx-deployment-65bbdb5f8-,Namespace:e2e-tests-deployment-jzshx,SelfLink:/api/v1/namespaces/e2e-tests-deployment-jzshx/pods/nginx-deployment-65bbdb5f8-95khm,UID:4a384751-91c4-11e9-bf44-fa6f350b29f0,ResourceVersion:100714,Generation:0,CreationTimestamp:2019-06-18 12:26:25 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 65bbdb5f8,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-65bbdb5f8 4a2a98be-91c4-11e9-bf44-fa6f350b29f0 0xc0012dbc50 0xc0012dbc51}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-g4nvn {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-g4nvn,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] [] [] [] [] {map[] map[]} [{default-token-g4nvn true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.72.74.149,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc0012dbcd0} {node.kubernetes.io/unreachable Exists NoExecute 0xc0012dbcf0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:25 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:25 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:25 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:25 +0000 UTC }],Message:,Reason:,HostIP:10.72.74.149,PodIP:,StartTime:2019-06-18 12:26:25 +0000 UTC,ContainerStatuses:[{nginx {ContainerStateWaiting{Reason:ContainerCreating,Message:,} nil nil} {nil nil nil} false 0 nginx:404 }],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} +Jun 18 12:26:29.846: INFO: Pod "nginx-deployment-65bbdb5f8-98hkp" is not available: +&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-65bbdb5f8-98hkp,GenerateName:nginx-deployment-65bbdb5f8-,Namespace:e2e-tests-deployment-jzshx,SelfLink:/api/v1/namespaces/e2e-tests-deployment-jzshx/pods/nginx-deployment-65bbdb5f8-98hkp,UID:4a2c6631-91c4-11e9-bf44-fa6f350b29f0,ResourceVersion:100687,Generation:0,CreationTimestamp:2019-06-18 12:26:25 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 65bbdb5f8,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-65bbdb5f8 4a2a98be-91c4-11e9-bf44-fa6f350b29f0 0xc0012dbde0 0xc0012dbde1}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-g4nvn {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-g4nvn,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] [] [] [] [] {map[] map[]} [{default-token-g4nvn true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.72.74.149,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc0012dbe70} {node.kubernetes.io/unreachable Exists NoExecute 0xc0012dbe90}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:25 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:25 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:25 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:25 +0000 UTC }],Message:,Reason:,HostIP:10.72.74.149,PodIP:,StartTime:2019-06-18 12:26:25 +0000 UTC,ContainerStatuses:[{nginx {ContainerStateWaiting{Reason:ContainerCreating,Message:,} nil nil} {nil nil nil} false 0 nginx:404 }],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} +Jun 18 12:26:29.846: INFO: Pod "nginx-deployment-65bbdb5f8-bh87k" is not available: +&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-65bbdb5f8-bh87k,GenerateName:nginx-deployment-65bbdb5f8-,Namespace:e2e-tests-deployment-jzshx,SelfLink:/api/v1/namespaces/e2e-tests-deployment-jzshx/pods/nginx-deployment-65bbdb5f8-bh87k,UID:4b821eb4-91c4-11e9-bf44-fa6f350b29f0,ResourceVersion:100812,Generation:0,CreationTimestamp:2019-06-18 12:26:27 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 65bbdb5f8,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-65bbdb5f8 4a2a98be-91c4-11e9-bf44-fa6f350b29f0 0xc0012dbf60 0xc0012dbf61}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-g4nvn {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-g4nvn,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] [] [] [] [] {map[] map[]} [{default-token-g4nvn true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.72.74.144,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc0012dbfe0} {node.kubernetes.io/unreachable Exists NoExecute 0xc00195e060}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:27 +0000 UTC }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} +Jun 18 12:26:29.846: INFO: Pod "nginx-deployment-65bbdb5f8-br7m9" is not available: +&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-65bbdb5f8-br7m9,GenerateName:nginx-deployment-65bbdb5f8-,Namespace:e2e-tests-deployment-jzshx,SelfLink:/api/v1/namespaces/e2e-tests-deployment-jzshx/pods/nginx-deployment-65bbdb5f8-br7m9,UID:4b7f9287-91c4-11e9-bf44-fa6f350b29f0,ResourceVersion:100828,Generation:0,CreationTimestamp:2019-06-18 12:26:27 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 65bbdb5f8,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-65bbdb5f8 4a2a98be-91c4-11e9-bf44-fa6f350b29f0 0xc00195e0d0 0xc00195e0d1}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-g4nvn {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-g4nvn,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] [] [] [] [] {map[] map[]} [{default-token-g4nvn true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.72.74.143,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc00195e1c0} {node.kubernetes.io/unreachable Exists NoExecute 0xc00195e1e0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:27 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:27 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:27 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:27 +0000 UTC }],Message:,Reason:,HostIP:10.72.74.143,PodIP:,StartTime:2019-06-18 12:26:27 +0000 UTC,ContainerStatuses:[{nginx {ContainerStateWaiting{Reason:ContainerCreating,Message:,} nil nil} {nil nil nil} false 0 nginx:404 }],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} +Jun 18 12:26:29.846: INFO: Pod "nginx-deployment-65bbdb5f8-cj62g" is not available: +&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-65bbdb5f8-cj62g,GenerateName:nginx-deployment-65bbdb5f8-,Namespace:e2e-tests-deployment-jzshx,SelfLink:/api/v1/namespaces/e2e-tests-deployment-jzshx/pods/nginx-deployment-65bbdb5f8-cj62g,UID:4b7fcfa6-91c4-11e9-bf44-fa6f350b29f0,ResourceVersion:100864,Generation:0,CreationTimestamp:2019-06-18 12:26:27 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 65bbdb5f8,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-65bbdb5f8 4a2a98be-91c4-11e9-bf44-fa6f350b29f0 0xc00195e310 0xc00195e311}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-g4nvn {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-g4nvn,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] [] [] [] [] {map[] map[]} [{default-token-g4nvn true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.72.74.143,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc00195e390} {node.kubernetes.io/unreachable Exists NoExecute 0xc00195e5a0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:27 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:27 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:27 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:27 +0000 UTC }],Message:,Reason:,HostIP:10.72.74.143,PodIP:,StartTime:2019-06-18 12:26:27 +0000 UTC,ContainerStatuses:[{nginx {ContainerStateWaiting{Reason:ContainerCreating,Message:,} nil nil} {nil nil nil} false 0 nginx:404 }],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} +Jun 18 12:26:29.846: INFO: Pod "nginx-deployment-65bbdb5f8-fwb76" is not available: +&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-65bbdb5f8-fwb76,GenerateName:nginx-deployment-65bbdb5f8-,Namespace:e2e-tests-deployment-jzshx,SelfLink:/api/v1/namespaces/e2e-tests-deployment-jzshx/pods/nginx-deployment-65bbdb5f8-fwb76,UID:4b7a2ca5-91c4-11e9-bf44-fa6f350b29f0,ResourceVersion:100802,Generation:0,CreationTimestamp:2019-06-18 12:26:27 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 65bbdb5f8,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-65bbdb5f8 4a2a98be-91c4-11e9-bf44-fa6f350b29f0 0xc00195e660 0xc00195e661}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-g4nvn {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-g4nvn,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] [] [] [] [] {map[] map[]} [{default-token-g4nvn true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.72.74.149,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc00195e6e0} {node.kubernetes.io/unreachable Exists NoExecute 0xc00195e7e0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:27 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:27 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:27 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:27 +0000 UTC }],Message:,Reason:,HostIP:10.72.74.149,PodIP:,StartTime:2019-06-18 12:26:27 +0000 UTC,ContainerStatuses:[{nginx {ContainerStateWaiting{Reason:ContainerCreating,Message:,} nil nil} {nil nil nil} false 0 nginx:404 }],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} +Jun 18 12:26:29.847: INFO: Pod "nginx-deployment-65bbdb5f8-h6rnn" is not available: +&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-65bbdb5f8-h6rnn,GenerateName:nginx-deployment-65bbdb5f8-,Namespace:e2e-tests-deployment-jzshx,SelfLink:/api/v1/namespaces/e2e-tests-deployment-jzshx/pods/nginx-deployment-65bbdb5f8-h6rnn,UID:4b7c8636-91c4-11e9-bf44-fa6f350b29f0,ResourceVersion:100807,Generation:0,CreationTimestamp:2019-06-18 12:26:27 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 65bbdb5f8,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-65bbdb5f8 4a2a98be-91c4-11e9-bf44-fa6f350b29f0 0xc00195e8a0 0xc00195e8a1}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-g4nvn {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-g4nvn,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] [] [] [] [] {map[] map[]} [{default-token-g4nvn true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.72.74.143,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc00195e9d0} {node.kubernetes.io/unreachable Exists NoExecute 0xc00195e9f0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:27 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:27 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:27 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:27 +0000 UTC }],Message:,Reason:,HostIP:10.72.74.143,PodIP:,StartTime:2019-06-18 12:26:27 +0000 UTC,ContainerStatuses:[{nginx {ContainerStateWaiting{Reason:ContainerCreating,Message:,} nil nil} {nil nil nil} false 0 nginx:404 }],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} +Jun 18 12:26:29.847: INFO: Pod "nginx-deployment-65bbdb5f8-scd8r" is not available: +&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-65bbdb5f8-scd8r,GenerateName:nginx-deployment-65bbdb5f8-,Namespace:e2e-tests-deployment-jzshx,SelfLink:/api/v1/namespaces/e2e-tests-deployment-jzshx/pods/nginx-deployment-65bbdb5f8-scd8r,UID:4b7fa7b3-91c4-11e9-bf44-fa6f350b29f0,ResourceVersion:100887,Generation:0,CreationTimestamp:2019-06-18 12:26:27 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 65bbdb5f8,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-65bbdb5f8 4a2a98be-91c4-11e9-bf44-fa6f350b29f0 0xc00195eb00 0xc00195eb01}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-g4nvn {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-g4nvn,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] [] [] [] [] {map[] map[]} [{default-token-g4nvn true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.72.74.144,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc00195f010} {node.kubernetes.io/unreachable Exists NoExecute 0xc00195f030}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:27 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:27 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:27 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:27 +0000 UTC }],Message:,Reason:,HostIP:10.72.74.144,PodIP:,StartTime:2019-06-18 12:26:27 +0000 UTC,ContainerStatuses:[{nginx {ContainerStateWaiting{Reason:ContainerCreating,Message:,} nil nil} {nil nil nil} false 0 nginx:404 }],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} +Jun 18 12:26:29.847: INFO: Pod "nginx-deployment-65bbdb5f8-shgr2" is not available: +&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-65bbdb5f8-shgr2,GenerateName:nginx-deployment-65bbdb5f8-,Namespace:e2e-tests-deployment-jzshx,SelfLink:/api/v1/namespaces/e2e-tests-deployment-jzshx/pods/nginx-deployment-65bbdb5f8-shgr2,UID:4b7fb6fc-91c4-11e9-bf44-fa6f350b29f0,ResourceVersion:100837,Generation:0,CreationTimestamp:2019-06-18 12:26:27 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 65bbdb5f8,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-65bbdb5f8 4a2a98be-91c4-11e9-bf44-fa6f350b29f0 0xc00195f410 0xc00195f411}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-g4nvn {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-g4nvn,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] [] [] [] [] {map[] map[]} [{default-token-g4nvn true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.72.74.149,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc00195f490} {node.kubernetes.io/unreachable Exists NoExecute 0xc00195f4b0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:27 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:27 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:27 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:27 +0000 UTC }],Message:,Reason:,HostIP:10.72.74.149,PodIP:,StartTime:2019-06-18 12:26:27 +0000 UTC,ContainerStatuses:[{nginx {ContainerStateWaiting{Reason:ContainerCreating,Message:,} nil nil} {nil nil nil} false 0 nginx:404 }],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} +Jun 18 12:26:29.847: INFO: Pod "nginx-deployment-65bbdb5f8-ssccf" is not available: +&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-65bbdb5f8-ssccf,GenerateName:nginx-deployment-65bbdb5f8-,Namespace:e2e-tests-deployment-jzshx,SelfLink:/api/v1/namespaces/e2e-tests-deployment-jzshx/pods/nginx-deployment-65bbdb5f8-ssccf,UID:4b7d003e-91c4-11e9-bf44-fa6f350b29f0,ResourceVersion:100818,Generation:0,CreationTimestamp:2019-06-18 12:26:27 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 65bbdb5f8,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-65bbdb5f8 4a2a98be-91c4-11e9-bf44-fa6f350b29f0 0xc00195f570 0xc00195f571}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-g4nvn {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-g4nvn,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] [] [] [] [] {map[] map[]} [{default-token-g4nvn true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.72.74.144,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc00195f620} {node.kubernetes.io/unreachable Exists NoExecute 0xc00195f670}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:27 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:27 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:27 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:27 +0000 UTC }],Message:,Reason:,HostIP:10.72.74.144,PodIP:,StartTime:2019-06-18 12:26:27 +0000 UTC,ContainerStatuses:[{nginx {ContainerStateWaiting{Reason:ContainerCreating,Message:,} nil nil} {nil nil nil} false 0 nginx:404 }],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} +Jun 18 12:26:29.847: INFO: Pod "nginx-deployment-65bbdb5f8-vx4tp" is not available: +&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-65bbdb5f8-vx4tp,GenerateName:nginx-deployment-65bbdb5f8-,Namespace:e2e-tests-deployment-jzshx,SelfLink:/api/v1/namespaces/e2e-tests-deployment-jzshx/pods/nginx-deployment-65bbdb5f8-vx4tp,UID:4a3a0012-91c4-11e9-bf44-fa6f350b29f0,ResourceVersion:100715,Generation:0,CreationTimestamp:2019-06-18 12:26:25 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 65bbdb5f8,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-65bbdb5f8 4a2a98be-91c4-11e9-bf44-fa6f350b29f0 0xc00195f760 0xc00195f761}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-g4nvn {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-g4nvn,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] [] [] [] [] {map[] map[]} [{default-token-g4nvn true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.72.74.144,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc00195f810} {node.kubernetes.io/unreachable Exists NoExecute 0xc00195f830}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:25 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:25 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:25 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:26:25 +0000 UTC }],Message:,Reason:,HostIP:10.72.74.144,PodIP:,StartTime:2019-06-18 12:26:25 +0000 UTC,ContainerStatuses:[{nginx {ContainerStateWaiting{Reason:ContainerCreating,Message:,} nil nil} {nil nil nil} false 0 nginx:404 }],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} +[AfterEach] [sig-apps] Deployment /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 19:37:08.530: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-projected-7fgn8" for this suite. -May 13 19:37:14.566: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 19:37:14.741: INFO: namespace: e2e-tests-projected-7fgn8, resource: bindings, ignored listing per whitelist -May 13 19:37:14.832: INFO: namespace e2e-tests-projected-7fgn8 deletion completed in 6.294169818s +Jun 18 12:26:29.847: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "e2e-tests-deployment-jzshx" for this suite. +Jun 18 12:26:39.915: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:26:40.513: INFO: namespace: e2e-tests-deployment-jzshx, resource: bindings, ignored listing per whitelist +Jun 18 12:26:40.667: INFO: namespace e2e-tests-deployment-jzshx deletion completed in 10.804576771s -• [SLOW TEST:8.716 seconds] -[sig-storage] Projected secret -/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_secret.go:34 - should be consumable from pods in volume with defaultMode set [NodeConformance] [Conformance] +• [SLOW TEST:19.919 seconds] +[sig-apps] Deployment +/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22 + deployment should support proportional scaling [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ -SSSSSSSSSSSS +SSSSSSSSSS ------------------------------ -[sig-storage] Projected downwardAPI - should set DefaultMode on files [NodeConformance] [Conformance] +[sig-cli] Kubectl client [k8s.io] Kubectl run job + should create a job from an image when restart is OnFailure [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -[BeforeEach] [sig-storage] Projected downwardAPI +[BeforeEach] [sig-cli] Kubectl client /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client -May 13 19:37:14.833: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -STEP: Building a namespace api object, basename projected -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-projected-b8bdc +Jun 18 12:26:40.667: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 +STEP: Building a namespace api object, basename kubectl +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-kubectl-2j42f STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [sig-storage] Projected downwardAPI - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:39 -[It] should set DefaultMode on files [NodeConformance] [Conformance] +[BeforeEach] [sig-cli] Kubectl client + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:243 +[BeforeEach] [k8s.io] Kubectl run job + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1454 +[It] should create a job from an image when restart is OnFailure [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -STEP: Creating a pod to test downward API volume plugin -May 13 19:37:15.114: INFO: Waiting up to 5m0s for pod "downwardapi-volume-82d431f8-75b6-11e9-8f67-2632f168be36" in namespace "e2e-tests-projected-b8bdc" to be "success or failure" -May 13 19:37:15.121: INFO: Pod "downwardapi-volume-82d431f8-75b6-11e9-8f67-2632f168be36": Phase="Pending", Reason="", readiness=false. Elapsed: 7.269859ms -May 13 19:37:17.142: INFO: Pod "downwardapi-volume-82d431f8-75b6-11e9-8f67-2632f168be36": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.028085372s -STEP: Saw pod success -May 13 19:37:17.142: INFO: Pod "downwardapi-volume-82d431f8-75b6-11e9-8f67-2632f168be36" satisfied condition "success or failure" -May 13 19:37:17.151: INFO: Trying to get logs from node 10.170.219.177 pod downwardapi-volume-82d431f8-75b6-11e9-8f67-2632f168be36 container client-container: -STEP: delete the pod -May 13 19:37:17.188: INFO: Waiting for pod downwardapi-volume-82d431f8-75b6-11e9-8f67-2632f168be36 to disappear -May 13 19:37:17.198: INFO: Pod downwardapi-volume-82d431f8-75b6-11e9-8f67-2632f168be36 no longer exists -[AfterEach] [sig-storage] Projected downwardAPI +STEP: running the image docker.io/library/nginx:1.14-alpine +Jun 18 12:26:41.130: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 run e2e-test-nginx-job --restart=OnFailure --generator=job/v1 --image=docker.io/library/nginx:1.14-alpine --namespace=e2e-tests-kubectl-2j42f' +Jun 18 12:26:41.511: INFO: stderr: "kubectl run --generator=job/v1 is DEPRECATED and will be removed in a future version. Use kubectl run --generator=run-pod/v1 or kubectl create instead.\n" +Jun 18 12:26:41.511: INFO: stdout: "job.batch/e2e-test-nginx-job created\n" +STEP: verifying the job e2e-test-nginx-job was created +[AfterEach] [k8s.io] Kubectl run job + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1459 +Jun 18 12:26:41.522: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 delete jobs e2e-test-nginx-job --namespace=e2e-tests-kubectl-2j42f' +Jun 18 12:26:41.694: INFO: stderr: "" +Jun 18 12:26:41.694: INFO: stdout: "job.batch \"e2e-test-nginx-job\" deleted\n" +[AfterEach] [sig-cli] Kubectl client /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 19:37:17.198: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-projected-b8bdc" for this suite. -May 13 19:37:23.249: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 19:37:23.471: INFO: namespace: e2e-tests-projected-b8bdc, resource: bindings, ignored listing per whitelist -May 13 19:37:23.557: INFO: namespace e2e-tests-projected-b8bdc deletion completed in 6.335506957s +Jun 18 12:26:41.694: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "e2e-tests-kubectl-2j42f" for this suite. +Jun 18 12:26:47.790: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:26:47.937: INFO: namespace: e2e-tests-kubectl-2j42f, resource: bindings, ignored listing per whitelist +Jun 18 12:26:48.271: INFO: namespace e2e-tests-kubectl-2j42f deletion completed in 6.54620729s -• [SLOW TEST:8.724 seconds] -[sig-storage] Projected downwardAPI -/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:33 - should set DefaultMode on files [NodeConformance] [Conformance] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 +• [SLOW TEST:7.603 seconds] +[sig-cli] Kubectl client +/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:22 + [k8s.io] Kubectl run job + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 + should create a job from an image when restart is OnFailure [Conformance] + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ -SSSSS +SSSSSSSSSSSSSSSS ------------------------------ -[sig-network] Proxy version v1 - should proxy logs on node using proxy subresource [Conformance] +[k8s.io] Variable Expansion + should allow substituting values in a container's args [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -[BeforeEach] version v1 +[BeforeEach] [k8s.io] Variable Expansion /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client -May 13 19:37:23.557: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -STEP: Building a namespace api object, basename proxy -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-proxy-l6774 +Jun 18 12:26:48.271: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 +STEP: Building a namespace api object, basename var-expansion +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-var-expansion-ph5vs STEP: Waiting for a default service account to be provisioned in namespace -[It] should proxy logs on node using proxy subresource [Conformance] +[It] should allow substituting values in a container's args [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -May 13 19:37:23.945: INFO: (0) /api/v1/nodes/10.170.219.151/proxy/logs/:
-alb/
-alternatives.log
-apt/... (200; 15.288773ms)
-May 13 19:37:23.954: INFO: (1) /api/v1/nodes/10.170.219.151/proxy/logs/: 
-alb/
-alternatives.log
-apt/... (200; 8.394234ms)
-May 13 19:37:23.962: INFO: (2) /api/v1/nodes/10.170.219.151/proxy/logs/: 
-alb/
-alternatives.log
-apt/... (200; 8.304928ms)
-May 13 19:37:23.971: INFO: (3) /api/v1/nodes/10.170.219.151/proxy/logs/: 
-alb/
-alternatives.log
-apt/... (200; 8.527692ms)
-May 13 19:37:23.979: INFO: (4) /api/v1/nodes/10.170.219.151/proxy/logs/: 
-alb/
-alternatives.log
-apt/... (200; 7.912522ms)
-May 13 19:37:23.988: INFO: (5) /api/v1/nodes/10.170.219.151/proxy/logs/: 
-alb/
-alternatives.log
-apt/... (200; 9.342698ms)
-May 13 19:37:23.997: INFO: (6) /api/v1/nodes/10.170.219.151/proxy/logs/: 
-alb/
-alternatives.log
-apt/... (200; 8.482543ms)
-May 13 19:37:24.005: INFO: (7) /api/v1/nodes/10.170.219.151/proxy/logs/: 
-alb/
-alternatives.log
-apt/... (200; 8.604469ms)
-May 13 19:37:24.014: INFO: (8) /api/v1/nodes/10.170.219.151/proxy/logs/: 
-alb/
-alternatives.log
-apt/... (200; 8.417351ms)
-May 13 19:37:24.022: INFO: (9) /api/v1/nodes/10.170.219.151/proxy/logs/: 
-alb/
-alternatives.log
-apt/... (200; 8.630109ms)
-May 13 19:37:24.031: INFO: (10) /api/v1/nodes/10.170.219.151/proxy/logs/: 
-alb/
-alternatives.log
-apt/... (200; 8.131685ms)
-May 13 19:37:24.039: INFO: (11) /api/v1/nodes/10.170.219.151/proxy/logs/: 
-alb/
-alternatives.log
-apt/... (200; 8.257885ms)
-May 13 19:37:24.048: INFO: (12) /api/v1/nodes/10.170.219.151/proxy/logs/: 
-alb/
-alternatives.log
-apt/... (200; 8.731738ms)
-May 13 19:37:24.057: INFO: (13) /api/v1/nodes/10.170.219.151/proxy/logs/: 
-alb/
-alternatives.log
-apt/... (200; 8.914792ms)
-May 13 19:37:24.066: INFO: (14) /api/v1/nodes/10.170.219.151/proxy/logs/: 
-alb/
-alternatives.log
-apt/... (200; 9.655386ms)
-May 13 19:37:24.075: INFO: (15) /api/v1/nodes/10.170.219.151/proxy/logs/: 
-alb/
-alternatives.log
-apt/... (200; 8.522576ms)
-May 13 19:37:24.084: INFO: (16) /api/v1/nodes/10.170.219.151/proxy/logs/: 
-alb/
-alternatives.log
-apt/... (200; 8.848382ms)
-May 13 19:37:24.092: INFO: (17) /api/v1/nodes/10.170.219.151/proxy/logs/: 
-alb/
-alternatives.log
-apt/... (200; 7.946529ms)
-May 13 19:37:24.101: INFO: (18) /api/v1/nodes/10.170.219.151/proxy/logs/: 
-alb/
-alternatives.log
-apt/... (200; 9.032523ms)
-May 13 19:37:24.109: INFO: (19) /api/v1/nodes/10.170.219.151/proxy/logs/: 
-alb/
-alternatives.log
-apt/... (200; 8.410069ms)
-[AfterEach] version v1
+STEP: Creating a pod to test substitution in container's args
+Jun 18 12:26:48.817: INFO: Waiting up to 5m0s for pod "var-expansion-5805e26d-91c4-11e9-bce2-ae54e022189f" in namespace "e2e-tests-var-expansion-ph5vs" to be "success or failure"
+Jun 18 12:26:48.832: INFO: Pod "var-expansion-5805e26d-91c4-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 15.513307ms
+Jun 18 12:26:50.848: INFO: Pod "var-expansion-5805e26d-91c4-11e9-bce2-ae54e022189f": Phase="Running", Reason="", readiness=true. Elapsed: 2.030574539s
+Jun 18 12:26:52.863: INFO: Pod "var-expansion-5805e26d-91c4-11e9-bce2-ae54e022189f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.046089782s
+STEP: Saw pod success
+Jun 18 12:26:52.863: INFO: Pod "var-expansion-5805e26d-91c4-11e9-bce2-ae54e022189f" satisfied condition "success or failure"
+Jun 18 12:26:52.878: INFO: Trying to get logs from node 10.72.74.149 pod var-expansion-5805e26d-91c4-11e9-bce2-ae54e022189f container dapi-container: 
+STEP: delete the pod
+Jun 18 12:26:53.052: INFO: Waiting for pod var-expansion-5805e26d-91c4-11e9-bce2-ae54e022189f to disappear
+Jun 18 12:26:53.066: INFO: Pod var-expansion-5805e26d-91c4-11e9-bce2-ae54e022189f no longer exists
+[AfterEach] [k8s.io] Variable Expansion
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
-May 13 19:37:24.109: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "e2e-tests-proxy-l6774" for this suite.
-May 13 19:37:30.155: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-May 13 19:37:30.287: INFO: namespace: e2e-tests-proxy-l6774, resource: bindings, ignored listing per whitelist
-May 13 19:37:30.473: INFO: namespace e2e-tests-proxy-l6774 deletion completed in 6.356934296s
+Jun 18 12:26:53.066: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-tests-var-expansion-ph5vs" for this suite.
+Jun 18 12:27:01.139: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 12:27:01.580: INFO: namespace: e2e-tests-var-expansion-ph5vs, resource: bindings, ignored listing per whitelist
+Jun 18 12:27:02.998: INFO: namespace e2e-tests-var-expansion-ph5vs deletion completed in 9.91084251s
 
-• [SLOW TEST:6.916 seconds]
-[sig-network] Proxy
-/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:22
-  version v1
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/proxy.go:56
-    should proxy logs on node using proxy subresource  [Conformance]
-    /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
+• [SLOW TEST:14.726 seconds]
+[k8s.io] Variable Expansion
+/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694
+  should allow substituting values in a container's args [NodeConformance] [Conformance]
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 ------------------------------
-SSSSSSS
+SSSSSSSSSS
 ------------------------------
-[sig-cli] Kubectl client [k8s.io] Kubectl expose 
-  should create services for rc  [Conformance]
+[sig-storage] Projected configMap 
+  should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-[BeforeEach] [sig-cli] Kubectl client
+[BeforeEach] [sig-storage] Projected configMap
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
 STEP: Creating a kubernetes client
-May 13 19:37:30.473: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423
-STEP: Building a namespace api object, basename kubectl
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-kubectl-c6mlp
+Jun 18 12:27:02.999: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
+STEP: Building a namespace api object, basename projected
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-projected-kllld
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:243
-[It] should create services for rc  [Conformance]
+[It] should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-STEP: creating Redis RC
-May 13 19:37:30.743: INFO: namespace e2e-tests-kubectl-c6mlp
-May 13 19:37:30.743: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 create -f - --namespace=e2e-tests-kubectl-c6mlp'
-May 13 19:37:31.022: INFO: stderr: ""
-May 13 19:37:31.022: INFO: stdout: "replicationcontroller/redis-master created\n"
-STEP: Waiting for Redis master to start.
-May 13 19:37:32.031: INFO: Selector matched 1 pods for map[app:redis]
-May 13 19:37:32.031: INFO: Found 0 / 1
-May 13 19:37:33.032: INFO: Selector matched 1 pods for map[app:redis]
-May 13 19:37:33.032: INFO: Found 1 / 1
-May 13 19:37:33.032: INFO: WaitFor completed with timeout 5m0s.  Pods found = 1 out of 1
-May 13 19:37:33.040: INFO: Selector matched 1 pods for map[app:redis]
-May 13 19:37:33.040: INFO: ForEach: Found 1 pods from the filter.  Now looping through them.
-May 13 19:37:33.040: INFO: wait on redis-master startup in e2e-tests-kubectl-c6mlp 
-May 13 19:37:33.040: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 logs redis-master-ftlmz redis-master --namespace=e2e-tests-kubectl-c6mlp'
-May 13 19:37:33.254: INFO: stderr: ""
-May 13 19:37:33.254: INFO: stdout: "                _._                                                  \n           _.-``__ ''-._                                             \n      _.-``    `.  `_.  ''-._           Redis 3.2.12 (35a5711f/0) 64 bit\n  .-`` .-```.  ```\\/    _.,_ ''-._                                   \n (    '      ,       .-`  | `,    )     Running in standalone mode\n |`-._`-...-` __...-.``-._|'` _.-'|     Port: 6379\n |    `-._   `._    /     _.-'    |     PID: 1\n  `-._    `-._  `-./  _.-'    _.-'                                   \n |`-._`-._    `-.__.-'    _.-'_.-'|                                  \n |    `-._`-._        _.-'_.-'    |           http://redis.io        \n  `-._    `-._`-.__.-'_.-'    _.-'                                   \n |`-._`-._    `-.__.-'    _.-'_.-'|                                  \n |    `-._`-._        _.-'_.-'    |                                  \n  `-._    `-._`-.__.-'_.-'    _.-'                                   \n      `-._    `-.__.-'    _.-'                                       \n          `-._        _.-'                                           \n              `-.__.-'                                               \n\n1:M 13 May 19:37:32.367 # WARNING: The TCP backlog setting of 511 cannot be enforced because /proc/sys/net/core/somaxconn is set to the lower value of 128.\n1:M 13 May 19:37:32.367 # Server started, Redis version 3.2.12\n1:M 13 May 19:37:32.367 # WARNING you have Transparent Huge Pages (THP) support enabled in your kernel. This will create latency and memory usage issues with Redis. To fix this issue run the command 'echo never > /sys/kernel/mm/transparent_hugepage/enabled' as root, and add it to your /etc/rc.local in order to retain the setting after a reboot. Redis must be restarted after THP is disabled.\n1:M 13 May 19:37:32.367 * The server is now ready to accept connections on port 6379\n"
-STEP: exposing RC
-May 13 19:37:33.254: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 expose rc redis-master --name=rm2 --port=1234 --target-port=6379 --namespace=e2e-tests-kubectl-c6mlp'
-May 13 19:37:33.381: INFO: stderr: ""
-May 13 19:37:33.381: INFO: stdout: "service/rm2 exposed\n"
-May 13 19:37:33.388: INFO: Service rm2 in namespace e2e-tests-kubectl-c6mlp found.
-STEP: exposing service
-May 13 19:37:35.402: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 expose service rm2 --name=rm3 --port=2345 --target-port=6379 --namespace=e2e-tests-kubectl-c6mlp'
-May 13 19:37:35.520: INFO: stderr: ""
-May 13 19:37:35.520: INFO: stdout: "service/rm3 exposed\n"
-May 13 19:37:35.528: INFO: Service rm3 in namespace e2e-tests-kubectl-c6mlp found.
-[AfterEach] [sig-cli] Kubectl client
+STEP: Creating configMap with name projected-configmap-test-volume-60c5420a-91c4-11e9-bce2-ae54e022189f
+STEP: Creating a pod to test consume configMaps
+Jun 18 12:27:03.512: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-60c80a9c-91c4-11e9-bce2-ae54e022189f" in namespace "e2e-tests-projected-kllld" to be "success or failure"
+Jun 18 12:27:03.527: INFO: Pod "pod-projected-configmaps-60c80a9c-91c4-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 14.995975ms
+Jun 18 12:27:05.542: INFO: Pod "pod-projected-configmaps-60c80a9c-91c4-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 2.030521065s
+Jun 18 12:27:07.557: INFO: Pod "pod-projected-configmaps-60c80a9c-91c4-11e9-bce2-ae54e022189f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.045081629s
+STEP: Saw pod success
+Jun 18 12:27:07.557: INFO: Pod "pod-projected-configmaps-60c80a9c-91c4-11e9-bce2-ae54e022189f" satisfied condition "success or failure"
+Jun 18 12:27:07.598: INFO: Trying to get logs from node 10.72.74.144 pod pod-projected-configmaps-60c80a9c-91c4-11e9-bce2-ae54e022189f container projected-configmap-volume-test: 
+STEP: delete the pod
+Jun 18 12:27:07.673: INFO: Waiting for pod pod-projected-configmaps-60c80a9c-91c4-11e9-bce2-ae54e022189f to disappear
+Jun 18 12:27:07.687: INFO: Pod pod-projected-configmaps-60c80a9c-91c4-11e9-bce2-ae54e022189f no longer exists
+[AfterEach] [sig-storage] Projected configMap
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
-May 13 19:37:37.541: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "e2e-tests-kubectl-c6mlp" for this suite.
-May 13 19:38:01.577: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-May 13 19:38:01.978: INFO: namespace: e2e-tests-kubectl-c6mlp, resource: bindings, ignored listing per whitelist
-May 13 19:38:02.030: INFO: namespace e2e-tests-kubectl-c6mlp deletion completed in 24.481093261s
+Jun 18 12:27:07.687: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-tests-projected-kllld" for this suite.
+Jun 18 12:27:17.759: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 12:27:17.876: INFO: namespace: e2e-tests-projected-kllld, resource: bindings, ignored listing per whitelist
+Jun 18 12:27:18.263: INFO: namespace e2e-tests-projected-kllld deletion completed in 10.556521334s
 
-• [SLOW TEST:31.558 seconds]
-[sig-cli] Kubectl client
-/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:22
-  [k8s.io] Kubectl expose
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694
-    should create services for rc  [Conformance]
-    /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
+• [SLOW TEST:15.265 seconds]
+[sig-storage] Projected configMap
+/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_configmap.go:34
+  should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance]
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 ------------------------------
-SSSSSSSSSSSSSSSSSSSSSSSSSS
+S
 ------------------------------
-[k8s.io] Container Runtime blackbox test when starting a container that exits 
-  should run with the expected status [NodeConformance] [Conformance]
+[sig-storage] Projected downwardAPI 
+  should provide container's cpu limit [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-[BeforeEach] [k8s.io] Container Runtime
+[BeforeEach] [sig-storage] Projected downwardAPI
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
 STEP: Creating a kubernetes client
-May 13 19:38:02.031: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423
-STEP: Building a namespace api object, basename container-runtime
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-container-runtime-2f899
+Jun 18 12:27:18.263: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
+STEP: Building a namespace api object, basename projected
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-projected-rk98z
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should run with the expected status [NodeConformance] [Conformance]
+[BeforeEach] [sig-storage] Projected downwardAPI
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:39
+[It] should provide container's cpu limit [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-STEP: Container 'terminate-cmd-rpa': should get the expected 'RestartCount'
-STEP: Container 'terminate-cmd-rpa': should get the expected 'Phase'
-STEP: Container 'terminate-cmd-rpa': should get the expected 'Ready' condition
-STEP: Container 'terminate-cmd-rpa': should get the expected 'State'
-STEP: Container 'terminate-cmd-rpa': should be possible to delete [NodeConformance]
-STEP: Container 'terminate-cmd-rpof': should get the expected 'RestartCount'
-STEP: Container 'terminate-cmd-rpof': should get the expected 'Phase'
-STEP: Container 'terminate-cmd-rpof': should get the expected 'Ready' condition
-STEP: Container 'terminate-cmd-rpof': should get the expected 'State'
-STEP: Container 'terminate-cmd-rpof': should be possible to delete [NodeConformance]
-STEP: Container 'terminate-cmd-rpn': should get the expected 'RestartCount'
-STEP: Container 'terminate-cmd-rpn': should get the expected 'Phase'
-STEP: Container 'terminate-cmd-rpn': should get the expected 'Ready' condition
-STEP: Container 'terminate-cmd-rpn': should get the expected 'State'
-STEP: Container 'terminate-cmd-rpn': should be possible to delete [NodeConformance]
-[AfterEach] [k8s.io] Container Runtime
+STEP: Creating a pod to test downward API volume plugin
+Jun 18 12:27:18.948: INFO: Waiting up to 5m0s for pod "downwardapi-volume-69fbb760-91c4-11e9-bce2-ae54e022189f" in namespace "e2e-tests-projected-rk98z" to be "success or failure"
+Jun 18 12:27:18.962: INFO: Pod "downwardapi-volume-69fbb760-91c4-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 14.043282ms
+Jun 18 12:27:20.993: INFO: Pod "downwardapi-volume-69fbb760-91c4-11e9-bce2-ae54e022189f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.044799922s
+STEP: Saw pod success
+Jun 18 12:27:20.993: INFO: Pod "downwardapi-volume-69fbb760-91c4-11e9-bce2-ae54e022189f" satisfied condition "success or failure"
+Jun 18 12:27:21.008: INFO: Trying to get logs from node 10.72.74.149 pod downwardapi-volume-69fbb760-91c4-11e9-bce2-ae54e022189f container client-container: 
+STEP: delete the pod
+Jun 18 12:27:21.088: INFO: Waiting for pod downwardapi-volume-69fbb760-91c4-11e9-bce2-ae54e022189f to disappear
+Jun 18 12:27:21.103: INFO: Pod downwardapi-volume-69fbb760-91c4-11e9-bce2-ae54e022189f no longer exists
+[AfterEach] [sig-storage] Projected downwardAPI
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
-May 13 19:38:26.447: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "e2e-tests-container-runtime-2f899" for this suite.
-May 13 19:38:32.483: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-May 13 19:38:32.664: INFO: namespace: e2e-tests-container-runtime-2f899, resource: bindings, ignored listing per whitelist
-May 13 19:38:32.713: INFO: namespace e2e-tests-container-runtime-2f899 deletion completed in 6.257935507s
+Jun 18 12:27:21.103: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-tests-projected-rk98z" for this suite.
+Jun 18 12:27:27.183: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 12:27:27.727: INFO: namespace: e2e-tests-projected-rk98z, resource: bindings, ignored listing per whitelist
+Jun 18 12:27:28.111: INFO: namespace e2e-tests-projected-rk98z deletion completed in 6.983552879s
 
-• [SLOW TEST:30.682 seconds]
-[k8s.io] Container Runtime
-/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694
-  blackbox test
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/runtime.go:37
-    when starting a container that exits
-    /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/runtime.go:38
-      should run with the expected status [NodeConformance] [Conformance]
-      /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
+• [SLOW TEST:9.847 seconds]
+[sig-storage] Projected downwardAPI
+/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:33
+  should provide container's cpu limit [NodeConformance] [Conformance]
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 ------------------------------
 SSS
 ------------------------------
-[k8s.io] Docker Containers 
-  should be able to override the image's default command (docker entrypoint) [NodeConformance] [Conformance]
+[sig-storage] EmptyDir volumes 
+  should support (non-root,0644,tmpfs) [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-[BeforeEach] [k8s.io] Docker Containers
+[BeforeEach] [sig-storage] EmptyDir volumes
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
 STEP: Creating a kubernetes client
-May 13 19:38:32.713: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423
-STEP: Building a namespace api object, basename containers
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-containers-b84tg
+Jun 18 12:27:28.111: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
+STEP: Building a namespace api object, basename emptydir
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-emptydir-7p77j
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should be able to override the image's default command (docker entrypoint) [NodeConformance] [Conformance]
+[It] should support (non-root,0644,tmpfs) [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-STEP: Creating a pod to test override command
-May 13 19:38:33.016: INFO: Waiting up to 5m0s for pod "client-containers-b143ad32-75b6-11e9-8f67-2632f168be36" in namespace "e2e-tests-containers-b84tg" to be "success or failure"
-May 13 19:38:33.024: INFO: Pod "client-containers-b143ad32-75b6-11e9-8f67-2632f168be36": Phase="Pending", Reason="", readiness=false. Elapsed: 7.12353ms
-May 13 19:38:35.044: INFO: Pod "client-containers-b143ad32-75b6-11e9-8f67-2632f168be36": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.027665009s
+STEP: Creating a pod to test emptydir 0644 on tmpfs
+Jun 18 12:27:28.584: INFO: Waiting up to 5m0s for pod "pod-6fb97d1e-91c4-11e9-bce2-ae54e022189f" in namespace "e2e-tests-emptydir-7p77j" to be "success or failure"
+Jun 18 12:27:28.598: INFO: Pod "pod-6fb97d1e-91c4-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 13.637576ms
+Jun 18 12:27:30.613: INFO: Pod "pod-6fb97d1e-91c4-11e9-bce2-ae54e022189f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.028116402s
 STEP: Saw pod success
-May 13 19:38:35.044: INFO: Pod "client-containers-b143ad32-75b6-11e9-8f67-2632f168be36" satisfied condition "success or failure"
-May 13 19:38:35.120: INFO: Trying to get logs from node 10.170.219.177 pod client-containers-b143ad32-75b6-11e9-8f67-2632f168be36 container test-container: 
+Jun 18 12:27:30.613: INFO: Pod "pod-6fb97d1e-91c4-11e9-bce2-ae54e022189f" satisfied condition "success or failure"
+Jun 18 12:27:30.627: INFO: Trying to get logs from node 10.72.74.143 pod pod-6fb97d1e-91c4-11e9-bce2-ae54e022189f container test-container: 
 STEP: delete the pod
-May 13 19:38:35.160: INFO: Waiting for pod client-containers-b143ad32-75b6-11e9-8f67-2632f168be36 to disappear
-May 13 19:38:35.167: INFO: Pod client-containers-b143ad32-75b6-11e9-8f67-2632f168be36 no longer exists
-[AfterEach] [k8s.io] Docker Containers
+Jun 18 12:27:30.697: INFO: Waiting for pod pod-6fb97d1e-91c4-11e9-bce2-ae54e022189f to disappear
+Jun 18 12:27:30.711: INFO: Pod pod-6fb97d1e-91c4-11e9-bce2-ae54e022189f no longer exists
+[AfterEach] [sig-storage] EmptyDir volumes
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
-May 13 19:38:35.167: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "e2e-tests-containers-b84tg" for this suite.
-May 13 19:38:43.202: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-May 13 19:38:43.262: INFO: namespace: e2e-tests-containers-b84tg, resource: bindings, ignored listing per whitelist
-May 13 19:38:43.465: INFO: namespace e2e-tests-containers-b84tg deletion completed in 8.290663869s
+Jun 18 12:27:30.711: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-tests-emptydir-7p77j" for this suite.
+Jun 18 12:27:36.784: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 12:27:37.309: INFO: namespace: e2e-tests-emptydir-7p77j, resource: bindings, ignored listing per whitelist
+Jun 18 12:27:37.351: INFO: namespace e2e-tests-emptydir-7p77j deletion completed in 6.619169407s
 
-• [SLOW TEST:10.752 seconds]
-[k8s.io] Docker Containers
-/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694
-  should be able to override the image's default command (docker entrypoint) [NodeConformance] [Conformance]
+• [SLOW TEST:9.240 seconds]
+[sig-storage] EmptyDir volumes
+/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:40
+  should support (non-root,0644,tmpfs) [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 ------------------------------
-SSSSSSS
+SSSSSS
 ------------------------------
-[sig-api-machinery] Namespaces [Serial] 
-  should ensure that all services are removed when a namespace is deleted [Conformance]
+[sig-api-machinery] Garbage collector 
+  should keep the rc around until all its pods are deleted if the deleteOptions says so [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-[BeforeEach] [sig-api-machinery] Namespaces [Serial]
+[BeforeEach] [sig-api-machinery] Garbage collector
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
 STEP: Creating a kubernetes client
-May 13 19:38:43.466: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423
-STEP: Building a namespace api object, basename namespaces
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-namespaces-bt6bn
+Jun 18 12:27:37.351: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
+STEP: Building a namespace api object, basename gc
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-gc-scznw
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should ensure that all services are removed when a namespace is deleted [Conformance]
+[It] should keep the rc around until all its pods are deleted if the deleteOptions says so [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-STEP: Creating a test namespace
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-nsdeletetest-4v4kq
-STEP: Waiting for a default service account to be provisioned in namespace
-STEP: Creating a service in the namespace
-STEP: Deleting the namespace
-STEP: Waiting for the namespace to be removed.
-STEP: Recreating the namespace
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-nsdeletetest-dfz9l
-STEP: Verifying there is no service in the namespace
-[AfterEach] [sig-api-machinery] Namespaces [Serial]
+STEP: create the rc
+STEP: delete the rc
+STEP: wait for the rc to be deleted
+STEP: Gathering metrics
+W0618 12:27:44.031213      17 metrics_grabber.go:81] Master node is not registered. Grabbing metrics from Scheduler, ControllerManager and ClusterAutoscaler is disabled.
+Jun 18 12:27:44.031: INFO: For apiserver_request_count:
+For apiserver_request_latencies_summary:
+For etcd_helper_cache_entry_count:
+For etcd_helper_cache_hit_count:
+For etcd_helper_cache_miss_count:
+For etcd_request_cache_add_latencies_summary:
+For etcd_request_cache_get_latencies_summary:
+For etcd_request_latencies_summary:
+For garbage_collector_attempt_to_delete_queue_latency:
+For garbage_collector_attempt_to_delete_work_duration:
+For garbage_collector_attempt_to_orphan_queue_latency:
+For garbage_collector_attempt_to_orphan_work_duration:
+For garbage_collector_dirty_processing_latency_microseconds:
+For garbage_collector_event_processing_latency_microseconds:
+For garbage_collector_graph_changes_queue_latency:
+For garbage_collector_graph_changes_work_duration:
+For garbage_collector_orphan_processing_latency_microseconds:
+For namespace_queue_latency:
+For namespace_queue_latency_sum:
+For namespace_queue_latency_count:
+For namespace_retries:
+For namespace_work_duration:
+For namespace_work_duration_sum:
+For namespace_work_duration_count:
+For function_duration_seconds:
+For errors_total:
+For evicted_pods_total:
+
+[AfterEach] [sig-api-machinery] Garbage collector
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
-May 13 19:38:50.211: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "e2e-tests-namespaces-bt6bn" for this suite.
-May 13 19:38:56.402: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-May 13 19:38:56.662: INFO: namespace: e2e-tests-namespaces-bt6bn, resource: bindings, ignored listing per whitelist
-May 13 19:38:56.734: INFO: namespace e2e-tests-namespaces-bt6bn deletion completed in 6.515934116s
-STEP: Destroying namespace "e2e-tests-nsdeletetest-4v4kq" for this suite.
-May 13 19:38:56.741: INFO: Namespace e2e-tests-nsdeletetest-4v4kq was already deleted
-STEP: Destroying namespace "e2e-tests-nsdeletetest-dfz9l" for this suite.
-May 13 19:39:02.769: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-May 13 19:39:02.898: INFO: namespace: e2e-tests-nsdeletetest-dfz9l, resource: bindings, ignored listing per whitelist
-May 13 19:39:03.092: INFO: namespace e2e-tests-nsdeletetest-dfz9l deletion completed in 6.350237105s
-
-• [SLOW TEST:19.626 seconds]
-[sig-api-machinery] Namespaces [Serial]
+Jun 18 12:27:44.031: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-tests-gc-scznw" for this suite.
+Jun 18 12:27:52.100: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 12:27:52.758: INFO: namespace: e2e-tests-gc-scznw, resource: bindings, ignored listing per whitelist
+Jun 18 12:27:52.908: INFO: namespace e2e-tests-gc-scznw deletion completed in 8.863667938s
+
+• [SLOW TEST:15.557 seconds]
+[sig-api-machinery] Garbage collector
 /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:22
-  should ensure that all services are removed when a namespace is deleted [Conformance]
+  should keep the rc around until all its pods are deleted if the deleteOptions says so [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 ------------------------------
-S
+SSSSSSSS
 ------------------------------
 [sig-storage] Downward API volume 
-  should set DefaultMode on files [NodeConformance] [Conformance]
+  should provide container's memory request [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 [BeforeEach] [sig-storage] Downward API volume
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
 STEP: Creating a kubernetes client
-May 13 19:39:03.092: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423
+Jun 18 12:27:52.909: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
 STEP: Building a namespace api object, basename downward-api
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-downward-api-zppqb
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-downward-api-txwmz
 STEP: Waiting for a default service account to be provisioned in namespace
 [BeforeEach] [sig-storage] Downward API volume
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:39
-[It] should set DefaultMode on files [NodeConformance] [Conformance]
+[It] should provide container's memory request [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 STEP: Creating a pod to test downward API volume plugin
-May 13 19:39:03.379: INFO: Waiting up to 5m0s for pod "downwardapi-volume-c35c4029-75b6-11e9-8f67-2632f168be36" in namespace "e2e-tests-downward-api-zppqb" to be "success or failure"
-May 13 19:39:03.386: INFO: Pod "downwardapi-volume-c35c4029-75b6-11e9-8f67-2632f168be36": Phase="Pending", Reason="", readiness=false. Elapsed: 7.469787ms
-May 13 19:39:05.395: INFO: Pod "downwardapi-volume-c35c4029-75b6-11e9-8f67-2632f168be36": Phase="Running", Reason="", readiness=true. Elapsed: 2.016237972s
-May 13 19:39:07.416: INFO: Pod "downwardapi-volume-c35c4029-75b6-11e9-8f67-2632f168be36": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.037360494s
+Jun 18 12:27:53.433: INFO: Waiting up to 5m0s for pod "downwardapi-volume-7e89930a-91c4-11e9-bce2-ae54e022189f" in namespace "e2e-tests-downward-api-txwmz" to be "success or failure"
+Jun 18 12:27:53.447: INFO: Pod "downwardapi-volume-7e89930a-91c4-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 13.684233ms
+Jun 18 12:27:55.461: INFO: Pod "downwardapi-volume-7e89930a-91c4-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 2.028142414s
+Jun 18 12:27:57.478: INFO: Pod "downwardapi-volume-7e89930a-91c4-11e9-bce2-ae54e022189f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.044655422s
 STEP: Saw pod success
-May 13 19:39:07.416: INFO: Pod "downwardapi-volume-c35c4029-75b6-11e9-8f67-2632f168be36" satisfied condition "success or failure"
-May 13 19:39:07.423: INFO: Trying to get logs from node 10.170.219.177 pod downwardapi-volume-c35c4029-75b6-11e9-8f67-2632f168be36 container client-container: 
+Jun 18 12:27:57.478: INFO: Pod "downwardapi-volume-7e89930a-91c4-11e9-bce2-ae54e022189f" satisfied condition "success or failure"
+Jun 18 12:27:57.496: INFO: Trying to get logs from node 10.72.74.149 pod downwardapi-volume-7e89930a-91c4-11e9-bce2-ae54e022189f container client-container: 
 STEP: delete the pod
-May 13 19:39:07.476: INFO: Waiting for pod downwardapi-volume-c35c4029-75b6-11e9-8f67-2632f168be36 to disappear
-May 13 19:39:07.483: INFO: Pod downwardapi-volume-c35c4029-75b6-11e9-8f67-2632f168be36 no longer exists
+Jun 18 12:27:57.632: INFO: Waiting for pod downwardapi-volume-7e89930a-91c4-11e9-bce2-ae54e022189f to disappear
+Jun 18 12:27:57.646: INFO: Pod downwardapi-volume-7e89930a-91c4-11e9-bce2-ae54e022189f no longer exists
 [AfterEach] [sig-storage] Downward API volume
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
-May 13 19:39:07.483: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "e2e-tests-downward-api-zppqb" for this suite.
-May 13 19:39:13.518: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-May 13 19:39:13.587: INFO: namespace: e2e-tests-downward-api-zppqb, resource: bindings, ignored listing per whitelist
-May 13 19:39:13.754: INFO: namespace e2e-tests-downward-api-zppqb deletion completed in 6.263650485s
+Jun 18 12:27:57.646: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-tests-downward-api-txwmz" for this suite.
+Jun 18 12:28:03.740: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 12:28:04.183: INFO: namespace: e2e-tests-downward-api-txwmz, resource: bindings, ignored listing per whitelist
+Jun 18 12:28:04.247: INFO: namespace e2e-tests-downward-api-txwmz deletion completed in 6.576714077s
 
-• [SLOW TEST:10.662 seconds]
+• [SLOW TEST:11.339 seconds]
 [sig-storage] Downward API volume
 /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:34
-  should set DefaultMode on files [NodeConformance] [Conformance]
+  should provide container's memory request [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 ------------------------------
-[sig-storage] ConfigMap 
-  should be consumable from pods in volume as non-root [NodeConformance] [Conformance]
+SSSSSS
+------------------------------
+[sig-storage] Downward API volume 
+  should update labels on modification [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-[BeforeEach] [sig-storage] ConfigMap
+[BeforeEach] [sig-storage] Downward API volume
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
 STEP: Creating a kubernetes client
-May 13 19:39:13.755: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423
-STEP: Building a namespace api object, basename configmap
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-configmap-bzhfc
+Jun 18 12:28:04.247: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
+STEP: Building a namespace api object, basename downward-api
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-downward-api-ntsxs
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should be consumable from pods in volume as non-root [NodeConformance] [Conformance]
+[BeforeEach] [sig-storage] Downward API volume
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:39
+[It] should update labels on modification [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-STEP: Creating configMap with name configmap-test-volume-ca11ea69-75b6-11e9-8f67-2632f168be36
-STEP: Creating a pod to test consume configMaps
-May 13 19:39:14.642: INFO: Waiting up to 5m0s for pod "pod-configmaps-ca135151-75b6-11e9-8f67-2632f168be36" in namespace "e2e-tests-configmap-bzhfc" to be "success or failure"
-May 13 19:39:14.727: INFO: Pod "pod-configmaps-ca135151-75b6-11e9-8f67-2632f168be36": Phase="Pending", Reason="", readiness=false. Elapsed: 84.919358ms
-May 13 19:39:16.750: INFO: Pod "pod-configmaps-ca135151-75b6-11e9-8f67-2632f168be36": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.107519838s
-STEP: Saw pod success
-May 13 19:39:16.750: INFO: Pod "pod-configmaps-ca135151-75b6-11e9-8f67-2632f168be36" satisfied condition "success or failure"
-May 13 19:39:16.827: INFO: Trying to get logs from node 10.170.219.153 pod pod-configmaps-ca135151-75b6-11e9-8f67-2632f168be36 container configmap-volume-test: 
-STEP: delete the pod
-May 13 19:39:16.866: INFO: Waiting for pod pod-configmaps-ca135151-75b6-11e9-8f67-2632f168be36 to disappear
-May 13 19:39:16.873: INFO: Pod pod-configmaps-ca135151-75b6-11e9-8f67-2632f168be36 no longer exists
-[AfterEach] [sig-storage] ConfigMap
+STEP: Creating the pod
+Jun 18 12:28:09.400: INFO: Successfully updated pod "labelsupdate85455b79-91c4-11e9-bce2-ae54e022189f"
+[AfterEach] [sig-storage] Downward API volume
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
-May 13 19:39:16.873: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "e2e-tests-configmap-bzhfc" for this suite.
-May 13 19:39:22.908: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-May 13 19:39:22.950: INFO: namespace: e2e-tests-configmap-bzhfc, resource: bindings, ignored listing per whitelist
-May 13 19:39:23.159: INFO: namespace e2e-tests-configmap-bzhfc deletion completed in 6.279244142s
+Jun 18 12:28:12.285: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-tests-downward-api-ntsxs" for this suite.
+Jun 18 12:28:36.352: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 12:28:36.973: INFO: namespace: e2e-tests-downward-api-ntsxs, resource: bindings, ignored listing per whitelist
+Jun 18 12:28:37.428: INFO: namespace e2e-tests-downward-api-ntsxs deletion completed in 25.123957633s
 
-• [SLOW TEST:9.404 seconds]
-[sig-storage] ConfigMap
-/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:33
-  should be consumable from pods in volume as non-root [NodeConformance] [Conformance]
+• [SLOW TEST:33.181 seconds]
+[sig-storage] Downward API volume
+/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:34
+  should update labels on modification [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 ------------------------------
-[sig-network] Networking Granular Checks: Pods 
-  should function for node-pod communication: http [NodeConformance] [Conformance]
+SSSSSSSS
+------------------------------
+[k8s.io] Pods 
+  should get a host IP [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-[BeforeEach] [sig-network] Networking
+[BeforeEach] [k8s.io] Pods
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
 STEP: Creating a kubernetes client
-May 13 19:39:23.160: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423
-STEP: Building a namespace api object, basename pod-network-test
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-pod-network-test-ms6f5
+Jun 18 12:28:37.429: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
+STEP: Building a namespace api object, basename pods
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-pods-ldj62
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should function for node-pod communication: http [NodeConformance] [Conformance]
+[BeforeEach] [k8s.io] Pods
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:132
+[It] should get a host IP [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-STEP: Performing setup for networking test in namespace e2e-tests-pod-network-test-ms6f5
-STEP: creating a selector
-STEP: Creating the service pods in kubernetes
-May 13 19:39:23.419: INFO: Waiting up to 10m0s for all (but 0) nodes to be schedulable
-STEP: Creating test pods
-May 13 19:39:43.642: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s --max-time 15 --connect-timeout 1 http://172.30.63.63:8080/hostName | grep -v '^\s*$'] Namespace:e2e-tests-pod-network-test-ms6f5 PodName:host-test-container-pod ContainerName:hostexec Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
-May 13 19:39:43.642: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423
-May 13 19:39:43.832: INFO: Found all expected endpoints: [netserver-0]
-May 13 19:39:43.840: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s --max-time 15 --connect-timeout 1 http://172.30.227.91:8080/hostName | grep -v '^\s*$'] Namespace:e2e-tests-pod-network-test-ms6f5 PodName:host-test-container-pod ContainerName:hostexec Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
-May 13 19:39:43.840: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423
-May 13 19:39:44.012: INFO: Found all expected endpoints: [netserver-1]
-May 13 19:39:44.020: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s --max-time 15 --connect-timeout 1 http://172.30.19.108:8080/hostName | grep -v '^\s*$'] Namespace:e2e-tests-pod-network-test-ms6f5 PodName:host-test-container-pod ContainerName:hostexec Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
-May 13 19:39:44.020: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423
-May 13 19:39:44.194: INFO: Found all expected endpoints: [netserver-2]
-[AfterEach] [sig-network] Networking
+STEP: creating pod
+Jun 18 12:28:40.019: INFO: Pod pod-hostip-991353cc-91c4-11e9-bce2-ae54e022189f has hostIP: 10.72.74.143
+[AfterEach] [k8s.io] Pods
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
-May 13 19:39:44.194: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "e2e-tests-pod-network-test-ms6f5" for this suite.
-May 13 19:40:08.232: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-May 13 19:40:08.493: INFO: namespace: e2e-tests-pod-network-test-ms6f5, resource: bindings, ignored listing per whitelist
-May 13 19:40:08.545: INFO: namespace e2e-tests-pod-network-test-ms6f5 deletion completed in 24.341758945s
+Jun 18 12:28:40.019: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-tests-pods-ldj62" for this suite.
+Jun 18 12:29:04.095: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 12:29:04.415: INFO: namespace: e2e-tests-pods-ldj62, resource: bindings, ignored listing per whitelist
+Jun 18 12:29:04.563: INFO: namespace e2e-tests-pods-ldj62 deletion completed in 24.518894698s
 
-• [SLOW TEST:45.385 seconds]
-[sig-network] Networking
-/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:25
-  Granular Checks: Pods
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:28
-    should function for node-pod communication: http [NodeConformance] [Conformance]
-    /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
+• [SLOW TEST:27.134 seconds]
+[k8s.io] Pods
+/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694
+  should get a host IP [NodeConformance] [Conformance]
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 ------------------------------
-S
+SSSSSSS
 ------------------------------
 [sig-storage] ConfigMap 
-  should be consumable from pods in volume with defaultMode set [NodeConformance] [Conformance]
+  should be consumable from pods in volume [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 [BeforeEach] [sig-storage] ConfigMap
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
 STEP: Creating a kubernetes client
-May 13 19:40:08.545: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423
+Jun 18 12:29:04.563: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
 STEP: Building a namespace api object, basename configmap
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-configmap-v5lgc
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-configmap-fglh2
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should be consumable from pods in volume with defaultMode set [NodeConformance] [Conformance]
+[It] should be consumable from pods in volume [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-STEP: Creating configMap with name configmap-test-volume-ea621e49-75b6-11e9-8f67-2632f168be36
+STEP: Creating configMap with name configmap-test-volume-a942ba31-91c4-11e9-bce2-ae54e022189f
 STEP: Creating a pod to test consume configMaps
-May 13 19:40:08.854: INFO: Waiting up to 5m0s for pod "pod-configmaps-ea636de2-75b6-11e9-8f67-2632f168be36" in namespace "e2e-tests-configmap-v5lgc" to be "success or failure"
-May 13 19:40:08.862: INFO: Pod "pod-configmaps-ea636de2-75b6-11e9-8f67-2632f168be36": Phase="Pending", Reason="", readiness=false. Elapsed: 8.043696ms
-May 13 19:40:10.870: INFO: Pod "pod-configmaps-ea636de2-75b6-11e9-8f67-2632f168be36": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.015663923s
+Jun 18 12:29:05.124: INFO: Waiting up to 5m0s for pod "pod-configmaps-a944c093-91c4-11e9-bce2-ae54e022189f" in namespace "e2e-tests-configmap-fglh2" to be "success or failure"
+Jun 18 12:29:05.138: INFO: Pod "pod-configmaps-a944c093-91c4-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 13.2729ms
+Jun 18 12:29:07.152: INFO: Pod "pod-configmaps-a944c093-91c4-11e9-bce2-ae54e022189f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.027837558s
 STEP: Saw pod success
-May 13 19:40:10.870: INFO: Pod "pod-configmaps-ea636de2-75b6-11e9-8f67-2632f168be36" satisfied condition "success or failure"
-May 13 19:40:10.879: INFO: Trying to get logs from node 10.170.219.177 pod pod-configmaps-ea636de2-75b6-11e9-8f67-2632f168be36 container configmap-volume-test: 
+Jun 18 12:29:07.152: INFO: Pod "pod-configmaps-a944c093-91c4-11e9-bce2-ae54e022189f" satisfied condition "success or failure"
+Jun 18 12:29:07.202: INFO: Trying to get logs from node 10.72.74.144 pod pod-configmaps-a944c093-91c4-11e9-bce2-ae54e022189f container configmap-volume-test: 
 STEP: delete the pod
-May 13 19:40:10.921: INFO: Waiting for pod pod-configmaps-ea636de2-75b6-11e9-8f67-2632f168be36 to disappear
-May 13 19:40:10.928: INFO: Pod pod-configmaps-ea636de2-75b6-11e9-8f67-2632f168be36 no longer exists
+Jun 18 12:29:07.277: INFO: Waiting for pod pod-configmaps-a944c093-91c4-11e9-bce2-ae54e022189f to disappear
+Jun 18 12:29:07.291: INFO: Pod pod-configmaps-a944c093-91c4-11e9-bce2-ae54e022189f no longer exists
 [AfterEach] [sig-storage] ConfigMap
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
-May 13 19:40:10.928: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "e2e-tests-configmap-v5lgc" for this suite.
-May 13 19:40:16.964: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-May 13 19:40:17.037: INFO: namespace: e2e-tests-configmap-v5lgc, resource: bindings, ignored listing per whitelist
-May 13 19:40:17.196: INFO: namespace e2e-tests-configmap-v5lgc deletion completed in 6.260380348s
+Jun 18 12:29:07.291: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-tests-configmap-fglh2" for this suite.
+Jun 18 12:29:13.366: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 12:29:13.770: INFO: namespace: e2e-tests-configmap-fglh2, resource: bindings, ignored listing per whitelist
+Jun 18 12:29:13.849: INFO: namespace e2e-tests-configmap-fglh2 deletion completed in 6.531625795s
+
+• [SLOW TEST:9.285 seconds]
+[sig-storage] ConfigMap
+/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:33
+  should be consumable from pods in volume [NodeConformance] [Conformance]
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
+------------------------------
+SSSSSSS
+------------------------------
+[sig-storage] EmptyDir wrapper volumes 
+  should not cause race condition when used for configmaps [Serial] [Slow] [Conformance]
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
+[BeforeEach] [sig-storage] EmptyDir wrapper volumes
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
+STEP: Creating a kubernetes client
+Jun 18 12:29:13.849: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
+STEP: Building a namespace api object, basename emptydir-wrapper
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-emptydir-wrapper-xq29x
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should not cause race condition when used for configmaps [Serial] [Slow] [Conformance]
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
+STEP: Creating 50 configmaps
+STEP: Creating RC which spawns configmap-volume pods
+Jun 18 12:29:15.021: INFO: Pod name wrapped-volume-race-af273b75-91c4-11e9-bce2-ae54e022189f: Found 0 pods out of 5
+Jun 18 12:29:20.046: INFO: Pod name wrapped-volume-race-af273b75-91c4-11e9-bce2-ae54e022189f: Found 5 pods out of 5
+STEP: Ensuring each pod is running
+STEP: deleting ReplicationController wrapped-volume-race-af273b75-91c4-11e9-bce2-ae54e022189f in namespace e2e-tests-emptydir-wrapper-xq29x, will wait for the garbage collector to delete the pods
+Jun 18 12:32:04.349: INFO: Deleting ReplicationController wrapped-volume-race-af273b75-91c4-11e9-bce2-ae54e022189f took: 34.141243ms
+Jun 18 12:32:04.451: INFO: Terminating ReplicationController wrapped-volume-race-af273b75-91c4-11e9-bce2-ae54e022189f pods took: 101.687484ms
+STEP: Creating RC which spawns configmap-volume pods
+Jun 18 12:32:47.114: INFO: Pod name wrapped-volume-race-2d160b3e-91c5-11e9-bce2-ae54e022189f: Found 0 pods out of 5
+Jun 18 12:32:52.185: INFO: Pod name wrapped-volume-race-2d160b3e-91c5-11e9-bce2-ae54e022189f: Found 5 pods out of 5
+STEP: Ensuring each pod is running
+STEP: deleting ReplicationController wrapped-volume-race-2d160b3e-91c5-11e9-bce2-ae54e022189f in namespace e2e-tests-emptydir-wrapper-xq29x, will wait for the garbage collector to delete the pods
+Jun 18 12:34:46.524: INFO: Deleting ReplicationController wrapped-volume-race-2d160b3e-91c5-11e9-bce2-ae54e022189f took: 30.768477ms
+Jun 18 12:34:46.624: INFO: Terminating ReplicationController wrapped-volume-race-2d160b3e-91c5-11e9-bce2-ae54e022189f pods took: 100.197721ms
+STEP: Creating RC which spawns configmap-volume pods
+Jun 18 12:35:26.305: INFO: Pod name wrapped-volume-race-8c700672-91c5-11e9-bce2-ae54e022189f: Found 0 pods out of 5
+Jun 18 12:35:31.334: INFO: Pod name wrapped-volume-race-8c700672-91c5-11e9-bce2-ae54e022189f: Found 5 pods out of 5
+STEP: Ensuring each pod is running
+STEP: deleting ReplicationController wrapped-volume-race-8c700672-91c5-11e9-bce2-ae54e022189f in namespace e2e-tests-emptydir-wrapper-xq29x, will wait for the garbage collector to delete the pods
+Jun 18 12:37:25.522: INFO: Deleting ReplicationController wrapped-volume-race-8c700672-91c5-11e9-bce2-ae54e022189f took: 26.697911ms
+Jun 18 12:37:25.622: INFO: Terminating ReplicationController wrapped-volume-race-8c700672-91c5-11e9-bce2-ae54e022189f pods took: 100.41775ms
+STEP: Cleaning up the configMaps
+[AfterEach] [sig-storage] EmptyDir wrapper volumes
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
+Jun 18 12:38:07.850: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-tests-emptydir-wrapper-xq29x" for this suite.
+Jun 18 12:38:15.921: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 12:38:16.071: INFO: namespace: e2e-tests-emptydir-wrapper-xq29x, resource: bindings, ignored listing per whitelist
+Jun 18 12:38:16.410: INFO: namespace e2e-tests-emptydir-wrapper-xq29x deletion completed in 8.540661816s
 
-• [SLOW TEST:8.651 seconds]
-[sig-storage] ConfigMap
-/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:33
-  should be consumable from pods in volume with defaultMode set [NodeConformance] [Conformance]
+• [SLOW TEST:542.562 seconds]
+[sig-storage] EmptyDir wrapper volumes
+/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:22
+  should not cause race condition when used for configmaps [Serial] [Slow] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 ------------------------------
-SSSSS
+SSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-api-machinery] Watchers 
-  should be able to start watching from a specific resource version [Conformance]
+[k8s.io] Kubelet when scheduling a busybox command that always fails in a pod 
+  should have an terminated reason [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-[BeforeEach] [sig-api-machinery] Watchers
+[BeforeEach] [k8s.io] Kubelet
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
 STEP: Creating a kubernetes client
-May 13 19:40:17.197: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423
-STEP: Building a namespace api object, basename watch
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-watch-ckcnm
+Jun 18 12:38:16.411: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
+STEP: Building a namespace api object, basename kubelet-test
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-kubelet-test-2gx7x
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should be able to start watching from a specific resource version [Conformance]
+[BeforeEach] [k8s.io] Kubelet
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:37
+[BeforeEach] when scheduling a busybox command that always fails in a pod
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:81
+[It] should have an terminated reason [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-STEP: creating a new configmap
-STEP: modifying the configmap once
-STEP: modifying the configmap a second time
-STEP: deleting the configmap
-STEP: creating a watch on configmaps from the resource version returned by the first update
-STEP: Expecting to observe notifications for all changes to the configmap after the first update
-May 13 19:40:17.524: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-resource-version,GenerateName:,Namespace:e2e-tests-watch-ckcnm,SelfLink:/api/v1/namespaces/e2e-tests-watch-ckcnm/configmaps/e2e-watch-test-resource-version,UID:ef8765bd-75b6-11e9-b786-da20024d205c,ResourceVersion:46098,Generation:0,CreationTimestamp:2019-05-13 19:40:17 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: from-resource-version,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},}
-May 13 19:40:17.524: INFO: Got : DELETED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-resource-version,GenerateName:,Namespace:e2e-tests-watch-ckcnm,SelfLink:/api/v1/namespaces/e2e-tests-watch-ckcnm/configmaps/e2e-watch-test-resource-version,UID:ef8765bd-75b6-11e9-b786-da20024d205c,ResourceVersion:46099,Generation:0,CreationTimestamp:2019-05-13 19:40:17 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: from-resource-version,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},}
-[AfterEach] [sig-api-machinery] Watchers
+[AfterEach] [k8s.io] Kubelet
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
-May 13 19:40:17.524: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "e2e-tests-watch-ckcnm" for this suite.
-May 13 19:40:23.557: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-May 13 19:40:23.877: INFO: namespace: e2e-tests-watch-ckcnm, resource: bindings, ignored listing per whitelist
-May 13 19:40:23.909: INFO: namespace e2e-tests-watch-ckcnm deletion completed in 6.378830927s
+Jun 18 12:38:21.013: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-tests-kubelet-test-2gx7x" for this suite.
+Jun 18 12:38:27.153: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 12:38:27.417: INFO: namespace: e2e-tests-kubelet-test-2gx7x, resource: bindings, ignored listing per whitelist
+Jun 18 12:38:27.662: INFO: namespace e2e-tests-kubelet-test-2gx7x deletion completed in 6.577470622s
 
-• [SLOW TEST:6.712 seconds]
-[sig-api-machinery] Watchers
-/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:22
-  should be able to start watching from a specific resource version [Conformance]
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
+• [SLOW TEST:11.251 seconds]
+[k8s.io] Kubelet
+/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694
+  when scheduling a busybox command that always fails in a pod
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:78
+    should have an terminated reason [NodeConformance] [Conformance]
+    /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 ------------------------------
-S
+SSSSS
 ------------------------------
-[sig-api-machinery] Garbage collector 
-  should not delete dependents that have both valid owner and owner that's waiting for dependents to be deleted [Conformance]
+[sig-storage] Projected configMap 
+  should be consumable from pods in volume as non-root [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-[BeforeEach] [sig-api-machinery] Garbage collector
+[BeforeEach] [sig-storage] Projected configMap
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
 STEP: Creating a kubernetes client
-May 13 19:40:23.909: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423
-STEP: Building a namespace api object, basename gc
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-gc-cwd8q
+Jun 18 12:38:27.663: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
+STEP: Building a namespace api object, basename projected
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-projected-npbjq
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should not delete dependents that have both valid owner and owner that's waiting for dependents to be deleted [Conformance]
+[It] should be consumable from pods in volume as non-root [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-STEP: create the rc1
-STEP: create the rc2
-STEP: set half of pods created by rc simpletest-rc-to-be-deleted to have rc simpletest-rc-to-stay as owner as well
-STEP: delete the rc simpletest-rc-to-be-deleted
-STEP: wait for the rc to be deleted
-STEP: Gathering metrics
-W0513 19:40:34.329427      16 metrics_grabber.go:81] Master node is not registered. Grabbing metrics from Scheduler, ControllerManager and ClusterAutoscaler is disabled.
-May 13 19:40:34.329: INFO: For apiserver_request_count:
-For apiserver_request_latencies_summary:
-For etcd_helper_cache_entry_count:
-For etcd_helper_cache_hit_count:
-For etcd_helper_cache_miss_count:
-For etcd_request_cache_add_latencies_summary:
-For etcd_request_cache_get_latencies_summary:
-For etcd_request_latencies_summary:
-For garbage_collector_attempt_to_delete_queue_latency:
-For garbage_collector_attempt_to_delete_work_duration:
-For garbage_collector_attempt_to_orphan_queue_latency:
-For garbage_collector_attempt_to_orphan_work_duration:
-For garbage_collector_dirty_processing_latency_microseconds:
-For garbage_collector_event_processing_latency_microseconds:
-For garbage_collector_graph_changes_queue_latency:
-For garbage_collector_graph_changes_work_duration:
-For garbage_collector_orphan_processing_latency_microseconds:
-For namespace_queue_latency:
-For namespace_queue_latency_sum:
-For namespace_queue_latency_count:
-For namespace_retries:
-For namespace_work_duration:
-For namespace_work_duration_sum:
-For namespace_work_duration_count:
-For function_duration_seconds:
-For errors_total:
-For evicted_pods_total:
-
-[AfterEach] [sig-api-machinery] Garbage collector
+STEP: Creating configMap with name projected-configmap-test-volume-f8d79ddb-91c5-11e9-bce2-ae54e022189f
+STEP: Creating a pod to test consume configMaps
+Jun 18 12:38:29.255: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-f8e6bdf6-91c5-11e9-bce2-ae54e022189f" in namespace "e2e-tests-projected-npbjq" to be "success or failure"
+Jun 18 12:38:29.273: INFO: Pod "pod-projected-configmaps-f8e6bdf6-91c5-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 18.208458ms
+Jun 18 12:38:31.290: INFO: Pod "pod-projected-configmaps-f8e6bdf6-91c5-11e9-bce2-ae54e022189f": Phase="Running", Reason="", readiness=true. Elapsed: 2.03468284s
+Jun 18 12:38:33.305: INFO: Pod "pod-projected-configmaps-f8e6bdf6-91c5-11e9-bce2-ae54e022189f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.049725339s
+STEP: Saw pod success
+Jun 18 12:38:33.305: INFO: Pod "pod-projected-configmaps-f8e6bdf6-91c5-11e9-bce2-ae54e022189f" satisfied condition "success or failure"
+Jun 18 12:38:33.319: INFO: Trying to get logs from node 10.72.74.143 pod pod-projected-configmaps-f8e6bdf6-91c5-11e9-bce2-ae54e022189f container projected-configmap-volume-test: 
+STEP: delete the pod
+Jun 18 12:38:33.469: INFO: Waiting for pod pod-projected-configmaps-f8e6bdf6-91c5-11e9-bce2-ae54e022189f to disappear
+Jun 18 12:38:33.484: INFO: Pod pod-projected-configmaps-f8e6bdf6-91c5-11e9-bce2-ae54e022189f no longer exists
+[AfterEach] [sig-storage] Projected configMap
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
-May 13 19:40:34.329: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "e2e-tests-gc-cwd8q" for this suite.
-May 13 19:40:42.364: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-May 13 19:40:42.513: INFO: namespace: e2e-tests-gc-cwd8q, resource: bindings, ignored listing per whitelist
-May 13 19:40:42.604: INFO: namespace e2e-tests-gc-cwd8q deletion completed in 8.267915179s
+Jun 18 12:38:33.484: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-tests-projected-npbjq" for this suite.
+Jun 18 12:38:39.560: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 12:38:39.822: INFO: namespace: e2e-tests-projected-npbjq, resource: bindings, ignored listing per whitelist
+Jun 18 12:38:40.048: INFO: namespace e2e-tests-projected-npbjq deletion completed in 6.543984896s
 
-• [SLOW TEST:18.694 seconds]
-[sig-api-machinery] Garbage collector
-/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:22
-  should not delete dependents that have both valid owner and owner that's waiting for dependents to be deleted [Conformance]
+• [SLOW TEST:12.386 seconds]
+[sig-storage] Projected configMap
+/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_configmap.go:34
+  should be consumable from pods in volume as non-root [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 ------------------------------
-S
+SSSSSSSSSSSS
 ------------------------------
-[sig-node] Downward API 
-  should provide default limits.cpu/memory from node allocatable [NodeConformance] [Conformance]
+[sig-storage] EmptyDir volumes 
+  volume on default medium should have the correct mode [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-[BeforeEach] [sig-node] Downward API
+[BeforeEach] [sig-storage] EmptyDir volumes
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
 STEP: Creating a kubernetes client
-May 13 19:40:42.604: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423
-STEP: Building a namespace api object, basename downward-api
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-downward-api-4q5qt
+Jun 18 12:38:40.049: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
+STEP: Building a namespace api object, basename emptydir
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-emptydir-tjrrz
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should provide default limits.cpu/memory from node allocatable [NodeConformance] [Conformance]
+[It] volume on default medium should have the correct mode [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-STEP: Creating a pod to test downward api env vars
-May 13 19:40:42.904: INFO: Waiting up to 5m0s for pod "downward-api-feaf12a2-75b6-11e9-8f67-2632f168be36" in namespace "e2e-tests-downward-api-4q5qt" to be "success or failure"
-May 13 19:40:42.912: INFO: Pod "downward-api-feaf12a2-75b6-11e9-8f67-2632f168be36": Phase="Pending", Reason="", readiness=false. Elapsed: 7.778802ms
-May 13 19:40:44.921: INFO: Pod "downward-api-feaf12a2-75b6-11e9-8f67-2632f168be36": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.016410548s
+STEP: Creating a pod to test emptydir volume type on node default medium
+Jun 18 12:38:40.910: INFO: Waiting up to 5m0s for pod "pod-0076d37c-91c6-11e9-bce2-ae54e022189f" in namespace "e2e-tests-emptydir-tjrrz" to be "success or failure"
+Jun 18 12:38:40.924: INFO: Pod "pod-0076d37c-91c6-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 14.043013ms
+Jun 18 12:38:42.939: INFO: Pod "pod-0076d37c-91c6-11e9-bce2-ae54e022189f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.028843248s
 STEP: Saw pod success
-May 13 19:40:44.921: INFO: Pod "downward-api-feaf12a2-75b6-11e9-8f67-2632f168be36" satisfied condition "success or failure"
-May 13 19:40:44.929: INFO: Trying to get logs from node 10.170.219.153 pod downward-api-feaf12a2-75b6-11e9-8f67-2632f168be36 container dapi-container: 
+Jun 18 12:38:42.939: INFO: Pod "pod-0076d37c-91c6-11e9-bce2-ae54e022189f" satisfied condition "success or failure"
+Jun 18 12:38:42.953: INFO: Trying to get logs from node 10.72.74.143 pod pod-0076d37c-91c6-11e9-bce2-ae54e022189f container test-container: 
 STEP: delete the pod
-May 13 19:40:45.020: INFO: Waiting for pod downward-api-feaf12a2-75b6-11e9-8f67-2632f168be36 to disappear
-May 13 19:40:45.031: INFO: Pod downward-api-feaf12a2-75b6-11e9-8f67-2632f168be36 no longer exists
-[AfterEach] [sig-node] Downward API
+Jun 18 12:38:43.027: INFO: Waiting for pod pod-0076d37c-91c6-11e9-bce2-ae54e022189f to disappear
+Jun 18 12:38:43.042: INFO: Pod pod-0076d37c-91c6-11e9-bce2-ae54e022189f no longer exists
+[AfterEach] [sig-storage] EmptyDir volumes
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
-May 13 19:40:45.031: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "e2e-tests-downward-api-4q5qt" for this suite.
-May 13 19:40:51.081: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-May 13 19:40:51.213: INFO: namespace: e2e-tests-downward-api-4q5qt, resource: bindings, ignored listing per whitelist
-May 13 19:40:51.363: INFO: namespace e2e-tests-downward-api-4q5qt deletion completed in 6.324245909s
+Jun 18 12:38:43.042: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-tests-emptydir-tjrrz" for this suite.
+Jun 18 12:38:49.132: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 12:38:49.486: INFO: namespace: e2e-tests-emptydir-tjrrz, resource: bindings, ignored listing per whitelist
+Jun 18 12:38:49.606: INFO: namespace e2e-tests-emptydir-tjrrz deletion completed in 6.545564852s
 
-• [SLOW TEST:8.759 seconds]
-[sig-node] Downward API
-/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downward_api.go:38
-  should provide default limits.cpu/memory from node allocatable [NodeConformance] [Conformance]
+• [SLOW TEST:9.558 seconds]
+[sig-storage] EmptyDir volumes
+/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:40
+  volume on default medium should have the correct mode [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 ------------------------------
-SSSSSSS
+S
 ------------------------------
-[k8s.io] Probing container 
-  should *not* be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance]
+[sig-cli] Kubectl client [k8s.io] Update Demo 
+  should scale a replication controller  [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-[BeforeEach] [k8s.io] Probing container
+[BeforeEach] [sig-cli] Kubectl client
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
 STEP: Creating a kubernetes client
-May 13 19:40:51.363: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423
-STEP: Building a namespace api object, basename container-probe
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-container-probe-glhcw
+Jun 18 12:38:49.608: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
+STEP: Building a namespace api object, basename kubectl
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-kubectl-g7bx9
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [k8s.io] Probing container
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/container_probe.go:48
-[It] should *not* be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance]
+[BeforeEach] [sig-cli] Kubectl client
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:243
+[BeforeEach] [k8s.io] Update Demo
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:295
+[It] should scale a replication controller  [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-STEP: Creating pod liveness-exec in namespace e2e-tests-container-probe-glhcw
-May 13 19:40:53.650: INFO: Started pod liveness-exec in namespace e2e-tests-container-probe-glhcw
-STEP: checking the pod's current state and verifying that restartCount is present
-May 13 19:40:53.658: INFO: Initial restart count of pod liveness-exec is 0
-STEP: deleting the pod
-[AfterEach] [k8s.io] Probing container
+STEP: creating a replication controller
+Jun 18 12:38:50.091: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 create -f - --namespace=e2e-tests-kubectl-g7bx9'
+Jun 18 12:38:50.568: INFO: stderr: ""
+Jun 18 12:38:50.568: INFO: stdout: "replicationcontroller/update-demo-nautilus created\n"
+STEP: waiting for all containers in name=update-demo pods to come up.
+Jun 18 12:38:50.568: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=e2e-tests-kubectl-g7bx9'
+Jun 18 12:38:50.697: INFO: stderr: ""
+Jun 18 12:38:50.697: INFO: stdout: "update-demo-nautilus-dcbrd update-demo-nautilus-hqm4n "
+Jun 18 12:38:50.697: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 get pods update-demo-nautilus-dcbrd -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=e2e-tests-kubectl-g7bx9'
+Jun 18 12:38:50.829: INFO: stderr: ""
+Jun 18 12:38:50.829: INFO: stdout: ""
+Jun 18 12:38:50.829: INFO: update-demo-nautilus-dcbrd is created but not running
+Jun 18 12:38:55.829: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=e2e-tests-kubectl-g7bx9'
+Jun 18 12:38:55.990: INFO: stderr: ""
+Jun 18 12:38:55.990: INFO: stdout: "update-demo-nautilus-dcbrd update-demo-nautilus-hqm4n "
+Jun 18 12:38:55.991: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 get pods update-demo-nautilus-dcbrd -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=e2e-tests-kubectl-g7bx9'
+Jun 18 12:38:56.128: INFO: stderr: ""
+Jun 18 12:38:56.128: INFO: stdout: "true"
+Jun 18 12:38:56.128: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 get pods update-demo-nautilus-dcbrd -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=e2e-tests-kubectl-g7bx9'
+Jun 18 12:38:56.260: INFO: stderr: ""
+Jun 18 12:38:56.260: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0"
+Jun 18 12:38:56.261: INFO: validating pod update-demo-nautilus-dcbrd
+Jun 18 12:38:56.295: INFO: got data: {
+  "image": "nautilus.jpg"
+}
+
+Jun 18 12:38:56.295: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg .
+Jun 18 12:38:56.295: INFO: update-demo-nautilus-dcbrd is verified up and running
+Jun 18 12:38:56.295: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 get pods update-demo-nautilus-hqm4n -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=e2e-tests-kubectl-g7bx9'
+Jun 18 12:38:56.426: INFO: stderr: ""
+Jun 18 12:38:56.426: INFO: stdout: "true"
+Jun 18 12:38:56.426: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 get pods update-demo-nautilus-hqm4n -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=e2e-tests-kubectl-g7bx9'
+Jun 18 12:38:56.548: INFO: stderr: ""
+Jun 18 12:38:56.548: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0"
+Jun 18 12:38:56.548: INFO: validating pod update-demo-nautilus-hqm4n
+Jun 18 12:38:56.581: INFO: got data: {
+  "image": "nautilus.jpg"
+}
+
+Jun 18 12:38:56.581: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg .
+Jun 18 12:38:56.581: INFO: update-demo-nautilus-hqm4n is verified up and running
+STEP: scaling down the replication controller
+Jun 18 12:38:56.583: INFO: scanned /root for discovery docs: 
+Jun 18 12:38:56.583: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 scale rc update-demo-nautilus --replicas=1 --timeout=5m --namespace=e2e-tests-kubectl-g7bx9'
+Jun 18 12:38:57.810: INFO: stderr: ""
+Jun 18 12:38:57.810: INFO: stdout: "replicationcontroller/update-demo-nautilus scaled\n"
+STEP: waiting for all containers in name=update-demo pods to come up.
+Jun 18 12:38:57.810: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=e2e-tests-kubectl-g7bx9'
+Jun 18 12:38:57.945: INFO: stderr: ""
+Jun 18 12:38:57.945: INFO: stdout: "update-demo-nautilus-dcbrd update-demo-nautilus-hqm4n "
+STEP: Replicas for name=update-demo: expected=1 actual=2
+Jun 18 12:39:02.946: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=e2e-tests-kubectl-g7bx9'
+Jun 18 12:39:03.122: INFO: stderr: ""
+Jun 18 12:39:03.122: INFO: stdout: "update-demo-nautilus-dcbrd update-demo-nautilus-hqm4n "
+STEP: Replicas for name=update-demo: expected=1 actual=2
+Jun 18 12:39:08.122: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=e2e-tests-kubectl-g7bx9'
+Jun 18 12:39:08.353: INFO: stderr: ""
+Jun 18 12:39:08.353: INFO: stdout: "update-demo-nautilus-hqm4n "
+Jun 18 12:39:08.353: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 get pods update-demo-nautilus-hqm4n -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=e2e-tests-kubectl-g7bx9'
+Jun 18 12:39:08.493: INFO: stderr: ""
+Jun 18 12:39:08.493: INFO: stdout: "true"
+Jun 18 12:39:08.493: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 get pods update-demo-nautilus-hqm4n -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=e2e-tests-kubectl-g7bx9'
+Jun 18 12:39:08.641: INFO: stderr: ""
+Jun 18 12:39:08.641: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0"
+Jun 18 12:39:08.641: INFO: validating pod update-demo-nautilus-hqm4n
+Jun 18 12:39:08.666: INFO: got data: {
+  "image": "nautilus.jpg"
+}
+
+Jun 18 12:39:08.666: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg .
+Jun 18 12:39:08.666: INFO: update-demo-nautilus-hqm4n is verified up and running
+STEP: scaling up the replication controller
+Jun 18 12:39:08.668: INFO: scanned /root for discovery docs: 
+Jun 18 12:39:08.668: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 scale rc update-demo-nautilus --replicas=2 --timeout=5m --namespace=e2e-tests-kubectl-g7bx9'
+Jun 18 12:39:09.890: INFO: stderr: ""
+Jun 18 12:39:09.890: INFO: stdout: "replicationcontroller/update-demo-nautilus scaled\n"
+STEP: waiting for all containers in name=update-demo pods to come up.
+Jun 18 12:39:09.890: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=e2e-tests-kubectl-g7bx9'
+Jun 18 12:39:10.036: INFO: stderr: ""
+Jun 18 12:39:10.036: INFO: stdout: "update-demo-nautilus-hqm4n update-demo-nautilus-plnh7 "
+Jun 18 12:39:10.036: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 get pods update-demo-nautilus-hqm4n -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=e2e-tests-kubectl-g7bx9'
+Jun 18 12:39:10.174: INFO: stderr: ""
+Jun 18 12:39:10.174: INFO: stdout: "true"
+Jun 18 12:39:10.175: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 get pods update-demo-nautilus-hqm4n -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=e2e-tests-kubectl-g7bx9'
+Jun 18 12:39:10.387: INFO: stderr: ""
+Jun 18 12:39:10.387: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0"
+Jun 18 12:39:10.387: INFO: validating pod update-demo-nautilus-hqm4n
+Jun 18 12:39:10.411: INFO: got data: {
+  "image": "nautilus.jpg"
+}
+
+Jun 18 12:39:10.411: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg .
+Jun 18 12:39:10.411: INFO: update-demo-nautilus-hqm4n is verified up and running
+Jun 18 12:39:10.411: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 get pods update-demo-nautilus-plnh7 -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=e2e-tests-kubectl-g7bx9'
+Jun 18 12:39:10.535: INFO: stderr: ""
+Jun 18 12:39:10.535: INFO: stdout: "true"
+Jun 18 12:39:10.535: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 get pods update-demo-nautilus-plnh7 -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=e2e-tests-kubectl-g7bx9'
+Jun 18 12:39:10.672: INFO: stderr: ""
+Jun 18 12:39:10.672: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0"
+Jun 18 12:39:10.672: INFO: validating pod update-demo-nautilus-plnh7
+Jun 18 12:39:10.707: INFO: got data: {
+  "image": "nautilus.jpg"
+}
+
+Jun 18 12:39:10.707: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg .
+Jun 18 12:39:10.707: INFO: update-demo-nautilus-plnh7 is verified up and running
+STEP: using delete to clean up resources
+Jun 18 12:39:10.707: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 delete --grace-period=0 --force -f - --namespace=e2e-tests-kubectl-g7bx9'
+Jun 18 12:39:10.886: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n"
+Jun 18 12:39:10.886: INFO: stdout: "replicationcontroller \"update-demo-nautilus\" force deleted\n"
+Jun 18 12:39:10.887: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 get rc,svc -l name=update-demo --no-headers --namespace=e2e-tests-kubectl-g7bx9'
+Jun 18 12:39:11.026: INFO: stderr: "No resources found.\n"
+Jun 18 12:39:11.026: INFO: stdout: ""
+Jun 18 12:39:11.026: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 get pods -l name=update-demo --namespace=e2e-tests-kubectl-g7bx9 -o go-template={{ range .items }}{{ if not .metadata.deletionTimestamp }}{{ .metadata.name }}{{ "\n" }}{{ end }}{{ end }}'
+Jun 18 12:39:11.167: INFO: stderr: ""
+Jun 18 12:39:11.167: INFO: stdout: ""
+[AfterEach] [sig-cli] Kubectl client
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
-May 13 19:44:55.491: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "e2e-tests-container-probe-glhcw" for this suite.
-May 13 19:45:01.547: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-May 13 19:45:01.626: INFO: namespace: e2e-tests-container-probe-glhcw, resource: bindings, ignored listing per whitelist
-May 13 19:45:01.783: INFO: namespace e2e-tests-container-probe-glhcw deletion completed in 6.263066514s
+Jun 18 12:39:11.168: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-tests-kubectl-g7bx9" for this suite.
+Jun 18 12:39:19.242: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 12:39:19.785: INFO: namespace: e2e-tests-kubectl-g7bx9, resource: bindings, ignored listing per whitelist
+Jun 18 12:39:20.227: INFO: namespace e2e-tests-kubectl-g7bx9 deletion completed in 9.036202232s
 
-• [SLOW TEST:250.419 seconds]
-[k8s.io] Probing container
-/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694
-  should *not* be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance]
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
+• [SLOW TEST:30.619 seconds]
+[sig-cli] Kubectl client
+/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:22
+  [k8s.io] Update Demo
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694
+    should scale a replication controller  [Conformance]
+    /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 ------------------------------
-[sig-api-machinery] Garbage collector 
-  should orphan pods created by rc if delete options say so [Conformance]
+SSSSSSSSSSSSS
+------------------------------
+[k8s.io] Kubelet when scheduling a read only busybox container 
+  should not write to root filesystem [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-[BeforeEach] [sig-api-machinery] Garbage collector
+[BeforeEach] [k8s.io] Kubelet
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
 STEP: Creating a kubernetes client
-May 13 19:45:01.783: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423
-STEP: Building a namespace api object, basename gc
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-gc-98znn
+Jun 18 12:39:20.227: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
+STEP: Building a namespace api object, basename kubelet-test
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-kubelet-test-s4h9k
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should orphan pods created by rc if delete options say so [Conformance]
+[BeforeEach] [k8s.io] Kubelet
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:37
+[It] should not write to root filesystem [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-STEP: create the rc
-STEP: delete the rc
-STEP: wait for the rc to be deleted
-STEP: wait for 30 seconds to see if the garbage collector mistakenly deletes the pods
-STEP: Gathering metrics
-W0513 19:45:42.237491      16 metrics_grabber.go:81] Master node is not registered. Grabbing metrics from Scheduler, ControllerManager and ClusterAutoscaler is disabled.
-May 13 19:45:42.237: INFO: For apiserver_request_count:
-For apiserver_request_latencies_summary:
-For etcd_helper_cache_entry_count:
-For etcd_helper_cache_hit_count:
-For etcd_helper_cache_miss_count:
-For etcd_request_cache_add_latencies_summary:
-For etcd_request_cache_get_latencies_summary:
-For etcd_request_latencies_summary:
-For garbage_collector_attempt_to_delete_queue_latency:
-For garbage_collector_attempt_to_delete_work_duration:
-For garbage_collector_attempt_to_orphan_queue_latency:
-For garbage_collector_attempt_to_orphan_work_duration:
-For garbage_collector_dirty_processing_latency_microseconds:
-For garbage_collector_event_processing_latency_microseconds:
-For garbage_collector_graph_changes_queue_latency:
-For garbage_collector_graph_changes_work_duration:
-For garbage_collector_orphan_processing_latency_microseconds:
-For namespace_queue_latency:
-For namespace_queue_latency_sum:
-For namespace_queue_latency_count:
-For namespace_retries:
-For namespace_work_duration:
-For namespace_work_duration_sum:
-For namespace_work_duration_count:
-For function_duration_seconds:
-For errors_total:
-For evicted_pods_total:
-
-[AfterEach] [sig-api-machinery] Garbage collector
+[AfterEach] [k8s.io] Kubelet
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
-May 13 19:45:42.237: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "e2e-tests-gc-98znn" for this suite.
-May 13 19:45:50.354: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-May 13 19:45:50.505: INFO: namespace: e2e-tests-gc-98znn, resource: bindings, ignored listing per whitelist
-May 13 19:45:50.638: INFO: namespace e2e-tests-gc-98znn deletion completed in 8.312472835s
+Jun 18 12:39:22.812: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-tests-kubelet-test-s4h9k" for this suite.
+Jun 18 12:40:04.879: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 12:40:05.115: INFO: namespace: e2e-tests-kubelet-test-s4h9k, resource: bindings, ignored listing per whitelist
+Jun 18 12:40:05.484: INFO: namespace e2e-tests-kubelet-test-s4h9k deletion completed in 42.653141106s
 
-• [SLOW TEST:48.855 seconds]
-[sig-api-machinery] Garbage collector
-/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:22
-  should orphan pods created by rc if delete options say so [Conformance]
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
+• [SLOW TEST:45.257 seconds]
+[k8s.io] Kubelet
+/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694
+  when scheduling a read only busybox container
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:186
+    should not write to root filesystem [NodeConformance] [Conformance]
+    /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 ------------------------------
-SSS
+SSSSSSSSSS
 ------------------------------
-[sig-node] ConfigMap 
-  should be consumable via environment variable [NodeConformance] [Conformance]
+[sig-node] Downward API 
+  should provide pod UID as env vars [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-[BeforeEach] [sig-node] ConfigMap
+[BeforeEach] [sig-node] Downward API
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
 STEP: Creating a kubernetes client
-May 13 19:45:50.638: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423
-STEP: Building a namespace api object, basename configmap
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-configmap-8bvpx
+Jun 18 12:40:05.485: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
+STEP: Building a namespace api object, basename downward-api
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-downward-api-5vhth
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should be consumable via environment variable [NodeConformance] [Conformance]
+[It] should provide pod UID as env vars [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-STEP: Creating configMap e2e-tests-configmap-8bvpx/configmap-test-b64771e4-75b7-11e9-8f67-2632f168be36
-STEP: Creating a pod to test consume configMaps
-May 13 19:45:50.937: INFO: Waiting up to 5m0s for pod "pod-configmaps-b648d078-75b7-11e9-8f67-2632f168be36" in namespace "e2e-tests-configmap-8bvpx" to be "success or failure"
-May 13 19:45:50.944: INFO: Pod "pod-configmaps-b648d078-75b7-11e9-8f67-2632f168be36": Phase="Pending", Reason="", readiness=false. Elapsed: 7.406289ms
-May 13 19:45:52.968: INFO: Pod "pod-configmaps-b648d078-75b7-11e9-8f67-2632f168be36": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.031175045s
+STEP: Creating a pod to test downward api env vars
+Jun 18 12:40:06.040: INFO: Waiting up to 5m0s for pod "downward-api-33349705-91c6-11e9-bce2-ae54e022189f" in namespace "e2e-tests-downward-api-5vhth" to be "success or failure"
+Jun 18 12:40:06.053: INFO: Pod "downward-api-33349705-91c6-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 13.020096ms
+Jun 18 12:40:08.491: INFO: Pod "downward-api-33349705-91c6-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 2.45096525s
+Jun 18 12:40:10.506: INFO: Pod "downward-api-33349705-91c6-11e9-bce2-ae54e022189f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.466661348s
 STEP: Saw pod success
-May 13 19:45:52.968: INFO: Pod "pod-configmaps-b648d078-75b7-11e9-8f67-2632f168be36" satisfied condition "success or failure"
-May 13 19:45:52.975: INFO: Trying to get logs from node 10.170.219.153 pod pod-configmaps-b648d078-75b7-11e9-8f67-2632f168be36 container env-test: 
+Jun 18 12:40:10.506: INFO: Pod "downward-api-33349705-91c6-11e9-bce2-ae54e022189f" satisfied condition "success or failure"
+Jun 18 12:40:10.522: INFO: Trying to get logs from node 10.72.74.149 pod downward-api-33349705-91c6-11e9-bce2-ae54e022189f container dapi-container: 
 STEP: delete the pod
-May 13 19:45:53.020: INFO: Waiting for pod pod-configmaps-b648d078-75b7-11e9-8f67-2632f168be36 to disappear
-May 13 19:45:53.029: INFO: Pod pod-configmaps-b648d078-75b7-11e9-8f67-2632f168be36 no longer exists
-[AfterEach] [sig-node] ConfigMap
+Jun 18 12:40:10.626: INFO: Waiting for pod downward-api-33349705-91c6-11e9-bce2-ae54e022189f to disappear
+Jun 18 12:40:10.641: INFO: Pod downward-api-33349705-91c6-11e9-bce2-ae54e022189f no longer exists
+[AfterEach] [sig-node] Downward API
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
-May 13 19:45:53.029: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "e2e-tests-configmap-8bvpx" for this suite.
-May 13 19:45:59.062: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-May 13 19:45:59.252: INFO: namespace: e2e-tests-configmap-8bvpx, resource: bindings, ignored listing per whitelist
-May 13 19:45:59.516: INFO: namespace e2e-tests-configmap-8bvpx deletion completed in 6.479436644s
+Jun 18 12:40:10.642: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-tests-downward-api-5vhth" for this suite.
+Jun 18 12:40:16.711: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 12:40:17.177: INFO: namespace: e2e-tests-downward-api-5vhth, resource: bindings, ignored listing per whitelist
+Jun 18 12:40:17.278: INFO: namespace e2e-tests-downward-api-5vhth deletion completed in 6.619107117s
 
-• [SLOW TEST:8.877 seconds]
-[sig-node] ConfigMap
-/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/configmap.go:31
-  should be consumable via environment variable [NodeConformance] [Conformance]
+• [SLOW TEST:11.793 seconds]
+[sig-node] Downward API
+/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downward_api.go:38
+  should provide pod UID as env vars [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 ------------------------------
-SSSSSSSSSSSSSSSS
+SSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-storage] EmptyDir volumes 
-  should support (root,0666,default) [NodeConformance] [Conformance]
+[sig-cli] Kubectl client [k8s.io] Kubectl logs 
+  should be able to retrieve and filter logs  [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-[BeforeEach] [sig-storage] EmptyDir volumes
+[BeforeEach] [sig-cli] Kubectl client
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
 STEP: Creating a kubernetes client
-May 13 19:45:59.516: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423
-STEP: Building a namespace api object, basename emptydir
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-emptydir-h2nwr
+Jun 18 12:40:17.279: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
+STEP: Building a namespace api object, basename kubectl
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-kubectl-bqgx7
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should support (root,0666,default) [NodeConformance] [Conformance]
+[BeforeEach] [sig-cli] Kubectl client
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:243
+[BeforeEach] [k8s.io] Kubectl logs
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1134
+STEP: creating an rc
+Jun 18 12:40:17.742: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 create -f - --namespace=e2e-tests-kubectl-bqgx7'
+Jun 18 12:40:18.049: INFO: stderr: ""
+Jun 18 12:40:18.049: INFO: stdout: "replicationcontroller/redis-master created\n"
+[It] should be able to retrieve and filter logs  [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-STEP: Creating a pod to test emptydir 0666 on node default medium
-May 13 19:45:59.816: INFO: Waiting up to 5m0s for pod "pod-bb9285e0-75b7-11e9-8f67-2632f168be36" in namespace "e2e-tests-emptydir-h2nwr" to be "success or failure"
-May 13 19:45:59.824: INFO: Pod "pod-bb9285e0-75b7-11e9-8f67-2632f168be36": Phase="Pending", Reason="", readiness=false. Elapsed: 8.364025ms
-May 13 19:46:01.833: INFO: Pod "pod-bb9285e0-75b7-11e9-8f67-2632f168be36": Phase="Pending", Reason="", readiness=false. Elapsed: 2.016939374s
-May 13 19:46:03.853: INFO: Pod "pod-bb9285e0-75b7-11e9-8f67-2632f168be36": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.037410145s
-STEP: Saw pod success
-May 13 19:46:03.853: INFO: Pod "pod-bb9285e0-75b7-11e9-8f67-2632f168be36" satisfied condition "success or failure"
-May 13 19:46:03.860: INFO: Trying to get logs from node 10.170.219.177 pod pod-bb9285e0-75b7-11e9-8f67-2632f168be36 container test-container: 
-STEP: delete the pod
-May 13 19:46:03.951: INFO: Waiting for pod pod-bb9285e0-75b7-11e9-8f67-2632f168be36 to disappear
-May 13 19:46:03.958: INFO: Pod pod-bb9285e0-75b7-11e9-8f67-2632f168be36 no longer exists
-[AfterEach] [sig-storage] EmptyDir volumes
+STEP: Waiting for Redis master to start.
+Jun 18 12:40:19.065: INFO: Selector matched 1 pods for map[app:redis]
+Jun 18 12:40:19.065: INFO: Found 0 / 1
+Jun 18 12:40:20.063: INFO: Selector matched 1 pods for map[app:redis]
+Jun 18 12:40:20.063: INFO: Found 0 / 1
+Jun 18 12:40:21.065: INFO: Selector matched 1 pods for map[app:redis]
+Jun 18 12:40:21.065: INFO: Found 1 / 1
+Jun 18 12:40:21.065: INFO: WaitFor completed with timeout 5m0s.  Pods found = 1 out of 1
+Jun 18 12:40:21.081: INFO: Selector matched 1 pods for map[app:redis]
+Jun 18 12:40:21.081: INFO: ForEach: Found 1 pods from the filter.  Now looping through them.
+STEP: checking for a matching strings
+Jun 18 12:40:21.081: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 logs redis-master-bhdmv redis-master --namespace=e2e-tests-kubectl-bqgx7'
+Jun 18 12:40:21.351: INFO: stderr: ""
+Jun 18 12:40:21.351: INFO: stdout: "                _._                                                  \n           _.-``__ ''-._                                             \n      _.-``    `.  `_.  ''-._           Redis 3.2.12 (35a5711f/0) 64 bit\n  .-`` .-```.  ```\\/    _.,_ ''-._                                   \n (    '      ,       .-`  | `,    )     Running in standalone mode\n |`-._`-...-` __...-.``-._|'` _.-'|     Port: 6379\n |    `-._   `._    /     _.-'    |     PID: 1\n  `-._    `-._  `-./  _.-'    _.-'                                   \n |`-._`-._    `-.__.-'    _.-'_.-'|                                  \n |    `-._`-._        _.-'_.-'    |           http://redis.io        \n  `-._    `-._`-.__.-'_.-'    _.-'                                   \n |`-._`-._    `-.__.-'    _.-'_.-'|                                  \n |    `-._`-._        _.-'_.-'    |                                  \n  `-._    `-._`-.__.-'_.-'    _.-'                                   \n      `-._    `-.__.-'    _.-'                                       \n          `-._        _.-'                                           \n              `-.__.-'                                               \n\n1:M 18 Jun 12:40:19.478 # WARNING: The TCP backlog setting of 511 cannot be enforced because /proc/sys/net/core/somaxconn is set to the lower value of 128.\n1:M 18 Jun 12:40:19.478 # Server started, Redis version 3.2.12\n1:M 18 Jun 12:40:19.478 # WARNING you have Transparent Huge Pages (THP) support enabled in your kernel. This will create latency and memory usage issues with Redis. To fix this issue run the command 'echo never > /sys/kernel/mm/transparent_hugepage/enabled' as root, and add it to your /etc/rc.local in order to retain the setting after a reboot. Redis must be restarted after THP is disabled.\n1:M 18 Jun 12:40:19.478 * The server is now ready to accept connections on port 6379\n"
+STEP: limiting log lines
+Jun 18 12:40:21.351: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 log redis-master-bhdmv redis-master --namespace=e2e-tests-kubectl-bqgx7 --tail=1'
+Jun 18 12:40:21.523: INFO: stderr: ""
+Jun 18 12:40:21.523: INFO: stdout: "1:M 18 Jun 12:40:19.478 * The server is now ready to accept connections on port 6379\n"
+STEP: limiting log bytes
+Jun 18 12:40:21.523: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 log redis-master-bhdmv redis-master --namespace=e2e-tests-kubectl-bqgx7 --limit-bytes=1'
+Jun 18 12:40:21.783: INFO: stderr: ""
+Jun 18 12:40:21.783: INFO: stdout: " "
+STEP: exposing timestamps
+Jun 18 12:40:21.783: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 log redis-master-bhdmv redis-master --namespace=e2e-tests-kubectl-bqgx7 --tail=1 --timestamps'
+Jun 18 12:40:21.928: INFO: stderr: ""
+Jun 18 12:40:21.928: INFO: stdout: "2019-06-18T12:40:19.479252347Z 1:M 18 Jun 12:40:19.478 * The server is now ready to accept connections on port 6379\n"
+STEP: restricting to a time range
+Jun 18 12:40:24.429: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 log redis-master-bhdmv redis-master --namespace=e2e-tests-kubectl-bqgx7 --since=1s'
+Jun 18 12:40:24.876: INFO: stderr: ""
+Jun 18 12:40:24.876: INFO: stdout: ""
+Jun 18 12:40:24.876: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 log redis-master-bhdmv redis-master --namespace=e2e-tests-kubectl-bqgx7 --since=24h'
+Jun 18 12:40:25.033: INFO: stderr: ""
+Jun 18 12:40:25.033: INFO: stdout: "                _._                                                  \n           _.-``__ ''-._                                             \n      _.-``    `.  `_.  ''-._           Redis 3.2.12 (35a5711f/0) 64 bit\n  .-`` .-```.  ```\\/    _.,_ ''-._                                   \n (    '      ,       .-`  | `,    )     Running in standalone mode\n |`-._`-...-` __...-.``-._|'` _.-'|     Port: 6379\n |    `-._   `._    /     _.-'    |     PID: 1\n  `-._    `-._  `-./  _.-'    _.-'                                   \n |`-._`-._    `-.__.-'    _.-'_.-'|                                  \n |    `-._`-._        _.-'_.-'    |           http://redis.io        \n  `-._    `-._`-.__.-'_.-'    _.-'                                   \n |`-._`-._    `-.__.-'    _.-'_.-'|                                  \n |    `-._`-._        _.-'_.-'    |                                  \n  `-._    `-._`-.__.-'_.-'    _.-'                                   \n      `-._    `-.__.-'    _.-'                                       \n          `-._        _.-'                                           \n              `-.__.-'                                               \n\n1:M 18 Jun 12:40:19.478 # WARNING: The TCP backlog setting of 511 cannot be enforced because /proc/sys/net/core/somaxconn is set to the lower value of 128.\n1:M 18 Jun 12:40:19.478 # Server started, Redis version 3.2.12\n1:M 18 Jun 12:40:19.478 # WARNING you have Transparent Huge Pages (THP) support enabled in your kernel. This will create latency and memory usage issues with Redis. To fix this issue run the command 'echo never > /sys/kernel/mm/transparent_hugepage/enabled' as root, and add it to your /etc/rc.local in order to retain the setting after a reboot. Redis must be restarted after THP is disabled.\n1:M 18 Jun 12:40:19.478 * The server is now ready to accept connections on port 6379\n"
+[AfterEach] [k8s.io] Kubectl logs
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1140
+STEP: using delete to clean up resources
+Jun 18 12:40:25.033: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 delete --grace-period=0 --force -f - --namespace=e2e-tests-kubectl-bqgx7'
+Jun 18 12:40:25.174: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n"
+Jun 18 12:40:25.174: INFO: stdout: "replicationcontroller \"redis-master\" force deleted\n"
+Jun 18 12:40:25.174: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 get rc,svc -l name=nginx --no-headers --namespace=e2e-tests-kubectl-bqgx7'
+Jun 18 12:40:25.324: INFO: stderr: "No resources found.\n"
+Jun 18 12:40:25.324: INFO: stdout: ""
+Jun 18 12:40:25.324: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 get pods -l name=nginx --namespace=e2e-tests-kubectl-bqgx7 -o go-template={{ range .items }}{{ if not .metadata.deletionTimestamp }}{{ .metadata.name }}{{ "\n" }}{{ end }}{{ end }}'
+Jun 18 12:40:25.469: INFO: stderr: ""
+Jun 18 12:40:25.469: INFO: stdout: ""
+[AfterEach] [sig-cli] Kubectl client
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
-May 13 19:46:03.958: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "e2e-tests-emptydir-h2nwr" for this suite.
-May 13 19:46:09.993: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-May 13 19:46:10.292: INFO: namespace: e2e-tests-emptydir-h2nwr, resource: bindings, ignored listing per whitelist
-May 13 19:46:10.331: INFO: namespace e2e-tests-emptydir-h2nwr deletion completed in 6.365471555s
+Jun 18 12:40:25.469: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-tests-kubectl-bqgx7" for this suite.
+Jun 18 12:40:33.794: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 12:40:33.932: INFO: namespace: e2e-tests-kubectl-bqgx7, resource: bindings, ignored listing per whitelist
+Jun 18 12:40:34.402: INFO: namespace e2e-tests-kubectl-bqgx7 deletion completed in 8.913366606s
 
-• [SLOW TEST:10.815 seconds]
-[sig-storage] EmptyDir volumes
-/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:40
-  should support (root,0666,default) [NodeConformance] [Conformance]
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
+• [SLOW TEST:17.124 seconds]
+[sig-cli] Kubectl client
+/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:22
+  [k8s.io] Kubectl logs
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694
+    should be able to retrieve and filter logs  [Conformance]
+    /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 ------------------------------
-SSSSSSSSSSSS
+SSSSS
 ------------------------------
-[sig-storage] EmptyDir volumes 
-  should support (root,0644,tmpfs) [NodeConformance] [Conformance]
+[sig-auth] ServiceAccounts 
+  should allow opting out of API token automount  [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-[BeforeEach] [sig-storage] EmptyDir volumes
+[BeforeEach] [sig-auth] ServiceAccounts
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
 STEP: Creating a kubernetes client
-May 13 19:46:10.332: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423
-STEP: Building a namespace api object, basename emptydir
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-emptydir-lffwg
+Jun 18 12:40:34.403: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
+STEP: Building a namespace api object, basename svcaccounts
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-svcaccounts-fwvsr
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should support (root,0644,tmpfs) [NodeConformance] [Conformance]
+[It] should allow opting out of API token automount  [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-STEP: Creating a pod to test emptydir 0644 on tmpfs
-May 13 19:46:10.640: INFO: Waiting up to 5m0s for pod "pod-c207a9d7-75b7-11e9-8f67-2632f168be36" in namespace "e2e-tests-emptydir-lffwg" to be "success or failure"
-May 13 19:46:10.649: INFO: Pod "pod-c207a9d7-75b7-11e9-8f67-2632f168be36": Phase="Pending", Reason="", readiness=false. Elapsed: 8.267344ms
-May 13 19:46:12.657: INFO: Pod "pod-c207a9d7-75b7-11e9-8f67-2632f168be36": Phase="Running", Reason="", readiness=true. Elapsed: 2.016946101s
-May 13 19:46:14.681: INFO: Pod "pod-c207a9d7-75b7-11e9-8f67-2632f168be36": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.041041615s
-STEP: Saw pod success
-May 13 19:46:14.681: INFO: Pod "pod-c207a9d7-75b7-11e9-8f67-2632f168be36" satisfied condition "success or failure"
-May 13 19:46:14.690: INFO: Trying to get logs from node 10.170.219.177 pod pod-c207a9d7-75b7-11e9-8f67-2632f168be36 container test-container: 
-STEP: delete the pod
-May 13 19:46:14.731: INFO: Waiting for pod pod-c207a9d7-75b7-11e9-8f67-2632f168be36 to disappear
-May 13 19:46:14.740: INFO: Pod pod-c207a9d7-75b7-11e9-8f67-2632f168be36 no longer exists
-[AfterEach] [sig-storage] EmptyDir volumes
+STEP: getting the auto-created API token
+Jun 18 12:40:38.757: INFO: created pod pod-service-account-defaultsa
+Jun 18 12:40:38.757: INFO: pod pod-service-account-defaultsa service account token volume mount: true
+Jun 18 12:40:38.775: INFO: created pod pod-service-account-mountsa
+Jun 18 12:40:38.775: INFO: pod pod-service-account-mountsa service account token volume mount: true
+Jun 18 12:40:38.807: INFO: created pod pod-service-account-nomountsa
+Jun 18 12:40:38.807: INFO: pod pod-service-account-nomountsa service account token volume mount: false
+Jun 18 12:40:38.828: INFO: created pod pod-service-account-defaultsa-mountspec
+Jun 18 12:40:38.828: INFO: pod pod-service-account-defaultsa-mountspec service account token volume mount: true
+Jun 18 12:40:38.844: INFO: created pod pod-service-account-mountsa-mountspec
+Jun 18 12:40:38.844: INFO: pod pod-service-account-mountsa-mountspec service account token volume mount: true
+Jun 18 12:40:38.863: INFO: created pod pod-service-account-nomountsa-mountspec
+Jun 18 12:40:38.863: INFO: pod pod-service-account-nomountsa-mountspec service account token volume mount: true
+Jun 18 12:40:38.880: INFO: created pod pod-service-account-defaultsa-nomountspec
+Jun 18 12:40:38.880: INFO: pod pod-service-account-defaultsa-nomountspec service account token volume mount: false
+Jun 18 12:40:38.898: INFO: created pod pod-service-account-mountsa-nomountspec
+Jun 18 12:40:38.898: INFO: pod pod-service-account-mountsa-nomountspec service account token volume mount: false
+Jun 18 12:40:38.920: INFO: created pod pod-service-account-nomountsa-nomountspec
+Jun 18 12:40:38.920: INFO: pod pod-service-account-nomountsa-nomountspec service account token volume mount: false
+[AfterEach] [sig-auth] ServiceAccounts
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
-May 13 19:46:14.740: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "e2e-tests-emptydir-lffwg" for this suite.
-May 13 19:46:20.848: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-May 13 19:46:20.993: INFO: namespace: e2e-tests-emptydir-lffwg, resource: bindings, ignored listing per whitelist
-May 13 19:46:21.140: INFO: namespace e2e-tests-emptydir-lffwg deletion completed in 6.3205196s
+Jun 18 12:40:38.920: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-tests-svcaccounts-fwvsr" for this suite.
+Jun 18 12:40:47.050: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 12:40:47.235: INFO: namespace: e2e-tests-svcaccounts-fwvsr, resource: bindings, ignored listing per whitelist
+Jun 18 12:40:47.562: INFO: namespace e2e-tests-svcaccounts-fwvsr deletion completed in 8.60923952s
 
-• [SLOW TEST:10.809 seconds]
-[sig-storage] EmptyDir volumes
-/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:40
-  should support (root,0644,tmpfs) [NodeConformance] [Conformance]
+• [SLOW TEST:13.160 seconds]
+[sig-auth] ServiceAccounts
+/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/auth/framework.go:22
+  should allow opting out of API token automount  [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 ------------------------------
-SSSSS
+SSSSSSSSSSS
 ------------------------------
-[sig-network] Services 
-  should provide secure master service  [Conformance]
+[sig-api-machinery] Watchers 
+  should be able to restart watching from the last resource version observed by the previous watch [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-[BeforeEach] [sig-network] Services
+[BeforeEach] [sig-api-machinery] Watchers
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
 STEP: Creating a kubernetes client
-May 13 19:46:21.141: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423
-STEP: Building a namespace api object, basename services
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-services-jlvsb
+Jun 18 12:40:47.563: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
+STEP: Building a namespace api object, basename watch
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-watch-r8s22
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-network] Services
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:85
-[It] should provide secure master service  [Conformance]
+[It] should be able to restart watching from the last resource version observed by the previous watch [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-[AfterEach] [sig-network] Services
+STEP: creating a watch on configmaps
+STEP: creating a new configmap
+STEP: modifying the configmap once
+STEP: closing the watch once it receives two notifications
+Jun 18 12:40:49.276: INFO: Got : ADDED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-watch-closed,GenerateName:,Namespace:e2e-tests-watch-r8s22,SelfLink:/api/v1/namespaces/e2e-tests-watch-r8s22/configmaps/e2e-watch-test-watch-closed,UID:4c444736-91c6-11e9-bf44-fa6f350b29f0,ResourceVersion:104495,Generation:0,CreationTimestamp:2019-06-18 12:40:48 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: watch-closed-and-restarted,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Data:map[string]string{},BinaryData:map[string][]byte{},}
+Jun 18 12:40:49.276: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-watch-closed,GenerateName:,Namespace:e2e-tests-watch-r8s22,SelfLink:/api/v1/namespaces/e2e-tests-watch-r8s22/configmaps/e2e-watch-test-watch-closed,UID:4c444736-91c6-11e9-bf44-fa6f350b29f0,ResourceVersion:104496,Generation:0,CreationTimestamp:2019-06-18 12:40:48 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: watch-closed-and-restarted,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Data:map[string]string{mutation: 1,},BinaryData:map[string][]byte{},}
+STEP: modifying the configmap a second time, while the watch is closed
+STEP: creating a new watch on configmaps from the last resource version observed by the first watch
+STEP: deleting the configmap
+STEP: Expecting to observe notifications for all changes to the configmap since the first watch closed
+Jun 18 12:40:49.334: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-watch-closed,GenerateName:,Namespace:e2e-tests-watch-r8s22,SelfLink:/api/v1/namespaces/e2e-tests-watch-r8s22/configmaps/e2e-watch-test-watch-closed,UID:4c444736-91c6-11e9-bf44-fa6f350b29f0,ResourceVersion:104497,Generation:0,CreationTimestamp:2019-06-18 12:40:48 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: watch-closed-and-restarted,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},}
+Jun 18 12:40:49.334: INFO: Got : DELETED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-watch-closed,GenerateName:,Namespace:e2e-tests-watch-r8s22,SelfLink:/api/v1/namespaces/e2e-tests-watch-r8s22/configmaps/e2e-watch-test-watch-closed,UID:4c444736-91c6-11e9-bf44-fa6f350b29f0,ResourceVersion:104498,Generation:0,CreationTimestamp:2019-06-18 12:40:48 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: watch-closed-and-restarted,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},}
+[AfterEach] [sig-api-machinery] Watchers
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
-May 13 19:46:21.473: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "e2e-tests-services-jlvsb" for this suite.
-May 13 19:46:27.507: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-May 13 19:46:27.557: INFO: namespace: e2e-tests-services-jlvsb, resource: bindings, ignored listing per whitelist
-May 13 19:46:27.935: INFO: namespace e2e-tests-services-jlvsb deletion completed in 6.45541309s
-[AfterEach] [sig-network] Services
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:90
+Jun 18 12:40:49.334: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-tests-watch-r8s22" for this suite.
+Jun 18 12:40:57.423: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 12:40:59.019: INFO: namespace: e2e-tests-watch-r8s22, resource: bindings, ignored listing per whitelist
+Jun 18 12:40:59.200: INFO: namespace e2e-tests-watch-r8s22 deletion completed in 9.843509494s
 
-• [SLOW TEST:6.795 seconds]
-[sig-network] Services
-/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:22
-  should provide secure master service  [Conformance]
+• [SLOW TEST:11.637 seconds]
+[sig-api-machinery] Watchers
+/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:22
+  should be able to restart watching from the last resource version observed by the previous watch [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 ------------------------------
-[sig-storage] EmptyDir volumes 
-  should support (non-root,0666,tmpfs) [NodeConformance] [Conformance]
+[sig-cli] Kubectl client [k8s.io] Proxy server 
+  should support --unix-socket=/path  [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-[BeforeEach] [sig-storage] EmptyDir volumes
+[BeforeEach] [sig-cli] Kubectl client
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
 STEP: Creating a kubernetes client
-May 13 19:46:27.936: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423
-STEP: Building a namespace api object, basename emptydir
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-emptydir-xsds4
+Jun 18 12:40:59.201: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
+STEP: Building a namespace api object, basename kubectl
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-kubectl-fhhjx
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should support (non-root,0666,tmpfs) [NodeConformance] [Conformance]
+[BeforeEach] [sig-cli] Kubectl client
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:243
+[It] should support --unix-socket=/path  [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-STEP: Creating a pod to test emptydir 0666 on tmpfs
-May 13 19:46:28.228: INFO: Waiting up to 5m0s for pod "pod-cc83162e-75b7-11e9-8f67-2632f168be36" in namespace "e2e-tests-emptydir-xsds4" to be "success or failure"
-May 13 19:46:28.237: INFO: Pod "pod-cc83162e-75b7-11e9-8f67-2632f168be36": Phase="Pending", Reason="", readiness=false. Elapsed: 8.478597ms
-May 13 19:46:30.245: INFO: Pod "pod-cc83162e-75b7-11e9-8f67-2632f168be36": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.016889938s
-STEP: Saw pod success
-May 13 19:46:30.245: INFO: Pod "pod-cc83162e-75b7-11e9-8f67-2632f168be36" satisfied condition "success or failure"
-May 13 19:46:30.252: INFO: Trying to get logs from node 10.170.219.153 pod pod-cc83162e-75b7-11e9-8f67-2632f168be36 container test-container: 
-STEP: delete the pod
-May 13 19:46:30.291: INFO: Waiting for pod pod-cc83162e-75b7-11e9-8f67-2632f168be36 to disappear
-May 13 19:46:30.298: INFO: Pod pod-cc83162e-75b7-11e9-8f67-2632f168be36 no longer exists
-[AfterEach] [sig-storage] EmptyDir volumes
+STEP: Starting the proxy
+Jun 18 12:41:00.632: INFO: Asynchronously running '/usr/local/bin/kubectl kubectl --kubeconfig=/tmp/kubeconfig-953583206 proxy --unix-socket=/tmp/kubectl-proxy-unix640593037/test'
+STEP: retrieving proxy /api/ output
+[AfterEach] [sig-cli] Kubectl client
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
-May 13 19:46:30.299: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "e2e-tests-emptydir-xsds4" for this suite.
-May 13 19:46:36.347: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-May 13 19:46:36.485: INFO: namespace: e2e-tests-emptydir-xsds4, resource: bindings, ignored listing per whitelist
-May 13 19:46:36.701: INFO: namespace e2e-tests-emptydir-xsds4 deletion completed in 6.395214572s
+Jun 18 12:41:00.707: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-tests-kubectl-fhhjx" for this suite.
+Jun 18 12:41:06.777: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 12:41:06.833: INFO: namespace: e2e-tests-kubectl-fhhjx, resource: bindings, ignored listing per whitelist
+Jun 18 12:41:07.290: INFO: namespace e2e-tests-kubectl-fhhjx deletion completed in 6.563286468s
 
-• [SLOW TEST:8.765 seconds]
-[sig-storage] EmptyDir volumes
-/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:40
-  should support (non-root,0666,tmpfs) [NodeConformance] [Conformance]
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
+• [SLOW TEST:8.090 seconds]
+[sig-cli] Kubectl client
+/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:22
+  [k8s.io] Proxy server
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694
+    should support --unix-socket=/path  [Conformance]
+    /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 ------------------------------
-SSSSSSSSSS
+S
 ------------------------------
-[sig-storage] ConfigMap 
-  should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance]
+[sig-storage] Downward API volume 
+  should set DefaultMode on files [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-[BeforeEach] [sig-storage] ConfigMap
+[BeforeEach] [sig-storage] Downward API volume
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
 STEP: Creating a kubernetes client
-May 13 19:46:36.702: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423
-STEP: Building a namespace api object, basename configmap
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-configmap-z5z64
+Jun 18 12:41:07.291: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
+STEP: Building a namespace api object, basename downward-api
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-downward-api-qbxc5
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance]
+[BeforeEach] [sig-storage] Downward API volume
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:39
+[It] should set DefaultMode on files [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-STEP: Creating configMap with name configmap-test-volume-d1bbd050-75b7-11e9-8f67-2632f168be36
-STEP: Creating a pod to test consume configMaps
-May 13 19:46:36.996: INFO: Waiting up to 5m0s for pod "pod-configmaps-d1bd190c-75b7-11e9-8f67-2632f168be36" in namespace "e2e-tests-configmap-z5z64" to be "success or failure"
-May 13 19:46:37.004: INFO: Pod "pod-configmaps-d1bd190c-75b7-11e9-8f67-2632f168be36": Phase="Pending", Reason="", readiness=false. Elapsed: 7.904391ms
-May 13 19:46:39.012: INFO: Pod "pod-configmaps-d1bd190c-75b7-11e9-8f67-2632f168be36": Phase="Pending", Reason="", readiness=false. Elapsed: 2.016089297s
-May 13 19:46:41.021: INFO: Pod "pod-configmaps-d1bd190c-75b7-11e9-8f67-2632f168be36": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.024957527s
+STEP: Creating a pod to test downward API volume plugin
+Jun 18 12:41:07.873: INFO: Waiting up to 5m0s for pod "downwardapi-volume-580f401e-91c6-11e9-bce2-ae54e022189f" in namespace "e2e-tests-downward-api-qbxc5" to be "success or failure"
+Jun 18 12:41:07.888: INFO: Pod "downwardapi-volume-580f401e-91c6-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 14.887166ms
+Jun 18 12:41:09.904: INFO: Pod "downwardapi-volume-580f401e-91c6-11e9-bce2-ae54e022189f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.030609487s
 STEP: Saw pod success
-May 13 19:46:41.021: INFO: Pod "pod-configmaps-d1bd190c-75b7-11e9-8f67-2632f168be36" satisfied condition "success or failure"
-May 13 19:46:41.029: INFO: Trying to get logs from node 10.170.219.153 pod pod-configmaps-d1bd190c-75b7-11e9-8f67-2632f168be36 container configmap-volume-test: 
+Jun 18 12:41:09.904: INFO: Pod "downwardapi-volume-580f401e-91c6-11e9-bce2-ae54e022189f" satisfied condition "success or failure"
+Jun 18 12:41:09.919: INFO: Trying to get logs from node 10.72.74.149 pod downwardapi-volume-580f401e-91c6-11e9-bce2-ae54e022189f container client-container: 
 STEP: delete the pod
-May 13 19:46:41.120: INFO: Waiting for pod pod-configmaps-d1bd190c-75b7-11e9-8f67-2632f168be36 to disappear
-May 13 19:46:41.129: INFO: Pod pod-configmaps-d1bd190c-75b7-11e9-8f67-2632f168be36 no longer exists
-[AfterEach] [sig-storage] ConfigMap
+Jun 18 12:41:10.006: INFO: Waiting for pod downwardapi-volume-580f401e-91c6-11e9-bce2-ae54e022189f to disappear
+Jun 18 12:41:10.020: INFO: Pod downwardapi-volume-580f401e-91c6-11e9-bce2-ae54e022189f no longer exists
+[AfterEach] [sig-storage] Downward API volume
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
-May 13 19:46:41.129: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "e2e-tests-configmap-z5z64" for this suite.
-May 13 19:46:47.176: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-May 13 19:46:47.387: INFO: namespace: e2e-tests-configmap-z5z64, resource: bindings, ignored listing per whitelist
-May 13 19:46:47.497: INFO: namespace e2e-tests-configmap-z5z64 deletion completed in 6.35999882s
+Jun 18 12:41:10.020: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-tests-downward-api-qbxc5" for this suite.
+Jun 18 12:41:18.114: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 12:41:18.605: INFO: namespace: e2e-tests-downward-api-qbxc5, resource: bindings, ignored listing per whitelist
+Jun 18 12:41:18.746: INFO: namespace e2e-tests-downward-api-qbxc5 deletion completed in 8.703851603s
 
-• [SLOW TEST:10.795 seconds]
-[sig-storage] ConfigMap
-/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:33
-  should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance]
+• [SLOW TEST:11.455 seconds]
+[sig-storage] Downward API volume
+/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:34
+  should set DefaultMode on files [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 ------------------------------
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+SSSS
 ------------------------------
-[sig-storage] Projected configMap 
-  should be consumable from pods in volume with defaultMode set [NodeConformance] [Conformance]
+[sig-storage] Subpath Atomic writer volumes 
+  should support subpaths with projected pod [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-[BeforeEach] [sig-storage] Projected configMap
+[BeforeEach] [sig-storage] Subpath
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
 STEP: Creating a kubernetes client
-May 13 19:46:47.501: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423
-STEP: Building a namespace api object, basename projected
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-projected-6mclt
+Jun 18 12:41:18.746: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
+STEP: Building a namespace api object, basename subpath
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-subpath-fqzhc
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should be consumable from pods in volume with defaultMode set [NodeConformance] [Conformance]
+[BeforeEach] Atomic writer volumes
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:38
+STEP: Setting up data
+[It] should support subpaths with projected pod [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-STEP: Creating configMap with name projected-configmap-test-volume-d82d8c89-75b7-11e9-8f67-2632f168be36
-STEP: Creating a pod to test consume configMaps
-May 13 19:46:47.809: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-d82efa67-75b7-11e9-8f67-2632f168be36" in namespace "e2e-tests-projected-6mclt" to be "success or failure"
-May 13 19:46:47.819: INFO: Pod "pod-projected-configmaps-d82efa67-75b7-11e9-8f67-2632f168be36": Phase="Pending", Reason="", readiness=false. Elapsed: 10.31962ms
-May 13 19:46:49.828: INFO: Pod "pod-projected-configmaps-d82efa67-75b7-11e9-8f67-2632f168be36": Phase="Running", Reason="", readiness=true. Elapsed: 2.01885252s
-May 13 19:46:51.837: INFO: Pod "pod-projected-configmaps-d82efa67-75b7-11e9-8f67-2632f168be36": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.027711716s
+STEP: Creating pod pod-subpath-test-projected-f57z
+STEP: Creating a pod to test atomic-volume-subpath
+Jun 18 12:41:19.350: INFO: Waiting up to 5m0s for pod "pod-subpath-test-projected-f57z" in namespace "e2e-tests-subpath-fqzhc" to be "success or failure"
+Jun 18 12:41:19.366: INFO: Pod "pod-subpath-test-projected-f57z": Phase="Pending", Reason="", readiness=false. Elapsed: 15.492815ms
+Jun 18 12:41:21.384: INFO: Pod "pod-subpath-test-projected-f57z": Phase="Pending", Reason="", readiness=false. Elapsed: 2.033799726s
+Jun 18 12:41:23.399: INFO: Pod "pod-subpath-test-projected-f57z": Phase="Running", Reason="", readiness=false. Elapsed: 4.048860302s
+Jun 18 12:41:26.117: INFO: Pod "pod-subpath-test-projected-f57z": Phase="Running", Reason="", readiness=false. Elapsed: 6.767355268s
+Jun 18 12:41:28.160: INFO: Pod "pod-subpath-test-projected-f57z": Phase="Running", Reason="", readiness=false. Elapsed: 8.809669844s
+Jun 18 12:41:30.174: INFO: Pod "pod-subpath-test-projected-f57z": Phase="Running", Reason="", readiness=false. Elapsed: 10.824325251s
+Jun 18 12:41:32.189: INFO: Pod "pod-subpath-test-projected-f57z": Phase="Running", Reason="", readiness=false. Elapsed: 12.83875514s
+Jun 18 12:41:34.203: INFO: Pod "pod-subpath-test-projected-f57z": Phase="Running", Reason="", readiness=false. Elapsed: 14.853447433s
+Jun 18 12:41:36.218: INFO: Pod "pod-subpath-test-projected-f57z": Phase="Running", Reason="", readiness=false. Elapsed: 16.867841855s
+Jun 18 12:41:38.253: INFO: Pod "pod-subpath-test-projected-f57z": Phase="Running", Reason="", readiness=false. Elapsed: 18.903163908s
+Jun 18 12:41:40.268: INFO: Pod "pod-subpath-test-projected-f57z": Phase="Running", Reason="", readiness=false. Elapsed: 20.918148648s
+Jun 18 12:41:42.286: INFO: Pod "pod-subpath-test-projected-f57z": Phase="Succeeded", Reason="", readiness=false. Elapsed: 22.935935489s
 STEP: Saw pod success
-May 13 19:46:51.837: INFO: Pod "pod-projected-configmaps-d82efa67-75b7-11e9-8f67-2632f168be36" satisfied condition "success or failure"
-May 13 19:46:51.932: INFO: Trying to get logs from node 10.170.219.177 pod pod-projected-configmaps-d82efa67-75b7-11e9-8f67-2632f168be36 container projected-configmap-volume-test: 
+Jun 18 12:41:42.286: INFO: Pod "pod-subpath-test-projected-f57z" satisfied condition "success or failure"
+Jun 18 12:41:42.300: INFO: Trying to get logs from node 10.72.74.143 pod pod-subpath-test-projected-f57z container test-container-subpath-projected-f57z: 
 STEP: delete the pod
-May 13 19:46:51.982: INFO: Waiting for pod pod-projected-configmaps-d82efa67-75b7-11e9-8f67-2632f168be36 to disappear
-May 13 19:46:51.989: INFO: Pod pod-projected-configmaps-d82efa67-75b7-11e9-8f67-2632f168be36 no longer exists
-[AfterEach] [sig-storage] Projected configMap
+Jun 18 12:41:42.383: INFO: Waiting for pod pod-subpath-test-projected-f57z to disappear
+Jun 18 12:41:42.397: INFO: Pod pod-subpath-test-projected-f57z no longer exists
+STEP: Deleting pod pod-subpath-test-projected-f57z
+Jun 18 12:41:42.397: INFO: Deleting pod "pod-subpath-test-projected-f57z" in namespace "e2e-tests-subpath-fqzhc"
+[AfterEach] [sig-storage] Subpath
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
-May 13 19:46:51.989: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "e2e-tests-projected-6mclt" for this suite.
-May 13 19:46:58.038: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-May 13 19:46:58.130: INFO: namespace: e2e-tests-projected-6mclt, resource: bindings, ignored listing per whitelist
-May 13 19:46:58.303: INFO: namespace e2e-tests-projected-6mclt deletion completed in 6.305661521s
+Jun 18 12:41:42.412: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-tests-subpath-fqzhc" for this suite.
+Jun 18 12:41:48.496: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 12:41:48.962: INFO: namespace: e2e-tests-subpath-fqzhc, resource: bindings, ignored listing per whitelist
+Jun 18 12:41:49.043: INFO: namespace e2e-tests-subpath-fqzhc deletion completed in 6.613984316s
 
-• [SLOW TEST:10.803 seconds]
-[sig-storage] Projected configMap
-/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_configmap.go:34
-  should be consumable from pods in volume with defaultMode set [NodeConformance] [Conformance]
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
+• [SLOW TEST:30.297 seconds]
+[sig-storage] Subpath
+/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:22
+  Atomic writer volumes
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:34
+    should support subpaths with projected pod [Conformance]
+    /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 ------------------------------
-SSSSS
+SSS
 ------------------------------
-[sig-auth] ServiceAccounts 
-  should allow opting out of API token automount  [Conformance]
+[sig-api-machinery] CustomResourceDefinition resources Simple CustomResourceDefinition 
+  creating/deleting custom resource definition objects works  [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-[BeforeEach] [sig-auth] ServiceAccounts
+[BeforeEach] [sig-api-machinery] CustomResourceDefinition resources
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
 STEP: Creating a kubernetes client
-May 13 19:46:58.304: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423
-STEP: Building a namespace api object, basename svcaccounts
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-svcaccounts-nm754
+Jun 18 12:41:49.044: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
+STEP: Building a namespace api object, basename custom-resource-definition
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-custom-resource-definition-j2bcl
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should allow opting out of API token automount  [Conformance]
+[It] creating/deleting custom resource definition objects works  [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-STEP: getting the auto-created API token
-May 13 19:46:59.164: INFO: created pod pod-service-account-defaultsa
-May 13 19:46:59.164: INFO: pod pod-service-account-defaultsa service account token volume mount: true
-May 13 19:46:59.175: INFO: created pod pod-service-account-mountsa
-May 13 19:46:59.175: INFO: pod pod-service-account-mountsa service account token volume mount: true
-May 13 19:46:59.185: INFO: created pod pod-service-account-nomountsa
-May 13 19:46:59.185: INFO: pod pod-service-account-nomountsa service account token volume mount: false
-May 13 19:46:59.194: INFO: created pod pod-service-account-defaultsa-mountspec
-May 13 19:46:59.194: INFO: pod pod-service-account-defaultsa-mountspec service account token volume mount: true
-May 13 19:46:59.203: INFO: created pod pod-service-account-mountsa-mountspec
-May 13 19:46:59.203: INFO: pod pod-service-account-mountsa-mountspec service account token volume mount: true
-May 13 19:46:59.214: INFO: created pod pod-service-account-nomountsa-mountspec
-May 13 19:46:59.214: INFO: pod pod-service-account-nomountsa-mountspec service account token volume mount: true
-May 13 19:46:59.223: INFO: created pod pod-service-account-defaultsa-nomountspec
-May 13 19:46:59.223: INFO: pod pod-service-account-defaultsa-nomountspec service account token volume mount: false
-May 13 19:46:59.232: INFO: created pod pod-service-account-mountsa-nomountspec
-May 13 19:46:59.232: INFO: pod pod-service-account-mountsa-nomountspec service account token volume mount: false
-May 13 19:46:59.241: INFO: created pod pod-service-account-nomountsa-nomountspec
-May 13 19:46:59.241: INFO: pod pod-service-account-nomountsa-nomountspec service account token volume mount: false
-[AfterEach] [sig-auth] ServiceAccounts
+Jun 18 12:41:49.525: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
+[AfterEach] [sig-api-machinery] CustomResourceDefinition resources
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
-May 13 19:46:59.241: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "e2e-tests-svcaccounts-nm754" for this suite.
-May 13 19:47:05.283: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-May 13 19:47:05.371: INFO: namespace: e2e-tests-svcaccounts-nm754, resource: bindings, ignored listing per whitelist
-May 13 19:47:05.534: INFO: namespace e2e-tests-svcaccounts-nm754 deletion completed in 6.284504016s
+Jun 18 12:41:50.724: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-tests-custom-resource-definition-j2bcl" for this suite.
+Jun 18 12:41:56.834: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 12:41:56.985: INFO: namespace: e2e-tests-custom-resource-definition-j2bcl, resource: bindings, ignored listing per whitelist
+Jun 18 12:41:57.396: INFO: namespace e2e-tests-custom-resource-definition-j2bcl deletion completed in 6.610856577s
 
-• [SLOW TEST:7.231 seconds]
-[sig-auth] ServiceAccounts
-/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/auth/framework.go:22
-  should allow opting out of API token automount  [Conformance]
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
+• [SLOW TEST:8.352 seconds]
+[sig-api-machinery] CustomResourceDefinition resources
+/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:22
+  Simple CustomResourceDefinition
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/custom_resource_definition.go:35
+    creating/deleting custom resource definition objects works  [Conformance]
+    /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 ------------------------------
-SSSS
+S
 ------------------------------
-[sig-storage] Secrets 
-  should be consumable from pods in volume with mappings [NodeConformance] [Conformance]
+[sig-api-machinery] Secrets 
+  should be consumable via the environment [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-[BeforeEach] [sig-storage] Secrets
+[BeforeEach] [sig-api-machinery] Secrets
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
 STEP: Creating a kubernetes client
-May 13 19:47:05.535: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423
+Jun 18 12:41:57.396: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
 STEP: Building a namespace api object, basename secrets
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-secrets-6dbfv
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-secrets-tbkw4
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should be consumable from pods in volume with mappings [NodeConformance] [Conformance]
+[It] should be consumable via the environment [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-STEP: Creating secret with name secret-test-map-e2edc438-75b7-11e9-8f67-2632f168be36
+STEP: creating secret e2e-tests-secrets-tbkw4/secret-test-75dbfaec-91c6-11e9-bce2-ae54e022189f
 STEP: Creating a pod to test consume secrets
-May 13 19:47:05.844: INFO: Waiting up to 5m0s for pod "pod-secrets-e2ef374b-75b7-11e9-8f67-2632f168be36" in namespace "e2e-tests-secrets-6dbfv" to be "success or failure"
-May 13 19:47:05.851: INFO: Pod "pod-secrets-e2ef374b-75b7-11e9-8f67-2632f168be36": Phase="Pending", Reason="", readiness=false. Elapsed: 6.959364ms
-May 13 19:47:07.859: INFO: Pod "pod-secrets-e2ef374b-75b7-11e9-8f67-2632f168be36": Phase="Running", Reason="", readiness=true. Elapsed: 2.015084349s
-May 13 19:47:09.880: INFO: Pod "pod-secrets-e2ef374b-75b7-11e9-8f67-2632f168be36": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.035975489s
+Jun 18 12:41:57.879: INFO: Waiting up to 5m0s for pod "pod-configmaps-75ddee40-91c6-11e9-bce2-ae54e022189f" in namespace "e2e-tests-secrets-tbkw4" to be "success or failure"
+Jun 18 12:41:57.892: INFO: Pod "pod-configmaps-75ddee40-91c6-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 13.249916ms
+Jun 18 12:41:59.925: INFO: Pod "pod-configmaps-75ddee40-91c6-11e9-bce2-ae54e022189f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.046462878s
 STEP: Saw pod success
-May 13 19:47:09.880: INFO: Pod "pod-secrets-e2ef374b-75b7-11e9-8f67-2632f168be36" satisfied condition "success or failure"
-May 13 19:47:09.888: INFO: Trying to get logs from node 10.170.219.177 pod pod-secrets-e2ef374b-75b7-11e9-8f67-2632f168be36 container secret-volume-test: 
+Jun 18 12:41:59.925: INFO: Pod "pod-configmaps-75ddee40-91c6-11e9-bce2-ae54e022189f" satisfied condition "success or failure"
+Jun 18 12:41:59.941: INFO: Trying to get logs from node 10.72.74.144 pod pod-configmaps-75ddee40-91c6-11e9-bce2-ae54e022189f container env-test: 
 STEP: delete the pod
-May 13 19:47:09.929: INFO: Waiting for pod pod-secrets-e2ef374b-75b7-11e9-8f67-2632f168be36 to disappear
-May 13 19:47:09.936: INFO: Pod pod-secrets-e2ef374b-75b7-11e9-8f67-2632f168be36 no longer exists
-[AfterEach] [sig-storage] Secrets
+Jun 18 12:42:00.017: INFO: Waiting for pod pod-configmaps-75ddee40-91c6-11e9-bce2-ae54e022189f to disappear
+Jun 18 12:42:00.037: INFO: Pod pod-configmaps-75ddee40-91c6-11e9-bce2-ae54e022189f no longer exists
+[AfterEach] [sig-api-machinery] Secrets
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
-May 13 19:47:09.936: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "e2e-tests-secrets-6dbfv" for this suite.
-May 13 19:47:15.974: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-May 13 19:47:16.086: INFO: namespace: e2e-tests-secrets-6dbfv, resource: bindings, ignored listing per whitelist
-May 13 19:47:16.278: INFO: namespace e2e-tests-secrets-6dbfv deletion completed in 6.334023844s
+Jun 18 12:42:00.037: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-tests-secrets-tbkw4" for this suite.
+Jun 18 12:42:08.107: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 12:42:08.365: INFO: namespace: e2e-tests-secrets-tbkw4, resource: bindings, ignored listing per whitelist
+Jun 18 12:42:08.636: INFO: namespace e2e-tests-secrets-tbkw4 deletion completed in 8.579104705s
 
-• [SLOW TEST:10.744 seconds]
-[sig-storage] Secrets
-/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/secrets_volume.go:34
-  should be consumable from pods in volume with mappings [NodeConformance] [Conformance]
+• [SLOW TEST:11.240 seconds]
+[sig-api-machinery] Secrets
+/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/secrets.go:32
+  should be consumable via the environment [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 ------------------------------
-SSSSSSSSSSSSSSSSSSSSSSSSSSSS
+S
 ------------------------------
-[sig-node] Downward API 
-  should provide host IP as an env var [NodeConformance] [Conformance]
+[k8s.io] Probing container 
+  with readiness probe that fails should never be ready and never restart [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-[BeforeEach] [sig-node] Downward API
+[BeforeEach] [k8s.io] Probing container
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
 STEP: Creating a kubernetes client
-May 13 19:47:16.279: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423
-STEP: Building a namespace api object, basename downward-api
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-downward-api-rmc5x
+Jun 18 12:42:08.637: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
+STEP: Building a namespace api object, basename container-probe
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-container-probe-bt9dv
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should provide host IP as an env var [NodeConformance] [Conformance]
+[BeforeEach] [k8s.io] Probing container
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/container_probe.go:48
+[It] with readiness probe that fails should never be ready and never restart [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-STEP: Creating a pod to test downward api env vars
-May 13 19:47:16.577: INFO: Waiting up to 5m0s for pod "downward-api-e954b79f-75b7-11e9-8f67-2632f168be36" in namespace "e2e-tests-downward-api-rmc5x" to be "success or failure"
-May 13 19:47:16.585: INFO: Pod "downward-api-e954b79f-75b7-11e9-8f67-2632f168be36": Phase="Pending", Reason="", readiness=false. Elapsed: 7.598266ms
-May 13 19:47:18.593: INFO: Pod "downward-api-e954b79f-75b7-11e9-8f67-2632f168be36": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.01611317s
-STEP: Saw pod success
-May 13 19:47:18.594: INFO: Pod "downward-api-e954b79f-75b7-11e9-8f67-2632f168be36" satisfied condition "success or failure"
-May 13 19:47:18.601: INFO: Trying to get logs from node 10.170.219.153 pod downward-api-e954b79f-75b7-11e9-8f67-2632f168be36 container dapi-container: 
-STEP: delete the pod
-May 13 19:47:18.641: INFO: Waiting for pod downward-api-e954b79f-75b7-11e9-8f67-2632f168be36 to disappear
-May 13 19:47:18.727: INFO: Pod downward-api-e954b79f-75b7-11e9-8f67-2632f168be36 no longer exists
-[AfterEach] [sig-node] Downward API
+[AfterEach] [k8s.io] Probing container
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
-May 13 19:47:18.727: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "e2e-tests-downward-api-rmc5x" for this suite.
-May 13 19:47:24.764: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-May 13 19:47:24.937: INFO: namespace: e2e-tests-downward-api-rmc5x, resource: bindings, ignored listing per whitelist
-May 13 19:47:25.073: INFO: namespace e2e-tests-downward-api-rmc5x deletion completed in 6.337942305s
+Jun 18 12:43:09.177: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-tests-container-probe-bt9dv" for this suite.
+Jun 18 12:43:33.256: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 12:43:33.572: INFO: namespace: e2e-tests-container-probe-bt9dv, resource: bindings, ignored listing per whitelist
+Jun 18 12:43:33.799: INFO: namespace e2e-tests-container-probe-bt9dv deletion completed in 24.599593906s
 
-• [SLOW TEST:8.794 seconds]
-[sig-node] Downward API
-/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downward_api.go:38
-  should provide host IP as an env var [NodeConformance] [Conformance]
+• [SLOW TEST:85.162 seconds]
+[k8s.io] Probing container
+/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694
+  with readiness probe that fails should never be ready and never restart [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 ------------------------------
 SSSSSS
 ------------------------------
-[sig-storage] Subpath Atomic writer volumes 
-  should support subpaths with configmap pod with mountPath of existing file [Conformance]
+[sig-storage] Secrets 
+  should be able to mount in a volume regardless of a different secret existing with same name in different namespace [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-[BeforeEach] [sig-storage] Subpath
+[BeforeEach] [sig-storage] Secrets
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
 STEP: Creating a kubernetes client
-May 13 19:47:25.074: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423
-STEP: Building a namespace api object, basename subpath
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-subpath-p5mb9
+Jun 18 12:43:33.799: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
+STEP: Building a namespace api object, basename secrets
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-secrets-gsxkf
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] Atomic writer volumes
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:38
-STEP: Setting up data
-[It] should support subpaths with configmap pod with mountPath of existing file [Conformance]
+[It] should be able to mount in a volume regardless of a different secret existing with same name in different namespace [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-STEP: Creating pod pod-subpath-test-configmap-gb6h
-STEP: Creating a pod to test atomic-volume-subpath
-May 13 19:47:25.420: INFO: Waiting up to 5m0s for pod "pod-subpath-test-configmap-gb6h" in namespace "e2e-tests-subpath-p5mb9" to be "success or failure"
-May 13 19:47:25.430: INFO: Pod "pod-subpath-test-configmap-gb6h": Phase="Pending", Reason="", readiness=false. Elapsed: 9.021553ms
-May 13 19:47:27.446: INFO: Pod "pod-subpath-test-configmap-gb6h": Phase="Pending", Reason="", readiness=false. Elapsed: 2.025497979s
-May 13 19:47:29.455: INFO: Pod "pod-subpath-test-configmap-gb6h": Phase="Running", Reason="", readiness=false. Elapsed: 4.034529607s
-May 13 19:47:31.476: INFO: Pod "pod-subpath-test-configmap-gb6h": Phase="Running", Reason="", readiness=false. Elapsed: 6.055482007s
-May 13 19:47:33.484: INFO: Pod "pod-subpath-test-configmap-gb6h": Phase="Running", Reason="", readiness=false. Elapsed: 8.063461487s
-May 13 19:47:35.493: INFO: Pod "pod-subpath-test-configmap-gb6h": Phase="Running", Reason="", readiness=false. Elapsed: 10.071950442s
-May 13 19:47:37.501: INFO: Pod "pod-subpath-test-configmap-gb6h": Phase="Running", Reason="", readiness=false. Elapsed: 12.08049209s
-May 13 19:47:39.512: INFO: Pod "pod-subpath-test-configmap-gb6h": Phase="Running", Reason="", readiness=false. Elapsed: 14.090989565s
-May 13 19:47:41.534: INFO: Pod "pod-subpath-test-configmap-gb6h": Phase="Running", Reason="", readiness=false. Elapsed: 16.113164438s
-May 13 19:47:43.547: INFO: Pod "pod-subpath-test-configmap-gb6h": Phase="Running", Reason="", readiness=false. Elapsed: 18.12594685s
-May 13 19:47:45.555: INFO: Pod "pod-subpath-test-configmap-gb6h": Phase="Running", Reason="", readiness=false. Elapsed: 20.134177391s
-May 13 19:47:47.564: INFO: Pod "pod-subpath-test-configmap-gb6h": Phase="Running", Reason="", readiness=false. Elapsed: 22.143052742s
-May 13 19:47:49.572: INFO: Pod "pod-subpath-test-configmap-gb6h": Phase="Succeeded", Reason="", readiness=false. Elapsed: 24.151368183s
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-secret-namespace-n698v
+STEP: Creating secret with name secret-test-af5aec88-91c6-11e9-bce2-ae54e022189f
+STEP: Creating a pod to test consume secrets
+Jun 18 12:43:34.572: INFO: Waiting up to 5m0s for pod "pod-secrets-af800d8f-91c6-11e9-bce2-ae54e022189f" in namespace "e2e-tests-secrets-gsxkf" to be "success or failure"
+Jun 18 12:43:34.591: INFO: Pod "pod-secrets-af800d8f-91c6-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 18.94049ms
+Jun 18 12:43:36.606: INFO: Pod "pod-secrets-af800d8f-91c6-11e9-bce2-ae54e022189f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.033861853s
 STEP: Saw pod success
-May 13 19:47:49.572: INFO: Pod "pod-subpath-test-configmap-gb6h" satisfied condition "success or failure"
-May 13 19:47:49.628: INFO: Trying to get logs from node 10.170.219.177 pod pod-subpath-test-configmap-gb6h container test-container-subpath-configmap-gb6h: 
+Jun 18 12:43:36.606: INFO: Pod "pod-secrets-af800d8f-91c6-11e9-bce2-ae54e022189f" satisfied condition "success or failure"
+Jun 18 12:43:36.624: INFO: Trying to get logs from node 10.72.74.143 pod pod-secrets-af800d8f-91c6-11e9-bce2-ae54e022189f container secret-volume-test: 
 STEP: delete the pod
-May 13 19:47:49.668: INFO: Waiting for pod pod-subpath-test-configmap-gb6h to disappear
-May 13 19:47:49.678: INFO: Pod pod-subpath-test-configmap-gb6h no longer exists
-STEP: Deleting pod pod-subpath-test-configmap-gb6h
-May 13 19:47:49.678: INFO: Deleting pod "pod-subpath-test-configmap-gb6h" in namespace "e2e-tests-subpath-p5mb9"
-[AfterEach] [sig-storage] Subpath
+Jun 18 12:43:36.709: INFO: Waiting for pod pod-secrets-af800d8f-91c6-11e9-bce2-ae54e022189f to disappear
+Jun 18 12:43:36.724: INFO: Pod pod-secrets-af800d8f-91c6-11e9-bce2-ae54e022189f no longer exists
+[AfterEach] [sig-storage] Secrets
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
-May 13 19:47:49.685: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "e2e-tests-subpath-p5mb9" for this suite.
-May 13 19:47:55.721: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-May 13 19:47:56.091: INFO: namespace: e2e-tests-subpath-p5mb9, resource: bindings, ignored listing per whitelist
-May 13 19:47:56.158: INFO: namespace e2e-tests-subpath-p5mb9 deletion completed in 6.465234122s
-
-• [SLOW TEST:31.084 seconds]
-[sig-storage] Subpath
-/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:22
-  Atomic writer volumes
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:34
-    should support subpaths with configmap pod with mountPath of existing file [Conformance]
-    /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
+Jun 18 12:43:36.724: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-tests-secrets-gsxkf" for this suite.
+Jun 18 12:43:42.884: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 12:43:43.071: INFO: namespace: e2e-tests-secrets-gsxkf, resource: bindings, ignored listing per whitelist
+Jun 18 12:43:43.391: INFO: namespace e2e-tests-secrets-gsxkf deletion completed in 6.643578016s
+STEP: Destroying namespace "e2e-tests-secret-namespace-n698v" for this suite.
+Jun 18 12:43:51.439: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 12:43:52.118: INFO: namespace: e2e-tests-secret-namespace-n698v, resource: bindings, ignored listing per whitelist
+Jun 18 12:43:52.173: INFO: namespace e2e-tests-secret-namespace-n698v deletion completed in 8.782295194s
+
+• [SLOW TEST:18.375 seconds]
+[sig-storage] Secrets
+/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/secrets_volume.go:34
+  should be able to mount in a volume regardless of a different secret existing with same name in different namespace [NodeConformance] [Conformance]
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 ------------------------------
-SSSSSSS
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-storage] EmptyDir volumes 
-  should support (non-root,0644,default) [NodeConformance] [Conformance]
+[sig-storage] Projected downwardAPI 
+  should update annotations on modification [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-[BeforeEach] [sig-storage] EmptyDir volumes
+[BeforeEach] [sig-storage] Projected downwardAPI
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
 STEP: Creating a kubernetes client
-May 13 19:47:56.158: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423
-STEP: Building a namespace api object, basename emptydir
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-emptydir-rrhgf
+Jun 18 12:43:52.174: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
+STEP: Building a namespace api object, basename projected
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-projected-s499m
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should support (non-root,0644,default) [NodeConformance] [Conformance]
+[BeforeEach] [sig-storage] Projected downwardAPI
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:39
+[It] should update annotations on modification [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-STEP: Creating a pod to test emptydir 0644 on node default medium
-May 13 19:47:56.449: INFO: Waiting up to 5m0s for pod "pod-01189750-75b8-11e9-8f67-2632f168be36" in namespace "e2e-tests-emptydir-rrhgf" to be "success or failure"
-May 13 19:47:56.530: INFO: Pod "pod-01189750-75b8-11e9-8f67-2632f168be36": Phase="Pending", Reason="", readiness=false. Elapsed: 81.517197ms
-May 13 19:47:58.539: INFO: Pod "pod-01189750-75b8-11e9-8f67-2632f168be36": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.089952903s
-STEP: Saw pod success
-May 13 19:47:58.539: INFO: Pod "pod-01189750-75b8-11e9-8f67-2632f168be36" satisfied condition "success or failure"
-May 13 19:47:58.547: INFO: Trying to get logs from node 10.170.219.153 pod pod-01189750-75b8-11e9-8f67-2632f168be36 container test-container: 
-STEP: delete the pod
-May 13 19:47:58.588: INFO: Waiting for pod pod-01189750-75b8-11e9-8f67-2632f168be36 to disappear
-May 13 19:47:58.598: INFO: Pod pod-01189750-75b8-11e9-8f67-2632f168be36 no longer exists
-[AfterEach] [sig-storage] EmptyDir volumes
+STEP: Creating the pod
+Jun 18 12:43:57.284: INFO: Successfully updated pod "annotationupdateba45f552-91c6-11e9-bce2-ae54e022189f"
+[AfterEach] [sig-storage] Projected downwardAPI
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
-May 13 19:47:58.598: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "e2e-tests-emptydir-rrhgf" for this suite.
-May 13 19:48:04.634: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-May 13 19:48:04.858: INFO: namespace: e2e-tests-emptydir-rrhgf, resource: bindings, ignored listing per whitelist
-May 13 19:48:04.971: INFO: namespace e2e-tests-emptydir-rrhgf deletion completed in 6.365083969s
+Jun 18 12:43:59.353: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-tests-projected-s499m" for this suite.
+Jun 18 12:44:23.497: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 12:44:23.668: INFO: namespace: e2e-tests-projected-s499m, resource: bindings, ignored listing per whitelist
+Jun 18 12:44:23.980: INFO: namespace e2e-tests-projected-s499m deletion completed in 24.607799233s
 
-• [SLOW TEST:8.813 seconds]
-[sig-storage] EmptyDir volumes
-/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:40
-  should support (non-root,0644,default) [NodeConformance] [Conformance]
+• [SLOW TEST:31.806 seconds]
+[sig-storage] Projected downwardAPI
+/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:33
+  should update annotations on modification [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 ------------------------------
-S
-------------------------------
 [sig-storage] ConfigMap 
-  should be consumable from pods in volume with mappings [NodeConformance] [Conformance]
+  should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 [BeforeEach] [sig-storage] ConfigMap
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
 STEP: Creating a kubernetes client
-May 13 19:48:04.971: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423
+Jun 18 12:44:23.980: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
 STEP: Building a namespace api object, basename configmap
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-configmap-chvth
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-configmap-9p99t
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should be consumable from pods in volume with mappings [NodeConformance] [Conformance]
+[It] should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-STEP: Creating configMap with name configmap-test-volume-map-0673ee15-75b8-11e9-8f67-2632f168be36
+STEP: Creating configMap with name configmap-test-volume-cd3aee55-91c6-11e9-bce2-ae54e022189f
 STEP: Creating a pod to test consume configMaps
-May 13 19:48:05.446: INFO: Waiting up to 5m0s for pod "pod-configmaps-0675692f-75b8-11e9-8f67-2632f168be36" in namespace "e2e-tests-configmap-chvth" to be "success or failure"
-May 13 19:48:05.454: INFO: Pod "pod-configmaps-0675692f-75b8-11e9-8f67-2632f168be36": Phase="Pending", Reason="", readiness=false. Elapsed: 8.38333ms
-May 13 19:48:07.463: INFO: Pod "pod-configmaps-0675692f-75b8-11e9-8f67-2632f168be36": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.016977816s
+Jun 18 12:44:24.472: INFO: Waiting up to 5m0s for pod "pod-configmaps-cd3cf9b9-91c6-11e9-bce2-ae54e022189f" in namespace "e2e-tests-configmap-9p99t" to be "success or failure"
+Jun 18 12:44:24.486: INFO: Pod "pod-configmaps-cd3cf9b9-91c6-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 13.683547ms
+Jun 18 12:44:26.501: INFO: Pod "pod-configmaps-cd3cf9b9-91c6-11e9-bce2-ae54e022189f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.029483086s
 STEP: Saw pod success
-May 13 19:48:07.463: INFO: Pod "pod-configmaps-0675692f-75b8-11e9-8f67-2632f168be36" satisfied condition "success or failure"
-May 13 19:48:07.471: INFO: Trying to get logs from node 10.170.219.177 pod pod-configmaps-0675692f-75b8-11e9-8f67-2632f168be36 container configmap-volume-test: 
+Jun 18 12:44:26.501: INFO: Pod "pod-configmaps-cd3cf9b9-91c6-11e9-bce2-ae54e022189f" satisfied condition "success or failure"
+Jun 18 12:44:26.598: INFO: Trying to get logs from node 10.72.74.149 pod pod-configmaps-cd3cf9b9-91c6-11e9-bce2-ae54e022189f container configmap-volume-test: 
 STEP: delete the pod
-May 13 19:48:07.512: INFO: Waiting for pod pod-configmaps-0675692f-75b8-11e9-8f67-2632f168be36 to disappear
-May 13 19:48:07.519: INFO: Pod pod-configmaps-0675692f-75b8-11e9-8f67-2632f168be36 no longer exists
+Jun 18 12:44:26.677: INFO: Waiting for pod pod-configmaps-cd3cf9b9-91c6-11e9-bce2-ae54e022189f to disappear
+Jun 18 12:44:26.693: INFO: Pod pod-configmaps-cd3cf9b9-91c6-11e9-bce2-ae54e022189f no longer exists
 [AfterEach] [sig-storage] ConfigMap
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
-May 13 19:48:07.519: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "e2e-tests-configmap-chvth" for this suite.
-May 13 19:48:13.570: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-May 13 19:48:13.661: INFO: namespace: e2e-tests-configmap-chvth, resource: bindings, ignored listing per whitelist
-May 13 19:48:13.815: INFO: namespace e2e-tests-configmap-chvth deletion completed in 6.289781074s
+Jun 18 12:44:26.693: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-tests-configmap-9p99t" for this suite.
+Jun 18 12:44:32.798: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 12:44:33.231: INFO: namespace: e2e-tests-configmap-9p99t, resource: bindings, ignored listing per whitelist
+Jun 18 12:44:33.386: INFO: namespace e2e-tests-configmap-9p99t deletion completed in 6.673476284s
 
-• [SLOW TEST:8.844 seconds]
+• [SLOW TEST:9.406 seconds]
 [sig-storage] ConfigMap
 /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:33
-  should be consumable from pods in volume with mappings [NodeConformance] [Conformance]
+  should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 ------------------------------
-SS
+S
 ------------------------------
-[sig-storage] EmptyDir volumes 
-  volume on tmpfs should have the correct mode [NodeConformance] [Conformance]
+[k8s.io] Variable Expansion 
+  should allow composing env vars into new env vars [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-[BeforeEach] [sig-storage] EmptyDir volumes
+[BeforeEach] [k8s.io] Variable Expansion
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
 STEP: Creating a kubernetes client
-May 13 19:48:13.816: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423
-STEP: Building a namespace api object, basename emptydir
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-emptydir-dc5kt
+Jun 18 12:44:33.388: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
+STEP: Building a namespace api object, basename var-expansion
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-var-expansion-85xj7
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] volume on tmpfs should have the correct mode [NodeConformance] [Conformance]
+[It] should allow composing env vars into new env vars [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-STEP: Creating a pod to test emptydir volume type on tmpfs
-May 13 19:48:14.114: INFO: Waiting up to 5m0s for pod "pod-0ba02a21-75b8-11e9-8f67-2632f168be36" in namespace "e2e-tests-emptydir-dc5kt" to be "success or failure"
-May 13 19:48:14.125: INFO: Pod "pod-0ba02a21-75b8-11e9-8f67-2632f168be36": Phase="Pending", Reason="", readiness=false. Elapsed: 11.081995ms
-May 13 19:48:16.134: INFO: Pod "pod-0ba02a21-75b8-11e9-8f67-2632f168be36": Phase="Pending", Reason="", readiness=false. Elapsed: 2.020029883s
-May 13 19:48:18.143: INFO: Pod "pod-0ba02a21-75b8-11e9-8f67-2632f168be36": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.028816742s
+STEP: Creating a pod to test env composition
+Jun 18 12:44:33.924: INFO: Waiting up to 5m0s for pod "var-expansion-d2df5c2b-91c6-11e9-bce2-ae54e022189f" in namespace "e2e-tests-var-expansion-85xj7" to be "success or failure"
+Jun 18 12:44:33.938: INFO: Pod "var-expansion-d2df5c2b-91c6-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 14.179002ms
+Jun 18 12:44:35.954: INFO: Pod "var-expansion-d2df5c2b-91c6-11e9-bce2-ae54e022189f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.030236549s
 STEP: Saw pod success
-May 13 19:48:18.143: INFO: Pod "pod-0ba02a21-75b8-11e9-8f67-2632f168be36" satisfied condition "success or failure"
-May 13 19:48:18.151: INFO: Trying to get logs from node 10.170.219.177 pod pod-0ba02a21-75b8-11e9-8f67-2632f168be36 container test-container: 
+Jun 18 12:44:35.954: INFO: Pod "var-expansion-d2df5c2b-91c6-11e9-bce2-ae54e022189f" satisfied condition "success or failure"
+Jun 18 12:44:35.968: INFO: Trying to get logs from node 10.72.74.143 pod var-expansion-d2df5c2b-91c6-11e9-bce2-ae54e022189f container dapi-container: 
 STEP: delete the pod
-May 13 19:48:18.191: INFO: Waiting for pod pod-0ba02a21-75b8-11e9-8f67-2632f168be36 to disappear
-May 13 19:48:18.198: INFO: Pod pod-0ba02a21-75b8-11e9-8f67-2632f168be36 no longer exists
-[AfterEach] [sig-storage] EmptyDir volumes
+Jun 18 12:44:36.084: INFO: Waiting for pod var-expansion-d2df5c2b-91c6-11e9-bce2-ae54e022189f to disappear
+Jun 18 12:44:36.098: INFO: Pod var-expansion-d2df5c2b-91c6-11e9-bce2-ae54e022189f no longer exists
+[AfterEach] [k8s.io] Variable Expansion
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
-May 13 19:48:18.198: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "e2e-tests-emptydir-dc5kt" for this suite.
-May 13 19:48:24.248: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-May 13 19:48:24.385: INFO: namespace: e2e-tests-emptydir-dc5kt, resource: bindings, ignored listing per whitelist
-May 13 19:48:24.582: INFO: namespace e2e-tests-emptydir-dc5kt deletion completed in 6.375489128s
+Jun 18 12:44:36.099: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-tests-var-expansion-85xj7" for this suite.
+Jun 18 12:44:42.173: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 12:44:42.422: INFO: namespace: e2e-tests-var-expansion-85xj7, resource: bindings, ignored listing per whitelist
+Jun 18 12:44:42.663: INFO: namespace e2e-tests-var-expansion-85xj7 deletion completed in 6.540842073s
 
-• [SLOW TEST:10.766 seconds]
-[sig-storage] EmptyDir volumes
-/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:40
-  volume on tmpfs should have the correct mode [NodeConformance] [Conformance]
+• [SLOW TEST:9.276 seconds]
+[k8s.io] Variable Expansion
+/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694
+  should allow composing env vars into new env vars [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 ------------------------------
-SS
-------------------------------
-[sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] 
-  Burst scaling should run to completion even with unhealthy pods [Conformance]
+[sig-storage] Projected secret 
+  should be consumable from pods in volume with mappings and Item Mode set [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-[BeforeEach] [sig-apps] StatefulSet
+[BeforeEach] [sig-storage] Projected secret
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
 STEP: Creating a kubernetes client
-May 13 19:48:24.582: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423
-STEP: Building a namespace api object, basename statefulset
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-statefulset-4n8l5
+Jun 18 12:44:42.664: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
+STEP: Building a namespace api object, basename projected
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-projected-sw4z8
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-apps] StatefulSet
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:59
-[BeforeEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic]
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:74
-STEP: Creating service test in namespace e2e-tests-statefulset-4n8l5
-[It] Burst scaling should run to completion even with unhealthy pods [Conformance]
+[It] should be consumable from pods in volume with mappings and Item Mode set [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-STEP: Creating stateful set ss in namespace e2e-tests-statefulset-4n8l5
-STEP: Waiting until all stateful set ss replicas will be running in namespace e2e-tests-statefulset-4n8l5
-May 13 19:48:24.996: INFO: Found 0 stateful pods, waiting for 1
-May 13 19:48:35.020: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true
-STEP: Confirming that stateful set scale up will not halt with unhealthy stateful pod
-May 13 19:48:35.029: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 exec --namespace=e2e-tests-statefulset-4n8l5 ss-0 -- /bin/sh -c mv -v /usr/share/nginx/html/index.html /tmp/ || true'
-May 13 19:48:35.313: INFO: stderr: ""
-May 13 19:48:35.313: INFO: stdout: "'/usr/share/nginx/html/index.html' -> '/tmp/index.html'\n"
-May 13 19:48:35.313: INFO: stdout of mv -v /usr/share/nginx/html/index.html /tmp/ || true on ss-0: '/usr/share/nginx/html/index.html' -> '/tmp/index.html'
-
-May 13 19:48:35.321: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=true
-May 13 19:48:45.342: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=false
-May 13 19:48:45.342: INFO: Waiting for statefulset status.replicas updated to 0
-May 13 19:48:45.369: INFO: POD   NODE            PHASE    GRACE  CONDITIONS
-May 13 19:48:45.369: INFO: ss-0  10.170.219.153  Running         [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:48:25 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:48:35 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:48:35 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:48:25 +0000 UTC  }]
-May 13 19:48:45.369: INFO: 
-May 13 19:48:45.369: INFO: StatefulSet ss has not reached scale 3, at 1
-May 13 19:48:46.377: INFO: Verifying statefulset ss doesn't scale past 3 for another 8.992316371s
-May 13 19:48:47.386: INFO: Verifying statefulset ss doesn't scale past 3 for another 7.984211635s
-May 13 19:48:48.395: INFO: Verifying statefulset ss doesn't scale past 3 for another 6.975341753s
-May 13 19:48:49.404: INFO: Verifying statefulset ss doesn't scale past 3 for another 5.966139026s
-May 13 19:48:50.413: INFO: Verifying statefulset ss doesn't scale past 3 for another 4.957365721s
-May 13 19:48:51.425: INFO: Verifying statefulset ss doesn't scale past 3 for another 3.948940175s
-May 13 19:48:52.520: INFO: Verifying statefulset ss doesn't scale past 3 for another 2.937002083s
-May 13 19:48:53.567: INFO: Verifying statefulset ss doesn't scale past 3 for another 1.841826943s
-May 13 19:48:54.575: INFO: Verifying statefulset ss doesn't scale past 3 for another 795.081253ms
-STEP: Scaling up stateful set ss to 3 replicas and waiting until all of them will be running in namespace e2e-tests-statefulset-4n8l5
-May 13 19:48:55.620: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 exec --namespace=e2e-tests-statefulset-4n8l5 ss-0 -- /bin/sh -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
-May 13 19:48:55.998: INFO: stderr: ""
-May 13 19:48:55.999: INFO: stdout: "'/tmp/index.html' -> '/usr/share/nginx/html/index.html'\n"
-May 13 19:48:55.999: INFO: stdout of mv -v /tmp/index.html /usr/share/nginx/html/ || true on ss-0: '/tmp/index.html' -> '/usr/share/nginx/html/index.html'
-
-May 13 19:48:55.999: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 exec --namespace=e2e-tests-statefulset-4n8l5 ss-1 -- /bin/sh -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
-May 13 19:48:56.274: INFO: stderr: "mv: can't rename '/tmp/index.html': No such file or directory\n"
-May 13 19:48:56.274: INFO: stdout: "'/tmp/index.html' -> '/usr/share/nginx/html/index.html'\n"
-May 13 19:48:56.274: INFO: stdout of mv -v /tmp/index.html /usr/share/nginx/html/ || true on ss-1: '/tmp/index.html' -> '/usr/share/nginx/html/index.html'
-
-May 13 19:48:56.274: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 exec --namespace=e2e-tests-statefulset-4n8l5 ss-2 -- /bin/sh -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
-May 13 19:48:56.753: INFO: stderr: "mv: can't rename '/tmp/index.html': No such file or directory\n"
-May 13 19:48:56.753: INFO: stdout: "'/tmp/index.html' -> '/usr/share/nginx/html/index.html'\n"
-May 13 19:48:56.753: INFO: stdout of mv -v /tmp/index.html /usr/share/nginx/html/ || true on ss-2: '/tmp/index.html' -> '/usr/share/nginx/html/index.html'
-
-May 13 19:48:56.762: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true
-May 13 19:48:56.762: INFO: Waiting for pod ss-1 to enter Running - Ready=true, currently Running - Ready=true
-May 13 19:48:56.762: INFO: Waiting for pod ss-2 to enter Running - Ready=true, currently Running - Ready=true
-STEP: Scale down will not halt with unhealthy stateful pod
-May 13 19:48:56.769: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 exec --namespace=e2e-tests-statefulset-4n8l5 ss-0 -- /bin/sh -c mv -v /usr/share/nginx/html/index.html /tmp/ || true'
-May 13 19:48:57.069: INFO: stderr: ""
-May 13 19:48:57.069: INFO: stdout: "'/usr/share/nginx/html/index.html' -> '/tmp/index.html'\n"
-May 13 19:48:57.069: INFO: stdout of mv -v /usr/share/nginx/html/index.html /tmp/ || true on ss-0: '/usr/share/nginx/html/index.html' -> '/tmp/index.html'
-
-May 13 19:48:57.070: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 exec --namespace=e2e-tests-statefulset-4n8l5 ss-1 -- /bin/sh -c mv -v /usr/share/nginx/html/index.html /tmp/ || true'
-May 13 19:48:57.320: INFO: stderr: ""
-May 13 19:48:57.320: INFO: stdout: "'/usr/share/nginx/html/index.html' -> '/tmp/index.html'\n"
-May 13 19:48:57.320: INFO: stdout of mv -v /usr/share/nginx/html/index.html /tmp/ || true on ss-1: '/usr/share/nginx/html/index.html' -> '/tmp/index.html'
-
-May 13 19:48:57.320: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 exec --namespace=e2e-tests-statefulset-4n8l5 ss-2 -- /bin/sh -c mv -v /usr/share/nginx/html/index.html /tmp/ || true'
-May 13 19:48:57.612: INFO: stderr: ""
-May 13 19:48:57.612: INFO: stdout: "'/usr/share/nginx/html/index.html' -> '/tmp/index.html'\n"
-May 13 19:48:57.612: INFO: stdout of mv -v /usr/share/nginx/html/index.html /tmp/ || true on ss-2: '/usr/share/nginx/html/index.html' -> '/tmp/index.html'
-
-May 13 19:48:57.612: INFO: Waiting for statefulset status.replicas updated to 0
-May 13 19:48:57.625: INFO: Waiting for stateful set status.readyReplicas to become 0, currently 3
-May 13 19:49:07.652: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=false
-May 13 19:49:07.652: INFO: Waiting for pod ss-1 to enter Running - Ready=false, currently Running - Ready=false
-May 13 19:49:07.652: INFO: Waiting for pod ss-2 to enter Running - Ready=false, currently Running - Ready=false
-May 13 19:49:07.672: INFO: POD   NODE            PHASE    GRACE  CONDITIONS
-May 13 19:49:07.672: INFO: ss-0  10.170.219.153  Running         [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:48:25 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:48:57 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:48:57 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:48:25 +0000 UTC  }]
-May 13 19:49:07.672: INFO: ss-1  10.170.219.177  Running         [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:48:45 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:48:58 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:48:58 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:48:45 +0000 UTC  }]
-May 13 19:49:07.672: INFO: ss-2  10.170.219.151  Running         [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:48:45 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:48:58 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:48:58 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:48:45 +0000 UTC  }]
-May 13 19:49:07.672: INFO: 
-May 13 19:49:07.672: INFO: StatefulSet ss has not reached scale 0, at 3
-May 13 19:49:08.680: INFO: POD   NODE            PHASE    GRACE  CONDITIONS
-May 13 19:49:08.680: INFO: ss-0  10.170.219.153  Running  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:48:25 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:48:57 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:48:57 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:48:25 +0000 UTC  }]
-May 13 19:49:08.680: INFO: ss-1  10.170.219.177  Running  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:48:45 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:48:58 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:48:58 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:48:45 +0000 UTC  }]
-May 13 19:49:08.680: INFO: ss-2  10.170.219.151  Running  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:48:45 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:48:58 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:48:58 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:48:45 +0000 UTC  }]
-May 13 19:49:08.680: INFO: 
-May 13 19:49:08.680: INFO: StatefulSet ss has not reached scale 0, at 3
-May 13 19:49:09.689: INFO: POD   NODE            PHASE    GRACE  CONDITIONS
-May 13 19:49:09.689: INFO: ss-0  10.170.219.153  Running  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:48:25 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:48:57 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:48:57 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:48:25 +0000 UTC  }]
-May 13 19:49:09.689: INFO: ss-2  10.170.219.151  Pending  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:48:45 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:48:58 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:48:58 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:48:45 +0000 UTC  }]
-May 13 19:49:09.689: INFO: 
-May 13 19:49:09.689: INFO: StatefulSet ss has not reached scale 0, at 2
-May 13 19:49:10.698: INFO: POD   NODE            PHASE    GRACE  CONDITIONS
-May 13 19:49:10.698: INFO: ss-0  10.170.219.153  Running  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:48:25 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:48:57 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:48:57 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:48:25 +0000 UTC  }]
-May 13 19:49:10.698: INFO: ss-2  10.170.219.151  Pending  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:48:45 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:48:58 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:48:58 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:48:45 +0000 UTC  }]
-May 13 19:49:10.698: INFO: 
-May 13 19:49:10.698: INFO: StatefulSet ss has not reached scale 0, at 2
-May 13 19:49:11.713: INFO: POD   NODE            PHASE    GRACE  CONDITIONS
-May 13 19:49:11.713: INFO: ss-0  10.170.219.153  Running  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:48:25 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:48:57 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:48:57 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:48:25 +0000 UTC  }]
-May 13 19:49:11.713: INFO: ss-2  10.170.219.151  Pending  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:48:45 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:48:58 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:48:58 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:48:45 +0000 UTC  }]
-May 13 19:49:11.713: INFO: 
-May 13 19:49:11.713: INFO: StatefulSet ss has not reached scale 0, at 2
-May 13 19:49:12.721: INFO: POD   NODE            PHASE    GRACE  CONDITIONS
-May 13 19:49:12.721: INFO: ss-0  10.170.219.153  Running  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:48:25 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:48:57 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:48:57 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:48:25 +0000 UTC  }]
-May 13 19:49:12.721: INFO: 
-May 13 19:49:12.721: INFO: StatefulSet ss has not reached scale 0, at 1
-May 13 19:49:13.732: INFO: POD   NODE            PHASE    GRACE  CONDITIONS
-May 13 19:49:13.732: INFO: ss-0  10.170.219.153  Running  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:48:25 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:48:57 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:48:57 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:48:25 +0000 UTC  }]
-May 13 19:49:13.732: INFO: 
-May 13 19:49:13.732: INFO: StatefulSet ss has not reached scale 0, at 1
-May 13 19:49:14.741: INFO: POD   NODE            PHASE    GRACE  CONDITIONS
-May 13 19:49:14.741: INFO: ss-0  10.170.219.153  Running  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:48:25 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:48:57 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:48:57 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:48:25 +0000 UTC  }]
-May 13 19:49:14.741: INFO: 
-May 13 19:49:14.741: INFO: StatefulSet ss has not reached scale 0, at 1
-May 13 19:49:15.749: INFO: POD   NODE            PHASE    GRACE  CONDITIONS
-May 13 19:49:15.749: INFO: ss-0  10.170.219.153  Running  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:48:25 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:48:57 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:48:57 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 19:48:25 +0000 UTC  }]
-May 13 19:49:15.750: INFO: 
-May 13 19:49:15.750: INFO: StatefulSet ss has not reached scale 0, at 1
-May 13 19:49:16.758: INFO: Verifying statefulset ss doesn't scale past 0 for another 914.500424ms
-STEP: Scaling down stateful set ss to 0 replicas and waiting until none of pods will run in namespacee2e-tests-statefulset-4n8l5
-May 13 19:49:17.782: INFO: Scaling statefulset ss to 0
-May 13 19:49:17.800: INFO: Waiting for statefulset status.replicas updated to 0
-[AfterEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic]
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:85
-May 13 19:49:17.805: INFO: Deleting all statefulset in ns e2e-tests-statefulset-4n8l5
-May 13 19:49:17.810: INFO: Scaling statefulset ss to 0
-May 13 19:49:17.832: INFO: Waiting for statefulset status.replicas updated to 0
-May 13 19:49:17.837: INFO: Deleting statefulset ss
-[AfterEach] [sig-apps] StatefulSet
+STEP: Creating projection with secret that has name projected-secret-test-map-d86097d7-91c6-11e9-bce2-ae54e022189f
+STEP: Creating a pod to test consume secrets
+Jun 18 12:44:43.167: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-d862b6b5-91c6-11e9-bce2-ae54e022189f" in namespace "e2e-tests-projected-sw4z8" to be "success or failure"
+Jun 18 12:44:43.181: INFO: Pod "pod-projected-secrets-d862b6b5-91c6-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 13.719758ms
+Jun 18 12:44:45.198: INFO: Pod "pod-projected-secrets-d862b6b5-91c6-11e9-bce2-ae54e022189f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.031119413s
+STEP: Saw pod success
+Jun 18 12:44:45.199: INFO: Pod "pod-projected-secrets-d862b6b5-91c6-11e9-bce2-ae54e022189f" satisfied condition "success or failure"
+Jun 18 12:44:45.212: INFO: Trying to get logs from node 10.72.74.144 pod pod-projected-secrets-d862b6b5-91c6-11e9-bce2-ae54e022189f container projected-secret-volume-test: 
+STEP: delete the pod
+Jun 18 12:44:45.285: INFO: Waiting for pod pod-projected-secrets-d862b6b5-91c6-11e9-bce2-ae54e022189f to disappear
+Jun 18 12:44:45.299: INFO: Pod pod-projected-secrets-d862b6b5-91c6-11e9-bce2-ae54e022189f no longer exists
+[AfterEach] [sig-storage] Projected secret
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
-May 13 19:49:17.863: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "e2e-tests-statefulset-4n8l5" for this suite.
-May 13 19:49:23.902: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-May 13 19:49:24.518: INFO: namespace: e2e-tests-statefulset-4n8l5, resource: bindings, ignored listing per whitelist
-May 13 19:49:24.522: INFO: namespace e2e-tests-statefulset-4n8l5 deletion completed in 6.651491215s
+Jun 18 12:44:45.299: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-tests-projected-sw4z8" for this suite.
+Jun 18 12:44:51.385: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 12:44:51.710: INFO: namespace: e2e-tests-projected-sw4z8, resource: bindings, ignored listing per whitelist
+Jun 18 12:44:51.976: INFO: namespace e2e-tests-projected-sw4z8 deletion completed in 6.654290333s
 
-• [SLOW TEST:59.941 seconds]
-[sig-apps] StatefulSet
-/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22
-  [k8s.io] Basic StatefulSet functionality [StatefulSetBasic]
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694
-    Burst scaling should run to completion even with unhealthy pods [Conformance]
-    /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
+• [SLOW TEST:9.312 seconds]
+[sig-storage] Projected secret
+/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_secret.go:34
+  should be consumable from pods in volume with mappings and Item Mode set [NodeConformance] [Conformance]
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 ------------------------------
-S
+SSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-cli] Kubectl client [k8s.io] Update Demo 
-  should do a rolling update of a replication controller  [Conformance]
+[sig-cli] Kubectl client [k8s.io] Kubectl run deployment 
+  should create a deployment from an image  [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 [BeforeEach] [sig-cli] Kubectl client
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
 STEP: Creating a kubernetes client
-May 13 19:49:24.523: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423
+Jun 18 12:44:51.980: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
 STEP: Building a namespace api object, basename kubectl
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-kubectl-pgzg2
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-kubectl-p76x4
 STEP: Waiting for a default service account to be provisioned in namespace
 [BeforeEach] [sig-cli] Kubectl client
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:243
-[BeforeEach] [k8s.io] Update Demo
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:295
-[It] should do a rolling update of a replication controller  [Conformance]
+[BeforeEach] [k8s.io] Kubectl run deployment
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1399
+[It] should create a deployment from an image  [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-STEP: creating the initial replication controller
-May 13 19:49:24.805: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 create -f - --namespace=e2e-tests-kubectl-pgzg2'
-May 13 19:49:25.114: INFO: stderr: ""
-May 13 19:49:25.114: INFO: stdout: "replicationcontroller/update-demo-nautilus created\n"
-STEP: waiting for all containers in name=update-demo pods to come up.
-May 13 19:49:25.114: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=e2e-tests-kubectl-pgzg2'
-May 13 19:49:25.224: INFO: stderr: ""
-May 13 19:49:25.224: INFO: stdout: "update-demo-nautilus-5vddr update-demo-nautilus-77b5n "
-May 13 19:49:25.224: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 get pods update-demo-nautilus-5vddr -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=e2e-tests-kubectl-pgzg2'
-May 13 19:49:25.332: INFO: stderr: ""
-May 13 19:49:25.332: INFO: stdout: ""
-May 13 19:49:25.332: INFO: update-demo-nautilus-5vddr is created but not running
-May 13 19:49:30.332: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=e2e-tests-kubectl-pgzg2'
-May 13 19:49:30.439: INFO: stderr: ""
-May 13 19:49:30.439: INFO: stdout: "update-demo-nautilus-5vddr update-demo-nautilus-77b5n "
-May 13 19:49:30.439: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 get pods update-demo-nautilus-5vddr -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=e2e-tests-kubectl-pgzg2'
-May 13 19:49:30.557: INFO: stderr: ""
-May 13 19:49:30.557: INFO: stdout: "true"
-May 13 19:49:30.557: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 get pods update-demo-nautilus-5vddr -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=e2e-tests-kubectl-pgzg2'
-May 13 19:49:30.663: INFO: stderr: ""
-May 13 19:49:30.663: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0"
-May 13 19:49:30.663: INFO: validating pod update-demo-nautilus-5vddr
-May 13 19:49:30.676: INFO: got data: {
-  "image": "nautilus.jpg"
-}
-
-May 13 19:49:30.676: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg .
-May 13 19:49:30.676: INFO: update-demo-nautilus-5vddr is verified up and running
-May 13 19:49:30.676: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 get pods update-demo-nautilus-77b5n -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=e2e-tests-kubectl-pgzg2'
-May 13 19:49:30.798: INFO: stderr: ""
-May 13 19:49:30.798: INFO: stdout: "true"
-May 13 19:49:30.798: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 get pods update-demo-nautilus-77b5n -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=e2e-tests-kubectl-pgzg2'
-May 13 19:49:30.899: INFO: stderr: ""
-May 13 19:49:30.899: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0"
-May 13 19:49:30.899: INFO: validating pod update-demo-nautilus-77b5n
-May 13 19:49:30.913: INFO: got data: {
-  "image": "nautilus.jpg"
-}
-
-May 13 19:49:30.913: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg .
-May 13 19:49:30.913: INFO: update-demo-nautilus-77b5n is verified up and running
-STEP: rolling-update to new replication controller
-May 13 19:49:30.914: INFO: scanned /root for discovery docs: 
-May 13 19:49:30.914: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 rolling-update update-demo-nautilus --update-period=1s -f - --namespace=e2e-tests-kubectl-pgzg2'
-May 13 19:49:53.653: INFO: stderr: "Command \"rolling-update\" is deprecated, use \"rollout\" instead\n"
-May 13 19:49:53.653: INFO: stdout: "Created update-demo-kitten\nScaling up update-demo-kitten from 0 to 2, scaling down update-demo-nautilus from 2 to 0 (keep 2 pods available, don't exceed 3 pods)\nScaling update-demo-kitten up to 1\nScaling update-demo-nautilus down to 1\nScaling update-demo-kitten up to 2\nScaling update-demo-nautilus down to 0\nUpdate succeeded. Deleting old controller: update-demo-nautilus\nRenaming update-demo-kitten to update-demo-nautilus\nreplicationcontroller/update-demo-nautilus rolling updated\n"
-STEP: waiting for all containers in name=update-demo pods to come up.
-May 13 19:49:53.653: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=e2e-tests-kubectl-pgzg2'
-May 13 19:49:53.757: INFO: stderr: ""
-May 13 19:49:53.757: INFO: stdout: "update-demo-kitten-pr4jj update-demo-kitten-zrqpg "
-May 13 19:49:53.757: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 get pods update-demo-kitten-pr4jj -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=e2e-tests-kubectl-pgzg2'
-May 13 19:49:53.865: INFO: stderr: ""
-May 13 19:49:53.865: INFO: stdout: "true"
-May 13 19:49:53.865: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 get pods update-demo-kitten-pr4jj -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=e2e-tests-kubectl-pgzg2'
-May 13 19:49:53.986: INFO: stderr: ""
-May 13 19:49:53.986: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/kitten:1.0"
-May 13 19:49:53.986: INFO: validating pod update-demo-kitten-pr4jj
-May 13 19:49:54.001: INFO: got data: {
-  "image": "kitten.jpg"
-}
-
-May 13 19:49:54.001: INFO: Unmarshalled json jpg/img => {kitten.jpg} , expecting kitten.jpg .
-May 13 19:49:54.001: INFO: update-demo-kitten-pr4jj is verified up and running
-May 13 19:49:54.001: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 get pods update-demo-kitten-zrqpg -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=e2e-tests-kubectl-pgzg2'
-May 13 19:49:54.109: INFO: stderr: ""
-May 13 19:49:54.109: INFO: stdout: "true"
-May 13 19:49:54.109: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 get pods update-demo-kitten-zrqpg -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=e2e-tests-kubectl-pgzg2'
-May 13 19:49:54.209: INFO: stderr: ""
-May 13 19:49:54.209: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/kitten:1.0"
-May 13 19:49:54.209: INFO: validating pod update-demo-kitten-zrqpg
-May 13 19:49:54.223: INFO: got data: {
-  "image": "kitten.jpg"
-}
-
-May 13 19:49:54.223: INFO: Unmarshalled json jpg/img => {kitten.jpg} , expecting kitten.jpg .
-May 13 19:49:54.223: INFO: update-demo-kitten-zrqpg is verified up and running
+STEP: running the image docker.io/library/nginx:1.14-alpine
+Jun 18 12:44:52.436: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 run e2e-test-nginx-deployment --image=docker.io/library/nginx:1.14-alpine --generator=deployment/v1beta1 --namespace=e2e-tests-kubectl-p76x4'
+Jun 18 12:44:52.568: INFO: stderr: "kubectl run --generator=deployment/v1beta1 is DEPRECATED and will be removed in a future version. Use kubectl run --generator=run-pod/v1 or kubectl create instead.\n"
+Jun 18 12:44:52.568: INFO: stdout: "deployment.extensions/e2e-test-nginx-deployment created\n"
+STEP: verifying the deployment e2e-test-nginx-deployment was created
+STEP: verifying the pod controlled by deployment e2e-test-nginx-deployment was created
+[AfterEach] [k8s.io] Kubectl run deployment
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1404
+Jun 18 12:44:56.615: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 delete deployment e2e-test-nginx-deployment --namespace=e2e-tests-kubectl-p76x4'
+Jun 18 12:44:56.798: INFO: stderr: ""
+Jun 18 12:44:56.798: INFO: stdout: "deployment.extensions \"e2e-test-nginx-deployment\" deleted\n"
 [AfterEach] [sig-cli] Kubectl client
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
-May 13 19:49:54.223: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "e2e-tests-kubectl-pgzg2" for this suite.
-May 13 19:50:18.259: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-May 13 19:50:18.397: INFO: namespace: e2e-tests-kubectl-pgzg2, resource: bindings, ignored listing per whitelist
-May 13 19:50:18.580: INFO: namespace e2e-tests-kubectl-pgzg2 deletion completed in 24.349268678s
+Jun 18 12:44:56.798: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-tests-kubectl-p76x4" for this suite.
+Jun 18 12:45:20.867: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 12:45:21.243: INFO: namespace: e2e-tests-kubectl-p76x4, resource: bindings, ignored listing per whitelist
+Jun 18 12:45:21.447: INFO: namespace e2e-tests-kubectl-p76x4 deletion completed in 24.627880076s
 
-• [SLOW TEST:54.058 seconds]
+• [SLOW TEST:29.467 seconds]
 [sig-cli] Kubectl client
 /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:22
-  [k8s.io] Update Demo
+  [k8s.io] Kubectl run deployment
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694
-    should do a rolling update of a replication controller  [Conformance]
+    should create a deployment from an image  [Conformance]
     /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 ------------------------------
-SSSSSSSSSSSSSS
+S
 ------------------------------
-[sig-storage] Subpath Atomic writer volumes 
-  should support subpaths with configmap pod [Conformance]
+[k8s.io] Kubelet when scheduling a busybox command in a pod 
+  should print the output to logs [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-[BeforeEach] [sig-storage] Subpath
+[BeforeEach] [k8s.io] Kubelet
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
 STEP: Creating a kubernetes client
-May 13 19:50:18.581: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423
-STEP: Building a namespace api object, basename subpath
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-subpath-w6ql2
+Jun 18 12:45:21.447: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
+STEP: Building a namespace api object, basename kubelet-test
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-kubelet-test-lp2qj
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] Atomic writer volumes
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:38
-STEP: Setting up data
-[It] should support subpaths with configmap pod [Conformance]
+[BeforeEach] [k8s.io] Kubelet
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:37
+[It] should print the output to logs [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-STEP: Creating pod pod-subpath-test-configmap-wfpf
-STEP: Creating a pod to test atomic-volume-subpath
-May 13 19:50:18.885: INFO: Waiting up to 5m0s for pod "pod-subpath-test-configmap-wfpf" in namespace "e2e-tests-subpath-w6ql2" to be "success or failure"
-May 13 19:50:18.892: INFO: Pod "pod-subpath-test-configmap-wfpf": Phase="Pending", Reason="", readiness=false. Elapsed: 7.176846ms
-May 13 19:50:20.900: INFO: Pod "pod-subpath-test-configmap-wfpf": Phase="Pending", Reason="", readiness=false. Elapsed: 2.015260407s
-May 13 19:50:22.909: INFO: Pod "pod-subpath-test-configmap-wfpf": Phase="Running", Reason="", readiness=false. Elapsed: 4.023581469s
-May 13 19:50:24.918: INFO: Pod "pod-subpath-test-configmap-wfpf": Phase="Running", Reason="", readiness=false. Elapsed: 6.033058687s
-May 13 19:50:26.938: INFO: Pod "pod-subpath-test-configmap-wfpf": Phase="Running", Reason="", readiness=false. Elapsed: 8.05341209s
-May 13 19:50:28.947: INFO: Pod "pod-subpath-test-configmap-wfpf": Phase="Running", Reason="", readiness=false. Elapsed: 10.061868565s
-May 13 19:50:30.959: INFO: Pod "pod-subpath-test-configmap-wfpf": Phase="Running", Reason="", readiness=false. Elapsed: 12.073587455s
-May 13 19:50:32.967: INFO: Pod "pod-subpath-test-configmap-wfpf": Phase="Running", Reason="", readiness=false. Elapsed: 14.081845352s
-May 13 19:50:34.975: INFO: Pod "pod-subpath-test-configmap-wfpf": Phase="Running", Reason="", readiness=false. Elapsed: 16.090466074s
-May 13 19:50:36.997: INFO: Pod "pod-subpath-test-configmap-wfpf": Phase="Running", Reason="", readiness=false. Elapsed: 18.111783855s
-May 13 19:50:39.005: INFO: Pod "pod-subpath-test-configmap-wfpf": Phase="Running", Reason="", readiness=false. Elapsed: 20.11998139s
-May 13 19:50:41.015: INFO: Pod "pod-subpath-test-configmap-wfpf": Phase="Running", Reason="", readiness=false. Elapsed: 22.129921764s
-May 13 19:50:43.023: INFO: Pod "pod-subpath-test-configmap-wfpf": Phase="Succeeded", Reason="", readiness=false. Elapsed: 24.138425502s
-STEP: Saw pod success
-May 13 19:50:43.023: INFO: Pod "pod-subpath-test-configmap-wfpf" satisfied condition "success or failure"
-May 13 19:50:43.031: INFO: Trying to get logs from node 10.170.219.153 pod pod-subpath-test-configmap-wfpf container test-container-subpath-configmap-wfpf: 
-STEP: delete the pod
-May 13 19:50:43.076: INFO: Waiting for pod pod-subpath-test-configmap-wfpf to disappear
-May 13 19:50:43.084: INFO: Pod pod-subpath-test-configmap-wfpf no longer exists
-STEP: Deleting pod pod-subpath-test-configmap-wfpf
-May 13 19:50:43.084: INFO: Deleting pod "pod-subpath-test-configmap-wfpf" in namespace "e2e-tests-subpath-w6ql2"
-[AfterEach] [sig-storage] Subpath
+[AfterEach] [k8s.io] Kubelet
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
-May 13 19:50:43.092: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "e2e-tests-subpath-w6ql2" for this suite.
-May 13 19:50:49.129: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-May 13 19:50:49.331: INFO: namespace: e2e-tests-subpath-w6ql2, resource: bindings, ignored listing per whitelist
-May 13 19:50:49.489: INFO: namespace e2e-tests-subpath-w6ql2 deletion completed in 6.38877072s
+Jun 18 12:45:23.994: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-tests-kubelet-test-lp2qj" for this suite.
+Jun 18 12:46:16.084: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 12:46:16.369: INFO: namespace: e2e-tests-kubelet-test-lp2qj, resource: bindings, ignored listing per whitelist
+Jun 18 12:46:16.665: INFO: namespace e2e-tests-kubelet-test-lp2qj deletion completed in 52.653466484s
 
-• [SLOW TEST:30.908 seconds]
-[sig-storage] Subpath
-/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:22
-  Atomic writer volumes
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:34
-    should support subpaths with configmap pod [Conformance]
+• [SLOW TEST:55.218 seconds]
+[k8s.io] Kubelet
+/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694
+  when scheduling a busybox command in a pod
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:40
+    should print the output to logs [NodeConformance] [Conformance]
     /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 ------------------------------
-SSSSSSSSSSSS
+SSSSSSS
 ------------------------------
-[sig-storage] Projected configMap 
-  should be consumable from pods in volume as non-root [NodeConformance] [Conformance]
+[sig-storage] EmptyDir volumes 
+  should support (root,0777,tmpfs) [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-[BeforeEach] [sig-storage] Projected configMap
+[BeforeEach] [sig-storage] EmptyDir volumes
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
 STEP: Creating a kubernetes client
-May 13 19:50:49.489: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423
-STEP: Building a namespace api object, basename projected
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-projected-wk59n
+Jun 18 12:46:16.666: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
+STEP: Building a namespace api object, basename emptydir
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-emptydir-nrsx2
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should be consumable from pods in volume as non-root [NodeConformance] [Conformance]
+[It] should support (root,0777,tmpfs) [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-STEP: Creating configMap with name projected-configmap-test-volume-6872250f-75b8-11e9-8f67-2632f168be36
-STEP: Creating a pod to test consume configMaps
-May 13 19:50:49.850: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-6873aa08-75b8-11e9-8f67-2632f168be36" in namespace "e2e-tests-projected-wk59n" to be "success or failure"
-May 13 19:50:49.859: INFO: Pod "pod-projected-configmaps-6873aa08-75b8-11e9-8f67-2632f168be36": Phase="Pending", Reason="", readiness=false. Elapsed: 8.278532ms
-May 13 19:50:51.870: INFO: Pod "pod-projected-configmaps-6873aa08-75b8-11e9-8f67-2632f168be36": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.019884982s
+STEP: Creating a pod to test emptydir 0777 on tmpfs
+Jun 18 12:46:17.162: INFO: Waiting up to 5m0s for pod "pod-10697156-91c7-11e9-bce2-ae54e022189f" in namespace "e2e-tests-emptydir-nrsx2" to be "success or failure"
+Jun 18 12:46:17.176: INFO: Pod "pod-10697156-91c7-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 14.076494ms
+Jun 18 12:46:19.190: INFO: Pod "pod-10697156-91c7-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 2.028208675s
+Jun 18 12:46:21.205: INFO: Pod "pod-10697156-91c7-11e9-bce2-ae54e022189f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.043087296s
 STEP: Saw pod success
-May 13 19:50:51.870: INFO: Pod "pod-projected-configmaps-6873aa08-75b8-11e9-8f67-2632f168be36" satisfied condition "success or failure"
-May 13 19:50:51.878: INFO: Trying to get logs from node 10.170.219.177 pod pod-projected-configmaps-6873aa08-75b8-11e9-8f67-2632f168be36 container projected-configmap-volume-test: 
+Jun 18 12:46:21.205: INFO: Pod "pod-10697156-91c7-11e9-bce2-ae54e022189f" satisfied condition "success or failure"
+Jun 18 12:46:21.220: INFO: Trying to get logs from node 10.72.74.144 pod pod-10697156-91c7-11e9-bce2-ae54e022189f container test-container: 
 STEP: delete the pod
-May 13 19:50:51.918: INFO: Waiting for pod pod-projected-configmaps-6873aa08-75b8-11e9-8f67-2632f168be36 to disappear
-May 13 19:50:51.925: INFO: Pod pod-projected-configmaps-6873aa08-75b8-11e9-8f67-2632f168be36 no longer exists
-[AfterEach] [sig-storage] Projected configMap
+Jun 18 12:46:21.300: INFO: Waiting for pod pod-10697156-91c7-11e9-bce2-ae54e022189f to disappear
+Jun 18 12:46:21.314: INFO: Pod pod-10697156-91c7-11e9-bce2-ae54e022189f no longer exists
+[AfterEach] [sig-storage] EmptyDir volumes
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
-May 13 19:50:51.925: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "e2e-tests-projected-wk59n" for this suite.
-May 13 19:50:57.973: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-May 13 19:50:58.685: INFO: namespace: e2e-tests-projected-wk59n, resource: bindings, ignored listing per whitelist
-May 13 19:50:58.819: INFO: namespace e2e-tests-projected-wk59n deletion completed in 6.885788297s
+Jun 18 12:46:21.314: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-tests-emptydir-nrsx2" for this suite.
+Jun 18 12:46:27.386: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 12:46:28.146: INFO: namespace: e2e-tests-emptydir-nrsx2, resource: bindings, ignored listing per whitelist
+Jun 18 12:46:28.456: INFO: namespace e2e-tests-emptydir-nrsx2 deletion completed in 7.120905195s
 
-• [SLOW TEST:9.329 seconds]
-[sig-storage] Projected configMap
-/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_configmap.go:34
-  should be consumable from pods in volume as non-root [NodeConformance] [Conformance]
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-------------------------------
-SS
-------------------------------
-[sig-storage] Subpath Atomic writer volumes 
-  should support subpaths with secret pod [Conformance]
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-[BeforeEach] [sig-storage] Subpath
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
-STEP: Creating a kubernetes client
-May 13 19:50:58.819: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423
-STEP: Building a namespace api object, basename subpath
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-subpath-qm6vt
-STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] Atomic writer volumes
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:38
-STEP: Setting up data
-[It] should support subpaths with secret pod [Conformance]
+• [SLOW TEST:11.790 seconds]
+[sig-storage] EmptyDir volumes
+/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:40
+  should support (root,0777,tmpfs) [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-STEP: Creating pod pod-subpath-test-secret-8lsl
-STEP: Creating a pod to test atomic-volume-subpath
-May 13 19:50:59.122: INFO: Waiting up to 5m0s for pod "pod-subpath-test-secret-8lsl" in namespace "e2e-tests-subpath-qm6vt" to be "success or failure"
-May 13 19:50:59.130: INFO: Pod "pod-subpath-test-secret-8lsl": Phase="Pending", Reason="", readiness=false. Elapsed: 8.543061ms
-May 13 19:51:01.138: INFO: Pod "pod-subpath-test-secret-8lsl": Phase="Pending", Reason="", readiness=false. Elapsed: 2.01646935s
-May 13 19:51:03.146: INFO: Pod "pod-subpath-test-secret-8lsl": Phase="Running", Reason="", readiness=false. Elapsed: 4.024811009s
-May 13 19:51:05.155: INFO: Pod "pod-subpath-test-secret-8lsl": Phase="Running", Reason="", readiness=false. Elapsed: 6.033014571s
-May 13 19:51:07.164: INFO: Pod "pod-subpath-test-secret-8lsl": Phase="Running", Reason="", readiness=false. Elapsed: 8.042165069s
-May 13 19:51:09.184: INFO: Pod "pod-subpath-test-secret-8lsl": Phase="Running", Reason="", readiness=false. Elapsed: 10.062576912s
-May 13 19:51:11.193: INFO: Pod "pod-subpath-test-secret-8lsl": Phase="Running", Reason="", readiness=false. Elapsed: 12.070973229s
-May 13 19:51:13.201: INFO: Pod "pod-subpath-test-secret-8lsl": Phase="Running", Reason="", readiness=false. Elapsed: 14.079338422s
-May 13 19:51:15.209: INFO: Pod "pod-subpath-test-secret-8lsl": Phase="Running", Reason="", readiness=false. Elapsed: 16.08779422s
-May 13 19:51:17.218: INFO: Pod "pod-subpath-test-secret-8lsl": Phase="Running", Reason="", readiness=false. Elapsed: 18.096325497s
-May 13 19:51:19.238: INFO: Pod "pod-subpath-test-secret-8lsl": Phase="Running", Reason="", readiness=false. Elapsed: 20.116455764s
-May 13 19:51:21.246: INFO: Pod "pod-subpath-test-secret-8lsl": Phase="Running", Reason="", readiness=false. Elapsed: 22.124810257s
-May 13 19:51:23.256: INFO: Pod "pod-subpath-test-secret-8lsl": Phase="Succeeded", Reason="", readiness=false. Elapsed: 24.134035928s
-STEP: Saw pod success
-May 13 19:51:23.256: INFO: Pod "pod-subpath-test-secret-8lsl" satisfied condition "success or failure"
-May 13 19:51:23.264: INFO: Trying to get logs from node 10.170.219.153 pod pod-subpath-test-secret-8lsl container test-container-subpath-secret-8lsl: 
-STEP: delete the pod
-May 13 19:51:23.420: INFO: Waiting for pod pod-subpath-test-secret-8lsl to disappear
-May 13 19:51:23.427: INFO: Pod pod-subpath-test-secret-8lsl no longer exists
-STEP: Deleting pod pod-subpath-test-secret-8lsl
-May 13 19:51:23.427: INFO: Deleting pod "pod-subpath-test-secret-8lsl" in namespace "e2e-tests-subpath-qm6vt"
-[AfterEach] [sig-storage] Subpath
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
-May 13 19:51:23.434: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "e2e-tests-subpath-qm6vt" for this suite.
-May 13 19:51:29.489: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-May 13 19:51:29.617: INFO: namespace: e2e-tests-subpath-qm6vt, resource: bindings, ignored listing per whitelist
-May 13 19:51:29.741: INFO: namespace e2e-tests-subpath-qm6vt deletion completed in 6.299127879s
-
-• [SLOW TEST:30.922 seconds]
-[sig-storage] Subpath
-/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:22
-  Atomic writer volumes
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:34
-    should support subpaths with secret pod [Conformance]
-    /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 ------------------------------
-S
+SSSSSSSSSSS
 ------------------------------
-[sig-storage] EmptyDir volumes 
-  should support (non-root,0644,tmpfs) [NodeConformance] [Conformance]
+[sig-apps] Deployment 
+  deployment should support rollover [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-[BeforeEach] [sig-storage] EmptyDir volumes
+[BeforeEach] [sig-apps] Deployment
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
 STEP: Creating a kubernetes client
-May 13 19:51:29.741: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423
-STEP: Building a namespace api object, basename emptydir
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-emptydir-fvw4p
+Jun 18 12:46:28.456: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
+STEP: Building a namespace api object, basename deployment
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-deployment-pvzpw
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should support (non-root,0644,tmpfs) [NodeConformance] [Conformance]
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-STEP: Creating a pod to test emptydir 0644 on tmpfs
-May 13 19:51:30.033: INFO: Waiting up to 5m0s for pod "pod-8066ac0a-75b8-11e9-8f67-2632f168be36" in namespace "e2e-tests-emptydir-fvw4p" to be "success or failure"
-May 13 19:51:30.042: INFO: Pod "pod-8066ac0a-75b8-11e9-8f67-2632f168be36": Phase="Pending", Reason="", readiness=false. Elapsed: 8.642363ms
-May 13 19:51:32.051: INFO: Pod "pod-8066ac0a-75b8-11e9-8f67-2632f168be36": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.017488705s
-STEP: Saw pod success
-May 13 19:51:32.051: INFO: Pod "pod-8066ac0a-75b8-11e9-8f67-2632f168be36" satisfied condition "success or failure"
-May 13 19:51:32.058: INFO: Trying to get logs from node 10.170.219.177 pod pod-8066ac0a-75b8-11e9-8f67-2632f168be36 container test-container: 
-STEP: delete the pod
-May 13 19:51:32.103: INFO: Waiting for pod pod-8066ac0a-75b8-11e9-8f67-2632f168be36 to disappear
-May 13 19:51:32.112: INFO: Pod pod-8066ac0a-75b8-11e9-8f67-2632f168be36 no longer exists
-[AfterEach] [sig-storage] EmptyDir volumes
+[BeforeEach] [sig-apps] Deployment
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:65
+[It] deployment should support rollover [Conformance]
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
+Jun 18 12:46:30.293: INFO: Pod name rollover-pod: Found 0 pods out of 1
+Jun 18 12:46:35.310: INFO: Pod name rollover-pod: Found 1 pods out of 1
+STEP: ensuring each pod is running
+Jun 18 12:46:35.310: INFO: Waiting for pods owned by replica set "test-rollover-controller" to become ready
+Jun 18 12:46:37.341: INFO: Creating deployment "test-rollover-deployment"
+Jun 18 12:46:37.374: INFO: Make sure deployment "test-rollover-deployment" performs scaling operations
+Jun 18 12:46:39.407: INFO: Check revision of new replica set for deployment "test-rollover-deployment"
+Jun 18 12:46:39.434: INFO: Ensure that both replica sets have 1 created replica
+Jun 18 12:46:39.461: INFO: Rollover old replica sets for deployment "test-rollover-deployment" with new image update
+Jun 18 12:46:39.491: INFO: Updating deployment test-rollover-deployment
+Jun 18 12:46:39.491: INFO: Wait deployment "test-rollover-deployment" to be observed by the deployment controller
+Jun 18 12:46:41.522: INFO: Wait for revision update of deployment "test-rollover-deployment" to 2
+Jun 18 12:46:41.600: INFO: Make sure deployment "test-rollover-deployment" is complete
+Jun 18 12:46:41.628: INFO: all replica sets need to contain the pod-template-hash label
+Jun 18 12:46:41.628: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:2, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63696458797, loc:(*time.Location)(0x7b33b80)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63696458797, loc:(*time.Location)(0x7b33b80)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63696458801, loc:(*time.Location)(0x7b33b80)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63696458797, loc:(*time.Location)(0x7b33b80)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-6b7f9d6597\" is progressing."}}, CollisionCount:(*int32)(nil)}
+Jun 18 12:46:44.311: INFO: all replica sets need to contain the pod-template-hash label
+Jun 18 12:46:44.311: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:2, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63696458797, loc:(*time.Location)(0x7b33b80)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63696458797, loc:(*time.Location)(0x7b33b80)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63696458801, loc:(*time.Location)(0x7b33b80)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63696458797, loc:(*time.Location)(0x7b33b80)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-6b7f9d6597\" is progressing."}}, CollisionCount:(*int32)(nil)}
+Jun 18 12:46:45.655: INFO: all replica sets need to contain the pod-template-hash label
+Jun 18 12:46:45.656: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:2, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63696458797, loc:(*time.Location)(0x7b33b80)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63696458797, loc:(*time.Location)(0x7b33b80)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63696458801, loc:(*time.Location)(0x7b33b80)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63696458797, loc:(*time.Location)(0x7b33b80)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-6b7f9d6597\" is progressing."}}, CollisionCount:(*int32)(nil)}
+Jun 18 12:46:47.698: INFO: all replica sets need to contain the pod-template-hash label
+Jun 18 12:46:47.699: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:2, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63696458797, loc:(*time.Location)(0x7b33b80)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63696458797, loc:(*time.Location)(0x7b33b80)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63696458801, loc:(*time.Location)(0x7b33b80)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63696458797, loc:(*time.Location)(0x7b33b80)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-6b7f9d6597\" is progressing."}}, CollisionCount:(*int32)(nil)}
+Jun 18 12:46:49.656: INFO: all replica sets need to contain the pod-template-hash label
+Jun 18 12:46:49.656: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:2, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63696458797, loc:(*time.Location)(0x7b33b80)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63696458797, loc:(*time.Location)(0x7b33b80)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63696458801, loc:(*time.Location)(0x7b33b80)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63696458797, loc:(*time.Location)(0x7b33b80)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-6b7f9d6597\" is progressing."}}, CollisionCount:(*int32)(nil)}
+Jun 18 12:46:51.658: INFO: 
+Jun 18 12:46:51.658: INFO: Ensure that both old replica sets have no replicas
+[AfterEach] [sig-apps] Deployment
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:59
+Jun 18 12:46:51.704: INFO: Deployment "test-rollover-deployment":
+&Deployment{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-rollover-deployment,GenerateName:,Namespace:e2e-tests-deployment-pvzpw,SelfLink:/apis/apps/v1/namespaces/e2e-tests-deployment-pvzpw/deployments/test-rollover-deployment,UID:1c75ac7e-91c7-11e9-bf44-fa6f350b29f0,ResourceVersion:105749,Generation:2,CreationTimestamp:2019-06-18 12:46:37 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: rollover-pod,},Annotations:map[string]string{deployment.kubernetes.io/revision: 2,},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Spec:DeploymentSpec{Replicas:*1,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: rollover-pod,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: rollover-pod,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[],Containers:[{redis gcr.io/kubernetes-e2e-test-images/redis:1.0 [] []  [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,},},Strategy:DeploymentStrategy{Type:RollingUpdate,RollingUpdate:&RollingUpdateDeployment{MaxUnavailable:0,MaxSurge:1,},},MinReadySeconds:10,RevisionHistoryLimit:*10,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:2,Replicas:1,UpdatedReplicas:1,AvailableReplicas:1,UnavailableReplicas:0,Conditions:[{Available True 2019-06-18 12:46:37 +0000 UTC 2019-06-18 12:46:37 +0000 UTC MinimumReplicasAvailable Deployment has minimum availability.} {Progressing True 2019-06-18 12:46:51 +0000 UTC 2019-06-18 12:46:37 +0000 UTC NewReplicaSetAvailable ReplicaSet "test-rollover-deployment-6b7f9d6597" has successfully progressed.}],ReadyReplicas:1,CollisionCount:nil,},}
+
+Jun 18 12:46:51.719: INFO: New ReplicaSet "test-rollover-deployment-6b7f9d6597" of Deployment "test-rollover-deployment":
+&ReplicaSet{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-rollover-deployment-6b7f9d6597,GenerateName:,Namespace:e2e-tests-deployment-pvzpw,SelfLink:/apis/apps/v1/namespaces/e2e-tests-deployment-pvzpw/replicasets/test-rollover-deployment-6b7f9d6597,UID:1dbce917-91c7-11e9-bf44-fa6f350b29f0,ResourceVersion:105740,Generation:2,CreationTimestamp:2019-06-18 12:46:39 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: rollover-pod,pod-template-hash: 6b7f9d6597,},Annotations:map[string]string{deployment.kubernetes.io/desired-replicas: 1,deployment.kubernetes.io/max-replicas: 2,deployment.kubernetes.io/revision: 2,},OwnerReferences:[{apps/v1 Deployment test-rollover-deployment 1c75ac7e-91c7-11e9-bf44-fa6f350b29f0 0xc002249367 0xc002249368}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:ReplicaSetSpec{Replicas:*1,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: rollover-pod,pod-template-hash: 6b7f9d6597,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: rollover-pod,pod-template-hash: 6b7f9d6597,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[],Containers:[{redis gcr.io/kubernetes-e2e-test-images/redis:1.0 [] []  [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,},},MinReadySeconds:10,},Status:ReplicaSetStatus{Replicas:1,FullyLabeledReplicas:1,ObservedGeneration:2,ReadyReplicas:1,AvailableReplicas:1,Conditions:[],},}
+Jun 18 12:46:51.719: INFO: All old ReplicaSets of Deployment "test-rollover-deployment":
+Jun 18 12:46:51.719: INFO: &ReplicaSet{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-rollover-controller,GenerateName:,Namespace:e2e-tests-deployment-pvzpw,SelfLink:/apis/apps/v1/namespaces/e2e-tests-deployment-pvzpw/replicasets/test-rollover-controller,UID:183d8251-91c7-11e9-bf44-fa6f350b29f0,ResourceVersion:105748,Generation:2,CreationTimestamp:2019-06-18 12:46:30 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: rollover-pod,pod: nginx,},Annotations:map[string]string{deployment.kubernetes.io/desired-replicas: 1,deployment.kubernetes.io/max-replicas: 2,},OwnerReferences:[{apps/v1 Deployment test-rollover-deployment 1c75ac7e-91c7-11e9-bf44-fa6f350b29f0 0xc0022491d7 0xc0022491d8}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:ReplicaSetSpec{Replicas:*0,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: rollover-pod,pod: nginx,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: rollover-pod,pod: nginx,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] []  [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,},},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:0,FullyLabeledReplicas:0,ObservedGeneration:2,ReadyReplicas:0,AvailableReplicas:0,Conditions:[],},}
+Jun 18 12:46:51.719: INFO: &ReplicaSet{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-rollover-deployment-6586df867b,GenerateName:,Namespace:e2e-tests-deployment-pvzpw,SelfLink:/apis/apps/v1/namespaces/e2e-tests-deployment-pvzpw/replicasets/test-rollover-deployment-6586df867b,UID:1c7e8a0e-91c7-11e9-bf44-fa6f350b29f0,ResourceVersion:105706,Generation:2,CreationTimestamp:2019-06-18 12:46:37 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: rollover-pod,pod-template-hash: 6586df867b,},Annotations:map[string]string{deployment.kubernetes.io/desired-replicas: 1,deployment.kubernetes.io/max-replicas: 2,deployment.kubernetes.io/revision: 1,},OwnerReferences:[{apps/v1 Deployment test-rollover-deployment 1c75ac7e-91c7-11e9-bf44-fa6f350b29f0 0xc002249297 0xc002249298}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:ReplicaSetSpec{Replicas:*0,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: rollover-pod,pod-template-hash: 6586df867b,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: rollover-pod,pod-template-hash: 6586df867b,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[],Containers:[{redis-slave gcr.io/google_samples/gb-redisslave:nonexistent [] []  [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,},},MinReadySeconds:10,},Status:ReplicaSetStatus{Replicas:0,FullyLabeledReplicas:0,ObservedGeneration:2,ReadyReplicas:0,AvailableReplicas:0,Conditions:[],},}
+Jun 18 12:46:51.734: INFO: Pod "test-rollover-deployment-6b7f9d6597-8gsnj" is available:
+&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-rollover-deployment-6b7f9d6597-8gsnj,GenerateName:test-rollover-deployment-6b7f9d6597-,Namespace:e2e-tests-deployment-pvzpw,SelfLink:/api/v1/namespaces/e2e-tests-deployment-pvzpw/pods/test-rollover-deployment-6b7f9d6597-8gsnj,UID:1dc5b87c-91c7-11e9-bf44-fa6f350b29f0,ResourceVersion:105721,Generation:0,CreationTimestamp:2019-06-18 12:46:39 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: rollover-pod,pod-template-hash: 6b7f9d6597,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet test-rollover-deployment-6b7f9d6597 1dbce917-91c7-11e9-bf44-fa6f350b29f0 0xc0021543f7 0xc0021543f8}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-rkgg7 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-rkgg7,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{redis gcr.io/kubernetes-e2e-test-images/redis:1.0 [] []  [] [] [] {map[] map[]} [{default-token-rkgg7 true /var/run/secrets/kubernetes.io/serviceaccount  }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.72.74.144,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists  NoExecute 0xc002154600} {node.kubernetes.io/unreachable Exists  NoExecute 0xc002154620}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Running,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:46:39 +0000 UTC  } {Ready True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:46:41 +0000 UTC  } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:46:41 +0000 UTC  } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:46:39 +0000 UTC  }],Message:,Reason:,HostIP:10.72.74.144,PodIP:172.30.114.9,StartTime:2019-06-18 12:46:39 +0000 UTC,ContainerStatuses:[{redis {nil ContainerStateRunning{StartedAt:2019-06-18 12:46:40 +0000 UTC,} nil} {nil nil nil} true 0 gcr.io/kubernetes-e2e-test-images/redis:1.0 gcr.io/kubernetes-e2e-test-images/redis@sha256:af4748d1655c08dc54d4be5182135395db9ce87aba2d4699b26b14ae197c5830 containerd://ee6d4938b0a456a3060d09834073d7f7189870a173aa6dccd61082f995a7c7b8}],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},}
+[AfterEach] [sig-apps] Deployment
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
-May 13 19:51:32.113: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "e2e-tests-emptydir-fvw4p" for this suite.
-May 13 19:51:38.157: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-May 13 19:51:38.309: INFO: namespace: e2e-tests-emptydir-fvw4p, resource: bindings, ignored listing per whitelist
-May 13 19:51:38.469: INFO: namespace e2e-tests-emptydir-fvw4p deletion completed in 6.348340353s
+Jun 18 12:46:51.735: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-tests-deployment-pvzpw" for this suite.
+Jun 18 12:46:59.801: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 12:47:00.332: INFO: namespace: e2e-tests-deployment-pvzpw, resource: bindings, ignored listing per whitelist
+Jun 18 12:47:00.490: INFO: namespace e2e-tests-deployment-pvzpw deletion completed in 8.736759518s
 
-• [SLOW TEST:8.728 seconds]
-[sig-storage] EmptyDir volumes
-/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:40
-  should support (non-root,0644,tmpfs) [NodeConformance] [Conformance]
+• [SLOW TEST:32.034 seconds]
+[sig-apps] Deployment
+/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22
+  deployment should support rollover [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 ------------------------------
-SSSSSSSSSSSS
+SSS
 ------------------------------
-[sig-cli] Kubectl client [k8s.io] Kubectl api-versions 
-  should check if v1 is in available api versions  [Conformance]
+[sig-storage] Subpath Atomic writer volumes 
+  should support subpaths with downward pod [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-[BeforeEach] [sig-cli] Kubectl client
+[BeforeEach] [sig-storage] Subpath
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
 STEP: Creating a kubernetes client
-May 13 19:51:38.471: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423
-STEP: Building a namespace api object, basename kubectl
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-kubectl-nzhq9
+Jun 18 12:47:00.490: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
+STEP: Building a namespace api object, basename subpath
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-subpath-x7flg
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:243
-[It] should check if v1 is in available api versions  [Conformance]
+[BeforeEach] Atomic writer volumes
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:38
+STEP: Setting up data
+[It] should support subpaths with downward pod [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-STEP: validating api versions
-May 13 19:51:38.749: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 api-versions'
-May 13 19:51:38.878: INFO: stderr: ""
-May 13 19:51:38.879: INFO: stdout: "admissionregistration.k8s.io/v1alpha1\nadmissionregistration.k8s.io/v1beta1\napiextensions.k8s.io/v1beta1\napiregistration.k8s.io/v1\napiregistration.k8s.io/v1beta1\napps/v1\napps/v1beta1\napps/v1beta2\nauthentication.k8s.io/v1\nauthentication.k8s.io/v1beta1\nauthorization.k8s.io/v1\nauthorization.k8s.io/v1beta1\nautoscaling/v1\nautoscaling/v2beta1\nautoscaling/v2beta2\nbatch/v1\nbatch/v1beta1\nbatch/v2alpha1\ncertificates.k8s.io/v1beta1\ncoordination.k8s.io/v1beta1\nevents.k8s.io/v1beta1\nextensions/v1beta1\nmetrics.k8s.io/v1beta1\nnetworking.k8s.io/v1\npolicy/v1beta1\nrbac.authorization.k8s.io/v1\nrbac.authorization.k8s.io/v1beta1\nscheduling.k8s.io/v1beta1\nstorage.k8s.io/v1\nstorage.k8s.io/v1beta1\nv1\n"
-[AfterEach] [sig-cli] Kubectl client
+STEP: Creating pod pod-subpath-test-downwardapi-7jfd
+STEP: Creating a pod to test atomic-volume-subpath
+Jun 18 12:47:00.990: INFO: Waiting up to 5m0s for pod "pod-subpath-test-downwardapi-7jfd" in namespace "e2e-tests-subpath-x7flg" to be "success or failure"
+Jun 18 12:47:01.006: INFO: Pod "pod-subpath-test-downwardapi-7jfd": Phase="Pending", Reason="", readiness=false. Elapsed: 15.79041ms
+Jun 18 12:47:03.021: INFO: Pod "pod-subpath-test-downwardapi-7jfd": Phase="Pending", Reason="", readiness=false. Elapsed: 2.03112706s
+Jun 18 12:47:05.036: INFO: Pod "pod-subpath-test-downwardapi-7jfd": Phase="Running", Reason="", readiness=false. Elapsed: 4.046073067s
+Jun 18 12:47:07.055: INFO: Pod "pod-subpath-test-downwardapi-7jfd": Phase="Running", Reason="", readiness=false. Elapsed: 6.064696983s
+Jun 18 12:47:09.089: INFO: Pod "pod-subpath-test-downwardapi-7jfd": Phase="Running", Reason="", readiness=false. Elapsed: 8.098673934s
+Jun 18 12:47:11.105: INFO: Pod "pod-subpath-test-downwardapi-7jfd": Phase="Running", Reason="", readiness=false. Elapsed: 10.114520585s
+Jun 18 12:47:13.120: INFO: Pod "pod-subpath-test-downwardapi-7jfd": Phase="Running", Reason="", readiness=false. Elapsed: 12.129517708s
+Jun 18 12:47:15.135: INFO: Pod "pod-subpath-test-downwardapi-7jfd": Phase="Running", Reason="", readiness=false. Elapsed: 14.144368402s
+Jun 18 12:47:17.150: INFO: Pod "pod-subpath-test-downwardapi-7jfd": Phase="Running", Reason="", readiness=false. Elapsed: 16.159650311s
+Jun 18 12:47:19.182: INFO: Pod "pod-subpath-test-downwardapi-7jfd": Phase="Running", Reason="", readiness=false. Elapsed: 18.191589585s
+Jun 18 12:47:21.198: INFO: Pod "pod-subpath-test-downwardapi-7jfd": Phase="Running", Reason="", readiness=false. Elapsed: 20.207295315s
+Jun 18 12:47:23.216: INFO: Pod "pod-subpath-test-downwardapi-7jfd": Phase="Running", Reason="", readiness=false. Elapsed: 22.225507535s
+Jun 18 12:47:25.231: INFO: Pod "pod-subpath-test-downwardapi-7jfd": Phase="Succeeded", Reason="", readiness=false. Elapsed: 24.240625439s
+STEP: Saw pod success
+Jun 18 12:47:25.231: INFO: Pod "pod-subpath-test-downwardapi-7jfd" satisfied condition "success or failure"
+Jun 18 12:47:25.247: INFO: Trying to get logs from node 10.72.74.149 pod pod-subpath-test-downwardapi-7jfd container test-container-subpath-downwardapi-7jfd: 
+STEP: delete the pod
+Jun 18 12:47:25.321: INFO: Waiting for pod pod-subpath-test-downwardapi-7jfd to disappear
+Jun 18 12:47:25.335: INFO: Pod pod-subpath-test-downwardapi-7jfd no longer exists
+STEP: Deleting pod pod-subpath-test-downwardapi-7jfd
+Jun 18 12:47:25.335: INFO: Deleting pod "pod-subpath-test-downwardapi-7jfd" in namespace "e2e-tests-subpath-x7flg"
+[AfterEach] [sig-storage] Subpath
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
-May 13 19:51:38.879: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "e2e-tests-kubectl-nzhq9" for this suite.
-May 13 19:51:44.917: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-May 13 19:51:45.143: INFO: namespace: e2e-tests-kubectl-nzhq9, resource: bindings, ignored listing per whitelist
-May 13 19:51:45.268: INFO: namespace e2e-tests-kubectl-nzhq9 deletion completed in 6.381483792s
+Jun 18 12:47:25.349: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-tests-subpath-x7flg" for this suite.
+Jun 18 12:47:33.417: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 12:47:33.575: INFO: namespace: e2e-tests-subpath-x7flg, resource: bindings, ignored listing per whitelist
+Jun 18 12:47:33.912: INFO: namespace e2e-tests-subpath-x7flg deletion completed in 8.544396817s
 
-• [SLOW TEST:6.797 seconds]
-[sig-cli] Kubectl client
-/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:22
-  [k8s.io] Kubectl api-versions
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694
-    should check if v1 is in available api versions  [Conformance]
+• [SLOW TEST:33.422 seconds]
+[sig-storage] Subpath
+/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:22
+  Atomic writer volumes
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:34
+    should support subpaths with downward pod [Conformance]
     /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 ------------------------------
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+SSSSSSSSSSSSS
 ------------------------------
-[sig-network] DNS 
-  should provide DNS for the cluster  [Conformance]
+[sig-storage] ConfigMap 
+  binary data should be reflected in volume [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-[BeforeEach] [sig-network] DNS
+[BeforeEach] [sig-storage] ConfigMap
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
 STEP: Creating a kubernetes client
-May 13 19:51:45.269: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423
-STEP: Building a namespace api object, basename dns
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-dns-5rb82
+Jun 18 12:47:33.912: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
+STEP: Building a namespace api object, basename configmap
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-configmap-nqbn2
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should provide DNS for the cluster  [Conformance]
+[It] binary data should be reflected in volume [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-STEP: Running these commands on wheezy: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search kubernetes.default A)" && test -n "$$check" && echo OK > /results/wheezy_udp@kubernetes.default;check="$$(dig +tcp +noall +answer +search kubernetes.default A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@kubernetes.default;check="$$(dig +notcp +noall +answer +search kubernetes.default.svc A)" && test -n "$$check" && echo OK > /results/wheezy_udp@kubernetes.default.svc;check="$$(dig +tcp +noall +answer +search kubernetes.default.svc A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@kubernetes.default.svc;check="$$(dig +notcp +noall +answer +search kubernetes.default.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_udp@kubernetes.default.svc.cluster.local;check="$$(dig +tcp +noall +answer +search kubernetes.default.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@kubernetes.default.svc.cluster.local;test -n "$$(getent hosts dns-querier-1.dns-test-service.e2e-tests-dns-5rb82.svc.cluster.local)" && echo OK > /results/wheezy_hosts@dns-querier-1.dns-test-service.e2e-tests-dns-5rb82.svc.cluster.local;test -n "$$(getent hosts dns-querier-1)" && echo OK > /results/wheezy_hosts@dns-querier-1;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".e2e-tests-dns-5rb82.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@PodARecord;sleep 1; done
-
-STEP: Running these commands on jessie: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search kubernetes.default A)" && test -n "$$check" && echo OK > /results/jessie_udp@kubernetes.default;check="$$(dig +tcp +noall +answer +search kubernetes.default A)" && test -n "$$check" && echo OK > /results/jessie_tcp@kubernetes.default;check="$$(dig +notcp +noall +answer +search kubernetes.default.svc A)" && test -n "$$check" && echo OK > /results/jessie_udp@kubernetes.default.svc;check="$$(dig +tcp +noall +answer +search kubernetes.default.svc A)" && test -n "$$check" && echo OK > /results/jessie_tcp@kubernetes.default.svc;check="$$(dig +notcp +noall +answer +search kubernetes.default.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_udp@kubernetes.default.svc.cluster.local;check="$$(dig +tcp +noall +answer +search kubernetes.default.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_tcp@kubernetes.default.svc.cluster.local;test -n "$$(getent hosts dns-querier-1.dns-test-service.e2e-tests-dns-5rb82.svc.cluster.local)" && echo OK > /results/jessie_hosts@dns-querier-1.dns-test-service.e2e-tests-dns-5rb82.svc.cluster.local;test -n "$$(getent hosts dns-querier-1)" && echo OK > /results/jessie_hosts@dns-querier-1;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".e2e-tests-dns-5rb82.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_tcp@PodARecord;sleep 1; done
-
-STEP: creating a pod to probe DNS
-STEP: submitting the pod to kubernetes
-STEP: retrieving the pod
-STEP: looking for the results for each expected name from probers
-May 13 19:51:50.045: INFO: DNS probes using e2e-tests-dns-5rb82/dns-test-89c6372b-75b8-11e9-8f67-2632f168be36 succeeded
-
-STEP: deleting the pod
-[AfterEach] [sig-network] DNS
+STEP: Creating configMap with name configmap-test-upd-3e76f649-91c7-11e9-bce2-ae54e022189f
+STEP: Creating the pod
+STEP: Waiting for pod with text data
+STEP: Waiting for pod with binary data
+[AfterEach] [sig-storage] ConfigMap
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
-May 13 19:51:50.071: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "e2e-tests-dns-5rb82" for this suite.
-May 13 19:51:56.106: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-May 13 19:51:56.282: INFO: namespace: e2e-tests-dns-5rb82, resource: bindings, ignored listing per whitelist
-May 13 19:51:56.364: INFO: namespace e2e-tests-dns-5rb82 deletion completed in 6.284869766s
+Jun 18 12:47:38.554: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-tests-configmap-nqbn2" for this suite.
+Jun 18 12:48:02.620: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 12:48:02.699: INFO: namespace: e2e-tests-configmap-nqbn2, resource: bindings, ignored listing per whitelist
+Jun 18 12:48:03.153: INFO: namespace e2e-tests-configmap-nqbn2 deletion completed in 24.579861794s
 
-• [SLOW TEST:11.094 seconds]
-[sig-network] DNS
-/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:22
-  should provide DNS for the cluster  [Conformance]
+• [SLOW TEST:29.240 seconds]
+[sig-storage] ConfigMap
+/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:33
+  binary data should be reflected in volume [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 ------------------------------
-S
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
 [sig-api-machinery] Watchers 
   should observe an object deletion if it stops meeting the requirements of the selector [Conformance]
@@ -7456,9 +7261,9 @@ S
 [BeforeEach] [sig-api-machinery] Watchers
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
 STEP: Creating a kubernetes client
-May 13 19:51:56.364: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423
+Jun 18 12:48:03.153: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
 STEP: Building a namespace api object, basename watch
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-watch-qg8j4
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-watch-zzv98
 STEP: Waiting for a default service account to be provisioned in namespace
 [It] should observe an object deletion if it stops meeting the requirements of the selector [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
@@ -7467,1485 +7272,1804 @@ STEP: creating a new configmap
 STEP: modifying the configmap once
 STEP: changing the label value of the configmap
 STEP: Expecting to observe a delete notification for the watched object
-May 13 19:51:56.682: INFO: Got : ADDED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-label-changed,GenerateName:,Namespace:e2e-tests-watch-qg8j4,SelfLink:/api/v1/namespaces/e2e-tests-watch-qg8j4/configmaps/e2e-watch-test-label-changed,UID:9045931d-75b8-11e9-b786-da20024d205c,ResourceVersion:49188,Generation:0,CreationTimestamp:2019-05-13 19:51:56 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: label-changed-and-restored,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Data:map[string]string{},BinaryData:map[string][]byte{},}
-May 13 19:51:56.682: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-label-changed,GenerateName:,Namespace:e2e-tests-watch-qg8j4,SelfLink:/api/v1/namespaces/e2e-tests-watch-qg8j4/configmaps/e2e-watch-test-label-changed,UID:9045931d-75b8-11e9-b786-da20024d205c,ResourceVersion:49189,Generation:0,CreationTimestamp:2019-05-13 19:51:56 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: label-changed-and-restored,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Data:map[string]string{mutation: 1,},BinaryData:map[string][]byte{},}
-May 13 19:51:56.682: INFO: Got : DELETED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-label-changed,GenerateName:,Namespace:e2e-tests-watch-qg8j4,SelfLink:/api/v1/namespaces/e2e-tests-watch-qg8j4/configmaps/e2e-watch-test-label-changed,UID:9045931d-75b8-11e9-b786-da20024d205c,ResourceVersion:49190,Generation:0,CreationTimestamp:2019-05-13 19:51:56 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: label-changed-and-restored,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Data:map[string]string{mutation: 1,},BinaryData:map[string][]byte{},}
+Jun 18 12:48:03.724: INFO: Got : ADDED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-label-changed,GenerateName:,Namespace:e2e-tests-watch-zzv98,SelfLink:/api/v1/namespaces/e2e-tests-watch-zzv98/configmaps/e2e-watch-test-label-changed,UID:4fe188fe-91c7-11e9-bf44-fa6f350b29f0,ResourceVersion:106055,Generation:0,CreationTimestamp:2019-06-18 12:48:03 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: label-changed-and-restored,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Data:map[string]string{},BinaryData:map[string][]byte{},}
+Jun 18 12:48:03.724: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-label-changed,GenerateName:,Namespace:e2e-tests-watch-zzv98,SelfLink:/api/v1/namespaces/e2e-tests-watch-zzv98/configmaps/e2e-watch-test-label-changed,UID:4fe188fe-91c7-11e9-bf44-fa6f350b29f0,ResourceVersion:106056,Generation:0,CreationTimestamp:2019-06-18 12:48:03 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: label-changed-and-restored,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Data:map[string]string{mutation: 1,},BinaryData:map[string][]byte{},}
+Jun 18 12:48:03.724: INFO: Got : DELETED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-label-changed,GenerateName:,Namespace:e2e-tests-watch-zzv98,SelfLink:/api/v1/namespaces/e2e-tests-watch-zzv98/configmaps/e2e-watch-test-label-changed,UID:4fe188fe-91c7-11e9-bf44-fa6f350b29f0,ResourceVersion:106057,Generation:0,CreationTimestamp:2019-06-18 12:48:03 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: label-changed-and-restored,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Data:map[string]string{mutation: 1,},BinaryData:map[string][]byte{},}
 STEP: modifying the configmap a second time
 STEP: Expecting not to observe a notification because the object no longer meets the selector's requirements
 STEP: changing the label value of the configmap back
 STEP: modifying the configmap a third time
 STEP: deleting the configmap
 STEP: Expecting to observe an add notification for the watched object when the label value was restored
-May 13 19:52:06.765: INFO: Got : ADDED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-label-changed,GenerateName:,Namespace:e2e-tests-watch-qg8j4,SelfLink:/api/v1/namespaces/e2e-tests-watch-qg8j4/configmaps/e2e-watch-test-label-changed,UID:9045931d-75b8-11e9-b786-da20024d205c,ResourceVersion:49208,Generation:0,CreationTimestamp:2019-05-13 19:51:56 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: label-changed-and-restored,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},}
-May 13 19:52:06.765: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-label-changed,GenerateName:,Namespace:e2e-tests-watch-qg8j4,SelfLink:/api/v1/namespaces/e2e-tests-watch-qg8j4/configmaps/e2e-watch-test-label-changed,UID:9045931d-75b8-11e9-b786-da20024d205c,ResourceVersion:49209,Generation:0,CreationTimestamp:2019-05-13 19:51:56 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: label-changed-and-restored,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Data:map[string]string{mutation: 3,},BinaryData:map[string][]byte{},}
-May 13 19:52:06.766: INFO: Got : DELETED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-label-changed,GenerateName:,Namespace:e2e-tests-watch-qg8j4,SelfLink:/api/v1/namespaces/e2e-tests-watch-qg8j4/configmaps/e2e-watch-test-label-changed,UID:9045931d-75b8-11e9-b786-da20024d205c,ResourceVersion:49210,Generation:0,CreationTimestamp:2019-05-13 19:51:56 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: label-changed-and-restored,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Data:map[string]string{mutation: 3,},BinaryData:map[string][]byte{},}
+Jun 18 12:48:13.842: INFO: Got : ADDED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-label-changed,GenerateName:,Namespace:e2e-tests-watch-zzv98,SelfLink:/api/v1/namespaces/e2e-tests-watch-zzv98/configmaps/e2e-watch-test-label-changed,UID:4fe188fe-91c7-11e9-bf44-fa6f350b29f0,ResourceVersion:106075,Generation:0,CreationTimestamp:2019-06-18 12:48:03 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: label-changed-and-restored,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},}
+Jun 18 12:48:13.842: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-label-changed,GenerateName:,Namespace:e2e-tests-watch-zzv98,SelfLink:/api/v1/namespaces/e2e-tests-watch-zzv98/configmaps/e2e-watch-test-label-changed,UID:4fe188fe-91c7-11e9-bf44-fa6f350b29f0,ResourceVersion:106076,Generation:0,CreationTimestamp:2019-06-18 12:48:03 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: label-changed-and-restored,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Data:map[string]string{mutation: 3,},BinaryData:map[string][]byte{},}
+Jun 18 12:48:13.842: INFO: Got : DELETED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-label-changed,GenerateName:,Namespace:e2e-tests-watch-zzv98,SelfLink:/api/v1/namespaces/e2e-tests-watch-zzv98/configmaps/e2e-watch-test-label-changed,UID:4fe188fe-91c7-11e9-bf44-fa6f350b29f0,ResourceVersion:106077,Generation:0,CreationTimestamp:2019-06-18 12:48:03 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: label-changed-and-restored,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Data:map[string]string{mutation: 3,},BinaryData:map[string][]byte{},}
 [AfterEach] [sig-api-machinery] Watchers
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
-May 13 19:52:06.766: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "e2e-tests-watch-qg8j4" for this suite.
-May 13 19:52:13.091: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-May 13 19:52:13.352: INFO: namespace: e2e-tests-watch-qg8j4, resource: bindings, ignored listing per whitelist
-May 13 19:52:13.382: INFO: namespace e2e-tests-watch-qg8j4 deletion completed in 6.608007725s
+Jun 18 12:48:13.843: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-tests-watch-zzv98" for this suite.
+Jun 18 12:48:21.985: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 12:48:22.388: INFO: namespace: e2e-tests-watch-zzv98, resource: bindings, ignored listing per whitelist
+Jun 18 12:48:22.546: INFO: namespace e2e-tests-watch-zzv98 deletion completed in 8.682920468s
 
-• [SLOW TEST:17.018 seconds]
+• [SLOW TEST:19.393 seconds]
 [sig-api-machinery] Watchers
 /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:22
   should observe an object deletion if it stops meeting the requirements of the selector [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 ------------------------------
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-storage] Subpath Atomic writer volumes 
-  should support subpaths with projected pod [Conformance]
+[sig-storage] Downward API volume 
+  should provide container's memory limit [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-[BeforeEach] [sig-storage] Subpath
+[BeforeEach] [sig-storage] Downward API volume
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
 STEP: Creating a kubernetes client
-May 13 19:52:13.382: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423
-STEP: Building a namespace api object, basename subpath
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-subpath-tz26c
+Jun 18 12:48:22.551: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
+STEP: Building a namespace api object, basename downward-api
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-downward-api-2rl55
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] Atomic writer volumes
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:38
-STEP: Setting up data
-[It] should support subpaths with projected pod [Conformance]
+[BeforeEach] [sig-storage] Downward API volume
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:39
+[It] should provide container's memory limit [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-STEP: Creating pod pod-subpath-test-projected-6rb5
-STEP: Creating a pod to test atomic-volume-subpath
-May 13 19:52:13.801: INFO: Waiting up to 5m0s for pod "pod-subpath-test-projected-6rb5" in namespace "e2e-tests-subpath-tz26c" to be "success or failure"
-May 13 19:52:13.815: INFO: Pod "pod-subpath-test-projected-6rb5": Phase="Pending", Reason="", readiness=false. Elapsed: 13.824851ms
-May 13 19:52:15.823: INFO: Pod "pod-subpath-test-projected-6rb5": Phase="Pending", Reason="", readiness=false. Elapsed: 2.022192867s
-May 13 19:52:17.843: INFO: Pod "pod-subpath-test-projected-6rb5": Phase="Running", Reason="", readiness=false. Elapsed: 4.042386873s
-May 13 19:52:19.920: INFO: Pod "pod-subpath-test-projected-6rb5": Phase="Running", Reason="", readiness=false. Elapsed: 6.118921612s
-May 13 19:52:21.928: INFO: Pod "pod-subpath-test-projected-6rb5": Phase="Running", Reason="", readiness=false. Elapsed: 8.127395905s
-May 13 19:52:23.937: INFO: Pod "pod-subpath-test-projected-6rb5": Phase="Running", Reason="", readiness=false. Elapsed: 10.1358199s
-May 13 19:52:25.944: INFO: Pod "pod-subpath-test-projected-6rb5": Phase="Running", Reason="", readiness=false. Elapsed: 12.143618048s
-May 13 19:52:27.969: INFO: Pod "pod-subpath-test-projected-6rb5": Phase="Running", Reason="", readiness=false. Elapsed: 14.168120401s
-May 13 19:52:29.978: INFO: Pod "pod-subpath-test-projected-6rb5": Phase="Running", Reason="", readiness=false. Elapsed: 16.177079967s
-May 13 19:52:32.020: INFO: Pod "pod-subpath-test-projected-6rb5": Phase="Running", Reason="", readiness=false. Elapsed: 18.218860916s
-May 13 19:52:34.028: INFO: Pod "pod-subpath-test-projected-6rb5": Phase="Running", Reason="", readiness=false. Elapsed: 20.227529823s
-May 13 19:52:36.037: INFO: Pod "pod-subpath-test-projected-6rb5": Phase="Running", Reason="", readiness=false. Elapsed: 22.235858008s
-May 13 19:52:38.059: INFO: Pod "pod-subpath-test-projected-6rb5": Phase="Succeeded", Reason="", readiness=false. Elapsed: 24.258126792s
+STEP: Creating a pod to test downward API volume plugin
+Jun 18 12:48:23.110: INFO: Waiting up to 5m0s for pod "downwardapi-volume-5b7b84af-91c7-11e9-bce2-ae54e022189f" in namespace "e2e-tests-downward-api-2rl55" to be "success or failure"
+Jun 18 12:48:23.125: INFO: Pod "downwardapi-volume-5b7b84af-91c7-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 14.28053ms
+Jun 18 12:48:25.157: INFO: Pod "downwardapi-volume-5b7b84af-91c7-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 2.046146533s
+Jun 18 12:48:27.171: INFO: Pod "downwardapi-volume-5b7b84af-91c7-11e9-bce2-ae54e022189f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.0607141s
 STEP: Saw pod success
-May 13 19:52:38.059: INFO: Pod "pod-subpath-test-projected-6rb5" satisfied condition "success or failure"
-May 13 19:52:38.067: INFO: Trying to get logs from node 10.170.219.177 pod pod-subpath-test-projected-6rb5 container test-container-subpath-projected-6rb5: 
+Jun 18 12:48:27.171: INFO: Pod "downwardapi-volume-5b7b84af-91c7-11e9-bce2-ae54e022189f" satisfied condition "success or failure"
+Jun 18 12:48:27.186: INFO: Trying to get logs from node 10.72.74.149 pod downwardapi-volume-5b7b84af-91c7-11e9-bce2-ae54e022189f container client-container: 
 STEP: delete the pod
-May 13 19:52:38.107: INFO: Waiting for pod pod-subpath-test-projected-6rb5 to disappear
-May 13 19:52:38.115: INFO: Pod pod-subpath-test-projected-6rb5 no longer exists
-STEP: Deleting pod pod-subpath-test-projected-6rb5
-May 13 19:52:38.115: INFO: Deleting pod "pod-subpath-test-projected-6rb5" in namespace "e2e-tests-subpath-tz26c"
-[AfterEach] [sig-storage] Subpath
+Jun 18 12:48:27.327: INFO: Waiting for pod downwardapi-volume-5b7b84af-91c7-11e9-bce2-ae54e022189f to disappear
+Jun 18 12:48:27.341: INFO: Pod downwardapi-volume-5b7b84af-91c7-11e9-bce2-ae54e022189f no longer exists
+[AfterEach] [sig-storage] Downward API volume
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
-May 13 19:52:38.123: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "e2e-tests-subpath-tz26c" for this suite.
-May 13 19:52:44.159: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-May 13 19:52:44.231: INFO: namespace: e2e-tests-subpath-tz26c, resource: bindings, ignored listing per whitelist
-May 13 19:52:44.407: INFO: namespace e2e-tests-subpath-tz26c deletion completed in 6.27630348s
+Jun 18 12:48:27.341: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-tests-downward-api-2rl55" for this suite.
+Jun 18 12:48:33.416: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 12:48:33.903: INFO: namespace: e2e-tests-downward-api-2rl55, resource: bindings, ignored listing per whitelist
+Jun 18 12:48:33.943: INFO: namespace e2e-tests-downward-api-2rl55 deletion completed in 6.577511248s
 
-• [SLOW TEST:31.025 seconds]
-[sig-storage] Subpath
-/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:22
-  Atomic writer volumes
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:34
-    should support subpaths with projected pod [Conformance]
-    /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
+• [SLOW TEST:11.393 seconds]
+[sig-storage] Downward API volume
+/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:34
+  should provide container's memory limit [NodeConformance] [Conformance]
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 ------------------------------
-SSSSSSSSSSSSSSS
+SSSSSSSSSS
 ------------------------------
-[sig-cli] Kubectl client [k8s.io] Kubectl label 
-  should update the label on a resource  [Conformance]
+[sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] 
+  should perform canary updates and phased rolling updates of template modifications [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-[BeforeEach] [sig-cli] Kubectl client
+[BeforeEach] [sig-apps] StatefulSet
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
 STEP: Creating a kubernetes client
-May 13 19:52:44.407: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423
-STEP: Building a namespace api object, basename kubectl
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-kubectl-79kkb
+Jun 18 12:48:33.944: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
+STEP: Building a namespace api object, basename statefulset
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-statefulset-z2gvj
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:243
-[BeforeEach] [k8s.io] Kubectl label
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1052
-STEP: creating the pod
-May 13 19:52:44.699: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 create -f - --namespace=e2e-tests-kubectl-79kkb'
-May 13 19:52:45.003: INFO: stderr: ""
-May 13 19:52:45.004: INFO: stdout: "pod/pause created\n"
-May 13 19:52:45.004: INFO: Waiting up to 5m0s for 1 pods to be running and ready: [pause]
-May 13 19:52:45.004: INFO: Waiting up to 5m0s for pod "pause" in namespace "e2e-tests-kubectl-79kkb" to be "running and ready"
-May 13 19:52:45.014: INFO: Pod "pause": Phase="Pending", Reason="", readiness=false. Elapsed: 10.512575ms
-May 13 19:52:47.023: INFO: Pod "pause": Phase="Running", Reason="", readiness=true. Elapsed: 2.01910176s
-May 13 19:52:47.023: INFO: Pod "pause" satisfied condition "running and ready"
-May 13 19:52:47.023: INFO: Wanted all 1 pods to be running and ready. Result: true. Pods: [pause]
-[It] should update the label on a resource  [Conformance]
+[BeforeEach] [sig-apps] StatefulSet
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:59
+[BeforeEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic]
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:74
+STEP: Creating service test in namespace e2e-tests-statefulset-z2gvj
+[It] should perform canary updates and phased rolling updates of template modifications [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-STEP: adding the label testing-label with value testing-label-value to a pod
-May 13 19:52:47.023: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 label pods pause testing-label=testing-label-value --namespace=e2e-tests-kubectl-79kkb'
-May 13 19:52:47.121: INFO: stderr: ""
-May 13 19:52:47.121: INFO: stdout: "pod/pause labeled\n"
-STEP: verifying the pod has the label testing-label with the value testing-label-value
-May 13 19:52:47.121: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 get pod pause -L testing-label --namespace=e2e-tests-kubectl-79kkb'
-May 13 19:52:47.228: INFO: stderr: ""
-May 13 19:52:47.228: INFO: stdout: "NAME    READY   STATUS    RESTARTS   AGE   TESTING-LABEL\npause   1/1     Running   0          3s    testing-label-value\n"
-STEP: removing the label testing-label of a pod
-May 13 19:52:47.228: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 label pods pause testing-label- --namespace=e2e-tests-kubectl-79kkb'
-May 13 19:52:47.336: INFO: stderr: ""
-May 13 19:52:47.336: INFO: stdout: "pod/pause labeled\n"
-STEP: verifying the pod doesn't have the label testing-label
-May 13 19:52:47.336: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 get pod pause -L testing-label --namespace=e2e-tests-kubectl-79kkb'
-May 13 19:52:47.431: INFO: stderr: ""
-May 13 19:52:47.431: INFO: stdout: "NAME    READY   STATUS    RESTARTS   AGE   TESTING-LABEL\npause   1/1     Running   0          3s    \n"
-[AfterEach] [k8s.io] Kubectl label
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1059
-STEP: using delete to clean up resources
-May 13 19:52:47.431: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 delete --grace-period=0 --force -f - --namespace=e2e-tests-kubectl-79kkb'
-May 13 19:52:47.553: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n"
-May 13 19:52:47.553: INFO: stdout: "pod \"pause\" force deleted\n"
-May 13 19:52:47.553: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 get rc,svc -l name=pause --no-headers --namespace=e2e-tests-kubectl-79kkb'
-May 13 19:52:47.658: INFO: stderr: "No resources found.\n"
-May 13 19:52:47.658: INFO: stdout: ""
-May 13 19:52:47.658: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 get pods -l name=pause --namespace=e2e-tests-kubectl-79kkb -o go-template={{ range .items }}{{ if not .metadata.deletionTimestamp }}{{ .metadata.name }}{{ "\n" }}{{ end }}{{ end }}'
-May 13 19:52:47.760: INFO: stderr: ""
-May 13 19:52:47.760: INFO: stdout: ""
-[AfterEach] [sig-cli] Kubectl client
+STEP: Creating a new StaefulSet
+Jun 18 12:48:34.457: INFO: Found 0 stateful pods, waiting for 3
+Jun 18 12:48:44.491: INFO: Waiting for pod ss2-0 to enter Running - Ready=true, currently Running - Ready=true
+Jun 18 12:48:44.491: INFO: Waiting for pod ss2-1 to enter Running - Ready=true, currently Running - Ready=true
+Jun 18 12:48:44.491: INFO: Waiting for pod ss2-2 to enter Running - Ready=true, currently Running - Ready=true
+STEP: Updating stateful set template: update image from docker.io/library/nginx:1.14-alpine to docker.io/library/nginx:1.15-alpine
+Jun 18 12:48:44.580: INFO: Updating stateful set ss2
+STEP: Creating a new revision
+STEP: Not applying an update when the partition is greater than the number of replicas
+STEP: Performing a canary update
+Jun 18 12:48:54.696: INFO: Updating stateful set ss2
+Jun 18 12:48:54.725: INFO: Waiting for Pod e2e-tests-statefulset-z2gvj/ss2-2 to have revision ss2-c79899b9 update revision ss2-787997d666
+STEP: Restoring Pods to the correct revision when they are deleted
+Jun 18 12:49:04.870: INFO: Found 2 stateful pods, waiting for 3
+Jun 18 12:49:14.903: INFO: Waiting for pod ss2-0 to enter Running - Ready=true, currently Running - Ready=true
+Jun 18 12:49:14.903: INFO: Waiting for pod ss2-1 to enter Running - Ready=true, currently Running - Ready=true
+Jun 18 12:49:14.903: INFO: Waiting for pod ss2-2 to enter Running - Ready=true, currently Running - Ready=true
+STEP: Performing a phased rolling update
+Jun 18 12:49:16.257: INFO: Updating stateful set ss2
+Jun 18 12:49:16.304: INFO: Waiting for Pod e2e-tests-statefulset-z2gvj/ss2-1 to have revision ss2-c79899b9 update revision ss2-787997d666
+Jun 18 12:49:26.391: INFO: Updating stateful set ss2
+Jun 18 12:49:26.426: INFO: Waiting for StatefulSet e2e-tests-statefulset-z2gvj/ss2 to complete update
+Jun 18 12:49:26.426: INFO: Waiting for Pod e2e-tests-statefulset-z2gvj/ss2-0 to have revision ss2-c79899b9 update revision ss2-787997d666
+Jun 18 12:49:36.478: INFO: Waiting for StatefulSet e2e-tests-statefulset-z2gvj/ss2 to complete update
+Jun 18 12:49:36.479: INFO: Waiting for Pod e2e-tests-statefulset-z2gvj/ss2-0 to have revision ss2-c79899b9 update revision ss2-787997d666
+[AfterEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic]
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:85
+Jun 18 12:49:46.460: INFO: Deleting all statefulset in ns e2e-tests-statefulset-z2gvj
+Jun 18 12:49:46.491: INFO: Scaling statefulset ss2 to 0
+Jun 18 12:50:06.559: INFO: Waiting for statefulset status.replicas updated to 0
+Jun 18 12:50:06.573: INFO: Deleting statefulset ss2
+[AfterEach] [sig-apps] StatefulSet
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
-May 13 19:52:47.760: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "e2e-tests-kubectl-79kkb" for this suite.
-May 13 19:52:53.797: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-May 13 19:52:53.941: INFO: namespace: e2e-tests-kubectl-79kkb, resource: bindings, ignored listing per whitelist
-May 13 19:52:54.062: INFO: namespace e2e-tests-kubectl-79kkb deletion completed in 6.292229533s
+Jun 18 12:50:07.379: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-tests-statefulset-z2gvj" for this suite.
+Jun 18 12:50:15.452: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 12:50:15.712: INFO: namespace: e2e-tests-statefulset-z2gvj, resource: bindings, ignored listing per whitelist
+Jun 18 12:50:16.011: INFO: namespace e2e-tests-statefulset-z2gvj deletion completed in 8.611406983s
 
-• [SLOW TEST:9.654 seconds]
-[sig-cli] Kubectl client
-/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:22
-  [k8s.io] Kubectl label
+• [SLOW TEST:102.067 seconds]
+[sig-apps] StatefulSet
+/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22
+  [k8s.io] Basic StatefulSet functionality [StatefulSetBasic]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694
-    should update the label on a resource  [Conformance]
+    should perform canary updates and phased rolling updates of template modifications [Conformance]
     /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 ------------------------------
-SSSS
-------------------------------
-[sig-storage] Projected secret 
-  should be consumable from pods in volume with mappings [NodeConformance] [Conformance]
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-[BeforeEach] [sig-storage] Projected secret
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
-STEP: Creating a kubernetes client
-May 13 19:52:54.062: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423
-STEP: Building a namespace api object, basename projected
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-projected-qzhjj
-STEP: Waiting for a default service account to be provisioned in namespace
-[It] should be consumable from pods in volume with mappings [NodeConformance] [Conformance]
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-STEP: Creating projection with secret that has name projected-secret-test-map-b2a76ecb-75b8-11e9-8f67-2632f168be36
-STEP: Creating a pod to test consume secrets
-May 13 19:52:54.444: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-b2b71d11-75b8-11e9-8f67-2632f168be36" in namespace "e2e-tests-projected-qzhjj" to be "success or failure"
-May 13 19:52:54.452: INFO: Pod "pod-projected-secrets-b2b71d11-75b8-11e9-8f67-2632f168be36": Phase="Pending", Reason="", readiness=false. Elapsed: 8.076596ms
-May 13 19:52:56.462: INFO: Pod "pod-projected-secrets-b2b71d11-75b8-11e9-8f67-2632f168be36": Phase="Running", Reason="", readiness=true. Elapsed: 2.017654026s
-May 13 19:52:58.470: INFO: Pod "pod-projected-secrets-b2b71d11-75b8-11e9-8f67-2632f168be36": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.026169653s
-STEP: Saw pod success
-May 13 19:52:58.470: INFO: Pod "pod-projected-secrets-b2b71d11-75b8-11e9-8f67-2632f168be36" satisfied condition "success or failure"
-May 13 19:52:58.478: INFO: Trying to get logs from node 10.170.219.177 pod pod-projected-secrets-b2b71d11-75b8-11e9-8f67-2632f168be36 container projected-secret-volume-test: 
-STEP: delete the pod
-May 13 19:52:58.520: INFO: Waiting for pod pod-projected-secrets-b2b71d11-75b8-11e9-8f67-2632f168be36 to disappear
-May 13 19:52:58.529: INFO: Pod pod-projected-secrets-b2b71d11-75b8-11e9-8f67-2632f168be36 no longer exists
-[AfterEach] [sig-storage] Projected secret
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
-May 13 19:52:58.529: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "e2e-tests-projected-qzhjj" for this suite.
-May 13 19:53:04.565: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-May 13 19:53:04.702: INFO: namespace: e2e-tests-projected-qzhjj, resource: bindings, ignored listing per whitelist
-May 13 19:53:04.848: INFO: namespace e2e-tests-projected-qzhjj deletion completed in 6.311259505s
-
-• [SLOW TEST:10.786 seconds]
-[sig-storage] Projected secret
-/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_secret.go:34
-  should be consumable from pods in volume with mappings [NodeConformance] [Conformance]
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-------------------------------
-SSSSSSSSSS
-------------------------------
-[sig-cli] Kubectl client [k8s.io] Kubectl run job 
-  should create a job from an image when restart is OnFailure  [Conformance]
+[sig-cli] Kubectl client [k8s.io] Kubectl api-versions 
+  should check if v1 is in available api versions  [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 [BeforeEach] [sig-cli] Kubectl client
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
 STEP: Creating a kubernetes client
-May 13 19:53:04.848: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423
+Jun 18 12:50:16.013: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
 STEP: Building a namespace api object, basename kubectl
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-kubectl-64d7k
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-kubectl-hj4qw
 STEP: Waiting for a default service account to be provisioned in namespace
 [BeforeEach] [sig-cli] Kubectl client
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:243
-[BeforeEach] [k8s.io] Kubectl run job
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1454
-[It] should create a job from an image when restart is OnFailure  [Conformance]
+[It] should check if v1 is in available api versions  [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-STEP: running the image docker.io/library/nginx:1.14-alpine
-May 13 19:53:05.223: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 run e2e-test-nginx-job --restart=OnFailure --generator=job/v1 --image=docker.io/library/nginx:1.14-alpine --namespace=e2e-tests-kubectl-64d7k'
-May 13 19:53:05.359: INFO: stderr: "kubectl run --generator=job/v1 is DEPRECATED and will be removed in a future version. Use kubectl run --generator=run-pod/v1 or kubectl create instead.\n"
-May 13 19:53:05.359: INFO: stdout: "job.batch/e2e-test-nginx-job created\n"
-STEP: verifying the job e2e-test-nginx-job was created
-[AfterEach] [k8s.io] Kubectl run job
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1459
-May 13 19:53:05.364: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 delete jobs e2e-test-nginx-job --namespace=e2e-tests-kubectl-64d7k'
-May 13 19:53:05.468: INFO: stderr: ""
-May 13 19:53:05.468: INFO: stdout: "job.batch \"e2e-test-nginx-job\" deleted\n"
+STEP: validating api versions
+Jun 18 12:50:16.462: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 api-versions'
+Jun 18 12:50:16.598: INFO: stderr: ""
+Jun 18 12:50:16.598: INFO: stdout: "admissionregistration.k8s.io/v1alpha1\nadmissionregistration.k8s.io/v1beta1\napiextensions.k8s.io/v1beta1\napiregistration.k8s.io/v1\napiregistration.k8s.io/v1beta1\napps/v1\napps/v1beta1\napps/v1beta2\nauthentication.k8s.io/v1\nauthentication.k8s.io/v1beta1\nauthorization.k8s.io/v1\nauthorization.k8s.io/v1beta1\nautoscaling/v1\nautoscaling/v2beta1\nautoscaling/v2beta2\nbatch/v1\nbatch/v1beta1\nbatch/v2alpha1\ncertificates.k8s.io/v1beta1\ncoordination.k8s.io/v1beta1\nevents.k8s.io/v1beta1\nextensions/v1beta1\nmetrics.k8s.io/v1beta1\nnetworking.k8s.io/v1\npolicy/v1beta1\nrbac.authorization.k8s.io/v1\nrbac.authorization.k8s.io/v1beta1\nscheduling.k8s.io/v1beta1\nstorage.k8s.io/v1\nstorage.k8s.io/v1beta1\nv1\n"
 [AfterEach] [sig-cli] Kubectl client
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
-May 13 19:53:05.468: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "e2e-tests-kubectl-64d7k" for this suite.
-May 13 19:53:27.504: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-May 13 19:53:27.635: INFO: namespace: e2e-tests-kubectl-64d7k, resource: bindings, ignored listing per whitelist
-May 13 19:53:27.886: INFO: namespace e2e-tests-kubectl-64d7k deletion completed in 22.411192476s
+Jun 18 12:50:16.598: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-tests-kubectl-hj4qw" for this suite.
+Jun 18 12:50:22.662: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 12:50:22.799: INFO: namespace: e2e-tests-kubectl-hj4qw, resource: bindings, ignored listing per whitelist
+Jun 18 12:50:23.242: INFO: namespace e2e-tests-kubectl-hj4qw deletion completed in 6.628130348s
 
-• [SLOW TEST:23.038 seconds]
+• [SLOW TEST:7.230 seconds]
 [sig-cli] Kubectl client
 /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:22
-  [k8s.io] Kubectl run job
+  [k8s.io] Kubectl api-versions
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694
-    should create a job from an image when restart is OnFailure  [Conformance]
+    should check if v1 is in available api versions  [Conformance]
     /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 ------------------------------
-SS
+SSSSSSS
 ------------------------------
-[sig-storage] Projected downwardAPI 
-  should provide podname only [NodeConformance] [Conformance]
+[sig-auth] ServiceAccounts 
+  should mount an API token into pods  [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-[BeforeEach] [sig-storage] Projected downwardAPI
+[BeforeEach] [sig-auth] ServiceAccounts
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
 STEP: Creating a kubernetes client
-May 13 19:53:27.886: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423
-STEP: Building a namespace api object, basename projected
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-projected-4mnk5
+Jun 18 12:50:23.243: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
+STEP: Building a namespace api object, basename svcaccounts
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-svcaccounts-w8bdc
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-storage] Projected downwardAPI
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:39
-[It] should provide podname only [NodeConformance] [Conformance]
+[It] should mount an API token into pods  [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-STEP: Creating a pod to test downward API volume plugin
-May 13 19:53:28.260: INFO: Waiting up to 5m0s for pod "downwardapi-volume-c6df10d5-75b8-11e9-8f67-2632f168be36" in namespace "e2e-tests-projected-4mnk5" to be "success or failure"
-May 13 19:53:28.268: INFO: Pod "downwardapi-volume-c6df10d5-75b8-11e9-8f67-2632f168be36": Phase="Pending", Reason="", readiness=false. Elapsed: 7.391072ms
-May 13 19:53:30.276: INFO: Pod "downwardapi-volume-c6df10d5-75b8-11e9-8f67-2632f168be36": Phase="Pending", Reason="", readiness=false. Elapsed: 2.01561409s
-May 13 19:53:32.284: INFO: Pod "downwardapi-volume-c6df10d5-75b8-11e9-8f67-2632f168be36": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.023976113s
+STEP: getting the auto-created API token
+STEP: Creating a pod to test consume service account token
+Jun 18 12:50:24.254: INFO: Waiting up to 5m0s for pod "pod-service-account-a3b08a3b-91c7-11e9-bce2-ae54e022189f-hf4nf" in namespace "e2e-tests-svcaccounts-w8bdc" to be "success or failure"
+Jun 18 12:50:24.268: INFO: Pod "pod-service-account-a3b08a3b-91c7-11e9-bce2-ae54e022189f-hf4nf": Phase="Pending", Reason="", readiness=false. Elapsed: 13.342654ms
+Jun 18 12:50:26.283: INFO: Pod "pod-service-account-a3b08a3b-91c7-11e9-bce2-ae54e022189f-hf4nf": Phase="Pending", Reason="", readiness=false. Elapsed: 2.028326579s
+Jun 18 12:50:28.304: INFO: Pod "pod-service-account-a3b08a3b-91c7-11e9-bce2-ae54e022189f-hf4nf": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.050114865s
 STEP: Saw pod success
-May 13 19:53:32.285: INFO: Pod "downwardapi-volume-c6df10d5-75b8-11e9-8f67-2632f168be36" satisfied condition "success or failure"
-May 13 19:53:32.292: INFO: Trying to get logs from node 10.170.219.153 pod downwardapi-volume-c6df10d5-75b8-11e9-8f67-2632f168be36 container client-container: 
+Jun 18 12:50:28.304: INFO: Pod "pod-service-account-a3b08a3b-91c7-11e9-bce2-ae54e022189f-hf4nf" satisfied condition "success or failure"
+Jun 18 12:50:28.319: INFO: Trying to get logs from node 10.72.74.144 pod pod-service-account-a3b08a3b-91c7-11e9-bce2-ae54e022189f-hf4nf container token-test: 
 STEP: delete the pod
-May 13 19:53:32.330: INFO: Waiting for pod downwardapi-volume-c6df10d5-75b8-11e9-8f67-2632f168be36 to disappear
-May 13 19:53:32.337: INFO: Pod downwardapi-volume-c6df10d5-75b8-11e9-8f67-2632f168be36 no longer exists
-[AfterEach] [sig-storage] Projected downwardAPI
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
-May 13 19:53:32.337: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "e2e-tests-projected-4mnk5" for this suite.
-May 13 19:53:38.372: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-May 13 19:53:38.501: INFO: namespace: e2e-tests-projected-4mnk5, resource: bindings, ignored listing per whitelist
-May 13 19:53:38.650: INFO: namespace e2e-tests-projected-4mnk5 deletion completed in 6.3052817s
-
-• [SLOW TEST:10.764 seconds]
-[sig-storage] Projected downwardAPI
-/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:33
-  should provide podname only [NodeConformance] [Conformance]
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-------------------------------
-SSS
-------------------------------
-[sig-apps] Daemon set [Serial] 
-  should run and stop simple daemon [Conformance]
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-[BeforeEach] [sig-apps] Daemon set [Serial]
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
-STEP: Creating a kubernetes client
-May 13 19:53:38.650: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423
-STEP: Building a namespace api object, basename daemonsets
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-daemonsets-k56rn
-STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-apps] Daemon set [Serial]
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:102
-[It] should run and stop simple daemon [Conformance]
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-STEP: Creating simple DaemonSet "daemon-set"
-STEP: Check that daemon pods launch on every node of the cluster.
-May 13 19:53:38.978: INFO: Number of nodes with available pods: 0
-May 13 19:53:38.978: INFO: Node 10.170.219.151 is running more than one daemon pod
-May 13 19:53:39.995: INFO: Number of nodes with available pods: 0
-May 13 19:53:39.995: INFO: Node 10.170.219.151 is running more than one daemon pod
-May 13 19:53:40.995: INFO: Number of nodes with available pods: 3
-May 13 19:53:40.995: INFO: Number of running nodes: 3, number of available pods: 3
-STEP: Stop a daemon pod, check that the daemon pod is revived.
-May 13 19:53:41.038: INFO: Number of nodes with available pods: 2
-May 13 19:53:41.038: INFO: Node 10.170.219.177 is running more than one daemon pod
-May 13 19:53:42.120: INFO: Number of nodes with available pods: 2
-May 13 19:53:42.120: INFO: Node 10.170.219.177 is running more than one daemon pod
-May 13 19:53:43.059: INFO: Number of nodes with available pods: 2
-May 13 19:53:43.059: INFO: Node 10.170.219.177 is running more than one daemon pod
-May 13 19:53:44.058: INFO: Number of nodes with available pods: 2
-May 13 19:53:44.058: INFO: Node 10.170.219.177 is running more than one daemon pod
-May 13 19:53:45.120: INFO: Number of nodes with available pods: 2
-May 13 19:53:45.120: INFO: Node 10.170.219.177 is running more than one daemon pod
-May 13 19:53:46.056: INFO: Number of nodes with available pods: 2
-May 13 19:53:46.056: INFO: Node 10.170.219.177 is running more than one daemon pod
-May 13 19:53:47.057: INFO: Number of nodes with available pods: 2
-May 13 19:53:47.057: INFO: Node 10.170.219.177 is running more than one daemon pod
-May 13 19:53:48.079: INFO: Number of nodes with available pods: 2
-May 13 19:53:48.079: INFO: Node 10.170.219.177 is running more than one daemon pod
-May 13 19:53:49.056: INFO: Number of nodes with available pods: 2
-May 13 19:53:49.056: INFO: Node 10.170.219.177 is running more than one daemon pod
-May 13 19:53:50.057: INFO: Number of nodes with available pods: 2
-May 13 19:53:50.057: INFO: Node 10.170.219.177 is running more than one daemon pod
-May 13 19:53:51.129: INFO: Number of nodes with available pods: 2
-May 13 19:53:51.129: INFO: Node 10.170.219.177 is running more than one daemon pod
-May 13 19:53:52.056: INFO: Number of nodes with available pods: 2
-May 13 19:53:52.056: INFO: Node 10.170.219.177 is running more than one daemon pod
-May 13 19:53:53.055: INFO: Number of nodes with available pods: 2
-May 13 19:53:53.055: INFO: Node 10.170.219.177 is running more than one daemon pod
-May 13 19:53:54.057: INFO: Number of nodes with available pods: 2
-May 13 19:53:54.057: INFO: Node 10.170.219.177 is running more than one daemon pod
-May 13 19:53:55.057: INFO: Number of nodes with available pods: 2
-May 13 19:53:55.057: INFO: Node 10.170.219.177 is running more than one daemon pod
-May 13 19:53:56.067: INFO: Number of nodes with available pods: 2
-May 13 19:53:56.067: INFO: Node 10.170.219.177 is running more than one daemon pod
-May 13 19:53:57.054: INFO: Number of nodes with available pods: 2
-May 13 19:53:57.055: INFO: Node 10.170.219.177 is running more than one daemon pod
-May 13 19:53:58.061: INFO: Number of nodes with available pods: 2
-May 13 19:53:58.061: INFO: Node 10.170.219.177 is running more than one daemon pod
-May 13 19:53:59.128: INFO: Number of nodes with available pods: 2
-May 13 19:53:59.128: INFO: Node 10.170.219.177 is running more than one daemon pod
-May 13 19:54:00.056: INFO: Number of nodes with available pods: 2
-May 13 19:54:00.056: INFO: Node 10.170.219.177 is running more than one daemon pod
-May 13 19:54:01.056: INFO: Number of nodes with available pods: 2
-May 13 19:54:01.056: INFO: Node 10.170.219.177 is running more than one daemon pod
-May 13 19:54:02.055: INFO: Number of nodes with available pods: 2
-May 13 19:54:02.056: INFO: Node 10.170.219.177 is running more than one daemon pod
-May 13 19:54:03.059: INFO: Number of nodes with available pods: 2
-May 13 19:54:03.059: INFO: Node 10.170.219.177 is running more than one daemon pod
-May 13 19:54:04.056: INFO: Number of nodes with available pods: 2
-May 13 19:54:04.056: INFO: Node 10.170.219.177 is running more than one daemon pod
-May 13 19:54:05.056: INFO: Number of nodes with available pods: 2
-May 13 19:54:05.056: INFO: Node 10.170.219.177 is running more than one daemon pod
-May 13 19:54:06.141: INFO: Number of nodes with available pods: 2
-May 13 19:54:06.141: INFO: Node 10.170.219.177 is running more than one daemon pod
-May 13 19:54:07.060: INFO: Number of nodes with available pods: 2
-May 13 19:54:07.060: INFO: Node 10.170.219.177 is running more than one daemon pod
-May 13 19:54:08.120: INFO: Number of nodes with available pods: 2
-May 13 19:54:08.120: INFO: Node 10.170.219.177 is running more than one daemon pod
-May 13 19:54:09.055: INFO: Number of nodes with available pods: 2
-May 13 19:54:09.055: INFO: Node 10.170.219.177 is running more than one daemon pod
-May 13 19:54:10.056: INFO: Number of nodes with available pods: 2
-May 13 19:54:10.056: INFO: Node 10.170.219.177 is running more than one daemon pod
-May 13 19:54:11.060: INFO: Number of nodes with available pods: 2
-May 13 19:54:11.060: INFO: Node 10.170.219.177 is running more than one daemon pod
-May 13 19:54:12.532: INFO: Number of nodes with available pods: 2
-May 13 19:54:12.532: INFO: Node 10.170.219.177 is running more than one daemon pod
-May 13 19:54:13.055: INFO: Number of nodes with available pods: 2
-May 13 19:54:13.055: INFO: Node 10.170.219.177 is running more than one daemon pod
-May 13 19:54:14.055: INFO: Number of nodes with available pods: 2
-May 13 19:54:14.055: INFO: Node 10.170.219.177 is running more than one daemon pod
-May 13 19:54:15.060: INFO: Number of nodes with available pods: 2
-May 13 19:54:15.060: INFO: Node 10.170.219.177 is running more than one daemon pod
-May 13 19:54:16.131: INFO: Number of nodes with available pods: 2
-May 13 19:54:16.131: INFO: Node 10.170.219.177 is running more than one daemon pod
-May 13 19:54:17.071: INFO: Number of nodes with available pods: 2
-May 13 19:54:17.071: INFO: Node 10.170.219.177 is running more than one daemon pod
-May 13 19:54:18.055: INFO: Number of nodes with available pods: 2
-May 13 19:54:18.055: INFO: Node 10.170.219.177 is running more than one daemon pod
-May 13 19:54:19.058: INFO: Number of nodes with available pods: 2
-May 13 19:54:19.058: INFO: Node 10.170.219.177 is running more than one daemon pod
-May 13 19:54:20.055: INFO: Number of nodes with available pods: 2
-May 13 19:54:20.055: INFO: Node 10.170.219.177 is running more than one daemon pod
-May 13 19:54:21.056: INFO: Number of nodes with available pods: 2
-May 13 19:54:21.056: INFO: Node 10.170.219.177 is running more than one daemon pod
-May 13 19:54:22.055: INFO: Number of nodes with available pods: 2
-May 13 19:54:22.055: INFO: Node 10.170.219.177 is running more than one daemon pod
-May 13 19:54:23.060: INFO: Number of nodes with available pods: 3
-May 13 19:54:23.060: INFO: Number of running nodes: 3, number of available pods: 3
-[AfterEach] [sig-apps] Daemon set [Serial]
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:68
-STEP: Deleting DaemonSet "daemon-set"
-STEP: deleting DaemonSet.extensions daemon-set in namespace e2e-tests-daemonsets-k56rn, will wait for the garbage collector to delete the pods
-May 13 19:54:23.220: INFO: Deleting DaemonSet.extensions daemon-set took: 97.210624ms
-May 13 19:54:23.220: INFO: Terminating DaemonSet.extensions daemon-set pods took: 76.045µs
-May 13 19:55:06.343: INFO: Number of nodes with available pods: 0
-May 13 19:55:06.343: INFO: Number of running nodes: 0, number of available pods: 0
-May 13 19:55:06.350: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"selfLink":"/apis/apps/v1/namespaces/e2e-tests-daemonsets-k56rn/daemonsets","resourceVersion":"49842"},"items":null}
-
-May 13 19:55:06.360: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"selfLink":"/api/v1/namespaces/e2e-tests-daemonsets-k56rn/pods","resourceVersion":"49842"},"items":null}
-
-[AfterEach] [sig-apps] Daemon set [Serial]
+Jun 18 12:50:28.464: INFO: Waiting for pod pod-service-account-a3b08a3b-91c7-11e9-bce2-ae54e022189f-hf4nf to disappear
+Jun 18 12:50:28.480: INFO: Pod pod-service-account-a3b08a3b-91c7-11e9-bce2-ae54e022189f-hf4nf no longer exists
+STEP: Creating a pod to test consume service account root CA
+Jun 18 12:50:28.498: INFO: Waiting up to 5m0s for pod "pod-service-account-a3b08a3b-91c7-11e9-bce2-ae54e022189f-lxlps" in namespace "e2e-tests-svcaccounts-w8bdc" to be "success or failure"
+Jun 18 12:50:28.512: INFO: Pod "pod-service-account-a3b08a3b-91c7-11e9-bce2-ae54e022189f-lxlps": Phase="Pending", Reason="", readiness=false. Elapsed: 14.127458ms
+Jun 18 12:50:30.546: INFO: Pod "pod-service-account-a3b08a3b-91c7-11e9-bce2-ae54e022189f-lxlps": Phase="Pending", Reason="", readiness=false. Elapsed: 2.047955845s
+Jun 18 12:50:32.562: INFO: Pod "pod-service-account-a3b08a3b-91c7-11e9-bce2-ae54e022189f-lxlps": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.064304511s
+STEP: Saw pod success
+Jun 18 12:50:32.562: INFO: Pod "pod-service-account-a3b08a3b-91c7-11e9-bce2-ae54e022189f-lxlps" satisfied condition "success or failure"
+Jun 18 12:50:32.577: INFO: Trying to get logs from node 10.72.74.144 pod pod-service-account-a3b08a3b-91c7-11e9-bce2-ae54e022189f-lxlps container root-ca-test: 
+STEP: delete the pod
+Jun 18 12:50:32.650: INFO: Waiting for pod pod-service-account-a3b08a3b-91c7-11e9-bce2-ae54e022189f-lxlps to disappear
+Jun 18 12:50:32.664: INFO: Pod pod-service-account-a3b08a3b-91c7-11e9-bce2-ae54e022189f-lxlps no longer exists
+STEP: Creating a pod to test consume service account namespace
+Jun 18 12:50:32.681: INFO: Waiting up to 5m0s for pod "pod-service-account-a3b08a3b-91c7-11e9-bce2-ae54e022189f-7n5hk" in namespace "e2e-tests-svcaccounts-w8bdc" to be "success or failure"
+Jun 18 12:50:32.697: INFO: Pod "pod-service-account-a3b08a3b-91c7-11e9-bce2-ae54e022189f-7n5hk": Phase="Pending", Reason="", readiness=false. Elapsed: 16.105875ms
+Jun 18 12:50:34.873: INFO: Pod "pod-service-account-a3b08a3b-91c7-11e9-bce2-ae54e022189f-7n5hk": Phase="Pending", Reason="", readiness=false. Elapsed: 2.191852932s
+Jun 18 12:50:37.005: INFO: Pod "pod-service-account-a3b08a3b-91c7-11e9-bce2-ae54e022189f-7n5hk": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.32418989s
+STEP: Saw pod success
+Jun 18 12:50:37.005: INFO: Pod "pod-service-account-a3b08a3b-91c7-11e9-bce2-ae54e022189f-7n5hk" satisfied condition "success or failure"
+Jun 18 12:50:37.020: INFO: Trying to get logs from node 10.72.74.144 pod pod-service-account-a3b08a3b-91c7-11e9-bce2-ae54e022189f-7n5hk container namespace-test: 
+STEP: delete the pod
+Jun 18 12:50:37.102: INFO: Waiting for pod pod-service-account-a3b08a3b-91c7-11e9-bce2-ae54e022189f-7n5hk to disappear
+Jun 18 12:50:37.115: INFO: Pod pod-service-account-a3b08a3b-91c7-11e9-bce2-ae54e022189f-7n5hk no longer exists
+[AfterEach] [sig-auth] ServiceAccounts
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
-May 13 19:55:06.384: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "e2e-tests-daemonsets-k56rn" for this suite.
-May 13 19:55:12.419: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-May 13 19:55:12.535: INFO: namespace: e2e-tests-daemonsets-k56rn, resource: bindings, ignored listing per whitelist
-May 13 19:55:12.738: INFO: namespace e2e-tests-daemonsets-k56rn deletion completed in 6.347801582s
+Jun 18 12:50:37.115: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-tests-svcaccounts-w8bdc" for this suite.
+Jun 18 12:50:45.190: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 12:50:45.726: INFO: namespace: e2e-tests-svcaccounts-w8bdc, resource: bindings, ignored listing per whitelist
+Jun 18 12:50:45.751: INFO: namespace e2e-tests-svcaccounts-w8bdc deletion completed in 8.612077285s
 
-• [SLOW TEST:94.088 seconds]
-[sig-apps] Daemon set [Serial]
-/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22
-  should run and stop simple daemon [Conformance]
+• [SLOW TEST:22.509 seconds]
+[sig-auth] ServiceAccounts
+/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/auth/framework.go:22
+  should mount an API token into pods  [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 ------------------------------
-SSS
+S
 ------------------------------
 [sig-apps] ReplicationController 
-  should serve a basic image on each replica with a public image  [Conformance]
+  should release no longer matching pods [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 [BeforeEach] [sig-apps] ReplicationController
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
 STEP: Creating a kubernetes client
-May 13 19:55:12.739: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423
+Jun 18 12:50:45.752: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
 STEP: Building a namespace api object, basename replication-controller
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-replication-controller-hh2rm
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-replication-controller-jtl5n
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should serve a basic image on each replica with a public image  [Conformance]
+[It] should release no longer matching pods [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-STEP: Creating replication controller my-hostname-basic-0551a4b0-75b9-11e9-8f67-2632f168be36
-May 13 19:55:13.030: INFO: Pod name my-hostname-basic-0551a4b0-75b9-11e9-8f67-2632f168be36: Found 0 pods out of 1
-May 13 19:55:18.051: INFO: Pod name my-hostname-basic-0551a4b0-75b9-11e9-8f67-2632f168be36: Found 1 pods out of 1
-May 13 19:55:18.051: INFO: Ensuring all pods for ReplicationController "my-hostname-basic-0551a4b0-75b9-11e9-8f67-2632f168be36" are running
-May 13 19:55:18.059: INFO: Pod "my-hostname-basic-0551a4b0-75b9-11e9-8f67-2632f168be36-bbkpf" is running (conditions: [{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2019-05-13 19:55:13 +0000 UTC Reason: Message:} {Type:Ready Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2019-05-13 19:55:14 +0000 UTC Reason: Message:} {Type:ContainersReady Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2019-05-13 19:55:14 +0000 UTC Reason: Message:} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2019-05-13 19:55:13 +0000 UTC Reason: Message:}])
-May 13 19:55:18.059: INFO: Trying to dial the pod
-May 13 19:55:23.141: INFO: Controller my-hostname-basic-0551a4b0-75b9-11e9-8f67-2632f168be36: Got expected result from replica 1 [my-hostname-basic-0551a4b0-75b9-11e9-8f67-2632f168be36-bbkpf]: "my-hostname-basic-0551a4b0-75b9-11e9-8f67-2632f168be36-bbkpf", 1 of 1 required successes so far
+STEP: Given a ReplicationController is created
+STEP: When the matched label of one of its pods change
+Jun 18 12:50:46.588: INFO: Pod name pod-release: Found 1 pods out of 1
+STEP: Then the pod is released
 [AfterEach] [sig-apps] ReplicationController
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
-May 13 19:55:23.141: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "e2e-tests-replication-controller-hh2rm" for this suite.
-May 13 19:55:29.188: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-May 13 19:55:29.460: INFO: namespace: e2e-tests-replication-controller-hh2rm, resource: bindings, ignored listing per whitelist
-May 13 19:55:29.488: INFO: namespace e2e-tests-replication-controller-hh2rm deletion completed in 6.339455626s
+Jun 18 12:50:47.742: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-tests-replication-controller-jtl5n" for this suite.
+Jun 18 12:50:53.811: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 12:50:54.227: INFO: namespace: e2e-tests-replication-controller-jtl5n, resource: bindings, ignored listing per whitelist
+Jun 18 12:50:54.355: INFO: namespace e2e-tests-replication-controller-jtl5n deletion completed in 6.592482624s
 
-• [SLOW TEST:16.749 seconds]
+• [SLOW TEST:8.603 seconds]
 [sig-apps] ReplicationController
 /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22
-  should serve a basic image on each replica with a public image  [Conformance]
+  should release no longer matching pods [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 ------------------------------
-SSS
+SS
 ------------------------------
-[sig-storage] Downward API volume 
-  should provide container's memory request [NodeConformance] [Conformance]
+[sig-storage] Projected secret 
+  should be consumable from pods in volume as non-root with defaultMode and fsGroup set [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-[BeforeEach] [sig-storage] Downward API volume
+[BeforeEach] [sig-storage] Projected secret
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
 STEP: Creating a kubernetes client
-May 13 19:55:29.490: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423
-STEP: Building a namespace api object, basename downward-api
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-downward-api-t7ppg
+Jun 18 12:50:54.355: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
+STEP: Building a namespace api object, basename projected
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-projected-2nc6n
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-storage] Downward API volume
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:39
-[It] should provide container's memory request [NodeConformance] [Conformance]
+[It] should be consumable from pods in volume as non-root with defaultMode and fsGroup set [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-STEP: Creating a pod to test downward API volume plugin
-May 13 19:55:29.780: INFO: Waiting up to 5m0s for pod "downwardapi-volume-0f4d79c4-75b9-11e9-8f67-2632f168be36" in namespace "e2e-tests-downward-api-t7ppg" to be "success or failure"
-May 13 19:55:29.788: INFO: Pod "downwardapi-volume-0f4d79c4-75b9-11e9-8f67-2632f168be36": Phase="Pending", Reason="", readiness=false. Elapsed: 7.332019ms
-May 13 19:55:31.797: INFO: Pod "downwardapi-volume-0f4d79c4-75b9-11e9-8f67-2632f168be36": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.016028159s
+STEP: Creating projection with secret that has name projected-secret-test-b5f65992-91c7-11e9-bce2-ae54e022189f
+STEP: Creating a pod to test consume secrets
+Jun 18 12:50:54.921: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-b5f857a9-91c7-11e9-bce2-ae54e022189f" in namespace "e2e-tests-projected-2nc6n" to be "success or failure"
+Jun 18 12:50:54.935: INFO: Pod "pod-projected-secrets-b5f857a9-91c7-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 13.614617ms
+Jun 18 12:50:56.950: INFO: Pod "pod-projected-secrets-b5f857a9-91c7-11e9-bce2-ae54e022189f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.028433459s
 STEP: Saw pod success
-May 13 19:55:31.797: INFO: Pod "downwardapi-volume-0f4d79c4-75b9-11e9-8f67-2632f168be36" satisfied condition "success or failure"
-May 13 19:55:31.805: INFO: Trying to get logs from node 10.170.219.153 pod downwardapi-volume-0f4d79c4-75b9-11e9-8f67-2632f168be36 container client-container: 
+Jun 18 12:50:56.950: INFO: Pod "pod-projected-secrets-b5f857a9-91c7-11e9-bce2-ae54e022189f" satisfied condition "success or failure"
+Jun 18 12:50:56.964: INFO: Trying to get logs from node 10.72.74.143 pod pod-projected-secrets-b5f857a9-91c7-11e9-bce2-ae54e022189f container projected-secret-volume-test: 
 STEP: delete the pod
-May 13 19:55:31.846: INFO: Waiting for pod downwardapi-volume-0f4d79c4-75b9-11e9-8f67-2632f168be36 to disappear
-May 13 19:55:31.853: INFO: Pod downwardapi-volume-0f4d79c4-75b9-11e9-8f67-2632f168be36 no longer exists
-[AfterEach] [sig-storage] Downward API volume
+Jun 18 12:50:57.044: INFO: Waiting for pod pod-projected-secrets-b5f857a9-91c7-11e9-bce2-ae54e022189f to disappear
+Jun 18 12:50:57.057: INFO: Pod pod-projected-secrets-b5f857a9-91c7-11e9-bce2-ae54e022189f no longer exists
+[AfterEach] [sig-storage] Projected secret
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
-May 13 19:55:31.853: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "e2e-tests-downward-api-t7ppg" for this suite.
-May 13 19:55:37.890: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-May 13 19:55:37.949: INFO: namespace: e2e-tests-downward-api-t7ppg, resource: bindings, ignored listing per whitelist
-May 13 19:55:38.132: INFO: namespace e2e-tests-downward-api-t7ppg deletion completed in 6.271911473s
+Jun 18 12:50:57.057: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-tests-projected-2nc6n" for this suite.
+Jun 18 12:51:03.152: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 12:51:04.220: INFO: namespace: e2e-tests-projected-2nc6n, resource: bindings, ignored listing per whitelist
+Jun 18 12:51:04.620: INFO: namespace e2e-tests-projected-2nc6n deletion completed in 7.542171892s
 
-• [SLOW TEST:8.642 seconds]
-[sig-storage] Downward API volume
-/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:34
-  should provide container's memory request [NodeConformance] [Conformance]
+• [SLOW TEST:10.265 seconds]
+[sig-storage] Projected secret
+/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_secret.go:34
+  should be consumable from pods in volume as non-root with defaultMode and fsGroup set [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 ------------------------------
-SSSSSSSSSSSSSSSSS
+SSSSSSSSS
 ------------------------------
 [k8s.io] Probing container 
-  with readiness probe that fails should never be ready and never restart [NodeConformance] [Conformance]
+  with readiness probe should not be ready before initial delay and never restart [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 [BeforeEach] [k8s.io] Probing container
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
 STEP: Creating a kubernetes client
-May 13 19:55:38.133: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423
+Jun 18 12:51:04.622: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
 STEP: Building a namespace api object, basename container-probe
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-container-probe-m4mnx
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-container-probe-5x46v
 STEP: Waiting for a default service account to be provisioned in namespace
 [BeforeEach] [k8s.io] Probing container
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/container_probe.go:48
-[It] with readiness probe that fails should never be ready and never restart [NodeConformance] [Conformance]
+[It] with readiness probe should not be ready before initial delay and never restart [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
+Jun 18 12:51:33.199: INFO: Container started at 2019-06-18 12:51:08 +0000 UTC, pod became ready at 2019-06-18 12:51:31 +0000 UTC
 [AfterEach] [k8s.io] Probing container
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
-May 13 19:56:38.417: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "e2e-tests-container-probe-m4mnx" for this suite.
-May 13 19:57:02.452: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-May 13 19:57:02.744: INFO: namespace: e2e-tests-container-probe-m4mnx, resource: bindings, ignored listing per whitelist
-May 13 19:57:02.744: INFO: namespace e2e-tests-container-probe-m4mnx deletion completed in 24.319474875s
+Jun 18 12:51:33.199: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-tests-container-probe-5x46v" for this suite.
+Jun 18 12:51:57.264: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 12:51:57.326: INFO: namespace: e2e-tests-container-probe-5x46v, resource: bindings, ignored listing per whitelist
+Jun 18 12:51:57.773: INFO: namespace e2e-tests-container-probe-5x46v deletion completed in 24.555655445s
 
-• [SLOW TEST:84.612 seconds]
+• [SLOW TEST:53.151 seconds]
 [k8s.io] Probing container
 /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694
-  with readiness probe that fails should never be ready and never restart [NodeConformance] [Conformance]
+  with readiness probe should not be ready before initial delay and never restart [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 ------------------------------
-S
+[sig-storage] ConfigMap 
+  should be consumable from pods in volume with mappings [NodeConformance] [Conformance]
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
+[BeforeEach] [sig-storage] ConfigMap
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
+STEP: Creating a kubernetes client
+Jun 18 12:51:57.774: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
+STEP: Building a namespace api object, basename configmap
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-configmap-kw7q9
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should be consumable from pods in volume with mappings [NodeConformance] [Conformance]
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
+STEP: Creating configMap with name configmap-test-volume-map-dbb68e4c-91c7-11e9-bce2-ae54e022189f
+STEP: Creating a pod to test consume configMaps
+Jun 18 12:51:58.257: INFO: Waiting up to 5m0s for pod "pod-configmaps-dbb893a8-91c7-11e9-bce2-ae54e022189f" in namespace "e2e-tests-configmap-kw7q9" to be "success or failure"
+Jun 18 12:51:58.272: INFO: Pod "pod-configmaps-dbb893a8-91c7-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 14.468636ms
+Jun 18 12:52:00.304: INFO: Pod "pod-configmaps-dbb893a8-91c7-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 2.047001025s
+Jun 18 12:52:02.320: INFO: Pod "pod-configmaps-dbb893a8-91c7-11e9-bce2-ae54e022189f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.062795379s
+STEP: Saw pod success
+Jun 18 12:52:02.320: INFO: Pod "pod-configmaps-dbb893a8-91c7-11e9-bce2-ae54e022189f" satisfied condition "success or failure"
+Jun 18 12:52:02.384: INFO: Trying to get logs from node 10.72.74.144 pod pod-configmaps-dbb893a8-91c7-11e9-bce2-ae54e022189f container configmap-volume-test: 
+STEP: delete the pod
+Jun 18 12:52:02.464: INFO: Waiting for pod pod-configmaps-dbb893a8-91c7-11e9-bce2-ae54e022189f to disappear
+Jun 18 12:52:02.479: INFO: Pod pod-configmaps-dbb893a8-91c7-11e9-bce2-ae54e022189f no longer exists
+[AfterEach] [sig-storage] ConfigMap
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
+Jun 18 12:52:02.480: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-tests-configmap-kw7q9" for this suite.
+Jun 18 12:52:08.554: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 12:52:08.772: INFO: namespace: e2e-tests-configmap-kw7q9, resource: bindings, ignored listing per whitelist
+Jun 18 12:52:09.060: INFO: namespace e2e-tests-configmap-kw7q9 deletion completed in 6.560492233s
+
+• [SLOW TEST:11.286 seconds]
+[sig-storage] ConfigMap
+/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:33
+  should be consumable from pods in volume with mappings [NodeConformance] [Conformance]
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
+------------------------------
+SS
+------------------------------
+[sig-apps] Daemon set [Serial] 
+  should run and stop complex daemon [Conformance]
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
+[BeforeEach] [sig-apps] Daemon set [Serial]
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
+STEP: Creating a kubernetes client
+Jun 18 12:52:09.061: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
+STEP: Building a namespace api object, basename daemonsets
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-daemonsets-hcsms
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [sig-apps] Daemon set [Serial]
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:102
+[It] should run and stop complex daemon [Conformance]
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
+Jun 18 12:52:09.610: INFO: Creating daemon "daemon-set" with a node selector
+STEP: Initially, daemon pods should not be running on any nodes.
+Jun 18 12:52:09.640: INFO: Number of nodes with available pods: 0
+Jun 18 12:52:09.640: INFO: Number of running nodes: 0, number of available pods: 0
+STEP: Change node label to blue, check that daemon pod is launched.
+Jun 18 12:52:09.702: INFO: Number of nodes with available pods: 0
+Jun 18 12:52:09.702: INFO: Node 10.72.74.143 is running more than one daemon pod
+Jun 18 12:52:10.737: INFO: Number of nodes with available pods: 0
+Jun 18 12:52:10.737: INFO: Node 10.72.74.143 is running more than one daemon pod
+Jun 18 12:52:11.717: INFO: Number of nodes with available pods: 1
+Jun 18 12:52:11.717: INFO: Number of running nodes: 1, number of available pods: 1
+STEP: Update the node label to green, and wait for daemons to be unscheduled
+Jun 18 12:52:11.782: INFO: Number of nodes with available pods: 1
+Jun 18 12:52:11.782: INFO: Number of running nodes: 0, number of available pods: 1
+Jun 18 12:52:12.797: INFO: Number of nodes with available pods: 0
+Jun 18 12:52:12.797: INFO: Number of running nodes: 0, number of available pods: 0
+STEP: Update DaemonSet node selector to green, and change its update strategy to RollingUpdate
+Jun 18 12:52:12.899: INFO: Number of nodes with available pods: 0
+Jun 18 12:52:12.899: INFO: Node 10.72.74.143 is running more than one daemon pod
+Jun 18 12:52:13.914: INFO: Number of nodes with available pods: 0
+Jun 18 12:52:13.914: INFO: Node 10.72.74.143 is running more than one daemon pod
+Jun 18 12:52:14.914: INFO: Number of nodes with available pods: 0
+Jun 18 12:52:14.914: INFO: Node 10.72.74.143 is running more than one daemon pod
+Jun 18 12:52:15.915: INFO: Number of nodes with available pods: 0
+Jun 18 12:52:15.915: INFO: Node 10.72.74.143 is running more than one daemon pod
+Jun 18 12:52:16.913: INFO: Number of nodes with available pods: 0
+Jun 18 12:52:16.913: INFO: Node 10.72.74.143 is running more than one daemon pod
+Jun 18 12:52:17.914: INFO: Number of nodes with available pods: 0
+Jun 18 12:52:17.914: INFO: Node 10.72.74.143 is running more than one daemon pod
+Jun 18 12:52:18.914: INFO: Number of nodes with available pods: 0
+Jun 18 12:52:18.914: INFO: Node 10.72.74.143 is running more than one daemon pod
+Jun 18 12:52:19.914: INFO: Number of nodes with available pods: 0
+Jun 18 12:52:19.914: INFO: Node 10.72.74.143 is running more than one daemon pod
+Jun 18 12:52:20.932: INFO: Number of nodes with available pods: 0
+Jun 18 12:52:20.932: INFO: Node 10.72.74.143 is running more than one daemon pod
+Jun 18 12:52:21.914: INFO: Number of nodes with available pods: 0
+Jun 18 12:52:21.914: INFO: Node 10.72.74.143 is running more than one daemon pod
+Jun 18 12:52:22.914: INFO: Number of nodes with available pods: 0
+Jun 18 12:52:22.914: INFO: Node 10.72.74.143 is running more than one daemon pod
+Jun 18 12:52:23.914: INFO: Number of nodes with available pods: 0
+Jun 18 12:52:23.914: INFO: Node 10.72.74.143 is running more than one daemon pod
+Jun 18 12:52:24.915: INFO: Number of nodes with available pods: 0
+Jun 18 12:52:24.915: INFO: Node 10.72.74.143 is running more than one daemon pod
+Jun 18 12:52:25.914: INFO: Number of nodes with available pods: 0
+Jun 18 12:52:25.914: INFO: Node 10.72.74.143 is running more than one daemon pod
+Jun 18 12:52:26.914: INFO: Number of nodes with available pods: 0
+Jun 18 12:52:26.914: INFO: Node 10.72.74.143 is running more than one daemon pod
+Jun 18 12:52:27.914: INFO: Number of nodes with available pods: 0
+Jun 18 12:52:27.914: INFO: Node 10.72.74.143 is running more than one daemon pod
+Jun 18 12:52:28.914: INFO: Number of nodes with available pods: 0
+Jun 18 12:52:28.914: INFO: Node 10.72.74.143 is running more than one daemon pod
+Jun 18 12:52:29.915: INFO: Number of nodes with available pods: 0
+Jun 18 12:52:29.915: INFO: Node 10.72.74.143 is running more than one daemon pod
+Jun 18 12:52:30.915: INFO: Number of nodes with available pods: 0
+Jun 18 12:52:30.915: INFO: Node 10.72.74.143 is running more than one daemon pod
+Jun 18 12:52:31.943: INFO: Number of nodes with available pods: 0
+Jun 18 12:52:31.943: INFO: Node 10.72.74.143 is running more than one daemon pod
+Jun 18 12:52:32.916: INFO: Number of nodes with available pods: 0
+Jun 18 12:52:32.916: INFO: Node 10.72.74.143 is running more than one daemon pod
+Jun 18 12:52:33.914: INFO: Number of nodes with available pods: 0
+Jun 18 12:52:33.914: INFO: Node 10.72.74.143 is running more than one daemon pod
+Jun 18 12:52:34.914: INFO: Number of nodes with available pods: 0
+Jun 18 12:52:34.914: INFO: Node 10.72.74.143 is running more than one daemon pod
+Jun 18 12:52:35.916: INFO: Number of nodes with available pods: 0
+Jun 18 12:52:35.916: INFO: Node 10.72.74.143 is running more than one daemon pod
+Jun 18 12:52:36.913: INFO: Number of nodes with available pods: 0
+Jun 18 12:52:36.913: INFO: Node 10.72.74.143 is running more than one daemon pod
+Jun 18 12:52:37.916: INFO: Number of nodes with available pods: 0
+Jun 18 12:52:37.916: INFO: Node 10.72.74.143 is running more than one daemon pod
+Jun 18 12:52:38.985: INFO: Number of nodes with available pods: 0
+Jun 18 12:52:38.985: INFO: Node 10.72.74.143 is running more than one daemon pod
+Jun 18 12:52:39.915: INFO: Number of nodes with available pods: 0
+Jun 18 12:52:39.915: INFO: Node 10.72.74.143 is running more than one daemon pod
+Jun 18 12:52:40.918: INFO: Number of nodes with available pods: 0
+Jun 18 12:52:40.918: INFO: Node 10.72.74.143 is running more than one daemon pod
+Jun 18 12:52:41.914: INFO: Number of nodes with available pods: 0
+Jun 18 12:52:41.914: INFO: Node 10.72.74.143 is running more than one daemon pod
+Jun 18 12:52:42.939: INFO: Number of nodes with available pods: 0
+Jun 18 12:52:42.939: INFO: Node 10.72.74.143 is running more than one daemon pod
+Jun 18 12:52:43.914: INFO: Number of nodes with available pods: 0
+Jun 18 12:52:43.914: INFO: Node 10.72.74.143 is running more than one daemon pod
+Jun 18 12:52:44.915: INFO: Number of nodes with available pods: 0
+Jun 18 12:52:44.915: INFO: Node 10.72.74.143 is running more than one daemon pod
+Jun 18 12:52:45.913: INFO: Number of nodes with available pods: 0
+Jun 18 12:52:45.914: INFO: Node 10.72.74.143 is running more than one daemon pod
+Jun 18 12:52:46.914: INFO: Number of nodes with available pods: 0
+Jun 18 12:52:46.914: INFO: Node 10.72.74.143 is running more than one daemon pod
+Jun 18 12:52:47.918: INFO: Number of nodes with available pods: 0
+Jun 18 12:52:47.918: INFO: Node 10.72.74.143 is running more than one daemon pod
+Jun 18 12:52:48.920: INFO: Number of nodes with available pods: 0
+Jun 18 12:52:48.920: INFO: Node 10.72.74.143 is running more than one daemon pod
+Jun 18 12:52:49.914: INFO: Number of nodes with available pods: 0
+Jun 18 12:52:49.914: INFO: Node 10.72.74.143 is running more than one daemon pod
+Jun 18 12:52:50.915: INFO: Number of nodes with available pods: 0
+Jun 18 12:52:50.915: INFO: Node 10.72.74.143 is running more than one daemon pod
+Jun 18 12:52:51.984: INFO: Number of nodes with available pods: 0
+Jun 18 12:52:51.984: INFO: Node 10.72.74.143 is running more than one daemon pod
+Jun 18 12:52:52.914: INFO: Number of nodes with available pods: 0
+Jun 18 12:52:52.914: INFO: Node 10.72.74.143 is running more than one daemon pod
+Jun 18 12:52:53.931: INFO: Number of nodes with available pods: 0
+Jun 18 12:52:53.931: INFO: Node 10.72.74.143 is running more than one daemon pod
+Jun 18 12:52:54.915: INFO: Number of nodes with available pods: 0
+Jun 18 12:52:54.915: INFO: Node 10.72.74.143 is running more than one daemon pod
+Jun 18 12:52:55.914: INFO: Number of nodes with available pods: 0
+Jun 18 12:52:55.914: INFO: Node 10.72.74.143 is running more than one daemon pod
+Jun 18 12:52:56.914: INFO: Number of nodes with available pods: 0
+Jun 18 12:52:56.914: INFO: Node 10.72.74.143 is running more than one daemon pod
+Jun 18 12:52:57.914: INFO: Number of nodes with available pods: 1
+Jun 18 12:52:57.914: INFO: Number of running nodes: 1, number of available pods: 1
+[AfterEach] [sig-apps] Daemon set [Serial]
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:68
+STEP: Deleting DaemonSet "daemon-set"
+STEP: deleting DaemonSet.extensions daemon-set in namespace e2e-tests-daemonsets-hcsms, will wait for the garbage collector to delete the pods
+Jun 18 12:52:58.033: INFO: Deleting DaemonSet.extensions daemon-set took: 30.048381ms
+Jun 18 12:52:58.133: INFO: Terminating DaemonSet.extensions daemon-set pods took: 100.300231ms
+Jun 18 12:53:32.871: INFO: Number of nodes with available pods: 0
+Jun 18 12:53:32.871: INFO: Number of running nodes: 0, number of available pods: 0
+Jun 18 12:53:32.884: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"selfLink":"/apis/apps/v1/namespaces/e2e-tests-daemonsets-hcsms/daemonsets","resourceVersion":"107338"},"items":null}
+
+Jun 18 12:53:32.897: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"selfLink":"/api/v1/namespaces/e2e-tests-daemonsets-hcsms/pods","resourceVersion":"107338"},"items":null}
+
+[AfterEach] [sig-apps] Daemon set [Serial]
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
+Jun 18 12:53:32.980: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-tests-daemonsets-hcsms" for this suite.
+Jun 18 12:53:41.047: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 12:53:41.497: INFO: namespace: e2e-tests-daemonsets-hcsms, resource: bindings, ignored listing per whitelist
+Jun 18 12:53:41.517: INFO: namespace e2e-tests-daemonsets-hcsms deletion completed in 8.518706838s
+
+• [SLOW TEST:92.456 seconds]
+[sig-apps] Daemon set [Serial]
+/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22
+  should run and stop complex daemon [Conformance]
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
+------------------------------
+SSSSSSSSSSS
 ------------------------------
 [sig-storage] Downward API volume 
-  should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance]
+  should set mode on item file [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 [BeforeEach] [sig-storage] Downward API volume
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
 STEP: Creating a kubernetes client
-May 13 19:57:02.744: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423
+Jun 18 12:53:41.517: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
 STEP: Building a namespace api object, basename downward-api
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-downward-api-qcvkt
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-downward-api-jgr8h
 STEP: Waiting for a default service account to be provisioned in namespace
 [BeforeEach] [sig-storage] Downward API volume
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:39
-[It] should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance]
+[It] should set mode on item file [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 STEP: Creating a pod to test downward API volume plugin
-May 13 19:57:03.039: INFO: Waiting up to 5m0s for pod "downwardapi-volume-46e3a0ee-75b9-11e9-8f67-2632f168be36" in namespace "e2e-tests-downward-api-qcvkt" to be "success or failure"
-May 13 19:57:03.048: INFO: Pod "downwardapi-volume-46e3a0ee-75b9-11e9-8f67-2632f168be36": Phase="Pending", Reason="", readiness=false. Elapsed: 8.802663ms
-May 13 19:57:05.057: INFO: Pod "downwardapi-volume-46e3a0ee-75b9-11e9-8f67-2632f168be36": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.017823105s
+Jun 18 12:53:41.998: INFO: Waiting up to 5m0s for pod "downwardapi-volume-198e1312-91c8-11e9-bce2-ae54e022189f" in namespace "e2e-tests-downward-api-jgr8h" to be "success or failure"
+Jun 18 12:53:42.012: INFO: Pod "downwardapi-volume-198e1312-91c8-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 14.283073ms
+Jun 18 12:53:44.048: INFO: Pod "downwardapi-volume-198e1312-91c8-11e9-bce2-ae54e022189f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.050539154s
 STEP: Saw pod success
-May 13 19:57:05.057: INFO: Pod "downwardapi-volume-46e3a0ee-75b9-11e9-8f67-2632f168be36" satisfied condition "success or failure"
-May 13 19:57:05.066: INFO: Trying to get logs from node 10.170.219.153 pod downwardapi-volume-46e3a0ee-75b9-11e9-8f67-2632f168be36 container client-container: 
+Jun 18 12:53:44.048: INFO: Pod "downwardapi-volume-198e1312-91c8-11e9-bce2-ae54e022189f" satisfied condition "success or failure"
+Jun 18 12:53:44.063: INFO: Trying to get logs from node 10.72.74.143 pod downwardapi-volume-198e1312-91c8-11e9-bce2-ae54e022189f container client-container: 
 STEP: delete the pod
-May 13 19:57:05.119: INFO: Waiting for pod downwardapi-volume-46e3a0ee-75b9-11e9-8f67-2632f168be36 to disappear
-May 13 19:57:05.127: INFO: Pod downwardapi-volume-46e3a0ee-75b9-11e9-8f67-2632f168be36 no longer exists
+Jun 18 12:53:44.138: INFO: Waiting for pod downwardapi-volume-198e1312-91c8-11e9-bce2-ae54e022189f to disappear
+Jun 18 12:53:44.154: INFO: Pod downwardapi-volume-198e1312-91c8-11e9-bce2-ae54e022189f no longer exists
 [AfterEach] [sig-storage] Downward API volume
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
-May 13 19:57:05.127: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "e2e-tests-downward-api-qcvkt" for this suite.
-May 13 19:57:11.167: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-May 13 19:57:11.200: INFO: namespace: e2e-tests-downward-api-qcvkt, resource: bindings, ignored listing per whitelist
-May 13 19:57:11.424: INFO: namespace e2e-tests-downward-api-qcvkt deletion completed in 6.289128494s
+Jun 18 12:53:44.154: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-tests-downward-api-jgr8h" for this suite.
+Jun 18 12:53:50.231: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 12:53:50.625: INFO: namespace: e2e-tests-downward-api-jgr8h, resource: bindings, ignored listing per whitelist
+Jun 18 12:53:50.741: INFO: namespace e2e-tests-downward-api-jgr8h deletion completed in 6.565716731s
 
-• [SLOW TEST:8.680 seconds]
+• [SLOW TEST:9.224 seconds]
 [sig-storage] Downward API volume
 /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:34
-  should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance]
+  should set mode on item file [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 ------------------------------
-[sig-storage] Projected secret 
-  should be consumable from pods in volume with mappings and Item Mode set [NodeConformance] [Conformance]
+SSSSSSSSSSSSSS
+------------------------------
+[sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] 
+  Scaling should happen in predictable order and halt if any stateful pod is unhealthy [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-[BeforeEach] [sig-storage] Projected secret
+[BeforeEach] [sig-apps] StatefulSet
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
 STEP: Creating a kubernetes client
-May 13 19:57:11.424: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423
-STEP: Building a namespace api object, basename projected
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-projected-wbs8c
+Jun 18 12:53:50.741: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
+STEP: Building a namespace api object, basename statefulset
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-statefulset-g9244
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should be consumable from pods in volume with mappings and Item Mode set [NodeConformance] [Conformance]
+[BeforeEach] [sig-apps] StatefulSet
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:59
+[BeforeEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic]
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:74
+STEP: Creating service test in namespace e2e-tests-statefulset-g9244
+[It] Scaling should happen in predictable order and halt if any stateful pod is unhealthy [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-STEP: Creating projection with secret that has name projected-secret-test-map-4c0d14f3-75b9-11e9-8f67-2632f168be36
-STEP: Creating a pod to test consume secrets
-May 13 19:57:11.708: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-4c0e77a3-75b9-11e9-8f67-2632f168be36" in namespace "e2e-tests-projected-wbs8c" to be "success or failure"
-May 13 19:57:11.716: INFO: Pod "pod-projected-secrets-4c0e77a3-75b9-11e9-8f67-2632f168be36": Phase="Pending", Reason="", readiness=false. Elapsed: 7.932816ms
-May 13 19:57:13.738: INFO: Pod "pod-projected-secrets-4c0e77a3-75b9-11e9-8f67-2632f168be36": Phase="Running", Reason="", readiness=true. Elapsed: 2.030160128s
-May 13 19:57:15.747: INFO: Pod "pod-projected-secrets-4c0e77a3-75b9-11e9-8f67-2632f168be36": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.038424434s
-STEP: Saw pod success
-May 13 19:57:15.747: INFO: Pod "pod-projected-secrets-4c0e77a3-75b9-11e9-8f67-2632f168be36" satisfied condition "success or failure"
-May 13 19:57:15.755: INFO: Trying to get logs from node 10.170.219.153 pod pod-projected-secrets-4c0e77a3-75b9-11e9-8f67-2632f168be36 container projected-secret-volume-test: 
-STEP: delete the pod
-May 13 19:57:15.792: INFO: Waiting for pod pod-projected-secrets-4c0e77a3-75b9-11e9-8f67-2632f168be36 to disappear
-May 13 19:57:15.799: INFO: Pod pod-projected-secrets-4c0e77a3-75b9-11e9-8f67-2632f168be36 no longer exists
-[AfterEach] [sig-storage] Projected secret
+STEP: Initializing watcher for selector baz=blah,foo=bar
+STEP: Creating stateful set ss in namespace e2e-tests-statefulset-g9244
+STEP: Waiting until all stateful set ss replicas will be running in namespace e2e-tests-statefulset-g9244
+Jun 18 12:53:51.242: INFO: Found 0 stateful pods, waiting for 1
+Jun 18 12:54:01.276: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true
+STEP: Confirming that stateful set scale up will halt with unhealthy stateful pod
+Jun 18 12:54:01.291: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 exec --namespace=e2e-tests-statefulset-g9244 ss-0 -- /bin/sh -c mv -v /usr/share/nginx/html/index.html /tmp/ || true'
+Jun 18 12:54:01.904: INFO: stderr: ""
+Jun 18 12:54:01.904: INFO: stdout: "'/usr/share/nginx/html/index.html' -> '/tmp/index.html'\n"
+Jun 18 12:54:01.904: INFO: stdout of mv -v /usr/share/nginx/html/index.html /tmp/ || true on ss-0: '/usr/share/nginx/html/index.html' -> '/tmp/index.html'
+
+Jun 18 12:54:01.919: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=true
+Jun 18 12:54:11.954: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=false
+Jun 18 12:54:11.954: INFO: Waiting for statefulset status.replicas updated to 0
+Jun 18 12:54:12.018: INFO: Verifying statefulset ss doesn't scale past 1 for another 9.99999784s
+Jun 18 12:54:13.033: INFO: Verifying statefulset ss doesn't scale past 1 for another 8.981923918s
+Jun 18 12:54:14.049: INFO: Verifying statefulset ss doesn't scale past 1 for another 7.966832038s
+Jun 18 12:54:15.066: INFO: Verifying statefulset ss doesn't scale past 1 for another 6.951240246s
+Jun 18 12:54:16.082: INFO: Verifying statefulset ss doesn't scale past 1 for another 5.933718424s
+Jun 18 12:54:17.099: INFO: Verifying statefulset ss doesn't scale past 1 for another 4.917517053s
+Jun 18 12:54:18.115: INFO: Verifying statefulset ss doesn't scale past 1 for another 3.901253561s
+Jun 18 12:54:19.131: INFO: Verifying statefulset ss doesn't scale past 1 for another 2.884882785s
+Jun 18 12:54:20.147: INFO: Verifying statefulset ss doesn't scale past 1 for another 1.868517024s
+Jun 18 12:54:21.162: INFO: Verifying statefulset ss doesn't scale past 1 for another 853.20856ms
+STEP: Scaling up stateful set ss to 3 replicas and waiting until all of them will be running in namespace e2e-tests-statefulset-g9244
+Jun 18 12:54:22.200: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 exec --namespace=e2e-tests-statefulset-g9244 ss-0 -- /bin/sh -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
+Jun 18 12:54:22.642: INFO: stderr: ""
+Jun 18 12:54:22.642: INFO: stdout: "'/tmp/index.html' -> '/usr/share/nginx/html/index.html'\n"
+Jun 18 12:54:22.642: INFO: stdout of mv -v /tmp/index.html /usr/share/nginx/html/ || true on ss-0: '/tmp/index.html' -> '/usr/share/nginx/html/index.html'
+
+Jun 18 12:54:22.657: INFO: Found 1 stateful pods, waiting for 3
+Jun 18 12:54:32.690: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true
+Jun 18 12:54:32.690: INFO: Waiting for pod ss-1 to enter Running - Ready=true, currently Running - Ready=true
+Jun 18 12:54:32.690: INFO: Waiting for pod ss-2 to enter Running - Ready=true, currently Running - Ready=true
+STEP: Verifying that stateful set ss was scaled up in order
+STEP: Scale down will halt with unhealthy stateful pod
+Jun 18 12:54:32.714: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 exec --namespace=e2e-tests-statefulset-g9244 ss-0 -- /bin/sh -c mv -v /usr/share/nginx/html/index.html /tmp/ || true'
+Jun 18 12:54:33.192: INFO: stderr: ""
+Jun 18 12:54:33.192: INFO: stdout: "'/usr/share/nginx/html/index.html' -> '/tmp/index.html'\n"
+Jun 18 12:54:33.192: INFO: stdout of mv -v /usr/share/nginx/html/index.html /tmp/ || true on ss-0: '/usr/share/nginx/html/index.html' -> '/tmp/index.html'
+
+Jun 18 12:54:33.192: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 exec --namespace=e2e-tests-statefulset-g9244 ss-1 -- /bin/sh -c mv -v /usr/share/nginx/html/index.html /tmp/ || true'
+Jun 18 12:54:33.585: INFO: stderr: ""
+Jun 18 12:54:33.585: INFO: stdout: "'/usr/share/nginx/html/index.html' -> '/tmp/index.html'\n"
+Jun 18 12:54:33.585: INFO: stdout of mv -v /usr/share/nginx/html/index.html /tmp/ || true on ss-1: '/usr/share/nginx/html/index.html' -> '/tmp/index.html'
+
+Jun 18 12:54:33.585: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 exec --namespace=e2e-tests-statefulset-g9244 ss-2 -- /bin/sh -c mv -v /usr/share/nginx/html/index.html /tmp/ || true'
+Jun 18 12:54:34.164: INFO: stderr: ""
+Jun 18 12:54:34.164: INFO: stdout: "'/usr/share/nginx/html/index.html' -> '/tmp/index.html'\n"
+Jun 18 12:54:34.164: INFO: stdout of mv -v /usr/share/nginx/html/index.html /tmp/ || true on ss-2: '/usr/share/nginx/html/index.html' -> '/tmp/index.html'
+
+Jun 18 12:54:34.164: INFO: Waiting for statefulset status.replicas updated to 0
+Jun 18 12:54:34.178: INFO: Waiting for stateful set status.readyReplicas to become 0, currently 1
+Jun 18 12:54:44.284: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=false
+Jun 18 12:54:44.284: INFO: Waiting for pod ss-1 to enter Running - Ready=false, currently Running - Ready=false
+Jun 18 12:54:44.284: INFO: Waiting for pod ss-2 to enter Running - Ready=false, currently Running - Ready=false
+Jun 18 12:54:44.336: INFO: Verifying statefulset ss doesn't scale past 3 for another 9.999997996s
+Jun 18 12:54:45.351: INFO: Verifying statefulset ss doesn't scale past 3 for another 8.985299335s
+Jun 18 12:54:46.384: INFO: Verifying statefulset ss doesn't scale past 3 for another 7.970252935s
+Jun 18 12:54:47.400: INFO: Verifying statefulset ss doesn't scale past 3 for another 6.936546805s
+Jun 18 12:54:48.415: INFO: Verifying statefulset ss doesn't scale past 3 for another 5.921256515s
+Jun 18 12:54:49.432: INFO: Verifying statefulset ss doesn't scale past 3 for another 4.905479045s
+Jun 18 12:54:50.485: INFO: Verifying statefulset ss doesn't scale past 3 for another 3.888539774s
+Jun 18 12:54:51.500: INFO: Verifying statefulset ss doesn't scale past 3 for another 2.836201977s
+Jun 18 12:54:52.520: INFO: Verifying statefulset ss doesn't scale past 3 for another 1.820536858s
+Jun 18 12:54:53.537: INFO: Verifying statefulset ss doesn't scale past 3 for another 801.032582ms
+STEP: Scaling down stateful set ss to 0 replicas and waiting until none of pods will run in namespacee2e-tests-statefulset-g9244
+Jun 18 12:54:54.574: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 exec --namespace=e2e-tests-statefulset-g9244 ss-0 -- /bin/sh -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
+Jun 18 12:54:55.011: INFO: stderr: ""
+Jun 18 12:54:55.011: INFO: stdout: "'/tmp/index.html' -> '/usr/share/nginx/html/index.html'\n"
+Jun 18 12:54:55.011: INFO: stdout of mv -v /tmp/index.html /usr/share/nginx/html/ || true on ss-0: '/tmp/index.html' -> '/usr/share/nginx/html/index.html'
+
+Jun 18 12:54:55.011: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 exec --namespace=e2e-tests-statefulset-g9244 ss-1 -- /bin/sh -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
+Jun 18 12:54:55.467: INFO: stderr: ""
+Jun 18 12:54:55.467: INFO: stdout: "'/tmp/index.html' -> '/usr/share/nginx/html/index.html'\n"
+Jun 18 12:54:55.467: INFO: stdout of mv -v /tmp/index.html /usr/share/nginx/html/ || true on ss-1: '/tmp/index.html' -> '/usr/share/nginx/html/index.html'
+
+Jun 18 12:54:55.467: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 exec --namespace=e2e-tests-statefulset-g9244 ss-2 -- /bin/sh -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
+Jun 18 12:54:55.840: INFO: stderr: ""
+Jun 18 12:54:55.840: INFO: stdout: "'/tmp/index.html' -> '/usr/share/nginx/html/index.html'\n"
+Jun 18 12:54:55.840: INFO: stdout of mv -v /tmp/index.html /usr/share/nginx/html/ || true on ss-2: '/tmp/index.html' -> '/usr/share/nginx/html/index.html'
+
+Jun 18 12:54:55.840: INFO: Scaling statefulset ss to 0
+STEP: Verifying that stateful set ss was scaled down in reverse order
+[AfterEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic]
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:85
+Jun 18 12:55:25.925: INFO: Deleting all statefulset in ns e2e-tests-statefulset-g9244
+Jun 18 12:55:25.940: INFO: Scaling statefulset ss to 0
+Jun 18 12:55:26.015: INFO: Waiting for statefulset status.replicas updated to 0
+Jun 18 12:55:26.029: INFO: Deleting statefulset ss
+[AfterEach] [sig-apps] StatefulSet
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
-May 13 19:57:15.799: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "e2e-tests-projected-wbs8c" for this suite.
-May 13 19:57:21.837: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-May 13 19:57:22.040: INFO: namespace: e2e-tests-projected-wbs8c, resource: bindings, ignored listing per whitelist
-May 13 19:57:22.091: INFO: namespace e2e-tests-projected-wbs8c deletion completed in 6.284271551s
+Jun 18 12:55:26.090: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-tests-statefulset-g9244" for this suite.
+Jun 18 12:55:34.159: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 12:55:34.344: INFO: namespace: e2e-tests-statefulset-g9244, resource: bindings, ignored listing per whitelist
+Jun 18 12:55:34.646: INFO: namespace e2e-tests-statefulset-g9244 deletion completed in 8.534911717s
 
-• [SLOW TEST:10.667 seconds]
-[sig-storage] Projected secret
-/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_secret.go:34
-  should be consumable from pods in volume with mappings and Item Mode set [NodeConformance] [Conformance]
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
+• [SLOW TEST:103.904 seconds]
+[sig-apps] StatefulSet
+/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22
+  [k8s.io] Basic StatefulSet functionality [StatefulSetBasic]
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694
+    Scaling should happen in predictable order and halt if any stateful pod is unhealthy [Conformance]
+    /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 ------------------------------
-SSSSSSSSSSSS
+SSS
 ------------------------------
-[sig-api-machinery] Secrets 
-  should be consumable via the environment [NodeConformance] [Conformance]
+[sig-storage] EmptyDir volumes 
+  should support (non-root,0666,default) [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-[BeforeEach] [sig-api-machinery] Secrets
+[BeforeEach] [sig-storage] EmptyDir volumes
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
 STEP: Creating a kubernetes client
-May 13 19:57:22.092: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423
-STEP: Building a namespace api object, basename secrets
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-secrets-vkbr6
+Jun 18 12:55:34.646: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
+STEP: Building a namespace api object, basename emptydir
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-emptydir-bkq2d
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should be consumable via the environment [NodeConformance] [Conformance]
+[It] should support (non-root,0666,default) [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-STEP: creating secret e2e-tests-secrets-vkbr6/secret-test-5273bff6-75b9-11e9-8f67-2632f168be36
-STEP: Creating a pod to test consume secrets
-May 13 19:57:22.446: INFO: Waiting up to 5m0s for pod "pod-configmaps-52750d93-75b9-11e9-8f67-2632f168be36" in namespace "e2e-tests-secrets-vkbr6" to be "success or failure"
-May 13 19:57:22.453: INFO: Pod "pod-configmaps-52750d93-75b9-11e9-8f67-2632f168be36": Phase="Pending", Reason="", readiness=false. Elapsed: 7.332405ms
-May 13 19:57:24.478: INFO: Pod "pod-configmaps-52750d93-75b9-11e9-8f67-2632f168be36": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.031641361s
+STEP: Creating a pod to test emptydir 0666 on node default medium
+Jun 18 12:55:35.124: INFO: Waiting up to 5m0s for pod "pod-5cfbe5c1-91c8-11e9-bce2-ae54e022189f" in namespace "e2e-tests-emptydir-bkq2d" to be "success or failure"
+Jun 18 12:55:35.139: INFO: Pod "pod-5cfbe5c1-91c8-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 14.932131ms
+Jun 18 12:55:37.178: INFO: Pod "pod-5cfbe5c1-91c8-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 2.053398612s
+Jun 18 12:55:39.192: INFO: Pod "pod-5cfbe5c1-91c8-11e9-bce2-ae54e022189f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.067994446s
 STEP: Saw pod success
-May 13 19:57:24.478: INFO: Pod "pod-configmaps-52750d93-75b9-11e9-8f67-2632f168be36" satisfied condition "success or failure"
-May 13 19:57:24.485: INFO: Trying to get logs from node 10.170.219.153 pod pod-configmaps-52750d93-75b9-11e9-8f67-2632f168be36 container env-test: 
+Jun 18 12:55:39.192: INFO: Pod "pod-5cfbe5c1-91c8-11e9-bce2-ae54e022189f" satisfied condition "success or failure"
+Jun 18 12:55:39.207: INFO: Trying to get logs from node 10.72.74.144 pod pod-5cfbe5c1-91c8-11e9-bce2-ae54e022189f container test-container: 
 STEP: delete the pod
-May 13 19:57:24.526: INFO: Waiting for pod pod-configmaps-52750d93-75b9-11e9-8f67-2632f168be36 to disappear
-May 13 19:57:24.533: INFO: Pod pod-configmaps-52750d93-75b9-11e9-8f67-2632f168be36 no longer exists
-[AfterEach] [sig-api-machinery] Secrets
+Jun 18 12:55:39.281: INFO: Waiting for pod pod-5cfbe5c1-91c8-11e9-bce2-ae54e022189f to disappear
+Jun 18 12:55:39.296: INFO: Pod pod-5cfbe5c1-91c8-11e9-bce2-ae54e022189f no longer exists
+[AfterEach] [sig-storage] EmptyDir volumes
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
-May 13 19:57:24.533: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "e2e-tests-secrets-vkbr6" for this suite.
-May 13 19:57:30.683: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-May 13 19:57:30.733: INFO: namespace: e2e-tests-secrets-vkbr6, resource: bindings, ignored listing per whitelist
-May 13 19:57:30.919: INFO: namespace e2e-tests-secrets-vkbr6 deletion completed in 6.298993701s
+Jun 18 12:55:39.296: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-tests-emptydir-bkq2d" for this suite.
+Jun 18 12:55:45.370: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 12:55:45.468: INFO: namespace: e2e-tests-emptydir-bkq2d, resource: bindings, ignored listing per whitelist
+Jun 18 12:55:45.901: INFO: namespace e2e-tests-emptydir-bkq2d deletion completed in 6.585077841s
 
-• [SLOW TEST:8.827 seconds]
-[sig-api-machinery] Secrets
-/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/secrets.go:32
-  should be consumable via the environment [NodeConformance] [Conformance]
+• [SLOW TEST:11.255 seconds]
+[sig-storage] EmptyDir volumes
+/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:40
+  should support (non-root,0666,default) [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 ------------------------------
-SS
+SSSSSSSSS
 ------------------------------
-[sig-apps] ReplicationController 
-  should release no longer matching pods [Conformance]
+[sig-cli] Kubectl client [k8s.io] Kubectl describe 
+  should check if kubectl describe prints relevant information for rc and pods  [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-[BeforeEach] [sig-apps] ReplicationController
+[BeforeEach] [sig-cli] Kubectl client
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
 STEP: Creating a kubernetes client
-May 13 19:57:30.919: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423
-STEP: Building a namespace api object, basename replication-controller
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-replication-controller-w2csw
+Jun 18 12:55:45.902: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
+STEP: Building a namespace api object, basename kubectl
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-kubectl-mq6hg
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should release no longer matching pods [Conformance]
+[BeforeEach] [sig-cli] Kubectl client
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:243
+[It] should check if kubectl describe prints relevant information for rc and pods  [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-STEP: Given a ReplicationController is created
-STEP: When the matched label of one of its pods change
-May 13 19:57:31.229: INFO: Pod name pod-release: Found 1 pods out of 1
-STEP: Then the pod is released
-[AfterEach] [sig-apps] ReplicationController
+Jun 18 12:55:46.397: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 version --client'
+Jun 18 12:55:46.477: INFO: stderr: ""
+Jun 18 12:55:46.477: INFO: stdout: "Client Version: version.Info{Major:\"1\", Minor:\"13\", GitVersion:\"v1.13.0\", GitCommit:\"ddf47ac13c1a9483ea035a79cd7c10005ff21a6d\", GitTreeState:\"clean\", BuildDate:\"2018-12-03T21:04:45Z\", GoVersion:\"go1.11.2\", Compiler:\"gc\", Platform:\"linux/amd64\"}\n"
+Jun 18 12:55:46.483: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 create -f - --namespace=e2e-tests-kubectl-mq6hg'
+Jun 18 12:55:47.032: INFO: stderr: ""
+Jun 18 12:55:47.032: INFO: stdout: "replicationcontroller/redis-master created\n"
+Jun 18 12:55:47.032: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 create -f - --namespace=e2e-tests-kubectl-mq6hg'
+Jun 18 12:55:47.431: INFO: stderr: ""
+Jun 18 12:55:47.431: INFO: stdout: "service/redis-master created\n"
+STEP: Waiting for Redis master to start.
+Jun 18 12:55:48.445: INFO: Selector matched 1 pods for map[app:redis]
+Jun 18 12:55:48.445: INFO: Found 0 / 1
+Jun 18 12:55:49.447: INFO: Selector matched 1 pods for map[app:redis]
+Jun 18 12:55:49.447: INFO: Found 1 / 1
+Jun 18 12:55:49.447: INFO: WaitFor completed with timeout 5m0s.  Pods found = 1 out of 1
+Jun 18 12:55:49.462: INFO: Selector matched 1 pods for map[app:redis]
+Jun 18 12:55:49.462: INFO: ForEach: Found 1 pods from the filter.  Now looping through them.
+Jun 18 12:55:49.463: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 describe pod redis-master-wsfsn --namespace=e2e-tests-kubectl-mq6hg'
+Jun 18 12:55:49.629: INFO: stderr: ""
+Jun 18 12:55:49.629: INFO: stdout: "Name:               redis-master-wsfsn\nNamespace:          e2e-tests-kubectl-mq6hg\nPriority:           0\nPriorityClassName:  \nNode:               10.72.74.149/10.72.74.149\nStart Time:         Tue, 18 Jun 2019 12:55:47 +0000\nLabels:             app=redis\n                    role=master\nAnnotations:        kubernetes.io/psp: e2e-test-privileged-psp\nStatus:             Running\nIP:                 172.30.39.35\nControlled By:      ReplicationController/redis-master\nContainers:\n  redis-master:\n    Container ID:   containerd://00b8b539895762d3ca4d60eaed4ee5741b5b7d92cb5df37d7c9aa1abe3b25b44\n    Image:          gcr.io/kubernetes-e2e-test-images/redis:1.0\n    Image ID:       gcr.io/kubernetes-e2e-test-images/redis@sha256:af4748d1655c08dc54d4be5182135395db9ce87aba2d4699b26b14ae197c5830\n    Port:           6379/TCP\n    Host Port:      0/TCP\n    State:          Running\n      Started:      Tue, 18 Jun 2019 12:55:48 +0000\n    Ready:          True\n    Restart Count:  0\n    Environment:    \n    Mounts:\n      /var/run/secrets/kubernetes.io/serviceaccount from default-token-b8gws (ro)\nConditions:\n  Type              Status\n  Initialized       True \n  Ready             True \n  ContainersReady   True \n  PodScheduled      True \nVolumes:\n  default-token-b8gws:\n    Type:        Secret (a volume populated by a Secret)\n    SecretName:  default-token-b8gws\n    Optional:    false\nQoS Class:       BestEffort\nNode-Selectors:  \nTolerations:     node.kubernetes.io/not-ready:NoExecute for 300s\n                 node.kubernetes.io/unreachable:NoExecute for 300s\nEvents:\n  Type    Reason     Age   From                   Message\n  ----    ------     ----  ----                   -------\n  Normal  Scheduled  2s    default-scheduler      Successfully assigned e2e-tests-kubectl-mq6hg/redis-master-wsfsn to 10.72.74.149\n  Normal  Pulled     1s    kubelet, 10.72.74.149  Container image \"gcr.io/kubernetes-e2e-test-images/redis:1.0\" already present on machine\n  Normal  Created    1s    kubelet, 10.72.74.149  Created container\n  Normal  Started    1s    kubelet, 10.72.74.149  Started container\n"
+Jun 18 12:55:49.629: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 describe rc redis-master --namespace=e2e-tests-kubectl-mq6hg'
+Jun 18 12:55:50.559: INFO: stderr: ""
+Jun 18 12:55:50.559: INFO: stdout: "Name:         redis-master\nNamespace:    e2e-tests-kubectl-mq6hg\nSelector:     app=redis,role=master\nLabels:       app=redis\n              role=master\nAnnotations:  \nReplicas:     1 current / 1 desired\nPods Status:  1 Running / 0 Waiting / 0 Succeeded / 0 Failed\nPod Template:\n  Labels:  app=redis\n           role=master\n  Containers:\n   redis-master:\n    Image:        gcr.io/kubernetes-e2e-test-images/redis:1.0\n    Port:         6379/TCP\n    Host Port:    0/TCP\n    Environment:  \n    Mounts:       \n  Volumes:        \nEvents:\n  Type    Reason            Age   From                    Message\n  ----    ------            ----  ----                    -------\n  Normal  SuccessfulCreate  3s    replication-controller  Created pod: redis-master-wsfsn\n"
+Jun 18 12:55:50.559: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 describe service redis-master --namespace=e2e-tests-kubectl-mq6hg'
+Jun 18 12:55:50.737: INFO: stderr: ""
+Jun 18 12:55:50.737: INFO: stdout: "Name:              redis-master\nNamespace:         e2e-tests-kubectl-mq6hg\nLabels:            app=redis\n                   role=master\nAnnotations:       \nSelector:          app=redis,role=master\nType:              ClusterIP\nIP:                172.21.40.125\nPort:                6379/TCP\nTargetPort:        redis-server/TCP\nEndpoints:         172.30.39.35:6379\nSession Affinity:  None\nEvents:            \n"
+Jun 18 12:55:50.756: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 describe node 10.72.74.143'
+Jun 18 12:55:50.949: INFO: stderr: ""
+Jun 18 12:55:50.949: INFO: stdout: "Name:               10.72.74.143\nRoles:              \nLabels:             arch=amd64\n                    beta.kubernetes.io/arch=amd64\n                    beta.kubernetes.io/instance-type=b3c.4x16.encrypted\n                    beta.kubernetes.io/os=linux\n                    failure-domain.beta.kubernetes.io/region=eu-gb\n                    failure-domain.beta.kubernetes.io/zone=lon06\n                    ibm-cloud.kubernetes.io/encrypted-docker-data=true\n                    ibm-cloud.kubernetes.io/ha-worker=true\n                    ibm-cloud.kubernetes.io/iaas-provider=softlayer\n                    ibm-cloud.kubernetes.io/machine-type=b3c.4x16.encrypted\n                    ibm-cloud.kubernetes.io/os=UBUNTU_18_64\n                    ibm-cloud.kubernetes.io/sgx-enabled=false\n                    ibm-cloud.kubernetes.io/worker-pool-id=49a3e8d7011b436d9b4596ba0f279008-7e5ee57\n                    ibm-cloud.kubernetes.io/worker-version=1.13.7_1526\n                    kubernetes.io/hostname=10.72.74.143\n                    privateVLAN=2643595\n                    publicVLAN=2643593\nAnnotations:        node.alpha.kubernetes.io/ttl: 0\n                    volumes.kubernetes.io/controller-managed-attach-detach: true\nCreationTimestamp:  Mon, 17 Jun 2019 21:36:37 +0000\nTaints:             \nUnschedulable:      false\nConditions:\n  Type             Status  LastHeartbeatTime                 LastTransitionTime                Reason                       Message\n  ----             ------  -----------------                 ------------------                ------                       -------\n  MemoryPressure   False   Tue, 18 Jun 2019 12:55:47 +0000   Mon, 17 Jun 2019 21:36:37 +0000   KubeletHasSufficientMemory   kubelet has sufficient memory available\n  DiskPressure     False   Tue, 18 Jun 2019 12:55:47 +0000   Mon, 17 Jun 2019 21:36:37 +0000   KubeletHasNoDiskPressure     kubelet has no disk pressure\n  PIDPressure      False   Tue, 18 Jun 2019 12:55:47 +0000   Mon, 17 Jun 2019 21:36:37 +0000   KubeletHasSufficientPID      kubelet has sufficient PID available\n  Ready            True    Tue, 18 Jun 2019 12:55:47 +0000   Mon, 17 Jun 2019 21:36:47 +0000   KubeletReady                 kubelet is posting ready status. AppArmor enabled\nAddresses:\n  InternalIP:  10.72.74.143\n  ExternalIP:  158.176.111.60\n  Hostname:    10.72.74.143\nCapacity:\n cpu:                4\n ephemeral-storage:  102685624Ki\n hugepages-1Gi:      0\n hugepages-2Mi:      0\n memory:             16419916Ki\n pods:               110\nAllocatable:\n cpu:                3910m\n ephemeral-storage:  99892574949\n hugepages-1Gi:      0\n hugepages-2Mi:      0\n memory:             13627468Ki\n pods:               110\nSystem Info:\n Machine ID:                 c061a5f0b4cb45e69da8d3f656c657bf\n System UUID:                C5FEE2E3-75D6-CCA2-2125-F4191D25A4A8\n Boot ID:                    5da3ff2c-7d82-4839-b06f-1b7fc37e00de\n Kernel Version:             4.15.0-51-generic\n OS Image:                   Ubuntu 18.04.2 LTS\n Operating System:           linux\n Architecture:               amd64\n Container Runtime Version:  containerd://1.2.6\n Kubelet Version:            v1.13.7+IKS\n Kube-Proxy Version:         v1.13.7+IKS\nProviderID:                  ibm://d18c889395112a40d2f4e3065f237a7d///49a3e8d7011b436d9b4596ba0f279008/kube-lon06-cr49a3e8d7011b436d9b4596ba0f279008-w2\nNon-terminated Pods:         (15 in total)\n  Namespace                  Name                                                       CPU Requests  CPU Limits  Memory Requests  Memory Limits  AGE\n  ---------                  ----                                                       ------------  ----------  ---------------  -------------  ---\n  default                    test-k8s-e2e-pvg-master-verification                       0 (0%)        0 (0%)      0 (0%)           0 (0%)         82m\n  heptio-sonobuoy            sonobuoy                                                   0 (0%)        0 (0%)      0 (0%)           0 (0%)         82m\n  heptio-sonobuoy            sonobuoy-e2e-job-4f826760f7504668                          0 (0%)        0 (0%)      0 (0%)           0 (0%)         81m\n  heptio-sonobuoy            sonobuoy-systemd-logs-daemon-set-562f76bc52c447d0-zt85r    0 (0%)        0 (0%)      0 (0%)           0 (0%)         81m\n  kube-system                calico-kube-controllers-54d47c87f-kwkh9                    10m (0%)      0 (0%)      25Mi (0%)        0 (0%)         15h\n  kube-system                calico-node-fw2l9                                          250m (6%)     0 (0%)      80Mi (0%)        0 (0%)         15h\n  kube-system                coredns-5545c6ddc4-dxkvs                                   100m (2%)     0 (0%)      70Mi (0%)        400Mi (3%)     15h\n  kube-system                coredns-autoscaler-5c7646547d-dshx6                        20m (0%)      0 (0%)      10Mi (0%)        0 (0%)         15h\n  kube-system                ibm-file-plugin-bf4cc7987-jwdjh                            50m (1%)      200m (5%)   100Mi (0%)       0 (0%)         15h\n  kube-system                ibm-keepalived-watcher-5z7h2                               5m (0%)       0 (0%)      10Mi (0%)        0 (0%)         15h\n  kube-system                ibm-kube-fluentd-7spm2                                     25m (0%)      300m (7%)   150Mi (1%)       1600M (11%)    15h\n  kube-system                ibm-master-proxy-static-10.72.74.143                       25m (0%)      300m (7%)   32M (0%)         512M (3%)      15h\n  kube-system                ibm-storage-watcher-64989c44d-tp68k                        50m (1%)      200m (5%)   100Mi (0%)       0 (0%)         15h\n  kube-system                kubernetes-dashboard-6cf8b975c-prz8l                       50m (1%)      0 (0%)      100Mi (0%)       0 (0%)         15h\n  kube-system                vpn-7f677b8cb5-29tf9                                       5m (0%)       0 (0%)      5Mi (0%)         0 (0%)         15h\nAllocated resources:\n  (Total limits may be over 100 percent, i.e., overcommitted.)\n  Resource           Requests       Limits\n  --------           --------       ------\n  cpu                590m (15%)     1 (25%)\n  memory             696850Ki (5%)  2472100Ki (18%)\n  ephemeral-storage  0 (0%)         0 (0%)\nEvents:              \n"
+Jun 18 12:55:50.949: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 describe namespace e2e-tests-kubectl-mq6hg'
+Jun 18 12:55:51.112: INFO: stderr: ""
+Jun 18 12:55:51.112: INFO: stdout: "Name:         e2e-tests-kubectl-mq6hg\nLabels:       e2e-framework=kubectl\n              e2e-run=097fb1c0-91bd-11e9-bce2-ae54e022189f\nAnnotations:  \nStatus:       Active\n\nNo resource quota.\n\nNo resource limits.\n"
+[AfterEach] [sig-cli] Kubectl client
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
-May 13 19:57:32.266: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "e2e-tests-replication-controller-w2csw" for this suite.
-May 13 19:57:38.352: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-May 13 19:57:38.684: INFO: namespace: e2e-tests-replication-controller-w2csw, resource: bindings, ignored listing per whitelist
-May 13 19:57:38.766: INFO: namespace e2e-tests-replication-controller-w2csw deletion completed in 6.445500618s
+Jun 18 12:55:51.112: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-tests-kubectl-mq6hg" for this suite.
+Jun 18 12:56:15.184: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 12:56:16.107: INFO: namespace: e2e-tests-kubectl-mq6hg, resource: bindings, ignored listing per whitelist
+Jun 18 12:56:17.409: INFO: namespace e2e-tests-kubectl-mq6hg deletion completed in 26.278558433s
 
-• [SLOW TEST:7.847 seconds]
-[sig-apps] ReplicationController
-/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22
-  should release no longer matching pods [Conformance]
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
+• [SLOW TEST:31.507 seconds]
+[sig-cli] Kubectl client
+/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:22
+  [k8s.io] Kubectl describe
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694
+    should check if kubectl describe prints relevant information for rc and pods  [Conformance]
+    /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 ------------------------------
-SSSSSSSSS
+SSSSSSSSSSS
 ------------------------------
 [sig-storage] Projected configMap 
-  should be consumable from pods in volume [NodeConformance] [Conformance]
+  should be consumable from pods in volume with mappings as non-root [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 [BeforeEach] [sig-storage] Projected configMap
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
 STEP: Creating a kubernetes client
-May 13 19:57:38.766: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423
+Jun 18 12:56:17.410: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
 STEP: Building a namespace api object, basename projected
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-projected-g2ms6
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-projected-bjjw8
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should be consumable from pods in volume [NodeConformance] [Conformance]
+[It] should be consumable from pods in volume with mappings as non-root [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-STEP: Creating configMap with name projected-configmap-test-volume-5c5be95c-75b9-11e9-8f67-2632f168be36
+STEP: Creating configMap with name projected-configmap-test-volume-map-7678e739-91c8-11e9-bce2-ae54e022189f
 STEP: Creating a pod to test consume configMaps
-May 13 19:57:39.071: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-5c5dd29b-75b9-11e9-8f67-2632f168be36" in namespace "e2e-tests-projected-g2ms6" to be "success or failure"
-May 13 19:57:39.080: INFO: Pod "pod-projected-configmaps-5c5dd29b-75b9-11e9-8f67-2632f168be36": Phase="Pending", Reason="", readiness=false. Elapsed: 8.72581ms
-May 13 19:57:41.088: INFO: Pod "pod-projected-configmaps-5c5dd29b-75b9-11e9-8f67-2632f168be36": Phase="Pending", Reason="", readiness=false. Elapsed: 2.017158439s
-May 13 19:57:43.096: INFO: Pod "pod-projected-configmaps-5c5dd29b-75b9-11e9-8f67-2632f168be36": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.025501505s
+Jun 18 12:56:17.903: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-767b272d-91c8-11e9-bce2-ae54e022189f" in namespace "e2e-tests-projected-bjjw8" to be "success or failure"
+Jun 18 12:56:17.918: INFO: Pod "pod-projected-configmaps-767b272d-91c8-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 14.430282ms
+Jun 18 12:56:19.932: INFO: Pod "pod-projected-configmaps-767b272d-91c8-11e9-bce2-ae54e022189f": Phase="Running", Reason="", readiness=true. Elapsed: 2.029126472s
+Jun 18 12:56:21.970: INFO: Pod "pod-projected-configmaps-767b272d-91c8-11e9-bce2-ae54e022189f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.066872994s
 STEP: Saw pod success
-May 13 19:57:43.096: INFO: Pod "pod-projected-configmaps-5c5dd29b-75b9-11e9-8f67-2632f168be36" satisfied condition "success or failure"
-May 13 19:57:43.104: INFO: Trying to get logs from node 10.170.219.177 pod pod-projected-configmaps-5c5dd29b-75b9-11e9-8f67-2632f168be36 container projected-configmap-volume-test: 
+Jun 18 12:56:21.970: INFO: Pod "pod-projected-configmaps-767b272d-91c8-11e9-bce2-ae54e022189f" satisfied condition "success or failure"
+Jun 18 12:56:21.987: INFO: Trying to get logs from node 10.72.74.143 pod pod-projected-configmaps-767b272d-91c8-11e9-bce2-ae54e022189f container projected-configmap-volume-test: 
 STEP: delete the pod
-May 13 19:57:43.220: INFO: Waiting for pod pod-projected-configmaps-5c5dd29b-75b9-11e9-8f67-2632f168be36 to disappear
-May 13 19:57:43.228: INFO: Pod pod-projected-configmaps-5c5dd29b-75b9-11e9-8f67-2632f168be36 no longer exists
+Jun 18 12:56:22.085: INFO: Waiting for pod pod-projected-configmaps-767b272d-91c8-11e9-bce2-ae54e022189f to disappear
+Jun 18 12:56:22.101: INFO: Pod pod-projected-configmaps-767b272d-91c8-11e9-bce2-ae54e022189f no longer exists
 [AfterEach] [sig-storage] Projected configMap
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
-May 13 19:57:43.228: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "e2e-tests-projected-g2ms6" for this suite.
-May 13 19:57:49.265: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-May 13 19:57:49.408: INFO: namespace: e2e-tests-projected-g2ms6, resource: bindings, ignored listing per whitelist
-May 13 19:57:49.566: INFO: namespace e2e-tests-projected-g2ms6 deletion completed in 6.330140424s
+Jun 18 12:56:22.101: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-tests-projected-bjjw8" for this suite.
+Jun 18 12:56:30.168: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 12:56:30.561: INFO: namespace: e2e-tests-projected-bjjw8, resource: bindings, ignored listing per whitelist
+Jun 18 12:56:30.696: INFO: namespace e2e-tests-projected-bjjw8 deletion completed in 8.576851145s
 
-• [SLOW TEST:10.800 seconds]
+• [SLOW TEST:13.287 seconds]
 [sig-storage] Projected configMap
 /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_configmap.go:34
-  should be consumable from pods in volume [NodeConformance] [Conformance]
+  should be consumable from pods in volume with mappings as non-root [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 ------------------------------
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+SSSSSSSSSSSSSSSS
 ------------------------------
-[sig-cli] Kubectl client [k8s.io] Kubectl run default 
-  should create an rc or deployment from an image  [Conformance]
+[sig-api-machinery] Watchers 
+  should be able to start watching from a specific resource version [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-[BeforeEach] [sig-cli] Kubectl client
+[BeforeEach] [sig-api-machinery] Watchers
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
 STEP: Creating a kubernetes client
-May 13 19:57:49.567: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423
-STEP: Building a namespace api object, basename kubectl
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-kubectl-ckp8r
+Jun 18 12:56:30.696: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
+STEP: Building a namespace api object, basename watch
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-watch-9shpd
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:243
-[BeforeEach] [k8s.io] Kubectl run default
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1262
-[It] should create an rc or deployment from an image  [Conformance]
+[It] should be able to start watching from a specific resource version [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-STEP: running the image docker.io/library/nginx:1.14-alpine
-May 13 19:57:49.847: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 run e2e-test-nginx-deployment --image=docker.io/library/nginx:1.14-alpine --namespace=e2e-tests-kubectl-ckp8r'
-May 13 19:57:49.967: INFO: stderr: "kubectl run --generator=deployment/apps.v1 is DEPRECATED and will be removed in a future version. Use kubectl run --generator=run-pod/v1 or kubectl create instead.\n"
-May 13 19:57:49.967: INFO: stdout: "deployment.apps/e2e-test-nginx-deployment created\n"
-STEP: verifying the pod controlled by e2e-test-nginx-deployment gets created
-[AfterEach] [k8s.io] Kubectl run default
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1268
-May 13 19:57:51.983: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 delete deployment e2e-test-nginx-deployment --namespace=e2e-tests-kubectl-ckp8r'
-May 13 19:57:52.101: INFO: stderr: ""
-May 13 19:57:52.101: INFO: stdout: "deployment.extensions \"e2e-test-nginx-deployment\" deleted\n"
-[AfterEach] [sig-cli] Kubectl client
+STEP: creating a new configmap
+STEP: modifying the configmap once
+STEP: modifying the configmap a second time
+STEP: deleting the configmap
+STEP: creating a watch on configmaps from the resource version returned by the first update
+STEP: Expecting to observe notifications for all changes to the configmap after the first update
+Jun 18 12:56:31.980: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-resource-version,GenerateName:,Namespace:e2e-tests-watch-9shpd,SelfLink:/api/v1/namespaces/e2e-tests-watch-9shpd/configmaps/e2e-watch-test-resource-version,UID:7ed18a8b-91c8-11e9-bf44-fa6f350b29f0,ResourceVersion:108045,Generation:0,CreationTimestamp:2019-06-18 12:56:31 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: from-resource-version,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},}
+Jun 18 12:56:31.980: INFO: Got : DELETED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-resource-version,GenerateName:,Namespace:e2e-tests-watch-9shpd,SelfLink:/api/v1/namespaces/e2e-tests-watch-9shpd/configmaps/e2e-watch-test-resource-version,UID:7ed18a8b-91c8-11e9-bf44-fa6f350b29f0,ResourceVersion:108046,Generation:0,CreationTimestamp:2019-06-18 12:56:31 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: from-resource-version,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},}
+[AfterEach] [sig-api-machinery] Watchers
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
-May 13 19:57:52.101: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "e2e-tests-kubectl-ckp8r" for this suite.
-May 13 19:58:16.157: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-May 13 19:58:16.343: INFO: namespace: e2e-tests-kubectl-ckp8r, resource: bindings, ignored listing per whitelist
-May 13 19:58:16.439: INFO: namespace e2e-tests-kubectl-ckp8r deletion completed in 24.3189894s
+Jun 18 12:56:31.980: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-tests-watch-9shpd" for this suite.
+Jun 18 12:56:38.047: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 12:56:38.521: INFO: namespace: e2e-tests-watch-9shpd, resource: bindings, ignored listing per whitelist
+Jun 18 12:56:38.609: INFO: namespace e2e-tests-watch-9shpd deletion completed in 6.610391087s
 
-• [SLOW TEST:26.872 seconds]
-[sig-cli] Kubectl client
-/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:22
-  [k8s.io] Kubectl run default
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694
-    should create an rc or deployment from an image  [Conformance]
-    /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
+• [SLOW TEST:7.913 seconds]
+[sig-api-machinery] Watchers
+/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:22
+  should be able to start watching from a specific resource version [Conformance]
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
+------------------------------
+S
+------------------------------
+[sig-storage] EmptyDir volumes 
+  should support (non-root,0644,default) [NodeConformance] [Conformance]
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
+[BeforeEach] [sig-storage] EmptyDir volumes
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
+STEP: Creating a kubernetes client
+Jun 18 12:56:38.610: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
+STEP: Building a namespace api object, basename emptydir
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-emptydir-bqzcl
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should support (non-root,0644,default) [NodeConformance] [Conformance]
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
+STEP: Creating a pod to test emptydir 0644 on node default medium
+Jun 18 12:56:39.070: INFO: Waiting up to 5m0s for pod "pod-83192200-91c8-11e9-bce2-ae54e022189f" in namespace "e2e-tests-emptydir-bqzcl" to be "success or failure"
+Jun 18 12:56:39.086: INFO: Pod "pod-83192200-91c8-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 15.768357ms
+Jun 18 12:56:41.102: INFO: Pod "pod-83192200-91c8-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 2.031752491s
+Jun 18 12:56:43.136: INFO: Pod "pod-83192200-91c8-11e9-bce2-ae54e022189f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.066057037s
+STEP: Saw pod success
+Jun 18 12:56:43.136: INFO: Pod "pod-83192200-91c8-11e9-bce2-ae54e022189f" satisfied condition "success or failure"
+Jun 18 12:56:43.151: INFO: Trying to get logs from node 10.72.74.144 pod pod-83192200-91c8-11e9-bce2-ae54e022189f container test-container: 
+STEP: delete the pod
+Jun 18 12:56:43.234: INFO: Waiting for pod pod-83192200-91c8-11e9-bce2-ae54e022189f to disappear
+Jun 18 12:56:43.248: INFO: Pod pod-83192200-91c8-11e9-bce2-ae54e022189f no longer exists
+[AfterEach] [sig-storage] EmptyDir volumes
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
+Jun 18 12:56:43.248: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-tests-emptydir-bqzcl" for this suite.
+Jun 18 12:56:49.324: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 12:56:49.821: INFO: namespace: e2e-tests-emptydir-bqzcl, resource: bindings, ignored listing per whitelist
+Jun 18 12:56:49.847: INFO: namespace e2e-tests-emptydir-bqzcl deletion completed in 6.577390308s
+
+• [SLOW TEST:11.238 seconds]
+[sig-storage] EmptyDir volumes
+/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:40
+  should support (non-root,0644,default) [NodeConformance] [Conformance]
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 ------------------------------
-SSSSSSSSSSSSSSSSS
+SSSSSSS
 ------------------------------
 [sig-storage] Projected downwardAPI 
-  should provide container's memory request [NodeConformance] [Conformance]
+  should set mode on item file [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 [BeforeEach] [sig-storage] Projected downwardAPI
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
 STEP: Creating a kubernetes client
-May 13 19:58:16.439: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423
+Jun 18 12:56:49.849: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
 STEP: Building a namespace api object, basename projected
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-projected-ngxn8
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-projected-lzpbx
 STEP: Waiting for a default service account to be provisioned in namespace
 [BeforeEach] [sig-storage] Projected downwardAPI
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:39
-[It] should provide container's memory request [NodeConformance] [Conformance]
+[It] should set mode on item file [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 STEP: Creating a pod to test downward API volume plugin
-May 13 19:58:16.732: INFO: Waiting up to 5m0s for pod "downwardapi-volume-72d018be-75b9-11e9-8f67-2632f168be36" in namespace "e2e-tests-projected-ngxn8" to be "success or failure"
-May 13 19:58:16.740: INFO: Pod "downwardapi-volume-72d018be-75b9-11e9-8f67-2632f168be36": Phase="Pending", Reason="", readiness=false. Elapsed: 7.84252ms
-May 13 19:58:18.748: INFO: Pod "downwardapi-volume-72d018be-75b9-11e9-8f67-2632f168be36": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.015936379s
+Jun 18 12:56:50.346: INFO: Waiting up to 5m0s for pod "downwardapi-volume-89d14c3d-91c8-11e9-bce2-ae54e022189f" in namespace "e2e-tests-projected-lzpbx" to be "success or failure"
+Jun 18 12:56:50.359: INFO: Pod "downwardapi-volume-89d14c3d-91c8-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 13.142983ms
+Jun 18 12:56:53.132: INFO: Pod "downwardapi-volume-89d14c3d-91c8-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 2.785909937s
+Jun 18 12:56:55.171: INFO: Pod "downwardapi-volume-89d14c3d-91c8-11e9-bce2-ae54e022189f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.824849366s
 STEP: Saw pod success
-May 13 19:58:18.748: INFO: Pod "downwardapi-volume-72d018be-75b9-11e9-8f67-2632f168be36" satisfied condition "success or failure"
-May 13 19:58:18.756: INFO: Trying to get logs from node 10.170.219.153 pod downwardapi-volume-72d018be-75b9-11e9-8f67-2632f168be36 container client-container: 
+Jun 18 12:56:55.171: INFO: Pod "downwardapi-volume-89d14c3d-91c8-11e9-bce2-ae54e022189f" satisfied condition "success or failure"
+Jun 18 12:56:55.186: INFO: Trying to get logs from node 10.72.74.144 pod downwardapi-volume-89d14c3d-91c8-11e9-bce2-ae54e022189f container client-container: 
 STEP: delete the pod
-May 13 19:58:18.853: INFO: Waiting for pod downwardapi-volume-72d018be-75b9-11e9-8f67-2632f168be36 to disappear
-May 13 19:58:18.865: INFO: Pod downwardapi-volume-72d018be-75b9-11e9-8f67-2632f168be36 no longer exists
+Jun 18 12:56:55.263: INFO: Waiting for pod downwardapi-volume-89d14c3d-91c8-11e9-bce2-ae54e022189f to disappear
+Jun 18 12:56:55.277: INFO: Pod downwardapi-volume-89d14c3d-91c8-11e9-bce2-ae54e022189f no longer exists
 [AfterEach] [sig-storage] Projected downwardAPI
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
-May 13 19:58:18.865: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "e2e-tests-projected-ngxn8" for this suite.
-May 13 19:58:24.902: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-May 13 19:58:25.189: INFO: namespace: e2e-tests-projected-ngxn8, resource: bindings, ignored listing per whitelist
-May 13 19:58:25.224: INFO: namespace e2e-tests-projected-ngxn8 deletion completed in 6.350731671s
+Jun 18 12:56:55.277: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-tests-projected-lzpbx" for this suite.
+Jun 18 12:57:01.363: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 12:57:01.577: INFO: namespace: e2e-tests-projected-lzpbx, resource: bindings, ignored listing per whitelist
+Jun 18 12:57:01.854: INFO: namespace e2e-tests-projected-lzpbx deletion completed in 6.554293096s
 
-• [SLOW TEST:8.785 seconds]
+• [SLOW TEST:12.006 seconds]
 [sig-storage] Projected downwardAPI
 /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:33
-  should provide container's memory request [NodeConformance] [Conformance]
+  should set mode on item file [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 ------------------------------
 SS
 ------------------------------
-[sig-storage] Downward API volume 
-  should update labels on modification [NodeConformance] [Conformance]
+[sig-network] Proxy version v1 
+  should proxy logs on node with explicit kubelet port using proxy subresource  [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-[BeforeEach] [sig-storage] Downward API volume
+[BeforeEach] version v1
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
 STEP: Creating a kubernetes client
-May 13 19:58:25.224: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423
-STEP: Building a namespace api object, basename downward-api
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-downward-api-w2dx7
+Jun 18 12:57:01.855: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
+STEP: Building a namespace api object, basename proxy
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-proxy-bx2jn
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-storage] Downward API volume
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:39
-[It] should update labels on modification [NodeConformance] [Conformance]
+[It] should proxy logs on node with explicit kubelet port using proxy subresource  [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-STEP: Creating the pod
-May 13 19:58:28.067: INFO: Successfully updated pod "labelsupdate780b3f06-75b9-11e9-8f67-2632f168be36"
-[AfterEach] [sig-storage] Downward API volume
+Jun 18 12:57:02.377: INFO: (0) /api/v1/nodes/10.72.74.143:10250/proxy/logs/: 
+alternatives.log
+apt/
+at/
+alternatives.log
+apt/
+at/
+alternatives.log
+apt/
+at/
+alternatives.log
+apt/
+at/
+alternatives.log
+apt/
+at/
+alternatives.log
+apt/
+at/
+alternatives.log
+apt/
+at/
+alternatives.log
+apt/
+at/
+alternatives.log
+apt/
+at/
+alternatives.log
+apt/
+at/
+alternatives.log
+apt/
+at/
+alternatives.log
+apt/
+at/
+alternatives.log
+apt/
+at/
+alternatives.log
+apt/
+at/
+alternatives.log
+apt/
+at/
+alternatives.log
+apt/
+at/
+alternatives.log
+apt/
+at/
+alternatives.log
+apt/
+at/
+alternatives.log
+apt/
+at/
+alternatives.log
+apt/
+at/>> kubeConfig: /tmp/kubeconfig-555138423
-STEP: Building a namespace api object, basename downward-api
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-downward-api-rvmjf
+Jun 18 12:57:09.530: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
+STEP: Building a namespace api object, basename emptydir
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-emptydir-5kkxc
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-storage] Downward API volume
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:39
-[It] should provide container's cpu limit [NodeConformance] [Conformance]
+[It] should support (root,0666,default) [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-STEP: Creating a pod to test downward API volume plugin
-May 13 19:58:54.785: INFO: Waiting up to 5m0s for pod "downwardapi-volume-897e85e0-75b9-11e9-8f67-2632f168be36" in namespace "e2e-tests-downward-api-rvmjf" to be "success or failure"
-May 13 19:58:54.794: INFO: Pod "downwardapi-volume-897e85e0-75b9-11e9-8f67-2632f168be36": Phase="Pending", Reason="", readiness=false. Elapsed: 8.646249ms
-May 13 19:58:56.803: INFO: Pod "downwardapi-volume-897e85e0-75b9-11e9-8f67-2632f168be36": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.017610568s
+STEP: Creating a pod to test emptydir 0666 on node default medium
+Jun 18 12:57:10.063: INFO: Waiting up to 5m0s for pod "pod-95920aa7-91c8-11e9-bce2-ae54e022189f" in namespace "e2e-tests-emptydir-5kkxc" to be "success or failure"
+Jun 18 12:57:10.078: INFO: Pod "pod-95920aa7-91c8-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 15.586793ms
+Jun 18 12:57:12.093: INFO: Pod "pod-95920aa7-91c8-11e9-bce2-ae54e022189f": Phase="Running", Reason="", readiness=true. Elapsed: 2.03023011s
+Jun 18 12:57:14.108: INFO: Pod "pod-95920aa7-91c8-11e9-bce2-ae54e022189f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.045221686s
 STEP: Saw pod success
-May 13 19:58:56.803: INFO: Pod "downwardapi-volume-897e85e0-75b9-11e9-8f67-2632f168be36" satisfied condition "success or failure"
-May 13 19:58:56.810: INFO: Trying to get logs from node 10.170.219.153 pod downwardapi-volume-897e85e0-75b9-11e9-8f67-2632f168be36 container client-container: 
+Jun 18 12:57:14.108: INFO: Pod "pod-95920aa7-91c8-11e9-bce2-ae54e022189f" satisfied condition "success or failure"
+Jun 18 12:57:14.122: INFO: Trying to get logs from node 10.72.74.143 pod pod-95920aa7-91c8-11e9-bce2-ae54e022189f container test-container: 
 STEP: delete the pod
-May 13 19:58:56.848: INFO: Waiting for pod downwardapi-volume-897e85e0-75b9-11e9-8f67-2632f168be36 to disappear
-May 13 19:58:56.857: INFO: Pod downwardapi-volume-897e85e0-75b9-11e9-8f67-2632f168be36 no longer exists
-[AfterEach] [sig-storage] Downward API volume
+Jun 18 12:57:14.196: INFO: Waiting for pod pod-95920aa7-91c8-11e9-bce2-ae54e022189f to disappear
+Jun 18 12:57:14.212: INFO: Pod pod-95920aa7-91c8-11e9-bce2-ae54e022189f no longer exists
+[AfterEach] [sig-storage] EmptyDir volumes
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
-May 13 19:58:56.857: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "e2e-tests-downward-api-rvmjf" for this suite.
-May 13 19:59:02.947: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-May 13 19:59:03.029: INFO: namespace: e2e-tests-downward-api-rvmjf, resource: bindings, ignored listing per whitelist
-May 13 19:59:03.182: INFO: namespace e2e-tests-downward-api-rvmjf deletion completed in 6.262192467s
+Jun 18 12:57:14.212: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-tests-emptydir-5kkxc" for this suite.
+Jun 18 12:57:20.290: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 12:57:20.836: INFO: namespace: e2e-tests-emptydir-5kkxc, resource: bindings, ignored listing per whitelist
+Jun 18 12:57:20.851: INFO: namespace e2e-tests-emptydir-5kkxc deletion completed in 6.617532075s
 
-• [SLOW TEST:8.691 seconds]
-[sig-storage] Downward API volume
-/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:34
-  should provide container's cpu limit [NodeConformance] [Conformance]
+• [SLOW TEST:11.320 seconds]
+[sig-storage] EmptyDir volumes
+/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:40
+  should support (root,0666,default) [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 ------------------------------
-SSSSSSSSSSSSSSS
+SSSS
 ------------------------------
-[k8s.io] Pods 
-  should support remote command execution over websockets [NodeConformance] [Conformance]
+[sig-storage] Projected downwardAPI 
+  should provide container's cpu request [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-[BeforeEach] [k8s.io] Pods
+[BeforeEach] [sig-storage] Projected downwardAPI
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
 STEP: Creating a kubernetes client
-May 13 19:59:03.183: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423
-STEP: Building a namespace api object, basename pods
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-pods-h6vbc
+Jun 18 12:57:20.852: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
+STEP: Building a namespace api object, basename projected
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-projected-xbh2k
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [k8s.io] Pods
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:132
-[It] should support remote command execution over websockets [NodeConformance] [Conformance]
+[BeforeEach] [sig-storage] Projected downwardAPI
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:39
+[It] should provide container's cpu request [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-May 13 19:59:03.563: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423
-STEP: creating the pod
-STEP: submitting the pod to kubernetes
-[AfterEach] [k8s.io] Pods
+STEP: Creating a pod to test downward API volume plugin
+Jun 18 12:57:21.340: INFO: Waiting up to 5m0s for pod "downwardapi-volume-9c4a6454-91c8-11e9-bce2-ae54e022189f" in namespace "e2e-tests-projected-xbh2k" to be "success or failure"
+Jun 18 12:57:21.354: INFO: Pod "downwardapi-volume-9c4a6454-91c8-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 14.306969ms
+Jun 18 12:57:23.372: INFO: Pod "downwardapi-volume-9c4a6454-91c8-11e9-bce2-ae54e022189f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.032063502s
+STEP: Saw pod success
+Jun 18 12:57:23.372: INFO: Pod "downwardapi-volume-9c4a6454-91c8-11e9-bce2-ae54e022189f" satisfied condition "success or failure"
+Jun 18 12:57:23.386: INFO: Trying to get logs from node 10.72.74.149 pod downwardapi-volume-9c4a6454-91c8-11e9-bce2-ae54e022189f container client-container: 
+STEP: delete the pod
+Jun 18 12:57:23.462: INFO: Waiting for pod downwardapi-volume-9c4a6454-91c8-11e9-bce2-ae54e022189f to disappear
+Jun 18 12:57:23.476: INFO: Pod downwardapi-volume-9c4a6454-91c8-11e9-bce2-ae54e022189f no longer exists
+[AfterEach] [sig-storage] Projected downwardAPI
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
-May 13 19:59:05.762: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "e2e-tests-pods-h6vbc" for this suite.
-May 13 19:59:45.798: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-May 13 19:59:45.880: INFO: namespace: e2e-tests-pods-h6vbc, resource: bindings, ignored listing per whitelist
-May 13 19:59:46.049: INFO: namespace e2e-tests-pods-h6vbc deletion completed in 40.279156052s
+Jun 18 12:57:23.476: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-tests-projected-xbh2k" for this suite.
+Jun 18 12:57:29.567: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 12:57:29.982: INFO: namespace: e2e-tests-projected-xbh2k, resource: bindings, ignored listing per whitelist
+Jun 18 12:57:30.085: INFO: namespace e2e-tests-projected-xbh2k deletion completed in 6.58691693s
 
-• [SLOW TEST:42.866 seconds]
-[k8s.io] Pods
-/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694
-  should support remote command execution over websockets [NodeConformance] [Conformance]
+• [SLOW TEST:9.233 seconds]
+[sig-storage] Projected downwardAPI
+/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:33
+  should provide container's cpu request [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 ------------------------------
-SSSS
+SSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-cli] Kubectl client [k8s.io] Proxy server 
-  should support proxy with --port 0  [Conformance]
+[sig-scheduling] SchedulerPredicates [Serial] 
+  validates that NodeSelector is respected if not matching  [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-[BeforeEach] [sig-cli] Kubectl client
+[BeforeEach] [sig-scheduling] SchedulerPredicates [Serial]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
 STEP: Creating a kubernetes client
-May 13 19:59:46.050: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423
-STEP: Building a namespace api object, basename kubectl
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-kubectl-d24rb
+Jun 18 12:57:30.086: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
+STEP: Building a namespace api object, basename sched-pred
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-sched-pred-wwz6v
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:243
-[It] should support proxy with --port 0  [Conformance]
+[BeforeEach] [sig-scheduling] SchedulerPredicates [Serial]
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:79
+Jun 18 12:57:30.519: INFO: Waiting up to 1m0s for all (but 0) nodes to be ready
+Jun 18 12:57:30.553: INFO: Waiting for terminating namespaces to be deleted...
+Jun 18 12:57:30.566: INFO: 
+Logging pods the kubelet thinks is on node 10.72.74.143 before test
+Jun 18 12:57:30.630: INFO: calico-node-fw2l9 from kube-system started at 2019-06-17 21:36:37 +0000 UTC (1 container statuses recorded)
+Jun 18 12:57:30.630: INFO: 	Container calico-node ready: true, restart count 0
+Jun 18 12:57:30.630: INFO: coredns-autoscaler-5c7646547d-dshx6 from kube-system started at 2019-06-17 21:36:47 +0000 UTC (1 container statuses recorded)
+Jun 18 12:57:30.630: INFO: 	Container autoscaler ready: true, restart count 0
+Jun 18 12:57:30.630: INFO: ibm-master-proxy-static-10.72.74.143 from kube-system started at  (0 container statuses recorded)
+Jun 18 12:57:30.630: INFO: ibm-keepalived-watcher-5z7h2 from kube-system started at 2019-06-17 21:36:37 +0000 UTC (1 container statuses recorded)
+Jun 18 12:57:30.630: INFO: 	Container keepalived-watcher ready: true, restart count 0
+Jun 18 12:57:30.630: INFO: kubernetes-dashboard-6cf8b975c-prz8l from kube-system started at 2019-06-17 21:36:47 +0000 UTC (1 container statuses recorded)
+Jun 18 12:57:30.631: INFO: 	Container kubernetes-dashboard ready: true, restart count 0
+Jun 18 12:57:30.631: INFO: coredns-5545c6ddc4-dxkvs from kube-system started at 2019-06-17 21:36:47 +0000 UTC (1 container statuses recorded)
+Jun 18 12:57:30.631: INFO: 	Container coredns ready: true, restart count 0
+Jun 18 12:57:30.631: INFO: calico-kube-controllers-54d47c87f-kwkh9 from kube-system started at 2019-06-17 21:36:47 +0000 UTC (1 container statuses recorded)
+Jun 18 12:57:30.631: INFO: 	Container calico-kube-controllers ready: true, restart count 0
+Jun 18 12:57:30.631: INFO: sonobuoy-systemd-logs-daemon-set-562f76bc52c447d0-zt85r from heptio-sonobuoy started at 2019-06-18 11:33:59 +0000 UTC (2 container statuses recorded)
+Jun 18 12:57:30.631: INFO: 	Container sonobuoy-worker ready: true, restart count 1
+Jun 18 12:57:30.631: INFO: 	Container systemd-logs ready: true, restart count 1
+Jun 18 12:57:30.631: INFO: vpn-7f677b8cb5-29tf9 from kube-system started at 2019-06-17 21:36:47 +0000 UTC (1 container statuses recorded)
+Jun 18 12:57:30.631: INFO: 	Container vpn ready: true, restart count 0
+Jun 18 12:57:30.631: INFO: test-k8s-e2e-pvg-master-verification from default started at 2019-06-18 11:33:43 +0000 UTC (1 container statuses recorded)
+Jun 18 12:57:30.631: INFO: 	Container test-k8s-e2e-pvg-master-verification ready: true, restart count 0
+Jun 18 12:57:30.631: INFO: ibm-kube-fluentd-7spm2 from kube-system started at 2019-06-17 21:43:21 +0000 UTC (1 container statuses recorded)
+Jun 18 12:57:30.631: INFO: 	Container fluentd ready: true, restart count 0
+Jun 18 12:57:30.631: INFO: ibm-file-plugin-bf4cc7987-jwdjh from kube-system started at 2019-06-17 21:36:47 +0000 UTC (1 container statuses recorded)
+Jun 18 12:57:30.631: INFO: 	Container ibm-file-plugin-container ready: true, restart count 0
+Jun 18 12:57:30.631: INFO: ibm-storage-watcher-64989c44d-tp68k from kube-system started at 2019-06-17 21:36:47 +0000 UTC (1 container statuses recorded)
+Jun 18 12:57:30.631: INFO: 	Container ibm-storage-watcher-container ready: true, restart count 0
+Jun 18 12:57:30.631: INFO: sonobuoy from heptio-sonobuoy started at 2019-06-18 11:33:50 +0000 UTC (1 container statuses recorded)
+Jun 18 12:57:30.631: INFO: 	Container kube-sonobuoy ready: true, restart count 0
+Jun 18 12:57:30.631: INFO: sonobuoy-e2e-job-4f826760f7504668 from heptio-sonobuoy started at 2019-06-18 11:33:59 +0000 UTC (2 container statuses recorded)
+Jun 18 12:57:30.631: INFO: 	Container e2e ready: true, restart count 0
+Jun 18 12:57:30.631: INFO: 	Container sonobuoy-worker ready: true, restart count 0
+Jun 18 12:57:30.631: INFO: 
+Logging pods the kubelet thinks is on node 10.72.74.144 before test
+Jun 18 12:57:30.697: INFO: coredns-5545c6ddc4-4s87g from kube-system started at 2019-06-17 21:37:04 +0000 UTC (1 container statuses recorded)
+Jun 18 12:57:30.697: INFO: 	Container coredns ready: true, restart count 0
+Jun 18 12:57:30.697: INFO: ibm-kube-fluentd-g5hgb from kube-system started at 2019-06-17 21:43:21 +0000 UTC (1 container statuses recorded)
+Jun 18 12:57:30.697: INFO: 	Container fluentd ready: true, restart count 0
+Jun 18 12:57:30.697: INFO: sonobuoy-systemd-logs-daemon-set-562f76bc52c447d0-ct76c from heptio-sonobuoy started at 2019-06-18 11:33:59 +0000 UTC (2 container statuses recorded)
+Jun 18 12:57:30.697: INFO: 	Container sonobuoy-worker ready: true, restart count 1
+Jun 18 12:57:30.697: INFO: 	Container systemd-logs ready: true, restart count 1
+Jun 18 12:57:30.697: INFO: ibm-cloud-provider-ip-158-176-120-130-699ff5cfd-z4hhb from ibm-system started at 2019-06-17 21:40:39 +0000 UTC (1 container statuses recorded)
+Jun 18 12:57:30.697: INFO: 	Container ibm-cloud-provider-ip-158-176-120-130 ready: true, restart count 0
+Jun 18 12:57:30.697: INFO: public-cr49a3e8d7011b436d9b4596ba0f279008-alb1-778b7ff477-tpktg from kube-system started at 2019-06-17 21:41:04 +0000 UTC (4 container statuses recorded)
+Jun 18 12:57:30.697: INFO: 	Container ingress-auth-1 ready: true, restart count 0
+Jun 18 12:57:30.697: INFO: 	Container ingress-auth-2 ready: true, restart count 0
+Jun 18 12:57:30.697: INFO: 	Container ingress-auth-3 ready: true, restart count 0
+Jun 18 12:57:30.697: INFO: 	Container nginx-ingress ready: true, restart count 0
+Jun 18 12:57:30.697: INFO: ibm-master-proxy-static-10.72.74.144 from kube-system started at  (0 container statuses recorded)
+Jun 18 12:57:30.697: INFO: calico-node-rptvs from kube-system started at 2019-06-17 21:36:43 +0000 UTC (1 container statuses recorded)
+Jun 18 12:57:30.697: INFO: 	Container calico-node ready: true, restart count 0
+Jun 18 12:57:30.697: INFO: ibm-keepalived-watcher-drbmt from kube-system started at 2019-06-17 21:36:43 +0000 UTC (1 container statuses recorded)
+Jun 18 12:57:30.697: INFO: 	Container keepalived-watcher ready: true, restart count 0
+Jun 18 12:57:30.697: INFO: 
+Logging pods the kubelet thinks is on node 10.72.74.149 before test
+Jun 18 12:57:30.748: INFO: sonobuoy-systemd-logs-daemon-set-562f76bc52c447d0-btfpp from heptio-sonobuoy started at 2019-06-18 11:33:59 +0000 UTC (2 container statuses recorded)
+Jun 18 12:57:30.748: INFO: 	Container sonobuoy-worker ready: true, restart count 1
+Jun 18 12:57:30.748: INFO: 	Container systemd-logs ready: true, restart count 1
+Jun 18 12:57:30.748: INFO: ibm-cloud-provider-ip-158-176-120-130-699ff5cfd-td8hg from ibm-system started at 2019-06-17 21:40:39 +0000 UTC (1 container statuses recorded)
+Jun 18 12:57:30.748: INFO: 	Container ibm-cloud-provider-ip-158-176-120-130 ready: true, restart count 0
+Jun 18 12:57:30.748: INFO: public-cr49a3e8d7011b436d9b4596ba0f279008-alb1-778b7ff477-sxttq from kube-system started at 2019-06-17 21:41:04 +0000 UTC (4 container statuses recorded)
+Jun 18 12:57:30.748: INFO: 	Container ingress-auth-1 ready: true, restart count 0
+Jun 18 12:57:30.748: INFO: 	Container ingress-auth-2 ready: true, restart count 0
+Jun 18 12:57:30.748: INFO: 	Container ingress-auth-3 ready: true, restart count 0
+Jun 18 12:57:30.748: INFO: 	Container nginx-ingress ready: true, restart count 0
+Jun 18 12:57:30.748: INFO: ibm-master-proxy-static-10.72.74.149 from kube-system started at  (0 container statuses recorded)
+Jun 18 12:57:30.748: INFO: calico-node-4pqtj from kube-system started at 2019-06-17 21:36:50 +0000 UTC (1 container statuses recorded)
+Jun 18 12:57:30.748: INFO: 	Container calico-node ready: true, restart count 0
+Jun 18 12:57:30.748: INFO: metrics-server-6ccf788d5b-6gwxm from kube-system started at 2019-06-17 21:37:11 +0000 UTC (2 container statuses recorded)
+Jun 18 12:57:30.748: INFO: 	Container metrics-server ready: true, restart count 0
+Jun 18 12:57:30.748: INFO: 	Container metrics-server-nanny ready: true, restart count 0
+Jun 18 12:57:30.748: INFO: ibm-keepalived-watcher-6846v from kube-system started at 2019-06-17 21:36:50 +0000 UTC (1 container statuses recorded)
+Jun 18 12:57:30.748: INFO: 	Container keepalived-watcher ready: true, restart count 0
+Jun 18 12:57:30.748: INFO: ibm-kube-fluentd-c6kth from kube-system started at 2019-06-17 21:43:21 +0000 UTC (1 container statuses recorded)
+Jun 18 12:57:30.748: INFO: 	Container fluentd ready: true, restart count 0
+[It] validates that NodeSelector is respected if not matching  [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-STEP: starting the proxy server
-May 13 19:59:46.329: INFO: Asynchronously running '/usr/local/bin/kubectl kubectl --kubeconfig=/tmp/kubeconfig-555138423 proxy -p 0 --disable-filter'
-STEP: curling proxy /api/ output
-[AfterEach] [sig-cli] Kubectl client
+STEP: Trying to schedule Pod with nonempty NodeSelector.
+STEP: Considering event: 
+Type = [Warning], Name = [restricted-pod.15a94c4fa5f64d11], Reason = [FailedScheduling], Message = [0/3 nodes are available: 3 node(s) didn't match node selector.]
+[AfterEach] [sig-scheduling] SchedulerPredicates [Serial]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
-May 13 19:59:46.408: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "e2e-tests-kubectl-d24rb" for this suite.
-May 13 19:59:52.527: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-May 13 19:59:52.713: INFO: namespace: e2e-tests-kubectl-d24rb, resource: bindings, ignored listing per whitelist
-May 13 19:59:52.759: INFO: namespace e2e-tests-kubectl-d24rb deletion completed in 6.342470518s
+Jun 18 12:57:31.844: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-tests-sched-pred-wwz6v" for this suite.
+Jun 18 12:57:37.918: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 12:57:38.034: INFO: namespace: e2e-tests-sched-pred-wwz6v, resource: bindings, ignored listing per whitelist
+Jun 18 12:57:38.471: INFO: namespace e2e-tests-sched-pred-wwz6v deletion completed in 6.608375963s
+[AfterEach] [sig-scheduling] SchedulerPredicates [Serial]
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:70
 
-• [SLOW TEST:6.709 seconds]
-[sig-cli] Kubectl client
-/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:22
-  [k8s.io] Proxy server
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694
-    should support proxy with --port 0  [Conformance]
-    /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
+• [SLOW TEST:8.385 seconds]
+[sig-scheduling] SchedulerPredicates [Serial]
+/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/framework.go:22
+  validates that NodeSelector is respected if not matching  [Conformance]
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 ------------------------------
-SSSSSSSSSSSSSSS
+SSSSSSSSSSSSS
 ------------------------------
-[sig-node] Downward API 
-  should provide container's limits.cpu/memory and requests.cpu/memory as env vars [NodeConformance] [Conformance]
+[sig-storage] Projected secret 
+  optional updates should be reflected in volume [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-[BeforeEach] [sig-node] Downward API
+[BeforeEach] [sig-storage] Projected secret
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
 STEP: Creating a kubernetes client
-May 13 19:59:52.759: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423
-STEP: Building a namespace api object, basename downward-api
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-downward-api-cntzj
+Jun 18 12:57:38.473: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
+STEP: Building a namespace api object, basename projected
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-projected-f6dmh
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should provide container's limits.cpu/memory and requests.cpu/memory as env vars [NodeConformance] [Conformance]
+[It] optional updates should be reflected in volume [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-STEP: Creating a pod to test downward api env vars
-May 13 19:59:53.039: INFO: Waiting up to 5m0s for pod "downward-api-ac37c00d-75b9-11e9-8f67-2632f168be36" in namespace "e2e-tests-downward-api-cntzj" to be "success or failure"
-May 13 19:59:53.047: INFO: Pod "downward-api-ac37c00d-75b9-11e9-8f67-2632f168be36": Phase="Pending", Reason="", readiness=false. Elapsed: 8.092401ms
-May 13 19:59:55.055: INFO: Pod "downward-api-ac37c00d-75b9-11e9-8f67-2632f168be36": Phase="Running", Reason="", readiness=true. Elapsed: 2.01601739s
-May 13 19:59:57.064: INFO: Pod "downward-api-ac37c00d-75b9-11e9-8f67-2632f168be36": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.024859971s
-STEP: Saw pod success
-May 13 19:59:57.064: INFO: Pod "downward-api-ac37c00d-75b9-11e9-8f67-2632f168be36" satisfied condition "success or failure"
-May 13 19:59:57.072: INFO: Trying to get logs from node 10.170.219.153 pod downward-api-ac37c00d-75b9-11e9-8f67-2632f168be36 container dapi-container: 
-STEP: delete the pod
-May 13 19:59:57.110: INFO: Waiting for pod downward-api-ac37c00d-75b9-11e9-8f67-2632f168be36 to disappear
-May 13 19:59:57.117: INFO: Pod downward-api-ac37c00d-75b9-11e9-8f67-2632f168be36 no longer exists
-[AfterEach] [sig-node] Downward API
+STEP: Creating secret with name s-test-opt-del-a6cf857d-91c8-11e9-bce2-ae54e022189f
+STEP: Creating secret with name s-test-opt-upd-a6cf85e4-91c8-11e9-bce2-ae54e022189f
+STEP: Creating the pod
+STEP: Deleting secret s-test-opt-del-a6cf857d-91c8-11e9-bce2-ae54e022189f
+STEP: Updating secret s-test-opt-upd-a6cf85e4-91c8-11e9-bce2-ae54e022189f
+STEP: Creating secret with name s-test-opt-create-a6cf860c-91c8-11e9-bce2-ae54e022189f
+STEP: waiting to observe update in volume
+[AfterEach] [sig-storage] Projected secret
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
-May 13 19:59:57.117: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "e2e-tests-downward-api-cntzj" for this suite.
-May 13 20:00:03.227: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-May 13 20:00:03.452: INFO: namespace: e2e-tests-downward-api-cntzj, resource: bindings, ignored listing per whitelist
-May 13 20:00:03.466: INFO: namespace e2e-tests-downward-api-cntzj deletion completed in 6.341130274s
+Jun 18 12:58:46.830: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-tests-projected-f6dmh" for this suite.
+Jun 18 12:59:10.900: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 12:59:11.394: INFO: namespace: e2e-tests-projected-f6dmh, resource: bindings, ignored listing per whitelist
+Jun 18 12:59:11.407: INFO: namespace e2e-tests-projected-f6dmh deletion completed in 24.556094283s
 
-• [SLOW TEST:10.707 seconds]
-[sig-node] Downward API
-/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downward_api.go:38
-  should provide container's limits.cpu/memory and requests.cpu/memory as env vars [NodeConformance] [Conformance]
+• [SLOW TEST:92.935 seconds]
+[sig-storage] Projected secret
+/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_secret.go:34
+  optional updates should be reflected in volume [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 ------------------------------
 SS
 ------------------------------
 [sig-storage] Projected configMap 
-  should be consumable from pods in volume with mappings as non-root [NodeConformance] [Conformance]
+  updates should be reflected in volume [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 [BeforeEach] [sig-storage] Projected configMap
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
 STEP: Creating a kubernetes client
-May 13 20:00:03.466: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423
+Jun 18 12:59:11.409: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
 STEP: Building a namespace api object, basename projected
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-projected-6q52l
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-projected-swjlb
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should be consumable from pods in volume with mappings as non-root [NodeConformance] [Conformance]
+[It] updates should be reflected in volume [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-STEP: Creating configMap with name projected-configmap-test-volume-map-b29b2530-75b9-11e9-8f67-2632f168be36
-STEP: Creating a pod to test consume configMaps
-May 13 20:00:03.765: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-b29c5ad2-75b9-11e9-8f67-2632f168be36" in namespace "e2e-tests-projected-6q52l" to be "success or failure"
-May 13 20:00:03.773: INFO: Pod "pod-projected-configmaps-b29c5ad2-75b9-11e9-8f67-2632f168be36": Phase="Pending", Reason="", readiness=false. Elapsed: 8.373414ms
-May 13 20:00:05.781: INFO: Pod "pod-projected-configmaps-b29c5ad2-75b9-11e9-8f67-2632f168be36": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.016295649s
-STEP: Saw pod success
-May 13 20:00:05.781: INFO: Pod "pod-projected-configmaps-b29c5ad2-75b9-11e9-8f67-2632f168be36" satisfied condition "success or failure"
-May 13 20:00:05.789: INFO: Trying to get logs from node 10.170.219.177 pod pod-projected-configmaps-b29c5ad2-75b9-11e9-8f67-2632f168be36 container projected-configmap-volume-test: 
-STEP: delete the pod
-May 13 20:00:05.828: INFO: Waiting for pod pod-projected-configmaps-b29c5ad2-75b9-11e9-8f67-2632f168be36 to disappear
-May 13 20:00:05.835: INFO: Pod pod-projected-configmaps-b29c5ad2-75b9-11e9-8f67-2632f168be36 no longer exists
+STEP: Creating projection with configMap that has name projected-configmap-test-upd-de3294c8-91c8-11e9-bce2-ae54e022189f
+STEP: Creating the pod
+STEP: Updating configmap projected-configmap-test-upd-de3294c8-91c8-11e9-bce2-ae54e022189f
+STEP: waiting to observe update in volume
 [AfterEach] [sig-storage] Projected configMap
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
-May 13 20:00:05.835: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "e2e-tests-projected-6q52l" for this suite.
-May 13 20:00:12.248: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-May 13 20:00:12.471: INFO: namespace: e2e-tests-projected-6q52l, resource: bindings, ignored listing per whitelist
-May 13 20:00:12.590: INFO: namespace e2e-tests-projected-6q52l deletion completed in 6.747296669s
+Jun 18 12:59:16.101: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-tests-projected-swjlb" for this suite.
+Jun 18 12:59:40.167: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 12:59:42.453: INFO: namespace: e2e-tests-projected-swjlb, resource: bindings, ignored listing per whitelist
+Jun 18 12:59:42.567: INFO: namespace e2e-tests-projected-swjlb deletion completed in 26.448164028s
 
-• [SLOW TEST:9.124 seconds]
+• [SLOW TEST:31.158 seconds]
 [sig-storage] Projected configMap
 /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_configmap.go:34
-  should be consumable from pods in volume with mappings as non-root [NodeConformance] [Conformance]
+  updates should be reflected in volume [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 ------------------------------
-SSSSSSSSSSS
+SSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] 
-  Should recreate evicted statefulset [Conformance]
+[sig-storage] Downward API volume 
+  should provide container's cpu request [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-[BeforeEach] [sig-apps] StatefulSet
+[BeforeEach] [sig-storage] Downward API volume
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
 STEP: Creating a kubernetes client
-May 13 20:00:12.591: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423
-STEP: Building a namespace api object, basename statefulset
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-statefulset-4mhl5
+Jun 18 12:59:42.568: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
+STEP: Building a namespace api object, basename downward-api
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-downward-api-gq7kx
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-apps] StatefulSet
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:59
-[BeforeEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic]
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:74
-STEP: Creating service test in namespace e2e-tests-statefulset-4mhl5
-[It] Should recreate evicted statefulset [Conformance]
+[BeforeEach] [sig-storage] Downward API volume
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:39
+[It] should provide container's cpu request [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-STEP: Looking for a node to schedule stateful set and pod
-STEP: Creating pod with conflicting port in namespace e2e-tests-statefulset-4mhl5
-STEP: Creating statefulset with conflicting port in namespace e2e-tests-statefulset-4mhl5
-STEP: Waiting until pod test-pod will start running in namespace e2e-tests-statefulset-4mhl5
-STEP: Waiting until stateful pod ss-0 will be recreated and deleted at least once in namespace e2e-tests-statefulset-4mhl5
-May 13 20:00:16.929: INFO: Observed stateful pod in namespace: e2e-tests-statefulset-4mhl5, name: ss-0, uid: ba3b3aaa-75b9-11e9-83a8-7e0242378207, status phase: Pending. Waiting for statefulset controller to delete.
-May 13 20:00:17.116: INFO: Observed stateful pod in namespace: e2e-tests-statefulset-4mhl5, name: ss-0, uid: ba3b3aaa-75b9-11e9-83a8-7e0242378207, status phase: Failed. Waiting for statefulset controller to delete.
-May 13 20:00:17.126: INFO: Observed stateful pod in namespace: e2e-tests-statefulset-4mhl5, name: ss-0, uid: ba3b3aaa-75b9-11e9-83a8-7e0242378207, status phase: Failed. Waiting for statefulset controller to delete.
-May 13 20:00:17.132: INFO: Observed delete event for stateful pod ss-0 in namespace e2e-tests-statefulset-4mhl5
-STEP: Removing pod with conflicting port in namespace e2e-tests-statefulset-4mhl5
-STEP: Waiting when stateful pod ss-0 will be recreated in namespace e2e-tests-statefulset-4mhl5 and will be in running state
-[AfterEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic]
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:85
-May 13 20:00:21.180: INFO: Deleting all statefulset in ns e2e-tests-statefulset-4mhl5
-May 13 20:00:21.186: INFO: Scaling statefulset ss to 0
-May 13 20:00:41.214: INFO: Waiting for statefulset status.replicas updated to 0
-May 13 20:00:41.219: INFO: Deleting statefulset ss
-[AfterEach] [sig-apps] StatefulSet
+STEP: Creating a pod to test downward API volume plugin
+Jun 18 12:59:43.052: INFO: Waiting up to 5m0s for pod "downwardapi-volume-f0c24ed9-91c8-11e9-bce2-ae54e022189f" in namespace "e2e-tests-downward-api-gq7kx" to be "success or failure"
+Jun 18 12:59:43.069: INFO: Pod "downwardapi-volume-f0c24ed9-91c8-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 17.200487ms
+Jun 18 12:59:45.085: INFO: Pod "downwardapi-volume-f0c24ed9-91c8-11e9-bce2-ae54e022189f": Phase="Running", Reason="", readiness=true. Elapsed: 2.033140579s
+Jun 18 12:59:47.100: INFO: Pod "downwardapi-volume-f0c24ed9-91c8-11e9-bce2-ae54e022189f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.048400701s
+STEP: Saw pod success
+Jun 18 12:59:47.101: INFO: Pod "downwardapi-volume-f0c24ed9-91c8-11e9-bce2-ae54e022189f" satisfied condition "success or failure"
+Jun 18 12:59:47.115: INFO: Trying to get logs from node 10.72.74.149 pod downwardapi-volume-f0c24ed9-91c8-11e9-bce2-ae54e022189f container client-container: 
+STEP: delete the pod
+Jun 18 12:59:47.226: INFO: Waiting for pod downwardapi-volume-f0c24ed9-91c8-11e9-bce2-ae54e022189f to disappear
+Jun 18 12:59:47.241: INFO: Pod downwardapi-volume-f0c24ed9-91c8-11e9-bce2-ae54e022189f no longer exists
+[AfterEach] [sig-storage] Downward API volume
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
-May 13 20:00:41.259: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "e2e-tests-statefulset-4mhl5" for this suite.
-May 13 20:00:47.348: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-May 13 20:00:47.638: INFO: namespace: e2e-tests-statefulset-4mhl5, resource: bindings, ignored listing per whitelist
-May 13 20:00:47.661: INFO: namespace e2e-tests-statefulset-4mhl5 deletion completed in 6.393557254s
+Jun 18 12:59:47.241: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-tests-downward-api-gq7kx" for this suite.
+Jun 18 12:59:55.315: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 12:59:55.475: INFO: namespace: e2e-tests-downward-api-gq7kx, resource: bindings, ignored listing per whitelist
+Jun 18 12:59:55.907: INFO: namespace e2e-tests-downward-api-gq7kx deletion completed in 8.645403275s
 
-• [SLOW TEST:35.070 seconds]
-[sig-apps] StatefulSet
-/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22
-  [k8s.io] Basic StatefulSet functionality [StatefulSetBasic]
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694
-    Should recreate evicted statefulset [Conformance]
-    /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
+• [SLOW TEST:13.339 seconds]
+[sig-storage] Downward API volume
+/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:34
+  should provide container's cpu request [NodeConformance] [Conformance]
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 ------------------------------
-SSSSS
+SSSSSSSSSSSSSSSSS
+------------------------------
+[k8s.io] Container Runtime blackbox test when starting a container that exits 
+  should run with the expected status [NodeConformance] [Conformance]
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
+[BeforeEach] [k8s.io] Container Runtime
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
+STEP: Creating a kubernetes client
+Jun 18 12:59:55.907: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
+STEP: Building a namespace api object, basename container-runtime
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-container-runtime-djsdz
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should run with the expected status [NodeConformance] [Conformance]
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
+STEP: Container 'terminate-cmd-rpa': should get the expected 'RestartCount'
+STEP: Container 'terminate-cmd-rpa': should get the expected 'Phase'
+STEP: Container 'terminate-cmd-rpa': should get the expected 'Ready' condition
+STEP: Container 'terminate-cmd-rpa': should get the expected 'State'
+STEP: Container 'terminate-cmd-rpa': should be possible to delete [NodeConformance]
+STEP: Container 'terminate-cmd-rpof': should get the expected 'RestartCount'
+STEP: Container 'terminate-cmd-rpof': should get the expected 'Phase'
+STEP: Container 'terminate-cmd-rpof': should get the expected 'Ready' condition
+STEP: Container 'terminate-cmd-rpof': should get the expected 'State'
+STEP: Container 'terminate-cmd-rpof': should be possible to delete [NodeConformance]
+STEP: Container 'terminate-cmd-rpn': should get the expected 'RestartCount'
+STEP: Container 'terminate-cmd-rpn': should get the expected 'Phase'
+STEP: Container 'terminate-cmd-rpn': should get the expected 'Ready' condition
+STEP: Container 'terminate-cmd-rpn': should get the expected 'State'
+STEP: Container 'terminate-cmd-rpn': should be possible to delete [NodeConformance]
+[AfterEach] [k8s.io] Container Runtime
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
+Jun 18 13:00:19.545: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-tests-container-runtime-djsdz" for this suite.
+Jun 18 13:00:27.624: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 13:00:28.449: INFO: namespace: e2e-tests-container-runtime-djsdz, resource: bindings, ignored listing per whitelist
+Jun 18 13:00:28.495: INFO: namespace e2e-tests-container-runtime-djsdz deletion completed in 8.92023168s
+
+• [SLOW TEST:32.589 seconds]
+[k8s.io] Container Runtime
+/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694
+  blackbox test
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/runtime.go:37
+    when starting a container that exits
+    /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/runtime.go:38
+      should run with the expected status [NodeConformance] [Conformance]
+      /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
+------------------------------
+SSSSSSS
 ------------------------------
-[sig-storage] Downward API volume 
-  should provide container's cpu request [NodeConformance] [Conformance]
+[sig-storage] HostPath 
+  should give a volume the correct mode [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-[BeforeEach] [sig-storage] Downward API volume
+[BeforeEach] [sig-storage] HostPath
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
 STEP: Creating a kubernetes client
-May 13 20:00:47.661: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423
-STEP: Building a namespace api object, basename downward-api
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-downward-api-742sl
+Jun 18 13:00:28.496: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
+STEP: Building a namespace api object, basename hostpath
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-hostpath-9h78m
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-storage] Downward API volume
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:39
-[It] should provide container's cpu request [NodeConformance] [Conformance]
+[BeforeEach] [sig-storage] HostPath
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/host_path.go:37
+[It] should give a volume the correct mode [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-STEP: Creating a pod to test downward API volume plugin
-May 13 20:00:47.955: INFO: Waiting up to 5m0s for pod "downwardapi-volume-ccf30a9a-75b9-11e9-8f67-2632f168be36" in namespace "e2e-tests-downward-api-742sl" to be "success or failure"
-May 13 20:00:47.962: INFO: Pod "downwardapi-volume-ccf30a9a-75b9-11e9-8f67-2632f168be36": Phase="Pending", Reason="", readiness=false. Elapsed: 7.399791ms
-May 13 20:00:49.971: INFO: Pod "downwardapi-volume-ccf30a9a-75b9-11e9-8f67-2632f168be36": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.015669105s
+STEP: Creating a pod to test hostPath mode
+Jun 18 13:00:29.094: INFO: Waiting up to 5m0s for pod "pod-host-path-test" in namespace "e2e-tests-hostpath-9h78m" to be "success or failure"
+Jun 18 13:00:29.110: INFO: Pod "pod-host-path-test": Phase="Pending", Reason="", readiness=false. Elapsed: 16.35673ms
+Jun 18 13:00:31.124: INFO: Pod "pod-host-path-test": Phase="Running", Reason="", readiness=false. Elapsed: 2.030461035s
+Jun 18 13:00:33.139: INFO: Pod "pod-host-path-test": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.045781301s
 STEP: Saw pod success
-May 13 20:00:49.971: INFO: Pod "downwardapi-volume-ccf30a9a-75b9-11e9-8f67-2632f168be36" satisfied condition "success or failure"
-May 13 20:00:49.979: INFO: Trying to get logs from node 10.170.219.153 pod downwardapi-volume-ccf30a9a-75b9-11e9-8f67-2632f168be36 container client-container: 
+Jun 18 13:00:33.140: INFO: Pod "pod-host-path-test" satisfied condition "success or failure"
+Jun 18 13:00:33.154: INFO: Trying to get logs from node 10.72.74.149 pod pod-host-path-test container test-container-1: 
 STEP: delete the pod
-May 13 20:00:50.018: INFO: Waiting for pod downwardapi-volume-ccf30a9a-75b9-11e9-8f67-2632f168be36 to disappear
-May 13 20:00:50.025: INFO: Pod downwardapi-volume-ccf30a9a-75b9-11e9-8f67-2632f168be36 no longer exists
-[AfterEach] [sig-storage] Downward API volume
+Jun 18 13:00:33.228: INFO: Waiting for pod pod-host-path-test to disappear
+Jun 18 13:00:33.241: INFO: Pod pod-host-path-test no longer exists
+[AfterEach] [sig-storage] HostPath
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
-May 13 20:00:50.025: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "e2e-tests-downward-api-742sl" for this suite.
-May 13 20:00:56.062: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-May 13 20:00:56.268: INFO: namespace: e2e-tests-downward-api-742sl, resource: bindings, ignored listing per whitelist
-May 13 20:00:56.351: INFO: namespace e2e-tests-downward-api-742sl deletion completed in 6.318203541s
+Jun 18 13:00:33.241: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-tests-hostpath-9h78m" for this suite.
+Jun 18 13:00:39.315: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 13:00:39.582: INFO: namespace: e2e-tests-hostpath-9h78m, resource: bindings, ignored listing per whitelist
+Jun 18 13:00:39.986: INFO: namespace e2e-tests-hostpath-9h78m deletion completed in 6.72087531s
 
-• [SLOW TEST:8.690 seconds]
-[sig-storage] Downward API volume
-/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:34
-  should provide container's cpu request [NodeConformance] [Conformance]
+• [SLOW TEST:11.490 seconds]
+[sig-storage] HostPath
+/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/host_path.go:34
+  should give a volume the correct mode [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 ------------------------------
-SSSSSSSSS
+SSSSSSSSSSS
 ------------------------------
-[sig-api-machinery] Watchers 
-  should be able to restart watching from the last resource version observed by the previous watch [Conformance]
+[k8s.io] Kubelet when scheduling a busybox Pod with hostAliases 
+  should write entries to /etc/hosts [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-[BeforeEach] [sig-api-machinery] Watchers
+[BeforeEach] [k8s.io] Kubelet
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
 STEP: Creating a kubernetes client
-May 13 20:00:56.352: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423
-STEP: Building a namespace api object, basename watch
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-watch-mvcq4
+Jun 18 13:00:39.987: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
+STEP: Building a namespace api object, basename kubelet-test
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-kubelet-test-4wcjm
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should be able to restart watching from the last resource version observed by the previous watch [Conformance]
+[BeforeEach] [k8s.io] Kubelet
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:37
+[It] should write entries to /etc/hosts [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-STEP: creating a watch on configmaps
-STEP: creating a new configmap
-STEP: modifying the configmap once
-STEP: closing the watch once it receives two notifications
-May 13 20:00:56.692: INFO: Got : ADDED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-watch-closed,GenerateName:,Namespace:e2e-tests-watch-mvcq4,SelfLink:/api/v1/namespaces/e2e-tests-watch-mvcq4/configmaps/e2e-watch-test-watch-closed,UID:d2278eac-75b9-11e9-b786-da20024d205c,ResourceVersion:51297,Generation:0,CreationTimestamp:2019-05-13 20:00:56 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: watch-closed-and-restarted,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Data:map[string]string{},BinaryData:map[string][]byte{},}
-May 13 20:00:56.692: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-watch-closed,GenerateName:,Namespace:e2e-tests-watch-mvcq4,SelfLink:/api/v1/namespaces/e2e-tests-watch-mvcq4/configmaps/e2e-watch-test-watch-closed,UID:d2278eac-75b9-11e9-b786-da20024d205c,ResourceVersion:51298,Generation:0,CreationTimestamp:2019-05-13 20:00:56 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: watch-closed-and-restarted,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Data:map[string]string{mutation: 1,},BinaryData:map[string][]byte{},}
-STEP: modifying the configmap a second time, while the watch is closed
-STEP: creating a new watch on configmaps from the last resource version observed by the first watch
-STEP: deleting the configmap
-STEP: Expecting to observe notifications for all changes to the configmap since the first watch closed
-May 13 20:00:56.729: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-watch-closed,GenerateName:,Namespace:e2e-tests-watch-mvcq4,SelfLink:/api/v1/namespaces/e2e-tests-watch-mvcq4/configmaps/e2e-watch-test-watch-closed,UID:d2278eac-75b9-11e9-b786-da20024d205c,ResourceVersion:51299,Generation:0,CreationTimestamp:2019-05-13 20:00:56 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: watch-closed-and-restarted,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},}
-May 13 20:00:56.729: INFO: Got : DELETED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-watch-closed,GenerateName:,Namespace:e2e-tests-watch-mvcq4,SelfLink:/api/v1/namespaces/e2e-tests-watch-mvcq4/configmaps/e2e-watch-test-watch-closed,UID:d2278eac-75b9-11e9-b786-da20024d205c,ResourceVersion:51300,Generation:0,CreationTimestamp:2019-05-13 20:00:56 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: watch-closed-and-restarted,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},}
-[AfterEach] [sig-api-machinery] Watchers
+[AfterEach] [k8s.io] Kubelet
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
-May 13 20:00:56.729: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "e2e-tests-watch-mvcq4" for this suite.
-May 13 20:01:02.776: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-May 13 20:01:02.879: INFO: namespace: e2e-tests-watch-mvcq4, resource: bindings, ignored listing per whitelist
-May 13 20:01:03.070: INFO: namespace e2e-tests-watch-mvcq4 deletion completed in 6.334509854s
+Jun 18 13:00:42.769: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-tests-kubelet-test-4wcjm" for this suite.
+Jun 18 13:01:25.030: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 13:01:25.529: INFO: namespace: e2e-tests-kubelet-test-4wcjm, resource: bindings, ignored listing per whitelist
+Jun 18 13:01:25.555: INFO: namespace e2e-tests-kubelet-test-4wcjm deletion completed in 42.668901033s
 
-• [SLOW TEST:6.718 seconds]
-[sig-api-machinery] Watchers
+• [SLOW TEST:45.568 seconds]
+[k8s.io] Kubelet
+/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694
+  when scheduling a busybox Pod with hostAliases
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:136
+    should write entries to /etc/hosts [NodeConformance] [Conformance]
+    /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
+------------------------------
+SS
+------------------------------
+[sig-api-machinery] Namespaces [Serial] 
+  should ensure that all pods are removed when a namespace is deleted [Conformance]
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
+[BeforeEach] [sig-api-machinery] Namespaces [Serial]
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
+STEP: Creating a kubernetes client
+Jun 18 13:01:25.555: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
+STEP: Building a namespace api object, basename namespaces
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-namespaces-27dld
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should ensure that all pods are removed when a namespace is deleted [Conformance]
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
+STEP: Creating a test namespace
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-nsdeletetest-ljkcq
+STEP: Waiting for a default service account to be provisioned in namespace
+STEP: Creating a pod in the namespace
+STEP: Waiting for the pod to have running status
+STEP: Creating an uninitialized pod in the namespace
+STEP: Deleting the namespace
+STEP: Waiting for the namespace to be removed.
+Jun 18 13:01:35.058: INFO: error from create uninitialized namespace: Internal error occurred: object deleted while waiting for creation
+STEP: Recreating the namespace
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-nsdeletetest-cr5fv
+STEP: Verifying there are no pods in the namespace
+[AfterEach] [sig-api-machinery] Namespaces [Serial]
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
+Jun 18 13:01:52.714: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-tests-namespaces-27dld" for this suite.
+Jun 18 13:02:00.793: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 13:02:01.314: INFO: namespace: e2e-tests-namespaces-27dld, resource: bindings, ignored listing per whitelist
+Jun 18 13:02:01.458: INFO: namespace e2e-tests-namespaces-27dld deletion completed in 8.720019452s
+STEP: Destroying namespace "e2e-tests-nsdeletetest-ljkcq" for this suite.
+Jun 18 13:02:01.471: INFO: Namespace e2e-tests-nsdeletetest-ljkcq was already deleted
+STEP: Destroying namespace "e2e-tests-nsdeletetest-cr5fv" for this suite.
+Jun 18 13:02:07.530: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 13:02:07.607: INFO: namespace: e2e-tests-nsdeletetest-cr5fv, resource: bindings, ignored listing per whitelist
+Jun 18 13:02:08.142: INFO: namespace e2e-tests-nsdeletetest-cr5fv deletion completed in 6.670408983s
+
+• [SLOW TEST:42.587 seconds]
+[sig-api-machinery] Namespaces [Serial]
 /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:22
-  should be able to restart watching from the last resource version observed by the previous watch [Conformance]
+  should ensure that all pods are removed when a namespace is deleted [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 ------------------------------
-SSSS
-------------------------------
-[sig-storage] ConfigMap 
-  binary data should be reflected in volume [NodeConformance] [Conformance]
+[k8s.io] Kubelet when scheduling a busybox command that always fails in a pod 
+  should be possible to delete [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-[BeforeEach] [sig-storage] ConfigMap
+[BeforeEach] [k8s.io] Kubelet
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
 STEP: Creating a kubernetes client
-May 13 20:01:03.070: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423
-STEP: Building a namespace api object, basename configmap
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-configmap-z7ns7
+Jun 18 13:02:08.142: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
+STEP: Building a namespace api object, basename kubelet-test
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-kubelet-test-qrb9x
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] binary data should be reflected in volume [NodeConformance] [Conformance]
+[BeforeEach] [k8s.io] Kubelet
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:37
+[BeforeEach] when scheduling a busybox command that always fails in a pod
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:81
+[It] should be possible to delete [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-STEP: Creating configMap with name configmap-test-upd-d623afcd-75b9-11e9-8f67-2632f168be36
-STEP: Creating the pod
-STEP: Waiting for pod with text data
-STEP: Waiting for pod with binary data
-[AfterEach] [sig-storage] ConfigMap
+[AfterEach] [k8s.io] Kubelet
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
-May 13 20:01:07.432: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "e2e-tests-configmap-z7ns7" for this suite.
-May 13 20:01:31.469: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-May 13 20:01:31.630: INFO: namespace: e2e-tests-configmap-z7ns7, resource: bindings, ignored listing per whitelist
-May 13 20:01:31.802: INFO: namespace e2e-tests-configmap-z7ns7 deletion completed in 24.362249204s
+Jun 18 13:02:08.673: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-tests-kubelet-test-qrb9x" for this suite.
+Jun 18 13:02:14.747: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 13:02:15.238: INFO: namespace: e2e-tests-kubelet-test-qrb9x, resource: bindings, ignored listing per whitelist
+Jun 18 13:02:15.267: INFO: namespace e2e-tests-kubelet-test-qrb9x deletion completed in 6.573741496s
 
-• [SLOW TEST:28.732 seconds]
-[sig-storage] ConfigMap
-/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:33
-  binary data should be reflected in volume [NodeConformance] [Conformance]
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
+• [SLOW TEST:7.124 seconds]
+[k8s.io] Kubelet
+/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694
+  when scheduling a busybox command that always fails in a pod
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:78
+    should be possible to delete [NodeConformance] [Conformance]
+    /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 ------------------------------
-SSSS
+SSSSSSSSSSS
 ------------------------------
-[sig-cli] Kubectl client [k8s.io] Kubectl logs 
-  should be able to retrieve and filter logs  [Conformance]
+[sig-cli] Kubectl client [k8s.io] Guestbook application 
+  should create and stop a working application  [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 [BeforeEach] [sig-cli] Kubectl client
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
 STEP: Creating a kubernetes client
-May 13 20:01:31.804: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423
+Jun 18 13:02:15.267: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
 STEP: Building a namespace api object, basename kubectl
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-kubectl-l8677
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-kubectl-f9kg8
 STEP: Waiting for a default service account to be provisioned in namespace
 [BeforeEach] [sig-cli] Kubectl client
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:243
-[BeforeEach] [k8s.io] Kubectl logs
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1134
-STEP: creating an rc
-May 13 20:01:32.123: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 create -f - --namespace=e2e-tests-kubectl-l8677'
-May 13 20:01:32.534: INFO: stderr: ""
-May 13 20:01:32.534: INFO: stdout: "replicationcontroller/redis-master created\n"
-[It] should be able to retrieve and filter logs  [Conformance]
+[It] should create and stop a working application  [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-STEP: Waiting for Redis master to start.
-May 13 20:01:33.543: INFO: Selector matched 1 pods for map[app:redis]
-May 13 20:01:33.543: INFO: Found 0 / 1
-May 13 20:01:34.620: INFO: Selector matched 1 pods for map[app:redis]
-May 13 20:01:34.620: INFO: Found 1 / 1
-May 13 20:01:34.620: INFO: WaitFor completed with timeout 5m0s.  Pods found = 1 out of 1
-May 13 20:01:34.633: INFO: Selector matched 1 pods for map[app:redis]
-May 13 20:01:34.633: INFO: ForEach: Found 1 pods from the filter.  Now looping through them.
-STEP: checking for a matching strings
-May 13 20:01:34.633: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 logs redis-master-974kk redis-master --namespace=e2e-tests-kubectl-l8677'
-May 13 20:01:34.750: INFO: stderr: ""
-May 13 20:01:34.750: INFO: stdout: "                _._                                                  \n           _.-``__ ''-._                                             \n      _.-``    `.  `_.  ''-._           Redis 3.2.12 (35a5711f/0) 64 bit\n  .-`` .-```.  ```\\/    _.,_ ''-._                                   \n (    '      ,       .-`  | `,    )     Running in standalone mode\n |`-._`-...-` __...-.``-._|'` _.-'|     Port: 6379\n |    `-._   `._    /     _.-'    |     PID: 1\n  `-._    `-._  `-./  _.-'    _.-'                                   \n |`-._`-._    `-.__.-'    _.-'_.-'|                                  \n |    `-._`-._        _.-'_.-'    |           http://redis.io        \n  `-._    `-._`-.__.-'_.-'    _.-'                                   \n |`-._`-._    `-.__.-'    _.-'_.-'|                                  \n |    `-._`-._        _.-'_.-'    |                                  \n  `-._    `-._`-.__.-'_.-'    _.-'                                   \n      `-._    `-.__.-'    _.-'                                       \n          `-._        _.-'                                           \n              `-.__.-'                                               \n\n1:M 13 May 20:01:33.722 # WARNING: The TCP backlog setting of 511 cannot be enforced because /proc/sys/net/core/somaxconn is set to the lower value of 128.\n1:M 13 May 20:01:33.722 # Server started, Redis version 3.2.12\n1:M 13 May 20:01:33.722 # WARNING you have Transparent Huge Pages (THP) support enabled in your kernel. This will create latency and memory usage issues with Redis. To fix this issue run the command 'echo never > /sys/kernel/mm/transparent_hugepage/enabled' as root, and add it to your /etc/rc.local in order to retain the setting after a reboot. Redis must be restarted after THP is disabled.\n1:M 13 May 20:01:33.722 * The server is now ready to accept connections on port 6379\n"
-STEP: limiting log lines
-May 13 20:01:34.750: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 log redis-master-974kk redis-master --namespace=e2e-tests-kubectl-l8677 --tail=1'
-May 13 20:01:34.884: INFO: stderr: ""
-May 13 20:01:34.884: INFO: stdout: "1:M 13 May 20:01:33.722 * The server is now ready to accept connections on port 6379\n"
-STEP: limiting log bytes
-May 13 20:01:34.884: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 log redis-master-974kk redis-master --namespace=e2e-tests-kubectl-l8677 --limit-bytes=1'
-May 13 20:01:34.999: INFO: stderr: ""
-May 13 20:01:35.000: INFO: stdout: " "
-STEP: exposing timestamps
-May 13 20:01:35.000: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 log redis-master-974kk redis-master --namespace=e2e-tests-kubectl-l8677 --tail=1 --timestamps'
-May 13 20:01:35.123: INFO: stderr: ""
-May 13 20:01:35.123: INFO: stdout: "2019-05-13T20:01:33.722286286Z 1:M 13 May 20:01:33.722 * The server is now ready to accept connections on port 6379\n"
-STEP: restricting to a time range
-May 13 20:01:37.623: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 log redis-master-974kk redis-master --namespace=e2e-tests-kubectl-l8677 --since=1s'
-May 13 20:01:37.840: INFO: stderr: ""
-May 13 20:01:37.840: INFO: stdout: ""
-May 13 20:01:37.841: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 log redis-master-974kk redis-master --namespace=e2e-tests-kubectl-l8677 --since=24h'
-May 13 20:01:37.961: INFO: stderr: ""
-May 13 20:01:37.961: INFO: stdout: "                _._                                                  \n           _.-``__ ''-._                                             \n      _.-``    `.  `_.  ''-._           Redis 3.2.12 (35a5711f/0) 64 bit\n  .-`` .-```.  ```\\/    _.,_ ''-._                                   \n (    '      ,       .-`  | `,    )     Running in standalone mode\n |`-._`-...-` __...-.``-._|'` _.-'|     Port: 6379\n |    `-._   `._    /     _.-'    |     PID: 1\n  `-._    `-._  `-./  _.-'    _.-'                                   \n |`-._`-._    `-.__.-'    _.-'_.-'|                                  \n |    `-._`-._        _.-'_.-'    |           http://redis.io        \n  `-._    `-._`-.__.-'_.-'    _.-'                                   \n |`-._`-._    `-.__.-'    _.-'_.-'|                                  \n |    `-._`-._        _.-'_.-'    |                                  \n  `-._    `-._`-.__.-'_.-'    _.-'                                   \n      `-._    `-.__.-'    _.-'                                       \n          `-._        _.-'                                           \n              `-.__.-'                                               \n\n1:M 13 May 20:01:33.722 # WARNING: The TCP backlog setting of 511 cannot be enforced because /proc/sys/net/core/somaxconn is set to the lower value of 128.\n1:M 13 May 20:01:33.722 # Server started, Redis version 3.2.12\n1:M 13 May 20:01:33.722 # WARNING you have Transparent Huge Pages (THP) support enabled in your kernel. This will create latency and memory usage issues with Redis. To fix this issue run the command 'echo never > /sys/kernel/mm/transparent_hugepage/enabled' as root, and add it to your /etc/rc.local in order to retain the setting after a reboot. Redis must be restarted after THP is disabled.\n1:M 13 May 20:01:33.722 * The server is now ready to accept connections on port 6379\n"
-[AfterEach] [k8s.io] Kubectl logs
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1140
+STEP: creating all guestbook components
+Jun 18 13:02:15.790: INFO: apiVersion: v1
+kind: Service
+metadata:
+  name: redis-slave
+  labels:
+    app: redis
+    role: slave
+    tier: backend
+spec:
+  ports:
+  - port: 6379
+  selector:
+    app: redis
+    role: slave
+    tier: backend
+
+Jun 18 13:02:15.791: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 create -f - --namespace=e2e-tests-kubectl-f9kg8'
+Jun 18 13:02:16.123: INFO: stderr: ""
+Jun 18 13:02:16.123: INFO: stdout: "service/redis-slave created\n"
+Jun 18 13:02:16.123: INFO: apiVersion: v1
+kind: Service
+metadata:
+  name: redis-master
+  labels:
+    app: redis
+    role: master
+    tier: backend
+spec:
+  ports:
+  - port: 6379
+    targetPort: 6379
+  selector:
+    app: redis
+    role: master
+    tier: backend
+
+Jun 18 13:02:16.123: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 create -f - --namespace=e2e-tests-kubectl-f9kg8'
+Jun 18 13:02:16.488: INFO: stderr: ""
+Jun 18 13:02:16.488: INFO: stdout: "service/redis-master created\n"
+Jun 18 13:02:16.489: INFO: apiVersion: v1
+kind: Service
+metadata:
+  name: frontend
+  labels:
+    app: guestbook
+    tier: frontend
+spec:
+  # if your cluster supports it, uncomment the following to automatically create
+  # an external load-balanced IP for the frontend service.
+  # type: LoadBalancer
+  ports:
+  - port: 80
+  selector:
+    app: guestbook
+    tier: frontend
+
+Jun 18 13:02:16.489: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 create -f - --namespace=e2e-tests-kubectl-f9kg8'
+Jun 18 13:02:16.742: INFO: stderr: ""
+Jun 18 13:02:16.742: INFO: stdout: "service/frontend created\n"
+Jun 18 13:02:16.742: INFO: apiVersion: extensions/v1beta1
+kind: Deployment
+metadata:
+  name: frontend
+spec:
+  replicas: 3
+  template:
+    metadata:
+      labels:
+        app: guestbook
+        tier: frontend
+    spec:
+      containers:
+      - name: php-redis
+        image: gcr.io/google-samples/gb-frontend:v6
+        resources:
+          requests:
+            cpu: 100m
+            memory: 100Mi
+        env:
+        - name: GET_HOSTS_FROM
+          value: dns
+          # If your cluster config does not include a dns service, then to
+          # instead access environment variables to find service host
+          # info, comment out the 'value: dns' line above, and uncomment the
+          # line below:
+          # value: env
+        ports:
+        - containerPort: 80
+
+Jun 18 13:02:16.743: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 create -f - --namespace=e2e-tests-kubectl-f9kg8'
+Jun 18 13:02:16.962: INFO: stderr: ""
+Jun 18 13:02:16.962: INFO: stdout: "deployment.extensions/frontend created\n"
+Jun 18 13:02:16.963: INFO: apiVersion: extensions/v1beta1
+kind: Deployment
+metadata:
+  name: redis-master
+spec:
+  replicas: 1
+  template:
+    metadata:
+      labels:
+        app: redis
+        role: master
+        tier: backend
+    spec:
+      containers:
+      - name: master
+        image: gcr.io/kubernetes-e2e-test-images/redis:1.0
+        resources:
+          requests:
+            cpu: 100m
+            memory: 100Mi
+        ports:
+        - containerPort: 6379
+
+Jun 18 13:02:16.963: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 create -f - --namespace=e2e-tests-kubectl-f9kg8'
+Jun 18 13:02:17.245: INFO: stderr: ""
+Jun 18 13:02:17.245: INFO: stdout: "deployment.extensions/redis-master created\n"
+Jun 18 13:02:17.246: INFO: apiVersion: extensions/v1beta1
+kind: Deployment
+metadata:
+  name: redis-slave
+spec:
+  replicas: 2
+  template:
+    metadata:
+      labels:
+        app: redis
+        role: slave
+        tier: backend
+    spec:
+      containers:
+      - name: slave
+        image: gcr.io/google-samples/gb-redisslave:v3
+        resources:
+          requests:
+            cpu: 100m
+            memory: 100Mi
+        env:
+        - name: GET_HOSTS_FROM
+          value: dns
+          # If your cluster config does not include a dns service, then to
+          # instead access an environment variable to find the master
+          # service's host, comment out the 'value: dns' line above, and
+          # uncomment the line below:
+          # value: env
+        ports:
+        - containerPort: 6379
+
+Jun 18 13:02:17.246: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 create -f - --namespace=e2e-tests-kubectl-f9kg8'
+Jun 18 13:02:17.635: INFO: stderr: ""
+Jun 18 13:02:17.635: INFO: stdout: "deployment.extensions/redis-slave created\n"
+STEP: validating guestbook app
+Jun 18 13:02:17.635: INFO: Waiting for all frontend pods to be Running.
+Jun 18 13:02:37.686: INFO: Waiting for frontend to serve content.
+Jun 18 13:02:42.755: INFO: Failed to get response from guestbook. err: , response: 
+Fatal error: Uncaught exception 'Predis\Connection\ConnectionException' with message 'Connection timed out [tcp://redis-slave:6379]' in /usr/local/lib/php/Predis/Connection/AbstractConnection.php:155 +Stack trace: +#0 /usr/local/lib/php/Predis/Connection/StreamConnection.php(128): Predis\Connection\AbstractConnection->onConnectionError('Connection time...', 110) +#1 /usr/local/lib/php/Predis/Connection/StreamConnection.php(178): Predis\Connection\StreamConnection->createStreamSocket(Object(Predis\Connection\Parameters), 'tcp://redis-sla...', 4) +#2 /usr/local/lib/php/Predis/Connection/StreamConnection.php(100): Predis\Connection\StreamConnection->tcpStreamInitializer(Object(Predis\Connection\Parameters)) +#3 /usr/local/lib/php/Predis/Connection/AbstractConnection.php(81): Predis\Connection\StreamConnection->createResource() +#4 /usr/local/lib/php/Predis/Connection/StreamConnection.php(258): Predis\Connection\AbstractConnection->connect() +#5 /usr/local/lib/php/Predis/Connection/AbstractConnection.php(180): Predis\Connection\Stre in /usr/local/lib/php/Predis/Connection/AbstractConnection.php on line 155
+ +Jun 18 13:02:47.821: INFO: Trying to add a new entry to the guestbook. +Jun 18 13:02:47.867: INFO: Verifying that added entry can be retrieved. +STEP: using delete to clean up resources +Jun 18 13:02:47.921: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 delete --grace-period=0 --force -f - --namespace=e2e-tests-kubectl-f9kg8' +Jun 18 13:02:48.213: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" +Jun 18 13:02:48.213: INFO: stdout: "service \"redis-slave\" force deleted\n" STEP: using delete to clean up resources -May 13 20:01:37.961: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 delete --grace-period=0 --force -f - --namespace=e2e-tests-kubectl-l8677' -May 13 20:01:38.083: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" -May 13 20:01:38.083: INFO: stdout: "replicationcontroller \"redis-master\" force deleted\n" -May 13 20:01:38.083: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 get rc,svc -l name=nginx --no-headers --namespace=e2e-tests-kubectl-l8677' -May 13 20:01:38.189: INFO: stderr: "No resources found.\n" -May 13 20:01:38.189: INFO: stdout: "" -May 13 20:01:38.189: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 get pods -l name=nginx --namespace=e2e-tests-kubectl-l8677 -o go-template={{ range .items }}{{ if not .metadata.deletionTimestamp }}{{ .metadata.name }}{{ "\n" }}{{ end }}{{ end }}' -May 13 20:01:38.287: INFO: stderr: "" -May 13 20:01:38.287: INFO: stdout: "" +Jun 18 13:02:48.213: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 delete --grace-period=0 --force -f - --namespace=e2e-tests-kubectl-f9kg8' +Jun 18 13:02:48.464: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" +Jun 18 13:02:48.464: INFO: stdout: "service \"redis-master\" force deleted\n" +STEP: using delete to clean up resources +Jun 18 13:02:48.464: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 delete --grace-period=0 --force -f - --namespace=e2e-tests-kubectl-f9kg8' +Jun 18 13:02:48.638: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" +Jun 18 13:02:48.638: INFO: stdout: "service \"frontend\" force deleted\n" +STEP: using delete to clean up resources +Jun 18 13:02:48.639: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 delete --grace-period=0 --force -f - --namespace=e2e-tests-kubectl-f9kg8' +Jun 18 13:02:48.785: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" +Jun 18 13:02:48.785: INFO: stdout: "deployment.extensions \"frontend\" force deleted\n" +STEP: using delete to clean up resources +Jun 18 13:02:48.785: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 delete --grace-period=0 --force -f - --namespace=e2e-tests-kubectl-f9kg8' +Jun 18 13:02:48.964: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" +Jun 18 13:02:48.964: INFO: stdout: "deployment.extensions \"redis-master\" force deleted\n" +STEP: using delete to clean up resources +Jun 18 13:02:48.964: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 delete --grace-period=0 --force -f - --namespace=e2e-tests-kubectl-f9kg8' +Jun 18 13:02:49.115: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" +Jun 18 13:02:49.115: INFO: stdout: "deployment.extensions \"redis-slave\" force deleted\n" [AfterEach] [sig-cli] Kubectl client /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 20:01:38.287: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-kubectl-l8677" for this suite. -May 13 20:01:44.324: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 20:01:44.532: INFO: namespace: e2e-tests-kubectl-l8677, resource: bindings, ignored listing per whitelist -May 13 20:01:44.734: INFO: namespace e2e-tests-kubectl-l8677 deletion completed in 6.4388628s +Jun 18 13:02:49.115: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "e2e-tests-kubectl-f9kg8" for this suite. +Jun 18 13:03:31.194: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 13:03:31.648: INFO: namespace: e2e-tests-kubectl-f9kg8, resource: bindings, ignored listing per whitelist +Jun 18 13:03:31.673: INFO: namespace e2e-tests-kubectl-f9kg8 deletion completed in 42.53981067s -• [SLOW TEST:12.930 seconds] +• [SLOW TEST:76.406 seconds] [sig-cli] Kubectl client /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:22 - [k8s.io] Kubectl logs + [k8s.io] Guestbook application /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 - should be able to retrieve and filter logs [Conformance] + should create and stop a working application [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ -SSSSSSSSSSS ------------------------------- -[sig-storage] Downward API volume - should provide podname only [NodeConformance] [Conformance] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -[BeforeEach] [sig-storage] Downward API volume - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 -STEP: Creating a kubernetes client -May 13 20:01:44.734: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -STEP: Building a namespace api object, basename downward-api -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-downward-api-r6z8h -STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [sig-storage] Downward API volume - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:39 -[It] should provide podname only [NodeConformance] [Conformance] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -STEP: Creating a pod to test downward API volume plugin -May 13 20:01:45.035: INFO: Waiting up to 5m0s for pod "downwardapi-volume-eef892b0-75b9-11e9-8f67-2632f168be36" in namespace "e2e-tests-downward-api-r6z8h" to be "success or failure" -May 13 20:01:45.043: INFO: Pod "downwardapi-volume-eef892b0-75b9-11e9-8f67-2632f168be36": Phase="Pending", Reason="", readiness=false. Elapsed: 7.951622ms -May 13 20:01:47.053: INFO: Pod "downwardapi-volume-eef892b0-75b9-11e9-8f67-2632f168be36": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.018206066s -STEP: Saw pod success -May 13 20:01:47.053: INFO: Pod "downwardapi-volume-eef892b0-75b9-11e9-8f67-2632f168be36" satisfied condition "success or failure" -May 13 20:01:47.061: INFO: Trying to get logs from node 10.170.219.177 pod downwardapi-volume-eef892b0-75b9-11e9-8f67-2632f168be36 container client-container: -STEP: delete the pod -May 13 20:01:47.101: INFO: Waiting for pod downwardapi-volume-eef892b0-75b9-11e9-8f67-2632f168be36 to disappear -May 13 20:01:47.109: INFO: Pod downwardapi-volume-eef892b0-75b9-11e9-8f67-2632f168be36 no longer exists -[AfterEach] [sig-storage] Downward API volume - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 20:01:47.110: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-downward-api-r6z8h" for this suite. -May 13 20:01:53.151: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 20:01:53.414: INFO: namespace: e2e-tests-downward-api-r6z8h, resource: bindings, ignored listing per whitelist -May 13 20:01:53.437: INFO: namespace e2e-tests-downward-api-r6z8h deletion completed in 6.320109789s - -• [SLOW TEST:8.703 seconds] -[sig-storage] Downward API volume -/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:34 - should provide podname only [NodeConformance] [Conformance] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------- -SSSSSSSSS ------------------------------- -[sig-storage] Secrets - should be consumable from pods in volume with mappings and Item Mode set [NodeConformance] [Conformance] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -[BeforeEach] [sig-storage] Secrets - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 -STEP: Creating a kubernetes client -May 13 20:01:53.438: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -STEP: Building a namespace api object, basename secrets -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-secrets-fjv5b -STEP: Waiting for a default service account to be provisioned in namespace -[It] should be consumable from pods in volume with mappings and Item Mode set [NodeConformance] [Conformance] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -STEP: Creating secret with name secret-test-map-f42639a6-75b9-11e9-8f67-2632f168be36 -STEP: Creating a pod to test consume secrets -May 13 20:01:53.731: INFO: Waiting up to 5m0s for pod "pod-secrets-f427ad56-75b9-11e9-8f67-2632f168be36" in namespace "e2e-tests-secrets-fjv5b" to be "success or failure" -May 13 20:01:53.742: INFO: Pod "pod-secrets-f427ad56-75b9-11e9-8f67-2632f168be36": Phase="Pending", Reason="", readiness=false. Elapsed: 10.591861ms -May 13 20:01:55.749: INFO: Pod "pod-secrets-f427ad56-75b9-11e9-8f67-2632f168be36": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.01845387s -STEP: Saw pod success -May 13 20:01:55.749: INFO: Pod "pod-secrets-f427ad56-75b9-11e9-8f67-2632f168be36" satisfied condition "success or failure" -May 13 20:01:55.757: INFO: Trying to get logs from node 10.170.219.153 pod pod-secrets-f427ad56-75b9-11e9-8f67-2632f168be36 container secret-volume-test: -STEP: delete the pod -May 13 20:01:55.798: INFO: Waiting for pod pod-secrets-f427ad56-75b9-11e9-8f67-2632f168be36 to disappear -May 13 20:01:55.805: INFO: Pod pod-secrets-f427ad56-75b9-11e9-8f67-2632f168be36 no longer exists -[AfterEach] [sig-storage] Secrets - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 20:01:55.805: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-secrets-fjv5b" for this suite. -May 13 20:02:01.843: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 20:02:02.091: INFO: namespace: e2e-tests-secrets-fjv5b, resource: bindings, ignored listing per whitelist -May 13 20:02:02.138: INFO: namespace e2e-tests-secrets-fjv5b deletion completed in 6.324254546s - -• [SLOW TEST:8.700 seconds] -[sig-storage] Secrets -/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/secrets_volume.go:34 - should be consumable from pods in volume with mappings and Item Mode set [NodeConformance] [Conformance] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------- -S ------------------------------- -[sig-apps] Deployment - RollingUpdateDeployment should delete old pods and create new ones [Conformance] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -[BeforeEach] [sig-apps] Deployment - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 -STEP: Creating a kubernetes client -May 13 20:02:02.139: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -STEP: Building a namespace api object, basename deployment -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-deployment-vsgzr -STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [sig-apps] Deployment - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:65 -[It] RollingUpdateDeployment should delete old pods and create new ones [Conformance] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -May 13 20:02:02.425: INFO: Creating replica set "test-rolling-update-controller" (going to be adopted) -May 13 20:02:02.441: INFO: Pod name sample-pod: Found 0 pods out of 1 -May 13 20:02:07.450: INFO: Pod name sample-pod: Found 1 pods out of 1 -STEP: ensuring each pod is running -May 13 20:02:07.450: INFO: Creating deployment "test-rolling-update-deployment" -May 13 20:02:07.533: INFO: Ensuring deployment "test-rolling-update-deployment" gets the next revision from the one the adopted replica set "test-rolling-update-controller" has -May 13 20:02:07.547: INFO: new replicaset for deployment "test-rolling-update-deployment" is yet to be created -May 13 20:02:09.561: INFO: Ensuring status for deployment "test-rolling-update-deployment" is the expected -May 13 20:02:09.566: INFO: Ensuring deployment "test-rolling-update-deployment" has one old replica set (the one it adopted) -[AfterEach] [sig-apps] Deployment - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:59 -May 13 20:02:09.584: INFO: Deployment "test-rolling-update-deployment": -&Deployment{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-rolling-update-deployment,GenerateName:,Namespace:e2e-tests-deployment-vsgzr,SelfLink:/apis/apps/v1/namespaces/e2e-tests-deployment-vsgzr/deployments/test-rolling-update-deployment,UID:fc62864c-75b9-11e9-b786-da20024d205c,ResourceVersion:51644,Generation:1,CreationTimestamp:2019-05-13 20:02:07 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod,},Annotations:map[string]string{deployment.kubernetes.io/revision: 3546343826724305833,},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Spec:DeploymentSpec{Replicas:*1,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[],Containers:[{redis gcr.io/kubernetes-e2e-test-images/redis:1.0 [] [] [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,},},Strategy:DeploymentStrategy{Type:RollingUpdate,RollingUpdate:&RollingUpdateDeployment{MaxUnavailable:25%!,(MISSING)MaxSurge:25%!,(MISSING)},},MinReadySeconds:0,RevisionHistoryLimit:*10,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:1,Replicas:1,UpdatedReplicas:1,AvailableReplicas:1,UnavailableReplicas:0,Conditions:[{Available True 2019-05-13 20:02:07 +0000 UTC 2019-05-13 20:02:07 +0000 UTC MinimumReplicasAvailable Deployment has minimum availability.} {Progressing True 2019-05-13 20:02:08 +0000 UTC 2019-05-13 20:02:07 +0000 UTC NewReplicaSetAvailable ReplicaSet "test-rolling-update-deployment-68b55d7bc6" has successfully progressed.}],ReadyReplicas:1,CollisionCount:nil,},} - -May 13 20:02:09.591: INFO: New ReplicaSet "test-rolling-update-deployment-68b55d7bc6" of Deployment "test-rolling-update-deployment": -&ReplicaSet{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-rolling-update-deployment-68b55d7bc6,GenerateName:,Namespace:e2e-tests-deployment-vsgzr,SelfLink:/apis/apps/v1/namespaces/e2e-tests-deployment-vsgzr/replicasets/test-rolling-update-deployment-68b55d7bc6,UID:fc67b49b-75b9-11e9-83a8-7e0242378207,ResourceVersion:51635,Generation:1,CreationTimestamp:2019-05-13 20:02:07 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod,pod-template-hash: 68b55d7bc6,},Annotations:map[string]string{deployment.kubernetes.io/desired-replicas: 1,deployment.kubernetes.io/max-replicas: 2,deployment.kubernetes.io/revision: 3546343826724305833,},OwnerReferences:[{apps/v1 Deployment test-rolling-update-deployment fc62864c-75b9-11e9-b786-da20024d205c 0xc000d28117 0xc000d28118}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:ReplicaSetSpec{Replicas:*1,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod,pod-template-hash: 68b55d7bc6,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod,pod-template-hash: 68b55d7bc6,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[],Containers:[{redis gcr.io/kubernetes-e2e-test-images/redis:1.0 [] [] [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,},},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:1,FullyLabeledReplicas:1,ObservedGeneration:1,ReadyReplicas:1,AvailableReplicas:1,Conditions:[],},} -May 13 20:02:09.591: INFO: All old ReplicaSets of Deployment "test-rolling-update-deployment": -May 13 20:02:09.591: INFO: &ReplicaSet{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-rolling-update-controller,GenerateName:,Namespace:e2e-tests-deployment-vsgzr,SelfLink:/apis/apps/v1/namespaces/e2e-tests-deployment-vsgzr/replicasets/test-rolling-update-controller,UID:f959494b-75b9-11e9-b786-da20024d205c,ResourceVersion:51643,Generation:2,CreationTimestamp:2019-05-13 20:02:02 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod,pod: nginx,},Annotations:map[string]string{deployment.kubernetes.io/desired-replicas: 1,deployment.kubernetes.io/max-replicas: 2,deployment.kubernetes.io/revision: 3546343826724305832,},OwnerReferences:[{apps/v1 Deployment test-rolling-update-deployment fc62864c-75b9-11e9-b786-da20024d205c 0xc001df3f67 0xc001df3f68}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:ReplicaSetSpec{Replicas:*0,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod,pod: nginx,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod,pod: nginx,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,},},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:0,FullyLabeledReplicas:0,ObservedGeneration:2,ReadyReplicas:0,AvailableReplicas:0,Conditions:[],},} -May 13 20:02:09.599: INFO: Pod "test-rolling-update-deployment-68b55d7bc6-xxznx" is available: -&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-rolling-update-deployment-68b55d7bc6-xxznx,GenerateName:test-rolling-update-deployment-68b55d7bc6-,Namespace:e2e-tests-deployment-vsgzr,SelfLink:/api/v1/namespaces/e2e-tests-deployment-vsgzr/pods/test-rolling-update-deployment-68b55d7bc6-xxznx,UID:fc693028-75b9-11e9-83a8-7e0242378207,ResourceVersion:51634,Generation:0,CreationTimestamp:2019-05-13 20:02:07 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod,pod-template-hash: 68b55d7bc6,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet test-rolling-update-deployment-68b55d7bc6 fc67b49b-75b9-11e9-83a8-7e0242378207 0xc001c52987 0xc001c52988}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-c7v8v {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-c7v8v,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{redis gcr.io/kubernetes-e2e-test-images/redis:1.0 [] [] [] [] [] {map[] map[]} [{default-token-c7v8v true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.170.219.153,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc001c52a90} {node.kubernetes.io/unreachable Exists NoExecute 0xc001c52ab0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Running,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 20:02:07 +0000 UTC } {Ready True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 20:02:08 +0000 UTC } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 20:02:08 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-05-13 20:02:07 +0000 UTC }],Message:,Reason:,HostIP:10.170.219.153,PodIP:172.30.63.46,StartTime:2019-05-13 20:02:07 +0000 UTC,ContainerStatuses:[{redis {nil ContainerStateRunning{StartedAt:2019-05-13 20:02:08 +0000 UTC,} nil} {nil nil nil} true 0 gcr.io/kubernetes-e2e-test-images/redis:1.0 gcr.io/kubernetes-e2e-test-images/redis@sha256:af4748d1655c08dc54d4be5182135395db9ce87aba2d4699b26b14ae197c5830 containerd://cd38f4553f3bef5fc9fdd0114c85b6295708fd2b92d939e79ca76e8248a90c16}],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} -[AfterEach] [sig-apps] Deployment - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 20:02:09.599: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-deployment-vsgzr" for this suite. -May 13 20:02:15.638: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 20:02:15.857: INFO: namespace: e2e-tests-deployment-vsgzr, resource: bindings, ignored listing per whitelist -May 13 20:02:16.080: INFO: namespace e2e-tests-deployment-vsgzr deletion completed in 6.472995899s - -• [SLOW TEST:13.941 seconds] -[sig-apps] Deployment -/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22 - RollingUpdateDeployment should delete old pods and create new ones [Conformance] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------- -SSSSSSS ------------------------------- -[sig-storage] Projected combined - should project all components that make up the projection API [Projection][NodeConformance] [Conformance] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -[BeforeEach] [sig-storage] Projected combined - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 -STEP: Creating a kubernetes client -May 13 20:02:16.082: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -STEP: Building a namespace api object, basename projected -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-projected-2fqf8 -STEP: Waiting for a default service account to be provisioned in namespace -[It] should project all components that make up the projection API [Projection][NodeConformance] [Conformance] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -STEP: Creating configMap with name configmap-projected-all-test-volume-01a851a0-75ba-11e9-8f67-2632f168be36 -STEP: Creating secret with name secret-projected-all-test-volume-01a85185-75ba-11e9-8f67-2632f168be36 -STEP: Creating a pod to test Check all projections for projected volume plugin -May 13 20:02:16.401: INFO: Waiting up to 5m0s for pod "projected-volume-01a85147-75ba-11e9-8f67-2632f168be36" in namespace "e2e-tests-projected-2fqf8" to be "success or failure" -May 13 20:02:16.409: INFO: Pod "projected-volume-01a85147-75ba-11e9-8f67-2632f168be36": Phase="Pending", Reason="", readiness=false. Elapsed: 7.585263ms -May 13 20:02:18.417: INFO: Pod "projected-volume-01a85147-75ba-11e9-8f67-2632f168be36": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.015795469s -STEP: Saw pod success -May 13 20:02:18.417: INFO: Pod "projected-volume-01a85147-75ba-11e9-8f67-2632f168be36" satisfied condition "success or failure" -May 13 20:02:18.426: INFO: Trying to get logs from node 10.170.219.177 pod projected-volume-01a85147-75ba-11e9-8f67-2632f168be36 container projected-all-volume-test: -STEP: delete the pod -May 13 20:02:18.472: INFO: Waiting for pod projected-volume-01a85147-75ba-11e9-8f67-2632f168be36 to disappear -May 13 20:02:18.479: INFO: Pod projected-volume-01a85147-75ba-11e9-8f67-2632f168be36 no longer exists -[AfterEach] [sig-storage] Projected combined - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 20:02:18.479: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-projected-2fqf8" for this suite. -May 13 20:02:24.521: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 20:02:24.534: INFO: namespace: e2e-tests-projected-2fqf8, resource: bindings, ignored listing per whitelist -May 13 20:02:24.759: INFO: namespace e2e-tests-projected-2fqf8 deletion completed in 6.271633448s - -• [SLOW TEST:8.677 seconds] -[sig-storage] Projected combined -/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_combined.go:31 - should project all components that make up the projection API [Projection][NodeConformance] [Conformance] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------- -SSSSSSSSSS ------------------------------- -[k8s.io] InitContainer [NodeConformance] - should invoke init containers on a RestartNever pod [Conformance] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -[BeforeEach] [k8s.io] InitContainer [NodeConformance] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 -STEP: Creating a kubernetes client -May 13 20:02:24.759: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -STEP: Building a namespace api object, basename init-container -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-init-container-2jchj -STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [k8s.io] InitContainer [NodeConformance] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/init_container.go:43 -[It] should invoke init containers on a RestartNever pod [Conformance] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -STEP: creating the pod -May 13 20:02:25.034: INFO: PodSpec: initContainers in spec.initContainers -[AfterEach] [k8s.io] InitContainer [NodeConformance] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 20:02:29.955: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-init-container-2jchj" for this suite. -May 13 20:02:35.995: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 20:02:36.139: INFO: namespace: e2e-tests-init-container-2jchj, resource: bindings, ignored listing per whitelist -May 13 20:02:36.523: INFO: namespace e2e-tests-init-container-2jchj deletion completed in 6.558950599s - -• [SLOW TEST:11.764 seconds] -[k8s.io] InitContainer [NodeConformance] -/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 - should invoke init containers on a RestartNever pod [Conformance] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------- -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +SSSSSSSSSSSSSS ------------------------------ [sig-api-machinery] Garbage collector - should keep the rc around until all its pods are deleted if the deleteOptions says so [Conformance] + should not delete dependents that have both valid owner and owner that's waiting for dependents to be deleted [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-api-machinery] Garbage collector /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client -May 13 20:02:36.523: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 +Jun 18 13:03:31.674: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 STEP: Building a namespace api object, basename gc -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-gc-g7ls7 +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-gc-z4c8g STEP: Waiting for a default service account to be provisioned in namespace -[It] should keep the rc around until all its pods are deleted if the deleteOptions says so [Conformance] +[It] should not delete dependents that have both valid owner and owner that's waiting for dependents to be deleted [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -STEP: create the rc -STEP: delete the rc +STEP: create the rc1 +STEP: create the rc2 +STEP: set half of pods created by rc simpletest-rc-to-be-deleted to have rc simpletest-rc-to-stay as owner as well +STEP: delete the rc simpletest-rc-to-be-deleted STEP: wait for the rc to be deleted STEP: Gathering metrics -W0513 20:02:43.028261 16 metrics_grabber.go:81] Master node is not registered. Grabbing metrics from Scheduler, ControllerManager and ClusterAutoscaler is disabled. -May 13 20:02:43.028: INFO: For apiserver_request_count: +W0618 13:03:42.416296 17 metrics_grabber.go:81] Master node is not registered. Grabbing metrics from Scheduler, ControllerManager and ClusterAutoscaler is disabled. +Jun 18 13:03:42.416: INFO: For apiserver_request_count: For apiserver_request_latencies_summary: For etcd_helper_cache_entry_count: For etcd_helper_cache_hit_count: @@ -8968,826 +9092,778 @@ For namespace_queue_latency_count: For namespace_retries: For namespace_work_duration: For namespace_work_duration_sum: -For namespace_work_duration_count: -For function_duration_seconds: -For errors_total: -For evicted_pods_total: - -[AfterEach] [sig-api-machinery] Garbage collector - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 20:02:43.028: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-gc-g7ls7" for this suite. -May 13 20:02:51.076: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 20:02:51.369: INFO: namespace: e2e-tests-gc-g7ls7, resource: bindings, ignored listing per whitelist -May 13 20:02:51.532: INFO: namespace e2e-tests-gc-g7ls7 deletion completed in 8.498316354s - -• [SLOW TEST:15.009 seconds] -[sig-api-machinery] Garbage collector -/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:22 - should keep the rc around until all its pods are deleted if the deleteOptions says so [Conformance] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------- -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------- -[k8s.io] Container Lifecycle Hook when create a pod with lifecycle hook - should execute prestop http hook properly [NodeConformance] [Conformance] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -[BeforeEach] [k8s.io] Container Lifecycle Hook - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 -STEP: Creating a kubernetes client -May 13 20:02:51.533: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -STEP: Building a namespace api object, basename container-lifecycle-hook -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-container-lifecycle-hook-r2n9h -STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] when create a pod with lifecycle hook - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:61 -STEP: create the container to handle the HTTPGet hook request. -[It] should execute prestop http hook properly [NodeConformance] [Conformance] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -STEP: create the pod with lifecycle hook -STEP: delete the pod with lifecycle hook -May 13 20:02:55.911: INFO: Waiting for pod pod-with-prestop-http-hook to disappear -May 13 20:02:55.918: INFO: Pod pod-with-prestop-http-hook still exists -May 13 20:02:57.918: INFO: Waiting for pod pod-with-prestop-http-hook to disappear -May 13 20:02:57.927: INFO: Pod pod-with-prestop-http-hook still exists -May 13 20:02:59.918: INFO: Waiting for pod pod-with-prestop-http-hook to disappear -May 13 20:02:59.927: INFO: Pod pod-with-prestop-http-hook still exists -May 13 20:03:01.918: INFO: Waiting for pod pod-with-prestop-http-hook to disappear -May 13 20:03:01.927: INFO: Pod pod-with-prestop-http-hook still exists -May 13 20:03:03.918: INFO: Waiting for pod pod-with-prestop-http-hook to disappear -May 13 20:03:03.928: INFO: Pod pod-with-prestop-http-hook still exists -May 13 20:03:05.918: INFO: Waiting for pod pod-with-prestop-http-hook to disappear -May 13 20:03:05.939: INFO: Pod pod-with-prestop-http-hook still exists -May 13 20:03:07.918: INFO: Waiting for pod pod-with-prestop-http-hook to disappear -May 13 20:03:07.928: INFO: Pod pod-with-prestop-http-hook no longer exists -STEP: check prestop hook -[AfterEach] [k8s.io] Container Lifecycle Hook - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 20:03:07.947: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-container-lifecycle-hook-r2n9h" for this suite. -May 13 20:03:32.052: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 20:03:32.270: INFO: namespace: e2e-tests-container-lifecycle-hook-r2n9h, resource: bindings, ignored listing per whitelist -May 13 20:03:32.335: INFO: namespace e2e-tests-container-lifecycle-hook-r2n9h deletion completed in 24.380732071s - -• [SLOW TEST:40.802 seconds] -[k8s.io] Container Lifecycle Hook -/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 - when create a pod with lifecycle hook - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:40 - should execute prestop http hook properly [NodeConformance] [Conformance] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------- -SSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------- -[sig-apps] Daemon set [Serial] - should retry creating failed daemon pods [Conformance] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -[BeforeEach] [sig-apps] Daemon set [Serial] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 -STEP: Creating a kubernetes client -May 13 20:03:32.336: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -STEP: Building a namespace api object, basename daemonsets -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-daemonsets-cm8h2 -STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [sig-apps] Daemon set [Serial] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:102 -[It] should retry creating failed daemon pods [Conformance] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -STEP: Creating a simple DaemonSet "daemon-set" -STEP: Check that daemon pods launch on every node of the cluster. -May 13 20:03:32.666: INFO: Number of nodes with available pods: 0 -May 13 20:03:32.666: INFO: Node 10.170.219.151 is running more than one daemon pod -May 13 20:03:33.684: INFO: Number of nodes with available pods: 0 -May 13 20:03:33.684: INFO: Node 10.170.219.151 is running more than one daemon pod -May 13 20:03:34.682: INFO: Number of nodes with available pods: 3 -May 13 20:03:34.682: INFO: Number of running nodes: 3, number of available pods: 3 -STEP: Set a daemon pod's phase to 'Failed', check that the daemon pod is revived. -May 13 20:03:34.719: INFO: Number of nodes with available pods: 2 -May 13 20:03:34.719: INFO: Node 10.170.219.153 is running more than one daemon pod -May 13 20:03:35.735: INFO: Number of nodes with available pods: 2 -May 13 20:03:35.735: INFO: Node 10.170.219.153 is running more than one daemon pod -May 13 20:03:36.736: INFO: Number of nodes with available pods: 3 -May 13 20:03:36.736: INFO: Number of running nodes: 3, number of available pods: 3 -STEP: Wait for the failed daemon pod to be completely deleted. -[AfterEach] [sig-apps] Daemon set [Serial] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:68 -STEP: Deleting DaemonSet "daemon-set" -STEP: deleting DaemonSet.extensions daemon-set in namespace e2e-tests-daemonsets-cm8h2, will wait for the garbage collector to delete the pods -May 13 20:03:36.819: INFO: Deleting DaemonSet.extensions daemon-set took: 11.937268ms -May 13 20:03:36.919: INFO: Terminating DaemonSet.extensions daemon-set pods took: 100.293141ms -May 13 20:04:12.541: INFO: Number of nodes with available pods: 0 -May 13 20:04:12.541: INFO: Number of running nodes: 0, number of available pods: 0 -May 13 20:04:12.548: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"selfLink":"/apis/apps/v1/namespaces/e2e-tests-daemonsets-cm8h2/daemonsets","resourceVersion":"52447"},"items":null} - -May 13 20:04:12.555: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"selfLink":"/api/v1/namespaces/e2e-tests-daemonsets-cm8h2/pods","resourceVersion":"52447"},"items":null} - -[AfterEach] [sig-apps] Daemon set [Serial] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 20:04:12.578: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-daemonsets-cm8h2" for this suite. -May 13 20:04:18.617: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 20:04:19.092: INFO: namespace: e2e-tests-daemonsets-cm8h2, resource: bindings, ignored listing per whitelist -May 13 20:04:19.141: INFO: namespace e2e-tests-daemonsets-cm8h2 deletion completed in 6.557642783s - -• [SLOW TEST:46.805 seconds] -[sig-apps] Daemon set [Serial] -/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22 - should retry creating failed daemon pods [Conformance] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------- -SSSSSSSSS ------------------------------- -[sig-api-machinery] CustomResourceDefinition resources Simple CustomResourceDefinition - creating/deleting custom resource definition objects works [Conformance] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -[BeforeEach] [sig-api-machinery] CustomResourceDefinition resources - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 -STEP: Creating a kubernetes client -May 13 20:04:19.141: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -STEP: Building a namespace api object, basename custom-resource-definition -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-custom-resource-definition-qz292 -STEP: Waiting for a default service account to be provisioned in namespace -[It] creating/deleting custom resource definition objects works [Conformance] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -May 13 20:04:19.528: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -[AfterEach] [sig-api-machinery] CustomResourceDefinition resources - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 20:04:20.616: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-custom-resource-definition-qz292" for this suite. -May 13 20:04:26.858: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 20:04:26.963: INFO: namespace: e2e-tests-custom-resource-definition-qz292, resource: bindings, ignored listing per whitelist -May 13 20:04:27.333: INFO: namespace e2e-tests-custom-resource-definition-qz292 deletion completed in 6.708989062s - -• [SLOW TEST:8.192 seconds] -[sig-api-machinery] CustomResourceDefinition resources -/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:22 - Simple CustomResourceDefinition - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/custom_resource_definition.go:35 - creating/deleting custom resource definition objects works [Conformance] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------- -SSSSSSSSSSSSSSSSSSSSS ------------------------------- -[sig-storage] Projected secret - should be consumable from pods in volume [NodeConformance] [Conformance] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -[BeforeEach] [sig-storage] Projected secret - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 -STEP: Creating a kubernetes client -May 13 20:04:27.336: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -STEP: Building a namespace api object, basename projected -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-projected-4zktd -STEP: Waiting for a default service account to be provisioned in namespace -[It] should be consumable from pods in volume [NodeConformance] [Conformance] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -STEP: Creating projection with secret that has name projected-secret-test-4fe7ee08-75ba-11e9-8f67-2632f168be36 -STEP: Creating a pod to test consume secrets -May 13 20:04:27.736: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-4ff2f61b-75ba-11e9-8f67-2632f168be36" in namespace "e2e-tests-projected-4zktd" to be "success or failure" -May 13 20:04:27.744: INFO: Pod "pod-projected-secrets-4ff2f61b-75ba-11e9-8f67-2632f168be36": Phase="Pending", Reason="", readiness=false. Elapsed: 7.112537ms -May 13 20:04:29.752: INFO: Pod "pod-projected-secrets-4ff2f61b-75ba-11e9-8f67-2632f168be36": Phase="Running", Reason="", readiness=true. Elapsed: 2.01530147s -May 13 20:04:31.763: INFO: Pod "pod-projected-secrets-4ff2f61b-75ba-11e9-8f67-2632f168be36": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.026605478s -STEP: Saw pod success -May 13 20:04:31.763: INFO: Pod "pod-projected-secrets-4ff2f61b-75ba-11e9-8f67-2632f168be36" satisfied condition "success or failure" -May 13 20:04:31.771: INFO: Trying to get logs from node 10.170.219.177 pod pod-projected-secrets-4ff2f61b-75ba-11e9-8f67-2632f168be36 container projected-secret-volume-test: -STEP: delete the pod -May 13 20:04:31.816: INFO: Waiting for pod pod-projected-secrets-4ff2f61b-75ba-11e9-8f67-2632f168be36 to disappear -May 13 20:04:31.826: INFO: Pod pod-projected-secrets-4ff2f61b-75ba-11e9-8f67-2632f168be36 no longer exists -[AfterEach] [sig-storage] Projected secret +For namespace_work_duration_count: +For function_duration_seconds: +For errors_total: +For evicted_pods_total: + +[AfterEach] [sig-api-machinery] Garbage collector /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 20:04:31.826: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-projected-4zktd" for this suite. -May 13 20:04:37.863: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 20:04:38.002: INFO: namespace: e2e-tests-projected-4zktd, resource: bindings, ignored listing per whitelist -May 13 20:04:38.181: INFO: namespace e2e-tests-projected-4zktd deletion completed in 6.347752119s +Jun 18 13:03:42.416: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "e2e-tests-gc-z4c8g" for this suite. +Jun 18 13:03:50.476: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 13:03:50.841: INFO: namespace: e2e-tests-gc-z4c8g, resource: bindings, ignored listing per whitelist +Jun 18 13:03:50.943: INFO: namespace e2e-tests-gc-z4c8g deletion completed in 8.51323857s -• [SLOW TEST:10.846 seconds] -[sig-storage] Projected secret -/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_secret.go:34 - should be consumable from pods in volume [NodeConformance] [Conformance] +• [SLOW TEST:19.270 seconds] +[sig-api-machinery] Garbage collector +/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:22 + should not delete dependents that have both valid owner and owner that's waiting for dependents to be deleted [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ -SSS +SS ------------------------------ -[k8s.io] Docker Containers - should be able to override the image's default command and arguments [NodeConformance] [Conformance] +[sig-cli] Kubectl client [k8s.io] Kubectl rolling-update + should support rolling-update to same image [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -[BeforeEach] [k8s.io] Docker Containers +[BeforeEach] [sig-cli] Kubectl client /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client -May 13 20:04:38.182: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -STEP: Building a namespace api object, basename containers -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-containers-8h94c +Jun 18 13:03:50.944: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 +STEP: Building a namespace api object, basename kubectl +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-kubectl-cffwb STEP: Waiting for a default service account to be provisioned in namespace -[It] should be able to override the image's default command and arguments [NodeConformance] [Conformance] +[BeforeEach] [sig-cli] Kubectl client + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:243 +[BeforeEach] [k8s.io] Kubectl rolling-update + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1358 +[It] should support rolling-update to same image [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -STEP: Creating a pod to test override all -May 13 20:04:38.484: INFO: Waiting up to 5m0s for pod "client-containers-565adc94-75ba-11e9-8f67-2632f168be36" in namespace "e2e-tests-containers-8h94c" to be "success or failure" -May 13 20:04:38.492: INFO: Pod "client-containers-565adc94-75ba-11e9-8f67-2632f168be36": Phase="Pending", Reason="", readiness=false. Elapsed: 7.968175ms -May 13 20:04:40.520: INFO: Pod "client-containers-565adc94-75ba-11e9-8f67-2632f168be36": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.035577737s -STEP: Saw pod success -May 13 20:04:40.520: INFO: Pod "client-containers-565adc94-75ba-11e9-8f67-2632f168be36" satisfied condition "success or failure" -May 13 20:04:40.527: INFO: Trying to get logs from node 10.170.219.153 pod client-containers-565adc94-75ba-11e9-8f67-2632f168be36 container test-container: -STEP: delete the pod -May 13 20:04:40.581: INFO: Waiting for pod client-containers-565adc94-75ba-11e9-8f67-2632f168be36 to disappear -May 13 20:04:40.590: INFO: Pod client-containers-565adc94-75ba-11e9-8f67-2632f168be36 no longer exists -[AfterEach] [k8s.io] Docker Containers +STEP: running the image docker.io/library/nginx:1.14-alpine +Jun 18 13:03:51.450: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 run e2e-test-nginx-rc --image=docker.io/library/nginx:1.14-alpine --generator=run/v1 --namespace=e2e-tests-kubectl-cffwb' +Jun 18 13:03:52.283: INFO: stderr: "kubectl run --generator=run/v1 is DEPRECATED and will be removed in a future version. Use kubectl run --generator=run-pod/v1 or kubectl create instead.\n" +Jun 18 13:03:52.283: INFO: stdout: "replicationcontroller/e2e-test-nginx-rc created\n" +STEP: verifying the rc e2e-test-nginx-rc was created +STEP: rolling-update to same image controller +Jun 18 13:03:52.392: INFO: scanned /root for discovery docs: +Jun 18 13:03:52.392: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 rolling-update e2e-test-nginx-rc --update-period=1s --image=docker.io/library/nginx:1.14-alpine --image-pull-policy=IfNotPresent --namespace=e2e-tests-kubectl-cffwb' +Jun 18 13:04:08.438: INFO: stderr: "Command \"rolling-update\" is deprecated, use \"rollout\" instead\n" +Jun 18 13:04:08.438: INFO: stdout: "Created e2e-test-nginx-rc-5c3b0d70e485ce34b8d17399023285b6\nScaling up e2e-test-nginx-rc-5c3b0d70e485ce34b8d17399023285b6 from 0 to 1, scaling down e2e-test-nginx-rc from 1 to 0 (keep 1 pods available, don't exceed 2 pods)\nScaling e2e-test-nginx-rc-5c3b0d70e485ce34b8d17399023285b6 up to 1\nScaling e2e-test-nginx-rc down to 0\nUpdate succeeded. Deleting old controller: e2e-test-nginx-rc\nRenaming e2e-test-nginx-rc-5c3b0d70e485ce34b8d17399023285b6 to e2e-test-nginx-rc\nreplicationcontroller/e2e-test-nginx-rc rolling updated\n" +Jun 18 13:04:08.438: INFO: stdout: "Created e2e-test-nginx-rc-5c3b0d70e485ce34b8d17399023285b6\nScaling up e2e-test-nginx-rc-5c3b0d70e485ce34b8d17399023285b6 from 0 to 1, scaling down e2e-test-nginx-rc from 1 to 0 (keep 1 pods available, don't exceed 2 pods)\nScaling e2e-test-nginx-rc-5c3b0d70e485ce34b8d17399023285b6 up to 1\nScaling e2e-test-nginx-rc down to 0\nUpdate succeeded. Deleting old controller: e2e-test-nginx-rc\nRenaming e2e-test-nginx-rc-5c3b0d70e485ce34b8d17399023285b6 to e2e-test-nginx-rc\nreplicationcontroller/e2e-test-nginx-rc rolling updated\n" +STEP: waiting for all containers in run=e2e-test-nginx-rc pods to come up. +Jun 18 13:04:08.438: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l run=e2e-test-nginx-rc --namespace=e2e-tests-kubectl-cffwb' +Jun 18 13:04:08.577: INFO: stderr: "" +Jun 18 13:04:08.577: INFO: stdout: "e2e-test-nginx-rc-5c3b0d70e485ce34b8d17399023285b6-t5plg " +Jun 18 13:04:08.577: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 get pods e2e-test-nginx-rc-5c3b0d70e485ce34b8d17399023285b6-t5plg -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "e2e-test-nginx-rc") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=e2e-tests-kubectl-cffwb' +Jun 18 13:04:08.753: INFO: stderr: "" +Jun 18 13:04:08.753: INFO: stdout: "true" +Jun 18 13:04:08.753: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 get pods e2e-test-nginx-rc-5c3b0d70e485ce34b8d17399023285b6-t5plg -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "e2e-test-nginx-rc"}}{{.image}}{{end}}{{end}}{{end}} --namespace=e2e-tests-kubectl-cffwb' +Jun 18 13:04:08.902: INFO: stderr: "" +Jun 18 13:04:08.902: INFO: stdout: "docker.io/library/nginx:1.14-alpine" +Jun 18 13:04:08.902: INFO: e2e-test-nginx-rc-5c3b0d70e485ce34b8d17399023285b6-t5plg is verified up and running +[AfterEach] [k8s.io] Kubectl rolling-update + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1364 +Jun 18 13:04:08.902: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 delete rc e2e-test-nginx-rc --namespace=e2e-tests-kubectl-cffwb' +Jun 18 13:04:09.161: INFO: stderr: "" +Jun 18 13:04:09.161: INFO: stdout: "replicationcontroller \"e2e-test-nginx-rc\" deleted\n" +[AfterEach] [sig-cli] Kubectl client /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 20:04:40.591: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-containers-8h94c" for this suite. -May 13 20:04:46.628: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 20:04:46.686: INFO: namespace: e2e-tests-containers-8h94c, resource: bindings, ignored listing per whitelist -May 13 20:04:46.855: INFO: namespace e2e-tests-containers-8h94c deletion completed in 6.256348638s +Jun 18 13:04:09.161: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "e2e-tests-kubectl-cffwb" for this suite. +Jun 18 13:04:33.244: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 13:04:33.739: INFO: namespace: e2e-tests-kubectl-cffwb, resource: bindings, ignored listing per whitelist +Jun 18 13:04:33.765: INFO: namespace e2e-tests-kubectl-cffwb deletion completed in 24.581208099s -• [SLOW TEST:8.674 seconds] -[k8s.io] Docker Containers -/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 - should be able to override the image's default command and arguments [NodeConformance] [Conformance] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 +• [SLOW TEST:42.822 seconds] +[sig-cli] Kubectl client +/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:22 + [k8s.io] Kubectl rolling-update + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 + should support rolling-update to same image [Conformance] + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ -SSSS +SS ------------------------------ -[sig-storage] Secrets - should be consumable from pods in volume with defaultMode set [NodeConformance] [Conformance] +[sig-apps] Deployment + RollingUpdateDeployment should delete old pods and create new ones [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -[BeforeEach] [sig-storage] Secrets +[BeforeEach] [sig-apps] Deployment /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client -May 13 20:04:46.855: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -STEP: Building a namespace api object, basename secrets -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-secrets-tzfhp +Jun 18 13:04:33.765: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 +STEP: Building a namespace api object, basename deployment +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-deployment-68kjh STEP: Waiting for a default service account to be provisioned in namespace -[It] should be consumable from pods in volume with defaultMode set [NodeConformance] [Conformance] +[BeforeEach] [sig-apps] Deployment + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:65 +[It] RollingUpdateDeployment should delete old pods and create new ones [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -STEP: Creating secret with name secret-test-5b831681-75ba-11e9-8f67-2632f168be36 -STEP: Creating a pod to test consume secrets -May 13 20:04:47.143: INFO: Waiting up to 5m0s for pod "pod-secrets-5b848186-75ba-11e9-8f67-2632f168be36" in namespace "e2e-tests-secrets-tzfhp" to be "success or failure" -May 13 20:04:47.151: INFO: Pod "pod-secrets-5b848186-75ba-11e9-8f67-2632f168be36": Phase="Pending", Reason="", readiness=false. Elapsed: 7.809287ms -May 13 20:04:49.159: INFO: Pod "pod-secrets-5b848186-75ba-11e9-8f67-2632f168be36": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.015407416s -STEP: Saw pod success -May 13 20:04:49.159: INFO: Pod "pod-secrets-5b848186-75ba-11e9-8f67-2632f168be36" satisfied condition "success or failure" -May 13 20:04:49.166: INFO: Trying to get logs from node 10.170.219.177 pod pod-secrets-5b848186-75ba-11e9-8f67-2632f168be36 container secret-volume-test: -STEP: delete the pod -May 13 20:04:49.204: INFO: Waiting for pod pod-secrets-5b848186-75ba-11e9-8f67-2632f168be36 to disappear -May 13 20:04:49.211: INFO: Pod pod-secrets-5b848186-75ba-11e9-8f67-2632f168be36 no longer exists -[AfterEach] [sig-storage] Secrets +Jun 18 13:04:34.220: INFO: Creating replica set "test-rolling-update-controller" (going to be adopted) +Jun 18 13:04:34.249: INFO: Pod name sample-pod: Found 0 pods out of 1 +Jun 18 13:04:39.265: INFO: Pod name sample-pod: Found 1 pods out of 1 +STEP: ensuring each pod is running +Jun 18 13:04:39.265: INFO: Creating deployment "test-rolling-update-deployment" +Jun 18 13:04:39.280: INFO: Ensuring deployment "test-rolling-update-deployment" gets the next revision from the one the adopted replica set "test-rolling-update-controller" has +Jun 18 13:04:39.306: INFO: new replicaset for deployment "test-rolling-update-deployment" is yet to be created +Jun 18 13:04:41.336: INFO: Ensuring status for deployment "test-rolling-update-deployment" is the expected +Jun 18 13:04:41.349: INFO: Ensuring deployment "test-rolling-update-deployment" has one old replica set (the one it adopted) +[AfterEach] [sig-apps] Deployment + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:59 +Jun 18 13:04:41.393: INFO: Deployment "test-rolling-update-deployment": +&Deployment{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-rolling-update-deployment,GenerateName:,Namespace:e2e-tests-deployment-68kjh,SelfLink:/apis/apps/v1/namespaces/e2e-tests-deployment-68kjh/deployments/test-rolling-update-deployment,UID:a156030f-91c9-11e9-bf44-fa6f350b29f0,ResourceVersion:110169,Generation:1,CreationTimestamp:2019-06-18 13:04:39 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod,},Annotations:map[string]string{deployment.kubernetes.io/revision: 3546343826724305833,},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Spec:DeploymentSpec{Replicas:*1,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[],Containers:[{redis gcr.io/kubernetes-e2e-test-images/redis:1.0 [] [] [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,},},Strategy:DeploymentStrategy{Type:RollingUpdate,RollingUpdate:&RollingUpdateDeployment{MaxUnavailable:25%!,(MISSING)MaxSurge:25%!,(MISSING)},},MinReadySeconds:0,RevisionHistoryLimit:*10,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:1,Replicas:1,UpdatedReplicas:1,AvailableReplicas:1,UnavailableReplicas:0,Conditions:[{Available True 2019-06-18 13:04:39 +0000 UTC 2019-06-18 13:04:39 +0000 UTC MinimumReplicasAvailable Deployment has minimum availability.} {Progressing True 2019-06-18 13:04:40 +0000 UTC 2019-06-18 13:04:39 +0000 UTC NewReplicaSetAvailable ReplicaSet "test-rolling-update-deployment-68b55d7bc6" has successfully progressed.}],ReadyReplicas:1,CollisionCount:nil,},} + +Jun 18 13:04:41.407: INFO: New ReplicaSet "test-rolling-update-deployment-68b55d7bc6" of Deployment "test-rolling-update-deployment": +&ReplicaSet{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-rolling-update-deployment-68b55d7bc6,GenerateName:,Namespace:e2e-tests-deployment-68kjh,SelfLink:/apis/apps/v1/namespaces/e2e-tests-deployment-68kjh/replicasets/test-rolling-update-deployment-68b55d7bc6,UID:a15d2d5d-91c9-11e9-bf44-fa6f350b29f0,ResourceVersion:110159,Generation:1,CreationTimestamp:2019-06-18 13:04:39 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod,pod-template-hash: 68b55d7bc6,},Annotations:map[string]string{deployment.kubernetes.io/desired-replicas: 1,deployment.kubernetes.io/max-replicas: 2,deployment.kubernetes.io/revision: 3546343826724305833,},OwnerReferences:[{apps/v1 Deployment test-rolling-update-deployment a156030f-91c9-11e9-bf44-fa6f350b29f0 0xc001734c37 0xc001734c38}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:ReplicaSetSpec{Replicas:*1,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod,pod-template-hash: 68b55d7bc6,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod,pod-template-hash: 68b55d7bc6,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[],Containers:[{redis gcr.io/kubernetes-e2e-test-images/redis:1.0 [] [] [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,},},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:1,FullyLabeledReplicas:1,ObservedGeneration:1,ReadyReplicas:1,AvailableReplicas:1,Conditions:[],},} +Jun 18 13:04:41.407: INFO: All old ReplicaSets of Deployment "test-rolling-update-deployment": +Jun 18 13:04:41.407: INFO: &ReplicaSet{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-rolling-update-controller,GenerateName:,Namespace:e2e-tests-deployment-68kjh,SelfLink:/apis/apps/v1/namespaces/e2e-tests-deployment-68kjh/replicasets/test-rolling-update-controller,UID:9e544ba8-91c9-11e9-bf44-fa6f350b29f0,ResourceVersion:110168,Generation:2,CreationTimestamp:2019-06-18 13:04:34 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod,pod: nginx,},Annotations:map[string]string{deployment.kubernetes.io/desired-replicas: 1,deployment.kubernetes.io/max-replicas: 2,deployment.kubernetes.io/revision: 3546343826724305832,},OwnerReferences:[{apps/v1 Deployment test-rolling-update-deployment a156030f-91c9-11e9-bf44-fa6f350b29f0 0xc001734b77 0xc001734b78}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:ReplicaSetSpec{Replicas:*0,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod,pod: nginx,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod,pod: nginx,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,},},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:0,FullyLabeledReplicas:0,ObservedGeneration:2,ReadyReplicas:0,AvailableReplicas:0,Conditions:[],},} +Jun 18 13:04:41.422: INFO: Pod "test-rolling-update-deployment-68b55d7bc6-z4jxp" is available: +&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-rolling-update-deployment-68b55d7bc6-z4jxp,GenerateName:test-rolling-update-deployment-68b55d7bc6-,Namespace:e2e-tests-deployment-68kjh,SelfLink:/api/v1/namespaces/e2e-tests-deployment-68kjh/pods/test-rolling-update-deployment-68b55d7bc6-z4jxp,UID:a15f0a1e-91c9-11e9-bf44-fa6f350b29f0,ResourceVersion:110158,Generation:0,CreationTimestamp:2019-06-18 13:04:39 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod,pod-template-hash: 68b55d7bc6,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet test-rolling-update-deployment-68b55d7bc6 a15d2d5d-91c9-11e9-bf44-fa6f350b29f0 0xc001735c47 0xc001735c48}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-5mnfv {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-5mnfv,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{redis gcr.io/kubernetes-e2e-test-images/redis:1.0 [] [] [] [] [] {map[] map[]} [{default-token-5mnfv true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.72.74.144,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc001735db0} {node.kubernetes.io/unreachable Exists NoExecute 0xc001735dd0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Running,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 13:04:39 +0000 UTC } {Ready True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 13:04:40 +0000 UTC } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 13:04:40 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 13:04:39 +0000 UTC }],Message:,Reason:,HostIP:10.72.74.144,PodIP:172.30.114.36,StartTime:2019-06-18 13:04:39 +0000 UTC,ContainerStatuses:[{redis {nil ContainerStateRunning{StartedAt:2019-06-18 13:04:40 +0000 UTC,} nil} {nil nil nil} true 0 gcr.io/kubernetes-e2e-test-images/redis:1.0 gcr.io/kubernetes-e2e-test-images/redis@sha256:af4748d1655c08dc54d4be5182135395db9ce87aba2d4699b26b14ae197c5830 containerd://c227d3ae976f62825517120a33ec351cee4b39631c1fa2504294f9aeb188920d}],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} +[AfterEach] [sig-apps] Deployment /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 20:04:49.211: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-secrets-tzfhp" for this suite. -May 13 20:04:55.260: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 20:04:55.407: INFO: namespace: e2e-tests-secrets-tzfhp, resource: bindings, ignored listing per whitelist -May 13 20:04:55.541: INFO: namespace e2e-tests-secrets-tzfhp deletion completed in 6.322287677s +Jun 18 13:04:41.422: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "e2e-tests-deployment-68kjh" for this suite. +Jun 18 13:04:49.491: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 13:04:49.927: INFO: namespace: e2e-tests-deployment-68kjh, resource: bindings, ignored listing per whitelist +Jun 18 13:04:49.979: INFO: namespace e2e-tests-deployment-68kjh deletion completed in 8.538474209s -• [SLOW TEST:8.686 seconds] -[sig-storage] Secrets -/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/secrets_volume.go:34 - should be consumable from pods in volume with defaultMode set [NodeConformance] [Conformance] +• [SLOW TEST:16.214 seconds] +[sig-apps] Deployment +/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22 + RollingUpdateDeployment should delete old pods and create new ones [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ -SSSSSSSSSSSSS +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[k8s.io] Kubelet when scheduling a read only busybox container - should not write to root filesystem [NodeConformance] [Conformance] +[sig-network] Services + should serve multiport endpoints from pods [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -[BeforeEach] [k8s.io] Kubelet +[BeforeEach] [sig-network] Services /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client -May 13 20:04:55.543: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -STEP: Building a namespace api object, basename kubelet-test -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-kubelet-test-qm97t +Jun 18 13:04:49.979: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 +STEP: Building a namespace api object, basename services +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-services-pqfxq STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [k8s.io] Kubelet - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:37 -[It] should not write to root filesystem [NodeConformance] [Conformance] +[BeforeEach] [sig-network] Services + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:85 +[It] should serve multiport endpoints from pods [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -[AfterEach] [k8s.io] Kubelet +STEP: creating service multi-endpoint-test in namespace e2e-tests-services-pqfxq +STEP: waiting up to 3m0s for service multi-endpoint-test in namespace e2e-tests-services-pqfxq to expose endpoints map[] +Jun 18 13:04:50.525: INFO: Get endpoints failed (12.078189ms elapsed, ignoring for 5s): endpoints "multi-endpoint-test" not found +Jun 18 13:04:51.539: INFO: successfully validated that service multi-endpoint-test in namespace e2e-tests-services-pqfxq exposes endpoints map[] (1.025369365s elapsed) +STEP: Creating pod pod1 in namespace e2e-tests-services-pqfxq +STEP: waiting up to 3m0s for service multi-endpoint-test in namespace e2e-tests-services-pqfxq to expose endpoints map[pod1:[100]] +Jun 18 13:04:53.663: INFO: successfully validated that service multi-endpoint-test in namespace e2e-tests-services-pqfxq exposes endpoints map[pod1:[100]] (2.098026929s elapsed) +STEP: Creating pod pod2 in namespace e2e-tests-services-pqfxq +STEP: waiting up to 3m0s for service multi-endpoint-test in namespace e2e-tests-services-pqfxq to expose endpoints map[pod1:[100] pod2:[101]] +Jun 18 13:04:56.927: INFO: successfully validated that service multi-endpoint-test in namespace e2e-tests-services-pqfxq exposes endpoints map[pod1:[100] pod2:[101]] (3.247748534s elapsed) +STEP: Deleting pod pod1 in namespace e2e-tests-services-pqfxq +STEP: waiting up to 3m0s for service multi-endpoint-test in namespace e2e-tests-services-pqfxq to expose endpoints map[pod2:[101]] +Jun 18 13:04:56.976: INFO: successfully validated that service multi-endpoint-test in namespace e2e-tests-services-pqfxq exposes endpoints map[pod2:[101]] (27.081541ms elapsed) +STEP: Deleting pod pod2 in namespace e2e-tests-services-pqfxq +STEP: waiting up to 3m0s for service multi-endpoint-test in namespace e2e-tests-services-pqfxq to expose endpoints map[] +Jun 18 13:04:57.012: INFO: successfully validated that service multi-endpoint-test in namespace e2e-tests-services-pqfxq exposes endpoints map[] (11.872768ms elapsed) +[AfterEach] [sig-network] Services /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 20:04:57.935: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-kubelet-test-qm97t" for this suite. -May 13 20:05:51.975: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 20:05:52.027: INFO: namespace: e2e-tests-kubelet-test-qm97t, resource: bindings, ignored listing per whitelist -May 13 20:05:52.243: INFO: namespace e2e-tests-kubelet-test-qm97t deletion completed in 54.29948035s +Jun 18 13:04:57.079: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "e2e-tests-services-pqfxq" for this suite. +Jun 18 13:05:21.149: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 13:05:21.253: INFO: namespace: e2e-tests-services-pqfxq, resource: bindings, ignored listing per whitelist +Jun 18 13:05:21.642: INFO: namespace e2e-tests-services-pqfxq deletion completed in 24.543851995s +[AfterEach] [sig-network] Services + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:90 -• [SLOW TEST:56.700 seconds] -[k8s.io] Kubelet -/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 - when scheduling a read only busybox container - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:186 - should not write to root filesystem [NodeConformance] [Conformance] - /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 +• [SLOW TEST:31.662 seconds] +[sig-network] Services +/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:22 + should serve multiport endpoints from pods [Conformance] + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSSSS +SSS ------------------------------ -[sig-storage] Projected configMap - updates should be reflected in volume [NodeConformance] [Conformance] +[k8s.io] Probing container + should *not* be restarted with a /healthz http liveness probe [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -[BeforeEach] [sig-storage] Projected configMap +[BeforeEach] [k8s.io] Probing container /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client -May 13 20:05:52.244: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 -STEP: Building a namespace api object, basename projected -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-projected-zv4cp +Jun 18 13:05:21.642: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 +STEP: Building a namespace api object, basename container-probe +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-container-probe-bgkpg STEP: Waiting for a default service account to be provisioned in namespace -[It] updates should be reflected in volume [NodeConformance] [Conformance] +[BeforeEach] [k8s.io] Probing container + /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/container_probe.go:48 +[It] should *not* be restarted with a /healthz http liveness probe [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -STEP: Creating projection with configMap that has name projected-configmap-test-upd-828db74e-75ba-11e9-8f67-2632f168be36 -STEP: Creating the pod -STEP: Updating configmap projected-configmap-test-upd-828db74e-75ba-11e9-8f67-2632f168be36 -STEP: waiting to observe update in volume -[AfterEach] [sig-storage] Projected configMap +STEP: Creating pod liveness-http in namespace e2e-tests-container-probe-bgkpg +Jun 18 13:05:24.172: INFO: Started pod liveness-http in namespace e2e-tests-container-probe-bgkpg +STEP: checking the pod's current state and verifying that restartCount is present +Jun 18 13:05:24.188: INFO: Initial restart count of pod liveness-http is 0 +STEP: deleting the pod +[AfterEach] [k8s.io] Probing container /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 -May 13 20:05:56.777: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-tests-projected-zv4cp" for this suite. -May 13 20:06:18.849: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -May 13 20:06:18.927: INFO: namespace: e2e-tests-projected-zv4cp, resource: bindings, ignored listing per whitelist -May 13 20:06:19.101: INFO: namespace e2e-tests-projected-zv4cp deletion completed in 22.28158403s +Jun 18 13:09:25.752: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "e2e-tests-container-probe-bgkpg" for this suite. +Jun 18 13:09:31.879: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 13:09:32.066: INFO: namespace: e2e-tests-container-probe-bgkpg, resource: bindings, ignored listing per whitelist +Jun 18 13:09:32.448: INFO: namespace e2e-tests-container-probe-bgkpg deletion completed in 6.645606921s -• [SLOW TEST:26.858 seconds] -[sig-storage] Projected configMap -/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_configmap.go:34 - updates should be reflected in volume [NodeConformance] [Conformance] +• [SLOW TEST:250.807 seconds] +[k8s.io] Probing container +/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 + should *not* be restarted with a /healthz http liveness probe [NodeConformance] [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ -SSSSSSSSSSS +SSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-network] Proxy version v1 - should proxy logs on node with explicit kubelet port using proxy subresource [Conformance] + should proxy logs on node using proxy subresource [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] version v1 /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client -May 13 20:06:19.103: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423 +Jun 18 13:09:32.449: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206 STEP: Building a namespace api object, basename proxy -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-proxy-5c9hh +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-proxy-tq6sf STEP: Waiting for a default service account to be provisioned in namespace -[It] should proxy logs on node with explicit kubelet port using proxy subresource [Conformance] +[It] should proxy logs on node using proxy subresource [Conformance] /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 -May 13 20:06:19.446: INFO: (0) /api/v1/nodes/10.170.219.151:10250/proxy/logs/:
-alb/
+Jun 18 13:09:32.967: INFO: (0) /api/v1/nodes/10.72.74.143/proxy/logs/: 
 alternatives.log
-apt/... (200; 17.264338ms)
-May 13 20:06:19.455: INFO: (1) /api/v1/nodes/10.170.219.151:10250/proxy/logs/: 
-alb/
+apt/
+at/
 alternatives.log
-apt/... (200; 9.298611ms)
-May 13 20:06:19.465: INFO: (2) /api/v1/nodes/10.170.219.151:10250/proxy/logs/: 
-alb/
+apt/
+at/
 alternatives.log
-apt/... (200; 9.136183ms)
-May 13 20:06:19.473: INFO: (3) /api/v1/nodes/10.170.219.151:10250/proxy/logs/: 
-alb/
+apt/
+at/
 alternatives.log
-apt/... (200; 8.744928ms)
-May 13 20:06:19.482: INFO: (4) /api/v1/nodes/10.170.219.151:10250/proxy/logs/: 
-alb/
+apt/
+at/
 alternatives.log
-apt/... (200; 8.70234ms)
-May 13 20:06:19.491: INFO: (5) /api/v1/nodes/10.170.219.151:10250/proxy/logs/: 
-alb/
+apt/
+at/
 alternatives.log
-apt/... (200; 9.101358ms)
-May 13 20:06:19.502: INFO: (6) /api/v1/nodes/10.170.219.151:10250/proxy/logs/: 
-alb/
+apt/
+at/
 alternatives.log
-apt/... (200; 10.537645ms)
-May 13 20:06:19.510: INFO: (7) /api/v1/nodes/10.170.219.151:10250/proxy/logs/: 
-alb/
+apt/
+at/
 alternatives.log
-apt/... (200; 8.627342ms)
-May 13 20:06:19.519: INFO: (8) /api/v1/nodes/10.170.219.151:10250/proxy/logs/: 
-alb/
+apt/
+at/
 alternatives.log
-apt/... (200; 9.023938ms)
-May 13 20:06:19.528: INFO: (9) /api/v1/nodes/10.170.219.151:10250/proxy/logs/: 
-alb/
+apt/
+at/
 alternatives.log
-apt/... (200; 8.520474ms)
-May 13 20:06:19.538: INFO: (10) /api/v1/nodes/10.170.219.151:10250/proxy/logs/: 
-alb/
+apt/
+at/
 alternatives.log
-apt/... (200; 9.869383ms)
-May 13 20:06:19.547: INFO: (11) /api/v1/nodes/10.170.219.151:10250/proxy/logs/: 
-alb/
+apt/
+at/
 alternatives.log
-apt/... (200; 9.108429ms)
-May 13 20:06:19.556: INFO: (12) /api/v1/nodes/10.170.219.151:10250/proxy/logs/: 
-alb/
+apt/
+at/
 alternatives.log
-apt/... (200; 8.875025ms)
-May 13 20:06:19.565: INFO: (13) /api/v1/nodes/10.170.219.151:10250/proxy/logs/: 
-alb/
+apt/
+at/
 alternatives.log
-apt/... (200; 8.734339ms)
-May 13 20:06:19.574: INFO: (14) /api/v1/nodes/10.170.219.151:10250/proxy/logs/: 
-alb/
+apt/
+at/
 alternatives.log
-apt/... (200; 9.035131ms)
-May 13 20:06:19.582: INFO: (15) /api/v1/nodes/10.170.219.151:10250/proxy/logs/: 
-alb/
+apt/
+at/
 alternatives.log
-apt/... (200; 8.243345ms)
-May 13 20:06:19.590: INFO: (16) /api/v1/nodes/10.170.219.151:10250/proxy/logs/: 
-alb/
+apt/
+at/
 alternatives.log
-apt/... (200; 8.036773ms)
-May 13 20:06:19.601: INFO: (17) /api/v1/nodes/10.170.219.151:10250/proxy/logs/: 
-alb/
+apt/
+at/
 alternatives.log
-apt/... (200; 10.532091ms)
-May 13 20:06:19.610: INFO: (18) /api/v1/nodes/10.170.219.151:10250/proxy/logs/: 
-alb/
+apt/
+at/
 alternatives.log
-apt/... (200; 9.463155ms)
-May 13 20:06:19.620: INFO: (19) /api/v1/nodes/10.170.219.151:10250/proxy/logs/: 
-alb/
+apt/
+at/
 alternatives.log
-apt/... (200; 9.380762ms)
+apt/
+at/>> kubeConfig: /tmp/kubeconfig-555138423
-STEP: Building a namespace api object, basename configmap
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-configmap-xgwn2
+Jun 18 13:09:40.073: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
+STEP: Building a namespace api object, basename projected
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-projected-99vqr
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] updates should be reflected in volume [NodeConformance] [Conformance]
+[BeforeEach] [sig-storage] Projected downwardAPI
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:39
+[It] should update labels on modification [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-STEP: Creating configMap with name configmap-test-upd-9695cb7d-75ba-11e9-8f67-2632f168be36
 STEP: Creating the pod
-STEP: Updating configmap configmap-test-upd-9695cb7d-75ba-11e9-8f67-2632f168be36
-STEP: waiting to observe update in volume
-[AfterEach] [sig-storage] ConfigMap
+Jun 18 13:09:44.460: INFO: Successfully updated pod "labelsupdate5502ad75-91ca-11e9-bce2-ae54e022189f"
+[AfterEach] [sig-storage] Projected downwardAPI
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
-May 13 20:06:30.326: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "e2e-tests-configmap-xgwn2" for this suite.
-May 13 20:06:54.454: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-May 13 20:06:54.658: INFO: namespace: e2e-tests-configmap-xgwn2, resource: bindings, ignored listing per whitelist
-May 13 20:06:54.700: INFO: namespace e2e-tests-configmap-xgwn2 deletion completed in 24.280403818s
+Jun 18 13:09:48.565: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-tests-projected-99vqr" for this suite.
+Jun 18 13:10:12.638: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 13:10:13.070: INFO: namespace: e2e-tests-projected-99vqr, resource: bindings, ignored listing per whitelist
+Jun 18 13:10:13.244: INFO: namespace e2e-tests-projected-99vqr deletion completed in 24.660263931s
 
-• [SLOW TEST:28.739 seconds]
-[sig-storage] ConfigMap
-/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:33
-  updates should be reflected in volume [NodeConformance] [Conformance]
+• [SLOW TEST:33.171 seconds]
+[sig-storage] Projected downwardAPI
+/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:33
+  should update labels on modification [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 ------------------------------
-SSSS
+SSSSSSSSSSS
 ------------------------------
-[k8s.io] Kubelet when scheduling a busybox command that always fails in a pod 
-  should be possible to delete [NodeConformance] [Conformance]
+[sig-apps] ReplicaSet 
+  should serve a basic image on each replica with a public image  [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-[BeforeEach] [k8s.io] Kubelet
+[BeforeEach] [sig-apps] ReplicaSet
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
 STEP: Creating a kubernetes client
-May 13 20:06:54.702: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423
-STEP: Building a namespace api object, basename kubelet-test
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-kubelet-test-88b2w
+Jun 18 13:10:13.245: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
+STEP: Building a namespace api object, basename replicaset
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-replicaset-2lnks
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [k8s.io] Kubelet
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:37
-[BeforeEach] when scheduling a busybox command that always fails in a pod
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:81
-[It] should be possible to delete [NodeConformance] [Conformance]
+[It] should serve a basic image on each replica with a public image  [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-[AfterEach] [k8s.io] Kubelet
+Jun 18 13:10:13.687: INFO: Creating ReplicaSet my-hostname-basic-68a98c8a-91ca-11e9-bce2-ae54e022189f
+Jun 18 13:10:13.716: INFO: Pod name my-hostname-basic-68a98c8a-91ca-11e9-bce2-ae54e022189f: Found 0 pods out of 1
+Jun 18 13:10:18.751: INFO: Pod name my-hostname-basic-68a98c8a-91ca-11e9-bce2-ae54e022189f: Found 1 pods out of 1
+Jun 18 13:10:18.751: INFO: Ensuring a pod for ReplicaSet "my-hostname-basic-68a98c8a-91ca-11e9-bce2-ae54e022189f" is running
+Jun 18 13:10:18.766: INFO: Pod "my-hostname-basic-68a98c8a-91ca-11e9-bce2-ae54e022189f-w45gp" is running (conditions: [{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2019-06-18 13:10:13 +0000 UTC Reason: Message:} {Type:Ready Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2019-06-18 13:10:16 +0000 UTC Reason: Message:} {Type:ContainersReady Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2019-06-18 13:10:16 +0000 UTC Reason: Message:} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2019-06-18 13:10:13 +0000 UTC Reason: Message:}])
+Jun 18 13:10:18.766: INFO: Trying to dial the pod
+Jun 18 13:10:23.832: INFO: Controller my-hostname-basic-68a98c8a-91ca-11e9-bce2-ae54e022189f: Got expected result from replica 1 [my-hostname-basic-68a98c8a-91ca-11e9-bce2-ae54e022189f-w45gp]: "my-hostname-basic-68a98c8a-91ca-11e9-bce2-ae54e022189f-w45gp", 1 of 1 required successes so far
+[AfterEach] [sig-apps] ReplicaSet
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
-May 13 20:06:55.058: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "e2e-tests-kubelet-test-88b2w" for this suite.
-May 13 20:07:17.095: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-May 13 20:07:17.162: INFO: namespace: e2e-tests-kubelet-test-88b2w, resource: bindings, ignored listing per whitelist
-May 13 20:07:17.331: INFO: namespace e2e-tests-kubelet-test-88b2w deletion completed in 22.265671568s
+Jun 18 13:10:23.832: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-tests-replicaset-2lnks" for this suite.
+Jun 18 13:10:29.920: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 13:10:30.406: INFO: namespace: e2e-tests-replicaset-2lnks, resource: bindings, ignored listing per whitelist
+Jun 18 13:10:30.476: INFO: namespace e2e-tests-replicaset-2lnks deletion completed in 6.62366172s
 
-• [SLOW TEST:22.629 seconds]
-[k8s.io] Kubelet
-/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694
-  when scheduling a busybox command that always fails in a pod
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:78
-    should be possible to delete [NodeConformance] [Conformance]
-    /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
+• [SLOW TEST:17.231 seconds]
+[sig-apps] ReplicaSet
+/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22
+  should serve a basic image on each replica with a public image  [Conformance]
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 ------------------------------
-[sig-apps] Daemon set [Serial] 
-  should run and stop complex daemon [Conformance]
+S
+------------------------------
+[k8s.io] Docker Containers 
+  should be able to override the image's default command (docker entrypoint) [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-[BeforeEach] [sig-apps] Daemon set [Serial]
+[BeforeEach] [k8s.io] Docker Containers
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
 STEP: Creating a kubernetes client
-May 13 20:07:17.332: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423
-STEP: Building a namespace api object, basename daemonsets
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-daemonsets-d5pqr
+Jun 18 13:10:30.477: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
+STEP: Building a namespace api object, basename containers
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-containers-7cg66
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-apps] Daemon set [Serial]
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:102
-[It] should run and stop complex daemon [Conformance]
+[It] should be able to override the image's default command (docker entrypoint) [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-May 13 20:07:17.747: INFO: Creating daemon "daemon-set" with a node selector
-STEP: Initially, daemon pods should not be running on any nodes.
-May 13 20:07:17.763: INFO: Number of nodes with available pods: 0
-May 13 20:07:17.763: INFO: Number of running nodes: 0, number of available pods: 0
-STEP: Change node label to blue, check that daemon pod is launched.
-May 13 20:07:17.790: INFO: Number of nodes with available pods: 0
-May 13 20:07:17.790: INFO: Node 10.170.219.151 is running more than one daemon pod
-May 13 20:07:18.798: INFO: Number of nodes with available pods: 0
-May 13 20:07:18.798: INFO: Node 10.170.219.151 is running more than one daemon pod
-May 13 20:07:19.798: INFO: Number of nodes with available pods: 1
-May 13 20:07:19.798: INFO: Number of running nodes: 1, number of available pods: 1
-STEP: Update the node label to green, and wait for daemons to be unscheduled
-May 13 20:07:19.833: INFO: Number of nodes with available pods: 1
-May 13 20:07:19.833: INFO: Number of running nodes: 0, number of available pods: 1
-May 13 20:07:20.842: INFO: Number of nodes with available pods: 0
-May 13 20:07:20.842: INFO: Number of running nodes: 0, number of available pods: 0
-STEP: Update DaemonSet node selector to green, and change its update strategy to RollingUpdate
-May 13 20:07:20.860: INFO: Number of nodes with available pods: 0
-May 13 20:07:20.860: INFO: Node 10.170.219.151 is running more than one daemon pod
-May 13 20:07:21.882: INFO: Number of nodes with available pods: 0
-May 13 20:07:21.882: INFO: Node 10.170.219.151 is running more than one daemon pod
-May 13 20:07:22.920: INFO: Number of nodes with available pods: 0
-May 13 20:07:22.920: INFO: Node 10.170.219.151 is running more than one daemon pod
-May 13 20:07:23.883: INFO: Number of nodes with available pods: 0
-May 13 20:07:23.883: INFO: Node 10.170.219.151 is running more than one daemon pod
-May 13 20:07:24.868: INFO: Number of nodes with available pods: 0
-May 13 20:07:24.868: INFO: Node 10.170.219.151 is running more than one daemon pod
-May 13 20:07:25.868: INFO: Number of nodes with available pods: 0
-May 13 20:07:25.868: INFO: Node 10.170.219.151 is running more than one daemon pod
-May 13 20:07:26.868: INFO: Number of nodes with available pods: 0
-May 13 20:07:26.868: INFO: Node 10.170.219.151 is running more than one daemon pod
-May 13 20:07:27.920: INFO: Number of nodes with available pods: 0
-May 13 20:07:27.920: INFO: Node 10.170.219.151 is running more than one daemon pod
-May 13 20:07:28.879: INFO: Number of nodes with available pods: 0
-May 13 20:07:28.879: INFO: Node 10.170.219.151 is running more than one daemon pod
-May 13 20:07:29.870: INFO: Number of nodes with available pods: 0
-May 13 20:07:29.871: INFO: Node 10.170.219.151 is running more than one daemon pod
-May 13 20:07:30.888: INFO: Number of nodes with available pods: 0
-May 13 20:07:30.888: INFO: Node 10.170.219.151 is running more than one daemon pod
-May 13 20:07:31.874: INFO: Number of nodes with available pods: 0
-May 13 20:07:31.874: INFO: Node 10.170.219.151 is running more than one daemon pod
-May 13 20:07:32.893: INFO: Number of nodes with available pods: 0
-May 13 20:07:32.893: INFO: Node 10.170.219.151 is running more than one daemon pod
-May 13 20:07:33.868: INFO: Number of nodes with available pods: 0
-May 13 20:07:33.868: INFO: Node 10.170.219.151 is running more than one daemon pod
-May 13 20:07:34.930: INFO: Number of nodes with available pods: 0
-May 13 20:07:34.930: INFO: Node 10.170.219.151 is running more than one daemon pod
-May 13 20:07:35.873: INFO: Number of nodes with available pods: 0
-May 13 20:07:35.873: INFO: Node 10.170.219.151 is running more than one daemon pod
-May 13 20:07:36.880: INFO: Number of nodes with available pods: 0
-May 13 20:07:36.880: INFO: Node 10.170.219.151 is running more than one daemon pod
-May 13 20:07:37.868: INFO: Number of nodes with available pods: 0
-May 13 20:07:37.868: INFO: Node 10.170.219.151 is running more than one daemon pod
-May 13 20:07:38.870: INFO: Number of nodes with available pods: 0
-May 13 20:07:38.870: INFO: Node 10.170.219.151 is running more than one daemon pod
-May 13 20:07:39.868: INFO: Number of nodes with available pods: 0
-May 13 20:07:39.868: INFO: Node 10.170.219.151 is running more than one daemon pod
-May 13 20:07:40.868: INFO: Number of nodes with available pods: 0
-May 13 20:07:40.869: INFO: Node 10.170.219.151 is running more than one daemon pod
-May 13 20:07:41.869: INFO: Number of nodes with available pods: 0
-May 13 20:07:41.869: INFO: Node 10.170.219.151 is running more than one daemon pod
-May 13 20:07:42.868: INFO: Number of nodes with available pods: 0
-May 13 20:07:42.868: INFO: Node 10.170.219.151 is running more than one daemon pod
-May 13 20:07:43.893: INFO: Number of nodes with available pods: 0
-May 13 20:07:43.893: INFO: Node 10.170.219.151 is running more than one daemon pod
-May 13 20:07:44.868: INFO: Number of nodes with available pods: 0
-May 13 20:07:44.868: INFO: Node 10.170.219.151 is running more than one daemon pod
-May 13 20:07:45.868: INFO: Number of nodes with available pods: 0
-May 13 20:07:45.868: INFO: Node 10.170.219.151 is running more than one daemon pod
-May 13 20:07:46.868: INFO: Number of nodes with available pods: 0
-May 13 20:07:46.868: INFO: Node 10.170.219.151 is running more than one daemon pod
-May 13 20:07:47.920: INFO: Number of nodes with available pods: 0
-May 13 20:07:47.920: INFO: Node 10.170.219.151 is running more than one daemon pod
-May 13 20:07:49.681: INFO: Number of nodes with available pods: 0
-May 13 20:07:49.681: INFO: Node 10.170.219.151 is running more than one daemon pod
-May 13 20:07:49.868: INFO: Number of nodes with available pods: 0
-May 13 20:07:49.868: INFO: Node 10.170.219.151 is running more than one daemon pod
-May 13 20:07:50.869: INFO: Number of nodes with available pods: 0
-May 13 20:07:50.869: INFO: Node 10.170.219.151 is running more than one daemon pod
-May 13 20:07:51.872: INFO: Number of nodes with available pods: 0
-May 13 20:07:51.872: INFO: Node 10.170.219.151 is running more than one daemon pod
-May 13 20:07:52.869: INFO: Number of nodes with available pods: 0
-May 13 20:07:52.869: INFO: Node 10.170.219.151 is running more than one daemon pod
-May 13 20:07:53.868: INFO: Number of nodes with available pods: 0
-May 13 20:07:53.868: INFO: Node 10.170.219.151 is running more than one daemon pod
-May 13 20:07:54.881: INFO: Number of nodes with available pods: 0
-May 13 20:07:54.881: INFO: Node 10.170.219.151 is running more than one daemon pod
-May 13 20:07:55.874: INFO: Number of nodes with available pods: 1
-May 13 20:07:55.875: INFO: Number of running nodes: 1, number of available pods: 1
-[AfterEach] [sig-apps] Daemon set [Serial]
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:68
-STEP: Deleting DaemonSet "daemon-set"
-STEP: deleting DaemonSet.extensions daemon-set in namespace e2e-tests-daemonsets-d5pqr, will wait for the garbage collector to delete the pods
-May 13 20:07:55.958: INFO: Deleting DaemonSet.extensions daemon-set took: 10.529376ms
-May 13 20:07:56.058: INFO: Terminating DaemonSet.extensions daemon-set pods took: 100.20647ms
-May 13 20:08:29.278: INFO: Number of nodes with available pods: 0
-May 13 20:08:29.279: INFO: Number of running nodes: 0, number of available pods: 0
-May 13 20:08:29.286: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"selfLink":"/apis/apps/v1/namespaces/e2e-tests-daemonsets-d5pqr/daemonsets","resourceVersion":"53288"},"items":null}
-
-May 13 20:08:29.293: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"selfLink":"/api/v1/namespaces/e2e-tests-daemonsets-d5pqr/pods","resourceVersion":"53288"},"items":null}
-
-[AfterEach] [sig-apps] Daemon set [Serial]
+STEP: Creating a pod to test override command
+Jun 18 13:10:30.988: INFO: Waiting up to 5m0s for pod "client-containers-72f16aff-91ca-11e9-bce2-ae54e022189f" in namespace "e2e-tests-containers-7cg66" to be "success or failure"
+Jun 18 13:10:31.004: INFO: Pod "client-containers-72f16aff-91ca-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 16.010377ms
+Jun 18 13:10:33.018: INFO: Pod "client-containers-72f16aff-91ca-11e9-bce2-ae54e022189f": Phase="Running", Reason="", readiness=true. Elapsed: 2.030713314s
+Jun 18 13:10:35.033: INFO: Pod "client-containers-72f16aff-91ca-11e9-bce2-ae54e022189f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.045601134s
+STEP: Saw pod success
+Jun 18 13:10:35.033: INFO: Pod "client-containers-72f16aff-91ca-11e9-bce2-ae54e022189f" satisfied condition "success or failure"
+Jun 18 13:10:35.052: INFO: Trying to get logs from node 10.72.74.144 pod client-containers-72f16aff-91ca-11e9-bce2-ae54e022189f container test-container: 
+STEP: delete the pod
+Jun 18 13:10:35.143: INFO: Waiting for pod client-containers-72f16aff-91ca-11e9-bce2-ae54e022189f to disappear
+Jun 18 13:10:35.161: INFO: Pod client-containers-72f16aff-91ca-11e9-bce2-ae54e022189f no longer exists
+[AfterEach] [k8s.io] Docker Containers
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
-May 13 20:08:29.325: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "e2e-tests-daemonsets-d5pqr" for this suite.
-May 13 20:08:35.361: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-May 13 20:08:35.616: INFO: namespace: e2e-tests-daemonsets-d5pqr, resource: bindings, ignored listing per whitelist
-May 13 20:08:35.669: INFO: namespace e2e-tests-daemonsets-d5pqr deletion completed in 6.33831091s
+Jun 18 13:10:35.162: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-tests-containers-7cg66" for this suite.
+Jun 18 13:10:41.248: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 13:10:41.379: INFO: namespace: e2e-tests-containers-7cg66, resource: bindings, ignored listing per whitelist
+Jun 18 13:10:41.747: INFO: namespace e2e-tests-containers-7cg66 deletion completed in 6.566269952s
 
-• [SLOW TEST:78.337 seconds]
-[sig-apps] Daemon set [Serial]
-/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22
-  should run and stop complex daemon [Conformance]
+• [SLOW TEST:11.271 seconds]
+[k8s.io] Docker Containers
+/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694
+  should be able to override the image's default command (docker entrypoint) [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 ------------------------------
-SSSSS
+SSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-storage] Downward API volume 
-  should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance]
+[sig-storage] Projected secret 
+  should be consumable from pods in volume [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-[BeforeEach] [sig-storage] Downward API volume
+[BeforeEach] [sig-storage] Projected secret
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
 STEP: Creating a kubernetes client
-May 13 20:08:35.670: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423
-STEP: Building a namespace api object, basename downward-api
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-downward-api-82kcd
+Jun 18 13:10:41.750: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
+STEP: Building a namespace api object, basename projected
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-projected-wsgvr
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should be consumable from pods in volume [NodeConformance] [Conformance]
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
+STEP: Creating projection with secret that has name projected-secret-test-79a9faa7-91ca-11e9-bce2-ae54e022189f
+STEP: Creating a pod to test consume secrets
+Jun 18 13:10:42.252: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-79ac196e-91ca-11e9-bce2-ae54e022189f" in namespace "e2e-tests-projected-wsgvr" to be "success or failure"
+Jun 18 13:10:42.268: INFO: Pod "pod-projected-secrets-79ac196e-91ca-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 15.365547ms
+Jun 18 13:10:44.291: INFO: Pod "pod-projected-secrets-79ac196e-91ca-11e9-bce2-ae54e022189f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.038934287s
+STEP: Saw pod success
+Jun 18 13:10:44.291: INFO: Pod "pod-projected-secrets-79ac196e-91ca-11e9-bce2-ae54e022189f" satisfied condition "success or failure"
+Jun 18 13:10:44.306: INFO: Trying to get logs from node 10.72.74.149 pod pod-projected-secrets-79ac196e-91ca-11e9-bce2-ae54e022189f container projected-secret-volume-test: 
+STEP: delete the pod
+Jun 18 13:10:44.382: INFO: Waiting for pod pod-projected-secrets-79ac196e-91ca-11e9-bce2-ae54e022189f to disappear
+Jun 18 13:10:44.397: INFO: Pod pod-projected-secrets-79ac196e-91ca-11e9-bce2-ae54e022189f no longer exists
+[AfterEach] [sig-storage] Projected secret
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
+Jun 18 13:10:44.397: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-tests-projected-wsgvr" for this suite.
+Jun 18 13:10:50.469: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 13:10:50.792: INFO: namespace: e2e-tests-projected-wsgvr, resource: bindings, ignored listing per whitelist
+Jun 18 13:10:50.956: INFO: namespace e2e-tests-projected-wsgvr deletion completed in 6.535961565s
+
+• [SLOW TEST:9.205 seconds]
+[sig-storage] Projected secret
+/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_secret.go:34
+  should be consumable from pods in volume [NodeConformance] [Conformance]
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
+------------------------------
+[sig-storage] EmptyDir volumes 
+  volume on tmpfs should have the correct mode [NodeConformance] [Conformance]
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
+[BeforeEach] [sig-storage] EmptyDir volumes
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
+STEP: Creating a kubernetes client
+Jun 18 13:10:50.956: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
+STEP: Building a namespace api object, basename emptydir
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-emptydir-vcb6m
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-storage] Downward API volume
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:39
-[It] should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance]
+[It] volume on tmpfs should have the correct mode [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-STEP: Creating a pod to test downward API volume plugin
-May 13 20:08:36.039: INFO: Waiting up to 5m0s for pod "downwardapi-volume-e3f309b1-75ba-11e9-8f67-2632f168be36" in namespace "e2e-tests-downward-api-82kcd" to be "success or failure"
-May 13 20:08:36.048: INFO: Pod "downwardapi-volume-e3f309b1-75ba-11e9-8f67-2632f168be36": Phase="Pending", Reason="", readiness=false. Elapsed: 8.173881ms
-May 13 20:08:38.058: INFO: Pod "downwardapi-volume-e3f309b1-75ba-11e9-8f67-2632f168be36": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.018206904s
+STEP: Creating a pod to test emptydir volume type on tmpfs
+Jun 18 13:10:51.483: INFO: Waiting up to 5m0s for pod "pod-7f2cccfa-91ca-11e9-bce2-ae54e022189f" in namespace "e2e-tests-emptydir-vcb6m" to be "success or failure"
+Jun 18 13:10:51.499: INFO: Pod "pod-7f2cccfa-91ca-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 16.133297ms
+Jun 18 13:10:53.514: INFO: Pod "pod-7f2cccfa-91ca-11e9-bce2-ae54e022189f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.031192454s
 STEP: Saw pod success
-May 13 20:08:38.058: INFO: Pod "downwardapi-volume-e3f309b1-75ba-11e9-8f67-2632f168be36" satisfied condition "success or failure"
-May 13 20:08:38.065: INFO: Trying to get logs from node 10.170.219.153 pod downwardapi-volume-e3f309b1-75ba-11e9-8f67-2632f168be36 container client-container: 
+Jun 18 13:10:53.514: INFO: Pod "pod-7f2cccfa-91ca-11e9-bce2-ae54e022189f" satisfied condition "success or failure"
+Jun 18 13:10:53.528: INFO: Trying to get logs from node 10.72.74.143 pod pod-7f2cccfa-91ca-11e9-bce2-ae54e022189f container test-container: 
 STEP: delete the pod
-May 13 20:08:38.107: INFO: Waiting for pod downwardapi-volume-e3f309b1-75ba-11e9-8f67-2632f168be36 to disappear
-May 13 20:08:38.126: INFO: Pod downwardapi-volume-e3f309b1-75ba-11e9-8f67-2632f168be36 no longer exists
-[AfterEach] [sig-storage] Downward API volume
+Jun 18 13:10:53.604: INFO: Waiting for pod pod-7f2cccfa-91ca-11e9-bce2-ae54e022189f to disappear
+Jun 18 13:10:53.619: INFO: Pod pod-7f2cccfa-91ca-11e9-bce2-ae54e022189f no longer exists
+[AfterEach] [sig-storage] EmptyDir volumes
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
-May 13 20:08:38.127: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "e2e-tests-downward-api-82kcd" for this suite.
-May 13 20:08:44.163: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-May 13 20:08:44.456: INFO: namespace: e2e-tests-downward-api-82kcd, resource: bindings, ignored listing per whitelist
-May 13 20:08:44.505: INFO: namespace e2e-tests-downward-api-82kcd deletion completed in 6.370818392s
+Jun 18 13:10:53.619: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-tests-emptydir-vcb6m" for this suite.
+Jun 18 13:11:01.708: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 13:11:02.002: INFO: namespace: e2e-tests-emptydir-vcb6m, resource: bindings, ignored listing per whitelist
+Jun 18 13:11:02.238: INFO: namespace e2e-tests-emptydir-vcb6m deletion completed in 8.597695433s
 
-• [SLOW TEST:8.836 seconds]
-[sig-storage] Downward API volume
-/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:34
-  should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance]
+• [SLOW TEST:11.282 seconds]
+[sig-storage] EmptyDir volumes
+/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:40
+  volume on tmpfs should have the correct mode [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 ------------------------------
-SSSSSSSSSSS
+SSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-node] ConfigMap 
-  should be consumable via the environment [NodeConformance] [Conformance]
+[sig-storage] Projected secret 
+  should be consumable from pods in volume with defaultMode set [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-[BeforeEach] [sig-node] ConfigMap
+[BeforeEach] [sig-storage] Projected secret
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
 STEP: Creating a kubernetes client
-May 13 20:08:44.507: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423
-STEP: Building a namespace api object, basename configmap
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-configmap-kvtfd
+Jun 18 13:11:02.242: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
+STEP: Building a namespace api object, basename projected
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-projected-64mbg
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should be consumable via the environment [NodeConformance] [Conformance]
+[It] should be consumable from pods in volume with defaultMode set [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-STEP: Creating configMap e2e-tests-configmap-kvtfd/configmap-test-e929fc5b-75ba-11e9-8f67-2632f168be36
-STEP: Creating a pod to test consume configMaps
-May 13 20:08:44.797: INFO: Waiting up to 5m0s for pod "pod-configmaps-e92b5c06-75ba-11e9-8f67-2632f168be36" in namespace "e2e-tests-configmap-kvtfd" to be "success or failure"
-May 13 20:08:44.809: INFO: Pod "pod-configmaps-e92b5c06-75ba-11e9-8f67-2632f168be36": Phase="Pending", Reason="", readiness=false. Elapsed: 11.585756ms
-May 13 20:08:46.816: INFO: Pod "pod-configmaps-e92b5c06-75ba-11e9-8f67-2632f168be36": Phase="Pending", Reason="", readiness=false. Elapsed: 2.019531545s
-May 13 20:08:48.826: INFO: Pod "pod-configmaps-e92b5c06-75ba-11e9-8f67-2632f168be36": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.029234296s
+STEP: Creating projection with secret that has name projected-secret-test-85e11667-91ca-11e9-bce2-ae54e022189f
+STEP: Creating a pod to test consume secrets
+Jun 18 13:11:02.747: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-85e33269-91ca-11e9-bce2-ae54e022189f" in namespace "e2e-tests-projected-64mbg" to be "success or failure"
+Jun 18 13:11:02.763: INFO: Pod "pod-projected-secrets-85e33269-91ca-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 15.224363ms
+Jun 18 13:11:04.778: INFO: Pod "pod-projected-secrets-85e33269-91ca-11e9-bce2-ae54e022189f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.030963038s
 STEP: Saw pod success
-May 13 20:08:48.826: INFO: Pod "pod-configmaps-e92b5c06-75ba-11e9-8f67-2632f168be36" satisfied condition "success or failure"
-May 13 20:08:48.834: INFO: Trying to get logs from node 10.170.219.153 pod pod-configmaps-e92b5c06-75ba-11e9-8f67-2632f168be36 container env-test: 
+Jun 18 13:11:04.778: INFO: Pod "pod-projected-secrets-85e33269-91ca-11e9-bce2-ae54e022189f" satisfied condition "success or failure"
+Jun 18 13:11:04.792: INFO: Trying to get logs from node 10.72.74.144 pod pod-projected-secrets-85e33269-91ca-11e9-bce2-ae54e022189f container projected-secret-volume-test: 
 STEP: delete the pod
-May 13 20:08:48.875: INFO: Waiting for pod pod-configmaps-e92b5c06-75ba-11e9-8f67-2632f168be36 to disappear
-May 13 20:08:48.882: INFO: Pod pod-configmaps-e92b5c06-75ba-11e9-8f67-2632f168be36 no longer exists
-[AfterEach] [sig-node] ConfigMap
+Jun 18 13:11:04.955: INFO: Waiting for pod pod-projected-secrets-85e33269-91ca-11e9-bce2-ae54e022189f to disappear
+Jun 18 13:11:04.971: INFO: Pod pod-projected-secrets-85e33269-91ca-11e9-bce2-ae54e022189f no longer exists
+[AfterEach] [sig-storage] Projected secret
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
-May 13 20:08:48.883: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "e2e-tests-configmap-kvtfd" for this suite.
-May 13 20:08:54.920: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-May 13 20:08:55.358: INFO: namespace: e2e-tests-configmap-kvtfd, resource: bindings, ignored listing per whitelist
-May 13 20:08:55.460: INFO: namespace e2e-tests-configmap-kvtfd deletion completed in 6.569515609s
+Jun 18 13:11:04.971: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-tests-projected-64mbg" for this suite.
+Jun 18 13:11:11.044: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 13:11:11.130: INFO: namespace: e2e-tests-projected-64mbg, resource: bindings, ignored listing per whitelist
+Jun 18 13:11:11.557: INFO: namespace e2e-tests-projected-64mbg deletion completed in 6.567768169s
 
-• [SLOW TEST:10.953 seconds]
-[sig-node] ConfigMap
-/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/configmap.go:31
-  should be consumable via the environment [NodeConformance] [Conformance]
+• [SLOW TEST:9.316 seconds]
+[sig-storage] Projected secret
+/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_secret.go:34
+  should be consumable from pods in volume with defaultMode set [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 ------------------------------
 S
 ------------------------------
-[sig-cli] Kubectl client [k8s.io] Kubectl rolling-update 
-  should support rolling-update to same image  [Conformance]
+[sig-cli] Kubectl client [k8s.io] Kubectl expose 
+  should create services for rc  [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 [BeforeEach] [sig-cli] Kubectl client
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
 STEP: Creating a kubernetes client
-May 13 20:08:55.460: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423
+Jun 18 13:11:11.558: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
 STEP: Building a namespace api object, basename kubectl
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-kubectl-kgw8x
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-kubectl-7qm9r
 STEP: Waiting for a default service account to be provisioned in namespace
 [BeforeEach] [sig-cli] Kubectl client
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:243
-[BeforeEach] [k8s.io] Kubectl rolling-update
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1358
-[It] should support rolling-update to same image  [Conformance]
+[It] should create services for rc  [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-STEP: running the image docker.io/library/nginx:1.14-alpine
-May 13 20:08:55.729: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 run e2e-test-nginx-rc --image=docker.io/library/nginx:1.14-alpine --generator=run/v1 --namespace=e2e-tests-kubectl-kgw8x'
-May 13 20:08:55.847: INFO: stderr: "kubectl run --generator=run/v1 is DEPRECATED and will be removed in a future version. Use kubectl run --generator=run-pod/v1 or kubectl create instead.\n"
-May 13 20:08:55.847: INFO: stdout: "replicationcontroller/e2e-test-nginx-rc created\n"
-STEP: verifying the rc e2e-test-nginx-rc was created
-May 13 20:08:55.858: INFO: Waiting for rc e2e-test-nginx-rc to stabilize, generation 1 observed generation 0 spec.replicas 1 status.replicas 0
-May 13 20:08:55.877: INFO: Waiting for rc e2e-test-nginx-rc to stabilize, generation 1 observed generation 1 spec.replicas 1 status.replicas 0
-STEP: rolling-update to same image controller
-May 13 20:08:55.889: INFO: scanned /root for discovery docs: 
-May 13 20:08:55.889: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 rolling-update e2e-test-nginx-rc --update-period=1s --image=docker.io/library/nginx:1.14-alpine --image-pull-policy=IfNotPresent --namespace=e2e-tests-kubectl-kgw8x'
-May 13 20:09:11.804: INFO: stderr: "Command \"rolling-update\" is deprecated, use \"rollout\" instead\n"
-May 13 20:09:11.804: INFO: stdout: "Created e2e-test-nginx-rc-656ee7a55f091ff84ed8bf35f4c4f106\nScaling up e2e-test-nginx-rc-656ee7a55f091ff84ed8bf35f4c4f106 from 0 to 1, scaling down e2e-test-nginx-rc from 1 to 0 (keep 1 pods available, don't exceed 2 pods)\nScaling e2e-test-nginx-rc-656ee7a55f091ff84ed8bf35f4c4f106 up to 1\nScaling e2e-test-nginx-rc down to 0\nUpdate succeeded. Deleting old controller: e2e-test-nginx-rc\nRenaming e2e-test-nginx-rc-656ee7a55f091ff84ed8bf35f4c4f106 to e2e-test-nginx-rc\nreplicationcontroller/e2e-test-nginx-rc rolling updated\n"
-May 13 20:09:11.804: INFO: stdout: "Created e2e-test-nginx-rc-656ee7a55f091ff84ed8bf35f4c4f106\nScaling up e2e-test-nginx-rc-656ee7a55f091ff84ed8bf35f4c4f106 from 0 to 1, scaling down e2e-test-nginx-rc from 1 to 0 (keep 1 pods available, don't exceed 2 pods)\nScaling e2e-test-nginx-rc-656ee7a55f091ff84ed8bf35f4c4f106 up to 1\nScaling e2e-test-nginx-rc down to 0\nUpdate succeeded. Deleting old controller: e2e-test-nginx-rc\nRenaming e2e-test-nginx-rc-656ee7a55f091ff84ed8bf35f4c4f106 to e2e-test-nginx-rc\nreplicationcontroller/e2e-test-nginx-rc rolling updated\n"
-STEP: waiting for all containers in run=e2e-test-nginx-rc pods to come up.
-May 13 20:09:11.804: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l run=e2e-test-nginx-rc --namespace=e2e-tests-kubectl-kgw8x'
-May 13 20:09:11.911: INFO: stderr: ""
-May 13 20:09:11.911: INFO: stdout: "e2e-test-nginx-rc-656ee7a55f091ff84ed8bf35f4c4f106-xf7fm "
-May 13 20:09:11.911: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 get pods e2e-test-nginx-rc-656ee7a55f091ff84ed8bf35f4c4f106-xf7fm -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "e2e-test-nginx-rc") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=e2e-tests-kubectl-kgw8x'
-May 13 20:09:12.032: INFO: stderr: ""
-May 13 20:09:12.032: INFO: stdout: "true"
-May 13 20:09:12.032: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 get pods e2e-test-nginx-rc-656ee7a55f091ff84ed8bf35f4c4f106-xf7fm -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "e2e-test-nginx-rc"}}{{.image}}{{end}}{{end}}{{end}} --namespace=e2e-tests-kubectl-kgw8x'
-May 13 20:09:12.130: INFO: stderr: ""
-May 13 20:09:12.130: INFO: stdout: "docker.io/library/nginx:1.14-alpine"
-May 13 20:09:12.130: INFO: e2e-test-nginx-rc-656ee7a55f091ff84ed8bf35f4c4f106-xf7fm is verified up and running
-[AfterEach] [k8s.io] Kubectl rolling-update
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1364
-May 13 20:09:12.130: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 delete rc e2e-test-nginx-rc --namespace=e2e-tests-kubectl-kgw8x'
-May 13 20:09:12.250: INFO: stderr: ""
-May 13 20:09:12.250: INFO: stdout: "replicationcontroller \"e2e-test-nginx-rc\" deleted\n"
+STEP: creating Redis RC
+Jun 18 13:11:12.041: INFO: namespace e2e-tests-kubectl-7qm9r
+Jun 18 13:11:12.041: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 create -f - --namespace=e2e-tests-kubectl-7qm9r'
+Jun 18 13:11:12.443: INFO: stderr: ""
+Jun 18 13:11:12.443: INFO: stdout: "replicationcontroller/redis-master created\n"
+STEP: Waiting for Redis master to start.
+Jun 18 13:11:13.460: INFO: Selector matched 1 pods for map[app:redis]
+Jun 18 13:11:13.460: INFO: Found 0 / 1
+Jun 18 13:11:14.460: INFO: Selector matched 1 pods for map[app:redis]
+Jun 18 13:11:14.460: INFO: Found 0 / 1
+Jun 18 13:11:15.460: INFO: Selector matched 1 pods for map[app:redis]
+Jun 18 13:11:15.460: INFO: Found 1 / 1
+Jun 18 13:11:15.460: INFO: WaitFor completed with timeout 5m0s.  Pods found = 1 out of 1
+Jun 18 13:11:15.475: INFO: Selector matched 1 pods for map[app:redis]
+Jun 18 13:11:15.475: INFO: ForEach: Found 1 pods from the filter.  Now looping through them.
+Jun 18 13:11:15.475: INFO: wait on redis-master startup in e2e-tests-kubectl-7qm9r 
+Jun 18 13:11:15.475: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 logs redis-master-8kdbk redis-master --namespace=e2e-tests-kubectl-7qm9r'
+Jun 18 13:11:15.707: INFO: stderr: ""
+Jun 18 13:11:15.707: INFO: stdout: "                _._                                                  \n           _.-``__ ''-._                                             \n      _.-``    `.  `_.  ''-._           Redis 3.2.12 (35a5711f/0) 64 bit\n  .-`` .-```.  ```\\/    _.,_ ''-._                                   \n (    '      ,       .-`  | `,    )     Running in standalone mode\n |`-._`-...-` __...-.``-._|'` _.-'|     Port: 6379\n |    `-._   `._    /     _.-'    |     PID: 1\n  `-._    `-._  `-./  _.-'    _.-'                                   \n |`-._`-._    `-.__.-'    _.-'_.-'|                                  \n |    `-._`-._        _.-'_.-'    |           http://redis.io        \n  `-._    `-._`-.__.-'_.-'    _.-'                                   \n |`-._`-._    `-.__.-'    _.-'_.-'|                                  \n |    `-._`-._        _.-'_.-'    |                                  \n  `-._    `-._`-.__.-'_.-'    _.-'                                   \n      `-._    `-.__.-'    _.-'                                       \n          `-._        _.-'                                           \n              `-.__.-'                                               \n\n1:M 18 Jun 13:11:13.718 # WARNING: The TCP backlog setting of 511 cannot be enforced because /proc/sys/net/core/somaxconn is set to the lower value of 128.\n1:M 18 Jun 13:11:13.718 # Server started, Redis version 3.2.12\n1:M 18 Jun 13:11:13.718 # WARNING you have Transparent Huge Pages (THP) support enabled in your kernel. This will create latency and memory usage issues with Redis. To fix this issue run the command 'echo never > /sys/kernel/mm/transparent_hugepage/enabled' as root, and add it to your /etc/rc.local in order to retain the setting after a reboot. Redis must be restarted after THP is disabled.\n1:M 18 Jun 13:11:13.718 * The server is now ready to accept connections on port 6379\n"
+STEP: exposing RC
+Jun 18 13:11:15.707: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 expose rc redis-master --name=rm2 --port=1234 --target-port=6379 --namespace=e2e-tests-kubectl-7qm9r'
+Jun 18 13:11:15.985: INFO: stderr: ""
+Jun 18 13:11:15.985: INFO: stdout: "service/rm2 exposed\n"
+Jun 18 13:11:15.999: INFO: Service rm2 in namespace e2e-tests-kubectl-7qm9r found.
+STEP: exposing service
+Jun 18 13:11:18.025: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 expose service rm2 --name=rm3 --port=2345 --target-port=6379 --namespace=e2e-tests-kubectl-7qm9r'
+Jun 18 13:11:18.272: INFO: stderr: ""
+Jun 18 13:11:18.272: INFO: stdout: "service/rm3 exposed\n"
+Jun 18 13:11:18.288: INFO: Service rm3 in namespace e2e-tests-kubectl-7qm9r found.
 [AfterEach] [sig-cli] Kubectl client
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
-May 13 20:09:12.250: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "e2e-tests-kubectl-kgw8x" for this suite.
-May 13 20:09:18.356: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-May 13 20:09:18.744: INFO: namespace: e2e-tests-kubectl-kgw8x, resource: bindings, ignored listing per whitelist
-May 13 20:09:18.831: INFO: namespace e2e-tests-kubectl-kgw8x deletion completed in 6.502925597s
+Jun 18 13:11:20.314: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-tests-kubectl-7qm9r" for this suite.
+Jun 18 13:11:44.384: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 13:11:44.758: INFO: namespace: e2e-tests-kubectl-7qm9r, resource: bindings, ignored listing per whitelist
+Jun 18 13:11:44.914: INFO: namespace e2e-tests-kubectl-7qm9r deletion completed in 24.580441115s
 
-• [SLOW TEST:23.371 seconds]
+• [SLOW TEST:33.357 seconds]
 [sig-cli] Kubectl client
 /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:22
-  [k8s.io] Kubectl rolling-update
+  [k8s.io] Kubectl expose
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694
-    should support rolling-update to same image  [Conformance]
+    should create services for rc  [Conformance]
     /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 ------------------------------
-SSS
+SSSSSSSSSSSSSSS
 ------------------------------
-[k8s.io] Pods 
-  should get a host IP [NodeConformance] [Conformance]
+[k8s.io] Probing container 
+  should be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-[BeforeEach] [k8s.io] Pods
+[BeforeEach] [k8s.io] Probing container
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
 STEP: Creating a kubernetes client
-May 13 20:09:18.831: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423
-STEP: Building a namespace api object, basename pods
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-pods-ctk8q
+Jun 18 13:11:44.916: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
+STEP: Building a namespace api object, basename container-probe
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-container-probe-4drcz
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [k8s.io] Pods
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:132
-[It] should get a host IP [NodeConformance] [Conformance]
+[BeforeEach] [k8s.io] Probing container
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/container_probe.go:48
+[It] should be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-STEP: creating pod
-May 13 20:09:21.164: INFO: Pod pod-hostip-fda1c021-75ba-11e9-8f67-2632f168be36 has hostIP: 10.170.219.153
-[AfterEach] [k8s.io] Pods
+STEP: Creating pod liveness-exec in namespace e2e-tests-container-probe-4drcz
+Jun 18 13:11:47.432: INFO: Started pod liveness-exec in namespace e2e-tests-container-probe-4drcz
+STEP: checking the pod's current state and verifying that restartCount is present
+Jun 18 13:11:47.447: INFO: Initial restart count of pod liveness-exec is 0
+Jun 18 13:12:41.968: INFO: Restart count of pod e2e-tests-container-probe-4drcz/liveness-exec is now 1 (54.521413371s elapsed)
+STEP: deleting the pod
+[AfterEach] [k8s.io] Probing container
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
-May 13 20:09:21.164: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "e2e-tests-pods-ctk8q" for this suite.
-May 13 20:09:45.251: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-May 13 20:09:45.401: INFO: namespace: e2e-tests-pods-ctk8q, resource: bindings, ignored listing per whitelist
-May 13 20:09:45.554: INFO: namespace e2e-tests-pods-ctk8q deletion completed in 24.380793794s
+Jun 18 13:12:42.008: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-tests-container-probe-4drcz" for this suite.
+Jun 18 13:12:48.076: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 13:12:48.356: INFO: namespace: e2e-tests-container-probe-4drcz, resource: bindings, ignored listing per whitelist
+Jun 18 13:12:49.154: INFO: namespace e2e-tests-container-probe-4drcz deletion completed in 7.126708611s
 
-• [SLOW TEST:26.723 seconds]
-[k8s.io] Pods
+• [SLOW TEST:64.238 seconds]
+[k8s.io] Probing container
 /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694
-  should get a host IP [NodeConformance] [Conformance]
+  should be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 ------------------------------
-S
+SSSSSSSSSSSSSSSSSSSSS
+------------------------------
+[sig-apps] ReplicaSet 
+  should adopt matching pods on creation and release no longer matching pods [Conformance]
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
+[BeforeEach] [sig-apps] ReplicaSet
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
+STEP: Creating a kubernetes client
+Jun 18 13:12:49.155: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
+STEP: Building a namespace api object, basename replicaset
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-replicaset-w9x8w
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should adopt matching pods on creation and release no longer matching pods [Conformance]
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
+STEP: Given a Pod with a 'name' label pod-adoption-release is created
+STEP: When a replicaset with a matching selector is created
+STEP: Then the orphan pod is adopted
+STEP: When the matched label of one of its pods change
+Jun 18 13:12:54.942: INFO: Pod name pod-adoption-release: Found 1 pods out of 1
+STEP: Then the pod is released
+[AfterEach] [sig-apps] ReplicaSet
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
+Jun 18 13:12:54.987: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-tests-replicaset-w9x8w" for this suite.
+Jun 18 13:13:19.860: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 13:13:20.342: INFO: namespace: e2e-tests-replicaset-w9x8w, resource: bindings, ignored listing per whitelist
+Jun 18 13:13:20.455: INFO: namespace e2e-tests-replicaset-w9x8w deletion completed in 25.438237253s
+
+• [SLOW TEST:31.300 seconds]
+[sig-apps] ReplicaSet
+/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22
+  should adopt matching pods on creation and release no longer matching pods [Conformance]
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
+------------------------------
+SSSSSSSSSSSSSSS
+------------------------------
+[sig-storage] Subpath Atomic writer volumes 
+  should support subpaths with configmap pod [Conformance]
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
+[BeforeEach] [sig-storage] Subpath
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
+STEP: Creating a kubernetes client
+Jun 18 13:13:20.455: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
+STEP: Building a namespace api object, basename subpath
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-subpath-hjl5x
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] Atomic writer volumes
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:38
+STEP: Setting up data
+[It] should support subpaths with configmap pod [Conformance]
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
+STEP: Creating pod pod-subpath-test-configmap-xgb7
+STEP: Creating a pod to test atomic-volume-subpath
+Jun 18 13:13:21.044: INFO: Waiting up to 5m0s for pod "pod-subpath-test-configmap-xgb7" in namespace "e2e-tests-subpath-hjl5x" to be "success or failure"
+Jun 18 13:13:21.062: INFO: Pod "pod-subpath-test-configmap-xgb7": Phase="Pending", Reason="", readiness=false. Elapsed: 17.856378ms
+Jun 18 13:13:23.103: INFO: Pod "pod-subpath-test-configmap-xgb7": Phase="Pending", Reason="", readiness=false. Elapsed: 2.059169233s
+Jun 18 13:13:25.118: INFO: Pod "pod-subpath-test-configmap-xgb7": Phase="Running", Reason="", readiness=false. Elapsed: 4.074205887s
+Jun 18 13:13:27.133: INFO: Pod "pod-subpath-test-configmap-xgb7": Phase="Running", Reason="", readiness=false. Elapsed: 6.089090054s
+Jun 18 13:13:29.149: INFO: Pod "pod-subpath-test-configmap-xgb7": Phase="Running", Reason="", readiness=false. Elapsed: 8.104405156s
+Jun 18 13:13:31.181: INFO: Pod "pod-subpath-test-configmap-xgb7": Phase="Running", Reason="", readiness=false. Elapsed: 10.136699649s
+Jun 18 13:13:33.196: INFO: Pod "pod-subpath-test-configmap-xgb7": Phase="Running", Reason="", readiness=false. Elapsed: 12.151280534s
+Jun 18 13:13:35.210: INFO: Pod "pod-subpath-test-configmap-xgb7": Phase="Running", Reason="", readiness=false. Elapsed: 14.166165172s
+Jun 18 13:13:37.226: INFO: Pod "pod-subpath-test-configmap-xgb7": Phase="Running", Reason="", readiness=false. Elapsed: 16.182072685s
+Jun 18 13:13:39.334: INFO: Pod "pod-subpath-test-configmap-xgb7": Phase="Running", Reason="", readiness=false. Elapsed: 18.289816801s
+Jun 18 13:13:41.368: INFO: Pod "pod-subpath-test-configmap-xgb7": Phase="Running", Reason="", readiness=false. Elapsed: 20.323974771s
+Jun 18 13:13:43.384: INFO: Pod "pod-subpath-test-configmap-xgb7": Phase="Running", Reason="", readiness=false. Elapsed: 22.339604032s
+Jun 18 13:13:45.399: INFO: Pod "pod-subpath-test-configmap-xgb7": Phase="Succeeded", Reason="", readiness=false. Elapsed: 24.3543421s
+STEP: Saw pod success
+Jun 18 13:13:45.399: INFO: Pod "pod-subpath-test-configmap-xgb7" satisfied condition "success or failure"
+Jun 18 13:13:45.413: INFO: Trying to get logs from node 10.72.74.143 pod pod-subpath-test-configmap-xgb7 container test-container-subpath-configmap-xgb7: 
+STEP: delete the pod
+Jun 18 13:13:45.984: INFO: Waiting for pod pod-subpath-test-configmap-xgb7 to disappear
+Jun 18 13:13:46.000: INFO: Pod pod-subpath-test-configmap-xgb7 no longer exists
+STEP: Deleting pod pod-subpath-test-configmap-xgb7
+Jun 18 13:13:46.000: INFO: Deleting pod "pod-subpath-test-configmap-xgb7" in namespace "e2e-tests-subpath-hjl5x"
+[AfterEach] [sig-storage] Subpath
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
+Jun 18 13:13:46.015: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-tests-subpath-hjl5x" for this suite.
+Jun 18 13:13:52.105: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 13:13:52.680: INFO: namespace: e2e-tests-subpath-hjl5x, resource: bindings, ignored listing per whitelist
+Jun 18 13:13:52.829: INFO: namespace e2e-tests-subpath-hjl5x deletion completed in 6.795241037s
+
+• [SLOW TEST:32.374 seconds]
+[sig-storage] Subpath
+/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:22
+  Atomic writer volumes
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:34
+    should support subpaths with configmap pod [Conformance]
+    /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
+------------------------------
+SS
+------------------------------
+[sig-storage] Secrets 
+  should be consumable from pods in volume with mappings [NodeConformance] [Conformance]
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
+[BeforeEach] [sig-storage] Secrets
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
+STEP: Creating a kubernetes client
+Jun 18 13:13:52.831: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
+STEP: Building a namespace api object, basename secrets
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-secrets-g9jtw
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should be consumable from pods in volume with mappings [NodeConformance] [Conformance]
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
+STEP: Creating secret with name secret-test-map-eb9784f6-91ca-11e9-bce2-ae54e022189f
+STEP: Creating a pod to test consume secrets
+Jun 18 13:13:53.391: INFO: Waiting up to 5m0s for pod "pod-secrets-eb99a2e2-91ca-11e9-bce2-ae54e022189f" in namespace "e2e-tests-secrets-g9jtw" to be "success or failure"
+Jun 18 13:13:53.404: INFO: Pod "pod-secrets-eb99a2e2-91ca-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 13.287486ms
+Jun 18 13:13:55.420: INFO: Pod "pod-secrets-eb99a2e2-91ca-11e9-bce2-ae54e022189f": Phase="Running", Reason="", readiness=true. Elapsed: 2.028949109s
+Jun 18 13:13:57.435: INFO: Pod "pod-secrets-eb99a2e2-91ca-11e9-bce2-ae54e022189f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.044299778s
+STEP: Saw pod success
+Jun 18 13:13:57.435: INFO: Pod "pod-secrets-eb99a2e2-91ca-11e9-bce2-ae54e022189f" satisfied condition "success or failure"
+Jun 18 13:13:57.449: INFO: Trying to get logs from node 10.72.74.144 pod pod-secrets-eb99a2e2-91ca-11e9-bce2-ae54e022189f container secret-volume-test: 
+STEP: delete the pod
+Jun 18 13:13:57.584: INFO: Waiting for pod pod-secrets-eb99a2e2-91ca-11e9-bce2-ae54e022189f to disappear
+Jun 18 13:13:57.598: INFO: Pod pod-secrets-eb99a2e2-91ca-11e9-bce2-ae54e022189f no longer exists
+[AfterEach] [sig-storage] Secrets
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
+Jun 18 13:13:57.599: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-tests-secrets-g9jtw" for this suite.
+Jun 18 13:14:03.687: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 13:14:04.185: INFO: namespace: e2e-tests-secrets-g9jtw, resource: bindings, ignored listing per whitelist
+Jun 18 13:14:04.526: INFO: namespace e2e-tests-secrets-g9jtw deletion completed in 6.90896854s
+
+• [SLOW TEST:11.695 seconds]
+[sig-storage] Secrets
+/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/secrets_volume.go:34
+  should be consumable from pods in volume with mappings [NodeConformance] [Conformance]
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
+------------------------------
+SSSSSSSSSSS
 ------------------------------
 [sig-storage] Secrets 
   should be consumable from pods in volume [NodeConformance] [Conformance]
@@ -9795,1041 +9871,1028 @@ S
 [BeforeEach] [sig-storage] Secrets
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
 STEP: Creating a kubernetes client
-May 13 20:09:45.554: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423
+Jun 18 13:14:04.528: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
 STEP: Building a namespace api object, basename secrets
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-secrets-tc48m
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-secrets-rsfbj
 STEP: Waiting for a default service account to be provisioned in namespace
 [It] should be consumable from pods in volume [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-STEP: Creating secret with name secret-test-0d9cecd6-75bb-11e9-8f67-2632f168be36
+STEP: Creating secret with name secret-test-f28ab4ea-91ca-11e9-bce2-ae54e022189f
 STEP: Creating a pod to test consume secrets
-May 13 20:09:45.947: INFO: Waiting up to 5m0s for pod "pod-secrets-0d9e6160-75bb-11e9-8f67-2632f168be36" in namespace "e2e-tests-secrets-tc48m" to be "success or failure"
-May 13 20:09:45.955: INFO: Pod "pod-secrets-0d9e6160-75bb-11e9-8f67-2632f168be36": Phase="Pending", Reason="", readiness=false. Elapsed: 7.293741ms
-May 13 20:09:47.975: INFO: Pod "pod-secrets-0d9e6160-75bb-11e9-8f67-2632f168be36": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.02724385s
+Jun 18 13:14:05.052: INFO: Waiting up to 5m0s for pod "pod-secrets-f28cd59f-91ca-11e9-bce2-ae54e022189f" in namespace "e2e-tests-secrets-rsfbj" to be "success or failure"
+Jun 18 13:14:05.066: INFO: Pod "pod-secrets-f28cd59f-91ca-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 13.653521ms
+Jun 18 13:14:07.081: INFO: Pod "pod-secrets-f28cd59f-91ca-11e9-bce2-ae54e022189f": Phase="Running", Reason="", readiness=true. Elapsed: 2.028771815s
+Jun 18 13:14:09.096: INFO: Pod "pod-secrets-f28cd59f-91ca-11e9-bce2-ae54e022189f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.043620661s
 STEP: Saw pod success
-May 13 20:09:47.975: INFO: Pod "pod-secrets-0d9e6160-75bb-11e9-8f67-2632f168be36" satisfied condition "success or failure"
-May 13 20:09:48.020: INFO: Trying to get logs from node 10.170.219.177 pod pod-secrets-0d9e6160-75bb-11e9-8f67-2632f168be36 container secret-volume-test: 
+Jun 18 13:14:09.096: INFO: Pod "pod-secrets-f28cd59f-91ca-11e9-bce2-ae54e022189f" satisfied condition "success or failure"
+Jun 18 13:14:09.111: INFO: Trying to get logs from node 10.72.74.149 pod pod-secrets-f28cd59f-91ca-11e9-bce2-ae54e022189f container secret-volume-test: 
 STEP: delete the pod
-May 13 20:09:48.062: INFO: Waiting for pod pod-secrets-0d9e6160-75bb-11e9-8f67-2632f168be36 to disappear
-May 13 20:09:48.070: INFO: Pod pod-secrets-0d9e6160-75bb-11e9-8f67-2632f168be36 no longer exists
+Jun 18 13:14:09.223: INFO: Waiting for pod pod-secrets-f28cd59f-91ca-11e9-bce2-ae54e022189f to disappear
+Jun 18 13:14:09.237: INFO: Pod pod-secrets-f28cd59f-91ca-11e9-bce2-ae54e022189f no longer exists
 [AfterEach] [sig-storage] Secrets
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
-May 13 20:09:48.070: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "e2e-tests-secrets-tc48m" for this suite.
-May 13 20:09:54.105: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-May 13 20:09:54.423: INFO: namespace: e2e-tests-secrets-tc48m, resource: bindings, ignored listing per whitelist
-May 13 20:09:54.501: INFO: namespace e2e-tests-secrets-tc48m deletion completed in 6.423705141s
+Jun 18 13:14:09.237: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-tests-secrets-rsfbj" for this suite.
+Jun 18 13:14:15.336: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 13:14:15.906: INFO: namespace: e2e-tests-secrets-rsfbj, resource: bindings, ignored listing per whitelist
+Jun 18 13:14:15.959: INFO: namespace e2e-tests-secrets-rsfbj deletion completed in 6.700216563s
 
-• [SLOW TEST:8.947 seconds]
+• [SLOW TEST:11.432 seconds]
 [sig-storage] Secrets
 /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/secrets_volume.go:34
   should be consumable from pods in volume [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 ------------------------------
-SS
-------------------------------
-[k8s.io] [sig-node] Pods Extended [k8s.io] Pods Set QOS Class 
-  should be submitted and removed  [Conformance]
+[sig-network] Networking Granular Checks: Pods 
+  should function for intra-pod communication: http [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-[BeforeEach] [k8s.io] [sig-node] Pods Extended
+[BeforeEach] [sig-network] Networking
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
 STEP: Creating a kubernetes client
-May 13 20:09:54.501: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423
-STEP: Building a namespace api object, basename pods
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-pods-srbl6
+Jun 18 13:14:15.961: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
+STEP: Building a namespace api object, basename pod-network-test
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-pod-network-test-wr2nx
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [k8s.io] Pods Set QOS Class
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/node/pods.go:204
-[It] should be submitted and removed  [Conformance]
+[It] should function for intra-pod communication: http [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-STEP: creating the pod
-STEP: submitting the pod to kubernetes
-STEP: verifying QOS class is set on the pod
-[AfterEach] [k8s.io] [sig-node] Pods Extended
+STEP: Performing setup for networking test in namespace e2e-tests-pod-network-test-wr2nx
+STEP: creating a selector
+STEP: Creating the service pods in kubernetes
+Jun 18 13:14:16.491: INFO: Waiting up to 10m0s for all (but 0) nodes to be schedulable
+STEP: Creating test pods
+Jun 18 13:14:38.985: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s 'http://172.30.114.51:8080/dial?request=hostName&protocol=http&host=172.30.114.45&port=8080&tries=1'] Namespace:e2e-tests-pod-network-test-wr2nx PodName:host-test-container-pod ContainerName:hostexec Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
+Jun 18 13:14:38.985: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
+Jun 18 13:14:39.232: INFO: Waiting for endpoints: map[]
+Jun 18 13:14:39.246: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s 'http://172.30.114.51:8080/dial?request=hostName&protocol=http&host=172.30.39.53&port=8080&tries=1'] Namespace:e2e-tests-pod-network-test-wr2nx PodName:host-test-container-pod ContainerName:hostexec Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
+Jun 18 13:14:39.246: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
+Jun 18 13:14:39.486: INFO: Waiting for endpoints: map[]
+Jun 18 13:14:39.501: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s 'http://172.30.114.51:8080/dial?request=hostName&protocol=http&host=172.30.58.191&port=8080&tries=1'] Namespace:e2e-tests-pod-network-test-wr2nx PodName:host-test-container-pod ContainerName:hostexec Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
+Jun 18 13:14:39.501: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
+Jun 18 13:14:39.730: INFO: Waiting for endpoints: map[]
+[AfterEach] [sig-network] Networking
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
-May 13 20:09:54.795: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "e2e-tests-pods-srbl6" for this suite.
-May 13 20:10:16.830: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-May 13 20:10:17.055: INFO: namespace: e2e-tests-pods-srbl6, resource: bindings, ignored listing per whitelist
-May 13 20:10:17.182: INFO: namespace e2e-tests-pods-srbl6 deletion completed in 22.379752585s
+Jun 18 13:14:39.730: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-tests-pod-network-test-wr2nx" for this suite.
+Jun 18 13:15:03.798: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 13:15:04.073: INFO: namespace: e2e-tests-pod-network-test-wr2nx, resource: bindings, ignored listing per whitelist
+Jun 18 13:15:04.498: INFO: namespace e2e-tests-pod-network-test-wr2nx deletion completed in 24.749006702s
 
-• [SLOW TEST:22.681 seconds]
-[k8s.io] [sig-node] Pods Extended
-/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694
-  [k8s.io] Pods Set QOS Class
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694
-    should be submitted and removed  [Conformance]
+• [SLOW TEST:48.537 seconds]
+[sig-network] Networking
+/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:25
+  Granular Checks: Pods
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:28
+    should function for intra-pod communication: http [NodeConformance] [Conformance]
     /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 ------------------------------
-SSSSSSSSSSSSSS
+SSSSSSS
 ------------------------------
-[k8s.io] Probing container 
-  should *not* be restarted with a /healthz http liveness probe [NodeConformance] [Conformance]
+[sig-network] DNS 
+  should provide DNS for the cluster  [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-[BeforeEach] [k8s.io] Probing container
+[BeforeEach] [sig-network] DNS
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
 STEP: Creating a kubernetes client
-May 13 20:10:17.185: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423
-STEP: Building a namespace api object, basename container-probe
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-container-probe-mshcx
+Jun 18 13:15:04.499: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
+STEP: Building a namespace api object, basename dns
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-dns-j625h
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [k8s.io] Probing container
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/container_probe.go:48
-[It] should *not* be restarted with a /healthz http liveness probe [NodeConformance] [Conformance]
+[It] should provide DNS for the cluster  [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-STEP: Creating pod liveness-http in namespace e2e-tests-container-probe-mshcx
-May 13 20:10:19.497: INFO: Started pod liveness-http in namespace e2e-tests-container-probe-mshcx
-STEP: checking the pod's current state and verifying that restartCount is present
-May 13 20:10:19.504: INFO: Initial restart count of pod liveness-http is 0
+STEP: Running these commands on wheezy: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search kubernetes.default A)" && test -n "$$check" && echo OK > /results/wheezy_udp@kubernetes.default;check="$$(dig +tcp +noall +answer +search kubernetes.default A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@kubernetes.default;check="$$(dig +notcp +noall +answer +search kubernetes.default.svc A)" && test -n "$$check" && echo OK > /results/wheezy_udp@kubernetes.default.svc;check="$$(dig +tcp +noall +answer +search kubernetes.default.svc A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@kubernetes.default.svc;check="$$(dig +notcp +noall +answer +search kubernetes.default.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_udp@kubernetes.default.svc.cluster.local;check="$$(dig +tcp +noall +answer +search kubernetes.default.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@kubernetes.default.svc.cluster.local;test -n "$$(getent hosts dns-querier-1.dns-test-service.e2e-tests-dns-j625h.svc.cluster.local)" && echo OK > /results/wheezy_hosts@dns-querier-1.dns-test-service.e2e-tests-dns-j625h.svc.cluster.local;test -n "$$(getent hosts dns-querier-1)" && echo OK > /results/wheezy_hosts@dns-querier-1;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".e2e-tests-dns-j625h.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@PodARecord;sleep 1; done
+
+STEP: Running these commands on jessie: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search kubernetes.default A)" && test -n "$$check" && echo OK > /results/jessie_udp@kubernetes.default;check="$$(dig +tcp +noall +answer +search kubernetes.default A)" && test -n "$$check" && echo OK > /results/jessie_tcp@kubernetes.default;check="$$(dig +notcp +noall +answer +search kubernetes.default.svc A)" && test -n "$$check" && echo OK > /results/jessie_udp@kubernetes.default.svc;check="$$(dig +tcp +noall +answer +search kubernetes.default.svc A)" && test -n "$$check" && echo OK > /results/jessie_tcp@kubernetes.default.svc;check="$$(dig +notcp +noall +answer +search kubernetes.default.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_udp@kubernetes.default.svc.cluster.local;check="$$(dig +tcp +noall +answer +search kubernetes.default.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_tcp@kubernetes.default.svc.cluster.local;test -n "$$(getent hosts dns-querier-1.dns-test-service.e2e-tests-dns-j625h.svc.cluster.local)" && echo OK > /results/jessie_hosts@dns-querier-1.dns-test-service.e2e-tests-dns-j625h.svc.cluster.local;test -n "$$(getent hosts dns-querier-1)" && echo OK > /results/jessie_hosts@dns-querier-1;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".e2e-tests-dns-j625h.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_tcp@PodARecord;sleep 1; done
+
+STEP: creating a pod to probe DNS
+STEP: submitting the pod to kubernetes
+STEP: retrieving the pod
+STEP: looking for the results for each expected name from probers
+Jun 18 13:15:09.606: INFO: DNS probes using e2e-tests-dns-j625h/dns-test-1644f1a4-91cb-11e9-bce2-ae54e022189f succeeded
+
 STEP: deleting the pod
-[AfterEach] [k8s.io] Probing container
+[AfterEach] [sig-network] DNS
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
-May 13 20:14:22.273: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "e2e-tests-container-probe-mshcx" for this suite.
-May 13 20:14:28.358: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-May 13 20:14:28.600: INFO: namespace: e2e-tests-container-probe-mshcx, resource: bindings, ignored listing per whitelist
-May 13 20:14:28.655: INFO: namespace e2e-tests-container-probe-mshcx deletion completed in 6.327500146s
+Jun 18 13:15:09.647: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-tests-dns-j625h" for this suite.
+Jun 18 13:15:15.722: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 13:15:16.116: INFO: namespace: e2e-tests-dns-j625h, resource: bindings, ignored listing per whitelist
+Jun 18 13:15:16.205: INFO: namespace e2e-tests-dns-j625h deletion completed in 6.533811732s
 
-• [SLOW TEST:251.471 seconds]
-[k8s.io] Probing container
-/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694
-  should *not* be restarted with a /healthz http liveness probe [NodeConformance] [Conformance]
+• [SLOW TEST:11.706 seconds]
+[sig-network] DNS
+/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:22
+  should provide DNS for the cluster  [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 ------------------------------
+SSSSSSSSSSS
+------------------------------
 [sig-storage] EmptyDir volumes 
-  should support (root,0777,default) [NodeConformance] [Conformance]
+  should support (root,0644,default) [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 [BeforeEach] [sig-storage] EmptyDir volumes
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
 STEP: Creating a kubernetes client
-May 13 20:14:28.656: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423
+Jun 18 13:15:16.205: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
 STEP: Building a namespace api object, basename emptydir
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-emptydir-5rsmm
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-emptydir-7vtph
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should support (root,0777,default) [NodeConformance] [Conformance]
+[It] should support (root,0644,default) [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-STEP: Creating a pod to test emptydir 0777 on node default medium
-May 13 20:14:29.037: INFO: Waiting up to 5m0s for pod "pod-b65a7534-75bb-11e9-8f67-2632f168be36" in namespace "e2e-tests-emptydir-5rsmm" to be "success or failure"
-May 13 20:14:29.047: INFO: Pod "pod-b65a7534-75bb-11e9-8f67-2632f168be36": Phase="Pending", Reason="", readiness=false. Elapsed: 10.414405ms
-May 13 20:14:31.059: INFO: Pod "pod-b65a7534-75bb-11e9-8f67-2632f168be36": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.021794917s
+STEP: Creating a pod to test emptydir 0644 on node default medium
+Jun 18 13:15:16.679: INFO: Waiting up to 5m0s for pod "pod-1d3e76e0-91cb-11e9-bce2-ae54e022189f" in namespace "e2e-tests-emptydir-7vtph" to be "success or failure"
+Jun 18 13:15:16.693: INFO: Pod "pod-1d3e76e0-91cb-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 14.254873ms
+Jun 18 13:15:18.709: INFO: Pod "pod-1d3e76e0-91cb-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 2.030537471s
+Jun 18 13:15:20.725: INFO: Pod "pod-1d3e76e0-91cb-11e9-bce2-ae54e022189f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.04609604s
 STEP: Saw pod success
-May 13 20:14:31.059: INFO: Pod "pod-b65a7534-75bb-11e9-8f67-2632f168be36" satisfied condition "success or failure"
-May 13 20:14:31.067: INFO: Trying to get logs from node 10.170.219.153 pod pod-b65a7534-75bb-11e9-8f67-2632f168be36 container test-container: 
+Jun 18 13:15:20.725: INFO: Pod "pod-1d3e76e0-91cb-11e9-bce2-ae54e022189f" satisfied condition "success or failure"
+Jun 18 13:15:20.742: INFO: Trying to get logs from node 10.72.74.143 pod pod-1d3e76e0-91cb-11e9-bce2-ae54e022189f container test-container: 
 STEP: delete the pod
-May 13 20:14:31.111: INFO: Waiting for pod pod-b65a7534-75bb-11e9-8f67-2632f168be36 to disappear
-May 13 20:14:31.117: INFO: Pod pod-b65a7534-75bb-11e9-8f67-2632f168be36 no longer exists
+Jun 18 13:15:20.819: INFO: Waiting for pod pod-1d3e76e0-91cb-11e9-bce2-ae54e022189f to disappear
+Jun 18 13:15:20.834: INFO: Pod pod-1d3e76e0-91cb-11e9-bce2-ae54e022189f no longer exists
 [AfterEach] [sig-storage] EmptyDir volumes
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
-May 13 20:14:31.117: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "e2e-tests-emptydir-5rsmm" for this suite.
-May 13 20:14:37.153: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-May 13 20:14:37.432: INFO: namespace: e2e-tests-emptydir-5rsmm, resource: bindings, ignored listing per whitelist
-May 13 20:14:37.552: INFO: namespace e2e-tests-emptydir-5rsmm deletion completed in 6.426595809s
+Jun 18 13:15:20.834: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-tests-emptydir-7vtph" for this suite.
+Jun 18 13:15:28.900: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 13:15:29.196: INFO: namespace: e2e-tests-emptydir-7vtph, resource: bindings, ignored listing per whitelist
+Jun 18 13:15:29.414: INFO: namespace e2e-tests-emptydir-7vtph deletion completed in 8.561247385s
 
-• [SLOW TEST:8.896 seconds]
+• [SLOW TEST:13.209 seconds]
 [sig-storage] EmptyDir volumes
 /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:40
-  should support (root,0777,default) [NodeConformance] [Conformance]
+  should support (root,0644,default) [NodeConformance] [Conformance]
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
+------------------------------
+SSSSSS
+------------------------------
+[sig-node] Downward API 
+  should provide pod name, namespace and IP address as env vars [NodeConformance] [Conformance]
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
+[BeforeEach] [sig-node] Downward API
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
+STEP: Creating a kubernetes client
+Jun 18 13:15:29.415: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
+STEP: Building a namespace api object, basename downward-api
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-downward-api-phbft
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should provide pod name, namespace and IP address as env vars [NodeConformance] [Conformance]
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
+STEP: Creating a pod to test downward api env vars
+Jun 18 13:15:29.903: INFO: Waiting up to 5m0s for pod "downward-api-252000a3-91cb-11e9-bce2-ae54e022189f" in namespace "e2e-tests-downward-api-phbft" to be "success or failure"
+Jun 18 13:15:29.917: INFO: Pod "downward-api-252000a3-91cb-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 13.926631ms
+Jun 18 13:15:31.932: INFO: Pod "downward-api-252000a3-91cb-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 2.028378987s
+Jun 18 13:15:33.967: INFO: Pod "downward-api-252000a3-91cb-11e9-bce2-ae54e022189f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.063922355s
+STEP: Saw pod success
+Jun 18 13:15:33.967: INFO: Pod "downward-api-252000a3-91cb-11e9-bce2-ae54e022189f" satisfied condition "success or failure"
+Jun 18 13:15:33.983: INFO: Trying to get logs from node 10.72.74.149 pod downward-api-252000a3-91cb-11e9-bce2-ae54e022189f container dapi-container: 
+STEP: delete the pod
+Jun 18 13:15:34.062: INFO: Waiting for pod downward-api-252000a3-91cb-11e9-bce2-ae54e022189f to disappear
+Jun 18 13:15:34.077: INFO: Pod downward-api-252000a3-91cb-11e9-bce2-ae54e022189f no longer exists
+[AfterEach] [sig-node] Downward API
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
+Jun 18 13:15:34.077: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-tests-downward-api-phbft" for this suite.
+Jun 18 13:15:40.153: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 13:15:40.246: INFO: namespace: e2e-tests-downward-api-phbft, resource: bindings, ignored listing per whitelist
+Jun 18 13:15:40.701: INFO: namespace e2e-tests-downward-api-phbft deletion completed in 6.595767325s
+
+• [SLOW TEST:11.286 seconds]
+[sig-node] Downward API
+/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downward_api.go:38
+  should provide pod name, namespace and IP address as env vars [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 ------------------------------
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+S
 ------------------------------
-[sig-storage] Projected downwardAPI 
-  should provide container's cpu limit [NodeConformance] [Conformance]
+[sig-api-machinery] Secrets 
+  should be consumable from pods in env vars [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-[BeforeEach] [sig-storage] Projected downwardAPI
+[BeforeEach] [sig-api-machinery] Secrets
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
 STEP: Creating a kubernetes client
-May 13 20:14:37.552: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423
-STEP: Building a namespace api object, basename projected
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-projected-4ffsr
+Jun 18 13:15:40.703: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
+STEP: Building a namespace api object, basename secrets
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-secrets-knphn
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-storage] Projected downwardAPI
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:39
-[It] should provide container's cpu limit [NodeConformance] [Conformance]
+[It] should be consumable from pods in env vars [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-STEP: Creating a pod to test downward API volume plugin
-May 13 20:14:37.899: INFO: Waiting up to 5m0s for pod "downwardapi-volume-bba27aa9-75bb-11e9-8f67-2632f168be36" in namespace "e2e-tests-projected-4ffsr" to be "success or failure"
-May 13 20:14:37.911: INFO: Pod "downwardapi-volume-bba27aa9-75bb-11e9-8f67-2632f168be36": Phase="Pending", Reason="", readiness=false. Elapsed: 11.990284ms
-May 13 20:14:39.919: INFO: Pod "downwardapi-volume-bba27aa9-75bb-11e9-8f67-2632f168be36": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.019870367s
+STEP: Creating secret with name secret-test-2bddb6a0-91cb-11e9-bce2-ae54e022189f
+STEP: Creating a pod to test consume secrets
+Jun 18 13:15:41.225: INFO: Waiting up to 5m0s for pod "pod-secrets-2bdfd18e-91cb-11e9-bce2-ae54e022189f" in namespace "e2e-tests-secrets-knphn" to be "success or failure"
+Jun 18 13:15:41.240: INFO: Pod "pod-secrets-2bdfd18e-91cb-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 14.883454ms
+Jun 18 13:15:43.255: INFO: Pod "pod-secrets-2bdfd18e-91cb-11e9-bce2-ae54e022189f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.029102313s
 STEP: Saw pod success
-May 13 20:14:39.919: INFO: Pod "downwardapi-volume-bba27aa9-75bb-11e9-8f67-2632f168be36" satisfied condition "success or failure"
-May 13 20:14:39.928: INFO: Trying to get logs from node 10.170.219.153 pod downwardapi-volume-bba27aa9-75bb-11e9-8f67-2632f168be36 container client-container: 
+Jun 18 13:15:43.255: INFO: Pod "pod-secrets-2bdfd18e-91cb-11e9-bce2-ae54e022189f" satisfied condition "success or failure"
+Jun 18 13:15:43.284: INFO: Trying to get logs from node 10.72.74.149 pod pod-secrets-2bdfd18e-91cb-11e9-bce2-ae54e022189f container secret-env-test: 
 STEP: delete the pod
-May 13 20:14:39.968: INFO: Waiting for pod downwardapi-volume-bba27aa9-75bb-11e9-8f67-2632f168be36 to disappear
-May 13 20:14:39.974: INFO: Pod downwardapi-volume-bba27aa9-75bb-11e9-8f67-2632f168be36 no longer exists
-[AfterEach] [sig-storage] Projected downwardAPI
+Jun 18 13:15:43.367: INFO: Waiting for pod pod-secrets-2bdfd18e-91cb-11e9-bce2-ae54e022189f to disappear
+Jun 18 13:15:43.380: INFO: Pod pod-secrets-2bdfd18e-91cb-11e9-bce2-ae54e022189f no longer exists
+[AfterEach] [sig-api-machinery] Secrets
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
-May 13 20:14:39.975: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "e2e-tests-projected-4ffsr" for this suite.
-May 13 20:14:46.022: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-May 13 20:14:46.257: INFO: namespace: e2e-tests-projected-4ffsr, resource: bindings, ignored listing per whitelist
-May 13 20:14:46.350: INFO: namespace e2e-tests-projected-4ffsr deletion completed in 6.368171089s
+Jun 18 13:15:43.380: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-tests-secrets-knphn" for this suite.
+Jun 18 13:15:49.446: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 13:15:49.771: INFO: namespace: e2e-tests-secrets-knphn, resource: bindings, ignored listing per whitelist
+Jun 18 13:15:51.233: INFO: namespace e2e-tests-secrets-knphn deletion completed in 7.832938476s
 
-• [SLOW TEST:8.798 seconds]
-[sig-storage] Projected downwardAPI
-/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:33
-  should provide container's cpu limit [NodeConformance] [Conformance]
+• [SLOW TEST:10.530 seconds]
+[sig-api-machinery] Secrets
+/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/secrets.go:32
+  should be consumable from pods in env vars [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 ------------------------------
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+SSSSSSSS
 ------------------------------
-[sig-storage] Secrets 
-  should be able to mount in a volume regardless of a different secret existing with same name in different namespace [NodeConformance] [Conformance]
+[sig-api-machinery] Garbage collector 
+  should not be blocked by dependency circle [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-[BeforeEach] [sig-storage] Secrets
+[BeforeEach] [sig-api-machinery] Garbage collector
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
 STEP: Creating a kubernetes client
-May 13 20:14:46.351: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423
-STEP: Building a namespace api object, basename secrets
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-secrets-chxwn
+Jun 18 13:15:51.233: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
+STEP: Building a namespace api object, basename gc
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-gc-mlchq
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should be able to mount in a volume regardless of a different secret existing with same name in different namespace [NodeConformance] [Conformance]
+[It] should not be blocked by dependency circle [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-secret-namespace-mxrzk
-STEP: Creating secret with name secret-test-c0d82497-75bb-11e9-8f67-2632f168be36
-STEP: Creating a pod to test consume secrets
-May 13 20:14:46.869: INFO: Waiting up to 5m0s for pod "pod-secrets-c0fb5fda-75bb-11e9-8f67-2632f168be36" in namespace "e2e-tests-secrets-chxwn" to be "success or failure"
-May 13 20:14:46.876: INFO: Pod "pod-secrets-c0fb5fda-75bb-11e9-8f67-2632f168be36": Phase="Pending", Reason="", readiness=false. Elapsed: 7.50352ms
-May 13 20:14:48.885: INFO: Pod "pod-secrets-c0fb5fda-75bb-11e9-8f67-2632f168be36": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.015557265s
-STEP: Saw pod success
-May 13 20:14:48.885: INFO: Pod "pod-secrets-c0fb5fda-75bb-11e9-8f67-2632f168be36" satisfied condition "success or failure"
-May 13 20:14:48.892: INFO: Trying to get logs from node 10.170.219.153 pod pod-secrets-c0fb5fda-75bb-11e9-8f67-2632f168be36 container secret-volume-test: 
-STEP: delete the pod
-May 13 20:14:49.020: INFO: Waiting for pod pod-secrets-c0fb5fda-75bb-11e9-8f67-2632f168be36 to disappear
-May 13 20:14:49.030: INFO: Pod pod-secrets-c0fb5fda-75bb-11e9-8f67-2632f168be36 no longer exists
-[AfterEach] [sig-storage] Secrets
+Jun 18 13:15:51.920: INFO: pod1.ObjectMeta.OwnerReferences=[]v1.OwnerReference{v1.OwnerReference{APIVersion:"v1", Kind:"Pod", Name:"pod3", UID:"323fa8cc-91cb-11e9-bf44-fa6f350b29f0", Controller:(*bool)(0xc002371b16), BlockOwnerDeletion:(*bool)(0xc002371b17)}}
+Jun 18 13:15:51.937: INFO: pod2.ObjectMeta.OwnerReferences=[]v1.OwnerReference{v1.OwnerReference{APIVersion:"v1", Kind:"Pod", Name:"pod1", UID:"3239f005-91cb-11e9-bf44-fa6f350b29f0", Controller:(*bool)(0xc002548532), BlockOwnerDeletion:(*bool)(0xc002548533)}}
+Jun 18 13:15:51.953: INFO: pod3.ObjectMeta.OwnerReferences=[]v1.OwnerReference{v1.OwnerReference{APIVersion:"v1", Kind:"Pod", Name:"pod2", UID:"323cd931-91cb-11e9-bf44-fa6f350b29f0", Controller:(*bool)(0xc002371daa), BlockOwnerDeletion:(*bool)(0xc002371dab)}}
+[AfterEach] [sig-api-machinery] Garbage collector
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
-May 13 20:14:49.030: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "e2e-tests-secrets-chxwn" for this suite.
-May 13 20:14:55.067: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-May 13 20:14:55.380: INFO: namespace: e2e-tests-secrets-chxwn, resource: bindings, ignored listing per whitelist
-May 13 20:14:55.395: INFO: namespace e2e-tests-secrets-chxwn deletion completed in 6.357320184s
-STEP: Destroying namespace "e2e-tests-secret-namespace-mxrzk" for this suite.
-May 13 20:15:01.427: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-May 13 20:15:01.576: INFO: namespace: e2e-tests-secret-namespace-mxrzk, resource: bindings, ignored listing per whitelist
-May 13 20:15:01.816: INFO: namespace e2e-tests-secret-namespace-mxrzk deletion completed in 6.421426597s
-
-• [SLOW TEST:15.465 seconds]
-[sig-storage] Secrets
-/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/secrets_volume.go:34
-  should be able to mount in a volume regardless of a different secret existing with same name in different namespace [NodeConformance] [Conformance]
+Jun 18 13:15:57.014: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-tests-gc-mlchq" for this suite.
+Jun 18 13:16:05.131: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 13:16:05.168: INFO: namespace: e2e-tests-gc-mlchq, resource: bindings, ignored listing per whitelist
+Jun 18 13:16:05.632: INFO: namespace e2e-tests-gc-mlchq deletion completed in 8.599046468s
+
+• [SLOW TEST:14.399 seconds]
+[sig-api-machinery] Garbage collector
+/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:22
+  should not be blocked by dependency circle [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 ------------------------------
-SSSSSSSSSSSSSSS
+S
 ------------------------------
-[k8s.io] Probing container 
-  with readiness probe should not be ready before initial delay and never restart [NodeConformance] [Conformance]
+[k8s.io] Container Lifecycle Hook when create a pod with lifecycle hook 
+  should execute poststart exec hook properly [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-[BeforeEach] [k8s.io] Probing container
+[BeforeEach] [k8s.io] Container Lifecycle Hook
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
 STEP: Creating a kubernetes client
-May 13 20:15:01.819: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423
-STEP: Building a namespace api object, basename container-probe
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-container-probe-c2jf8
+Jun 18 13:16:05.632: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
+STEP: Building a namespace api object, basename container-lifecycle-hook
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-container-lifecycle-hook-zf7f5
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [k8s.io] Probing container
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/container_probe.go:48
-[It] with readiness probe should not be ready before initial delay and never restart [NodeConformance] [Conformance]
+[BeforeEach] when create a pod with lifecycle hook
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:61
+STEP: create the container to handle the HTTPGet hook request.
+[It] should execute poststart exec hook properly [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-May 13 20:15:22.220: INFO: Container started at 2019-05-13 20:15:03 +0000 UTC, pod became ready at 2019-05-13 20:15:20 +0000 UTC
-[AfterEach] [k8s.io] Probing container
+STEP: create the pod with lifecycle hook
+STEP: check poststart hook
+STEP: delete the pod with lifecycle hook
+Jun 18 13:16:10.408: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear
+Jun 18 13:16:10.422: INFO: Pod pod-with-poststart-exec-hook still exists
+Jun 18 13:16:12.423: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear
+Jun 18 13:16:12.501: INFO: Pod pod-with-poststart-exec-hook still exists
+Jun 18 13:16:14.423: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear
+Jun 18 13:16:14.438: INFO: Pod pod-with-poststart-exec-hook still exists
+Jun 18 13:16:16.423: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear
+Jun 18 13:16:16.439: INFO: Pod pod-with-poststart-exec-hook still exists
+Jun 18 13:16:18.423: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear
+Jun 18 13:16:18.454: INFO: Pod pod-with-poststart-exec-hook still exists
+Jun 18 13:16:20.423: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear
+Jun 18 13:16:20.437: INFO: Pod pod-with-poststart-exec-hook still exists
+Jun 18 13:16:22.423: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear
+Jun 18 13:16:22.438: INFO: Pod pod-with-poststart-exec-hook still exists
+Jun 18 13:16:24.423: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear
+Jun 18 13:16:24.438: INFO: Pod pod-with-poststart-exec-hook still exists
+Jun 18 13:16:26.423: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear
+Jun 18 13:16:26.438: INFO: Pod pod-with-poststart-exec-hook still exists
+Jun 18 13:16:28.423: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear
+Jun 18 13:16:28.438: INFO: Pod pod-with-poststart-exec-hook still exists
+Jun 18 13:16:30.423: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear
+Jun 18 13:16:30.454: INFO: Pod pod-with-poststart-exec-hook no longer exists
+[AfterEach] [k8s.io] Container Lifecycle Hook
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
-May 13 20:15:22.220: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "e2e-tests-container-probe-c2jf8" for this suite.
-May 13 20:15:44.258: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-May 13 20:15:44.351: INFO: namespace: e2e-tests-container-probe-c2jf8, resource: bindings, ignored listing per whitelist
-May 13 20:15:44.489: INFO: namespace e2e-tests-container-probe-c2jf8 deletion completed in 22.261130375s
+Jun 18 13:16:30.455: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-tests-container-lifecycle-hook-zf7f5" for this suite.
+Jun 18 13:16:54.612: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 13:16:55.130: INFO: namespace: e2e-tests-container-lifecycle-hook-zf7f5, resource: bindings, ignored listing per whitelist
+Jun 18 13:16:55.182: INFO: namespace e2e-tests-container-lifecycle-hook-zf7f5 deletion completed in 24.707416018s
 
-• [SLOW TEST:42.671 seconds]
-[k8s.io] Probing container
+• [SLOW TEST:49.550 seconds]
+[k8s.io] Container Lifecycle Hook
 /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694
-  with readiness probe should not be ready before initial delay and never restart [NodeConformance] [Conformance]
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
+  when create a pod with lifecycle hook
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:40
+    should execute poststart exec hook properly [NodeConformance] [Conformance]
+    /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 ------------------------------
-SSSS
+SSSSSSSSSSSSSS
 ------------------------------
-[sig-storage] ConfigMap 
-  should be consumable from pods in volume [NodeConformance] [Conformance]
+[k8s.io] InitContainer [NodeConformance] 
+  should invoke init containers on a RestartAlways pod [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-[BeforeEach] [sig-storage] ConfigMap
+[BeforeEach] [k8s.io] InitContainer [NodeConformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
 STEP: Creating a kubernetes client
-May 13 20:15:44.489: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423
-STEP: Building a namespace api object, basename configmap
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-configmap-r9fqs
+Jun 18 13:16:55.183: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
+STEP: Building a namespace api object, basename init-container
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-init-container-dg268
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should be consumable from pods in volume [NodeConformance] [Conformance]
+[BeforeEach] [k8s.io] InitContainer [NodeConformance]
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/init_container.go:43
+[It] should invoke init containers on a RestartAlways pod [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-STEP: Creating configMap with name configmap-test-volume-e3896e31-75bb-11e9-8f67-2632f168be36
-STEP: Creating a pod to test consume configMaps
-May 13 20:15:44.852: INFO: Waiting up to 5m0s for pod "pod-configmaps-e38acc3f-75bb-11e9-8f67-2632f168be36" in namespace "e2e-tests-configmap-r9fqs" to be "success or failure"
-May 13 20:15:44.984: INFO: Pod "pod-configmaps-e38acc3f-75bb-11e9-8f67-2632f168be36": Phase="Pending", Reason="", readiness=false. Elapsed: 132.309598ms
-May 13 20:15:46.992: INFO: Pod "pod-configmaps-e38acc3f-75bb-11e9-8f67-2632f168be36": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.139894166s
-STEP: Saw pod success
-May 13 20:15:46.992: INFO: Pod "pod-configmaps-e38acc3f-75bb-11e9-8f67-2632f168be36" satisfied condition "success or failure"
-May 13 20:15:46.999: INFO: Trying to get logs from node 10.170.219.177 pod pod-configmaps-e38acc3f-75bb-11e9-8f67-2632f168be36 container configmap-volume-test: 
-STEP: delete the pod
-May 13 20:15:47.048: INFO: Waiting for pod pod-configmaps-e38acc3f-75bb-11e9-8f67-2632f168be36 to disappear
-May 13 20:15:47.058: INFO: Pod pod-configmaps-e38acc3f-75bb-11e9-8f67-2632f168be36 no longer exists
-[AfterEach] [sig-storage] ConfigMap
+STEP: creating the pod
+Jun 18 13:16:55.667: INFO: PodSpec: initContainers in spec.initContainers
+[AfterEach] [k8s.io] InitContainer [NodeConformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
-May 13 20:15:47.058: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "e2e-tests-configmap-r9fqs" for this suite.
-May 13 20:15:53.109: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-May 13 20:15:53.371: INFO: namespace: e2e-tests-configmap-r9fqs, resource: bindings, ignored listing per whitelist
-May 13 20:15:53.428: INFO: namespace e2e-tests-configmap-r9fqs deletion completed in 6.362596071s
+Jun 18 13:17:01.185: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-tests-init-container-dg268" for this suite.
+Jun 18 13:17:27.269: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 13:17:27.652: INFO: namespace: e2e-tests-init-container-dg268, resource: bindings, ignored listing per whitelist
+Jun 18 13:17:27.806: INFO: namespace e2e-tests-init-container-dg268 deletion completed in 26.601647431s
 
-• [SLOW TEST:8.939 seconds]
-[sig-storage] ConfigMap
-/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:33
-  should be consumable from pods in volume [NodeConformance] [Conformance]
+• [SLOW TEST:32.624 seconds]
+[k8s.io] InitContainer [NodeConformance]
+/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694
+  should invoke init containers on a RestartAlways pod [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 ------------------------------
-[k8s.io] Variable Expansion 
-  should allow composing env vars into new env vars [NodeConformance] [Conformance]
+SSSS
+------------------------------
+[k8s.io] Pods 
+  should allow activeDeadlineSeconds to be updated [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-[BeforeEach] [k8s.io] Variable Expansion
+[BeforeEach] [k8s.io] Pods
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
 STEP: Creating a kubernetes client
-May 13 20:15:53.428: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423
-STEP: Building a namespace api object, basename var-expansion
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-var-expansion-dzc9t
+Jun 18 13:17:27.807: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
+STEP: Building a namespace api object, basename pods
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-pods-z42lz
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should allow composing env vars into new env vars [NodeConformance] [Conformance]
+[BeforeEach] [k8s.io] Pods
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:132
+[It] should allow activeDeadlineSeconds to be updated [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-STEP: Creating a pod to test env composition
-May 13 20:15:53.726: INFO: Waiting up to 5m0s for pod "var-expansion-e8d4d3a7-75bb-11e9-8f67-2632f168be36" in namespace "e2e-tests-var-expansion-dzc9t" to be "success or failure"
-May 13 20:15:53.733: INFO: Pod "var-expansion-e8d4d3a7-75bb-11e9-8f67-2632f168be36": Phase="Pending", Reason="", readiness=false. Elapsed: 7.498608ms
-May 13 20:15:55.741: INFO: Pod "var-expansion-e8d4d3a7-75bb-11e9-8f67-2632f168be36": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.015254516s
-STEP: Saw pod success
-May 13 20:15:55.741: INFO: Pod "var-expansion-e8d4d3a7-75bb-11e9-8f67-2632f168be36" satisfied condition "success or failure"
-May 13 20:15:55.748: INFO: Trying to get logs from node 10.170.219.153 pod var-expansion-e8d4d3a7-75bb-11e9-8f67-2632f168be36 container dapi-container: 
-STEP: delete the pod
-May 13 20:15:55.846: INFO: Waiting for pod var-expansion-e8d4d3a7-75bb-11e9-8f67-2632f168be36 to disappear
-May 13 20:15:55.853: INFO: Pod var-expansion-e8d4d3a7-75bb-11e9-8f67-2632f168be36 no longer exists
-[AfterEach] [k8s.io] Variable Expansion
+STEP: creating the pod
+STEP: submitting the pod to kubernetes
+STEP: verifying the pod is in kubernetes
+STEP: updating the pod
+Jun 18 13:17:30.984: INFO: Successfully updated pod "pod-update-activedeadlineseconds-6bbd242d-91cb-11e9-bce2-ae54e022189f"
+Jun 18 13:17:30.984: INFO: Waiting up to 5m0s for pod "pod-update-activedeadlineseconds-6bbd242d-91cb-11e9-bce2-ae54e022189f" in namespace "e2e-tests-pods-z42lz" to be "terminated due to deadline exceeded"
+Jun 18 13:17:31.001: INFO: Pod "pod-update-activedeadlineseconds-6bbd242d-91cb-11e9-bce2-ae54e022189f": Phase="Running", Reason="", readiness=true. Elapsed: 16.786835ms
+Jun 18 13:17:33.091: INFO: Pod "pod-update-activedeadlineseconds-6bbd242d-91cb-11e9-bce2-ae54e022189f": Phase="Running", Reason="", readiness=true. Elapsed: 2.106513917s
+Jun 18 13:17:35.106: INFO: Pod "pod-update-activedeadlineseconds-6bbd242d-91cb-11e9-bce2-ae54e022189f": Phase="Failed", Reason="DeadlineExceeded", readiness=false. Elapsed: 4.121750271s
+Jun 18 13:17:35.106: INFO: Pod "pod-update-activedeadlineseconds-6bbd242d-91cb-11e9-bce2-ae54e022189f" satisfied condition "terminated due to deadline exceeded"
+[AfterEach] [k8s.io] Pods
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
-May 13 20:15:55.853: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "e2e-tests-var-expansion-dzc9t" for this suite.
-May 13 20:16:01.892: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-May 13 20:16:02.013: INFO: namespace: e2e-tests-var-expansion-dzc9t, resource: bindings, ignored listing per whitelist
-May 13 20:16:02.189: INFO: namespace e2e-tests-var-expansion-dzc9t deletion completed in 6.326041024s
+Jun 18 13:17:35.107: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-tests-pods-z42lz" for this suite.
+Jun 18 13:17:41.178: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 13:17:41.215: INFO: namespace: e2e-tests-pods-z42lz, resource: bindings, ignored listing per whitelist
+Jun 18 13:17:41.668: INFO: namespace e2e-tests-pods-z42lz deletion completed in 6.542006554s
 
-• [SLOW TEST:8.760 seconds]
-[k8s.io] Variable Expansion
+• [SLOW TEST:13.862 seconds]
+[k8s.io] Pods
 /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694
-  should allow composing env vars into new env vars [NodeConformance] [Conformance]
+  should allow activeDeadlineSeconds to be updated [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 ------------------------------
-SS
-------------------------------
-[sig-cli] Kubectl client [k8s.io] Kubectl run rc 
-  should create an rc from an image  [Conformance]
+[sig-storage] Secrets 
+  should be consumable from pods in volume with defaultMode set [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-[BeforeEach] [sig-cli] Kubectl client
+[BeforeEach] [sig-storage] Secrets
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
 STEP: Creating a kubernetes client
-May 13 20:16:02.189: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423
-STEP: Building a namespace api object, basename kubectl
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-kubectl-vvr62
+Jun 18 13:17:41.668: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
+STEP: Building a namespace api object, basename secrets
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-secrets-rr7wj
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:243
-[BeforeEach] [k8s.io] Kubectl run rc
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1298
-[It] should create an rc from an image  [Conformance]
+[It] should be consumable from pods in volume with defaultMode set [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-STEP: running the image docker.io/library/nginx:1.14-alpine
-May 13 20:16:02.543: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 run e2e-test-nginx-rc --image=docker.io/library/nginx:1.14-alpine --generator=run/v1 --namespace=e2e-tests-kubectl-vvr62'
-May 13 20:16:02.870: INFO: stderr: "kubectl run --generator=run/v1 is DEPRECATED and will be removed in a future version. Use kubectl run --generator=run-pod/v1 or kubectl create instead.\n"
-May 13 20:16:02.870: INFO: stdout: "replicationcontroller/e2e-test-nginx-rc created\n"
-STEP: verifying the rc e2e-test-nginx-rc was created
-STEP: verifying the pod controlled by rc e2e-test-nginx-rc was created
-STEP: confirm that you can get logs from an rc
-May 13 20:16:04.907: INFO: Waiting up to 5m0s for 1 pods to be running and ready: [e2e-test-nginx-rc-bvrtz]
-May 13 20:16:04.907: INFO: Waiting up to 5m0s for pod "e2e-test-nginx-rc-bvrtz" in namespace "e2e-tests-kubectl-vvr62" to be "running and ready"
-May 13 20:16:04.914: INFO: Pod "e2e-test-nginx-rc-bvrtz": Phase="Running", Reason="", readiness=true. Elapsed: 7.466876ms
-May 13 20:16:04.914: INFO: Pod "e2e-test-nginx-rc-bvrtz" satisfied condition "running and ready"
-May 13 20:16:04.914: INFO: Wanted all 1 pods to be running and ready. Result: true. Pods: [e2e-test-nginx-rc-bvrtz]
-May 13 20:16:04.914: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 logs rc/e2e-test-nginx-rc --namespace=e2e-tests-kubectl-vvr62'
-May 13 20:16:05.232: INFO: stderr: ""
-May 13 20:16:05.232: INFO: stdout: ""
-[AfterEach] [k8s.io] Kubectl run rc
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1303
-May 13 20:16:05.232: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 delete rc e2e-test-nginx-rc --namespace=e2e-tests-kubectl-vvr62'
-May 13 20:16:05.365: INFO: stderr: ""
-May 13 20:16:05.365: INFO: stdout: "replicationcontroller \"e2e-test-nginx-rc\" deleted\n"
-[AfterEach] [sig-cli] Kubectl client
+STEP: Creating secret with name secret-test-73f8089a-91cb-11e9-bce2-ae54e022189f
+STEP: Creating a pod to test consume secrets
+Jun 18 13:17:42.195: INFO: Waiting up to 5m0s for pod "pod-secrets-73fa6ecf-91cb-11e9-bce2-ae54e022189f" in namespace "e2e-tests-secrets-rr7wj" to be "success or failure"
+Jun 18 13:17:42.211: INFO: Pod "pod-secrets-73fa6ecf-91cb-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 15.911754ms
+Jun 18 13:17:44.226: INFO: Pod "pod-secrets-73fa6ecf-91cb-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 2.031425931s
+Jun 18 13:17:46.240: INFO: Pod "pod-secrets-73fa6ecf-91cb-11e9-bce2-ae54e022189f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.045620983s
+STEP: Saw pod success
+Jun 18 13:17:46.240: INFO: Pod "pod-secrets-73fa6ecf-91cb-11e9-bce2-ae54e022189f" satisfied condition "success or failure"
+Jun 18 13:17:46.258: INFO: Trying to get logs from node 10.72.74.144 pod pod-secrets-73fa6ecf-91cb-11e9-bce2-ae54e022189f container secret-volume-test: 
+STEP: delete the pod
+Jun 18 13:17:46.423: INFO: Waiting for pod pod-secrets-73fa6ecf-91cb-11e9-bce2-ae54e022189f to disappear
+Jun 18 13:17:46.437: INFO: Pod pod-secrets-73fa6ecf-91cb-11e9-bce2-ae54e022189f no longer exists
+[AfterEach] [sig-storage] Secrets
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
-May 13 20:16:05.365: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "e2e-tests-kubectl-vvr62" for this suite.
-May 13 20:16:29.401: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-May 13 20:16:29.640: INFO: namespace: e2e-tests-kubectl-vvr62, resource: bindings, ignored listing per whitelist
-May 13 20:16:29.651: INFO: namespace e2e-tests-kubectl-vvr62 deletion completed in 24.277262501s
+Jun 18 13:17:46.438: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-tests-secrets-rr7wj" for this suite.
+Jun 18 13:17:52.512: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 13:17:53.550: INFO: namespace: e2e-tests-secrets-rr7wj, resource: bindings, ignored listing per whitelist
+Jun 18 13:17:53.909: INFO: namespace e2e-tests-secrets-rr7wj deletion completed in 7.449275272s
 
-• [SLOW TEST:27.462 seconds]
-[sig-cli] Kubectl client
-/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:22
-  [k8s.io] Kubectl run rc
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694
-    should create an rc from an image  [Conformance]
-    /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-------------------------------
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+• [SLOW TEST:12.241 seconds]
+[sig-storage] Secrets
+/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/secrets_volume.go:34
+  should be consumable from pods in volume with defaultMode set [NodeConformance] [Conformance]
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 ------------------------------
 [k8s.io] Pods 
-  should allow activeDeadlineSeconds to be updated [NodeConformance] [Conformance]
+  should be updated [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 [BeforeEach] [k8s.io] Pods
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
 STEP: Creating a kubernetes client
-May 13 20:16:29.652: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423
+Jun 18 13:17:53.910: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
 STEP: Building a namespace api object, basename pods
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-pods-c9jvv
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-pods-5khw2
 STEP: Waiting for a default service account to be provisioned in namespace
 [BeforeEach] [k8s.io] Pods
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:132
-[It] should allow activeDeadlineSeconds to be updated [NodeConformance] [Conformance]
+[It] should be updated [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 STEP: creating the pod
 STEP: submitting the pod to kubernetes
 STEP: verifying the pod is in kubernetes
 STEP: updating the pod
-May 13 20:16:32.621: INFO: Successfully updated pod "pod-update-activedeadlineseconds-fe7d78cb-75bb-11e9-8f67-2632f168be36"
-May 13 20:16:32.621: INFO: Waiting up to 5m0s for pod "pod-update-activedeadlineseconds-fe7d78cb-75bb-11e9-8f67-2632f168be36" in namespace "e2e-tests-pods-c9jvv" to be "terminated due to deadline exceeded"
-May 13 20:16:32.630: INFO: Pod "pod-update-activedeadlineseconds-fe7d78cb-75bb-11e9-8f67-2632f168be36": Phase="Running", Reason="", readiness=true. Elapsed: 9.330431ms
-May 13 20:16:34.639: INFO: Pod "pod-update-activedeadlineseconds-fe7d78cb-75bb-11e9-8f67-2632f168be36": Phase="Running", Reason="", readiness=true. Elapsed: 2.017776709s
-May 13 20:16:36.647: INFO: Pod "pod-update-activedeadlineseconds-fe7d78cb-75bb-11e9-8f67-2632f168be36": Phase="Failed", Reason="DeadlineExceeded", readiness=false. Elapsed: 4.026009315s
-May 13 20:16:36.647: INFO: Pod "pod-update-activedeadlineseconds-fe7d78cb-75bb-11e9-8f67-2632f168be36" satisfied condition "terminated due to deadline exceeded"
+Jun 18 13:17:58.999: INFO: Successfully updated pod "pod-update-7b4271ce-91cb-11e9-bce2-ae54e022189f"
+STEP: verifying the updated pod is in kubernetes
+Jun 18 13:17:59.099: INFO: Pod update OK
 [AfterEach] [k8s.io] Pods
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
-May 13 20:16:36.647: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "e2e-tests-pods-c9jvv" for this suite.
-May 13 20:16:42.685: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-May 13 20:16:42.915: INFO: namespace: e2e-tests-pods-c9jvv, resource: bindings, ignored listing per whitelist
-May 13 20:16:42.924: INFO: namespace e2e-tests-pods-c9jvv deletion completed in 6.268454428s
+Jun 18 13:17:59.099: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-tests-pods-5khw2" for this suite.
+Jun 18 13:18:23.171: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 13:18:24.320: INFO: namespace: e2e-tests-pods-5khw2, resource: bindings, ignored listing per whitelist
+Jun 18 13:18:24.820: INFO: namespace e2e-tests-pods-5khw2 deletion completed in 25.699431957s
 
-• [SLOW TEST:13.272 seconds]
+• [SLOW TEST:30.911 seconds]
 [k8s.io] Pods
 /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694
-  should allow activeDeadlineSeconds to be updated [NodeConformance] [Conformance]
+  should be updated [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 ------------------------------
-S
+SSSSSSSS
 ------------------------------
-[sig-storage] EmptyDir volumes 
-  should support (non-root,0777,tmpfs) [NodeConformance] [Conformance]
+[sig-storage] ConfigMap 
+  should be consumable from pods in volume with mappings as non-root [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-[BeforeEach] [sig-storage] EmptyDir volumes
+[BeforeEach] [sig-storage] ConfigMap
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
 STEP: Creating a kubernetes client
-May 13 20:16:42.925: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423
-STEP: Building a namespace api object, basename emptydir
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-emptydir-65vck
+Jun 18 13:18:24.821: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
+STEP: Building a namespace api object, basename configmap
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-configmap-dk2dd
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should support (non-root,0777,tmpfs) [NodeConformance] [Conformance]
+[It] should be consumable from pods in volume with mappings as non-root [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-STEP: Creating a pod to test emptydir 0777 on tmpfs
-May 13 20:16:43.223: INFO: Waiting up to 5m0s for pod "pod-065585ae-75bc-11e9-8f67-2632f168be36" in namespace "e2e-tests-emptydir-65vck" to be "success or failure"
-May 13 20:16:43.231: INFO: Pod "pod-065585ae-75bc-11e9-8f67-2632f168be36": Phase="Pending", Reason="", readiness=false. Elapsed: 8.156224ms
-May 13 20:16:45.239: INFO: Pod "pod-065585ae-75bc-11e9-8f67-2632f168be36": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.015874661s
+STEP: Creating configMap with name configmap-test-volume-map-8dabeea2-91cb-11e9-bce2-ae54e022189f
+STEP: Creating a pod to test consume configMaps
+Jun 18 13:18:25.314: INFO: Waiting up to 5m0s for pod "pod-configmaps-8dade6db-91cb-11e9-bce2-ae54e022189f" in namespace "e2e-tests-configmap-dk2dd" to be "success or failure"
+Jun 18 13:18:25.335: INFO: Pod "pod-configmaps-8dade6db-91cb-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 21.263003ms
+Jun 18 13:18:27.350: INFO: Pod "pod-configmaps-8dade6db-91cb-11e9-bce2-ae54e022189f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.035770714s
 STEP: Saw pod success
-May 13 20:16:45.239: INFO: Pod "pod-065585ae-75bc-11e9-8f67-2632f168be36" satisfied condition "success or failure"
-May 13 20:16:45.246: INFO: Trying to get logs from node 10.170.219.153 pod pod-065585ae-75bc-11e9-8f67-2632f168be36 container test-container: 
+Jun 18 13:18:27.350: INFO: Pod "pod-configmaps-8dade6db-91cb-11e9-bce2-ae54e022189f" satisfied condition "success or failure"
+Jun 18 13:18:27.365: INFO: Trying to get logs from node 10.72.74.143 pod pod-configmaps-8dade6db-91cb-11e9-bce2-ae54e022189f container configmap-volume-test: 
 STEP: delete the pod
-May 13 20:16:45.345: INFO: Waiting for pod pod-065585ae-75bc-11e9-8f67-2632f168be36 to disappear
-May 13 20:16:45.353: INFO: Pod pod-065585ae-75bc-11e9-8f67-2632f168be36 no longer exists
-[AfterEach] [sig-storage] EmptyDir volumes
+Jun 18 13:18:27.447: INFO: Waiting for pod pod-configmaps-8dade6db-91cb-11e9-bce2-ae54e022189f to disappear
+Jun 18 13:18:27.462: INFO: Pod pod-configmaps-8dade6db-91cb-11e9-bce2-ae54e022189f no longer exists
+[AfterEach] [sig-storage] ConfigMap
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
-May 13 20:16:45.353: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "e2e-tests-emptydir-65vck" for this suite.
-May 13 20:16:51.391: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-May 13 20:16:51.489: INFO: namespace: e2e-tests-emptydir-65vck, resource: bindings, ignored listing per whitelist
-May 13 20:16:51.620: INFO: namespace e2e-tests-emptydir-65vck deletion completed in 6.258516218s
+Jun 18 13:18:27.462: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-tests-configmap-dk2dd" for this suite.
+Jun 18 13:18:33.550: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 13:18:33.739: INFO: namespace: e2e-tests-configmap-dk2dd, resource: bindings, ignored listing per whitelist
+Jun 18 13:18:34.040: INFO: namespace e2e-tests-configmap-dk2dd deletion completed in 6.556857498s
 
-• [SLOW TEST:8.696 seconds]
-[sig-storage] EmptyDir volumes
-/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:40
-  should support (non-root,0777,tmpfs) [NodeConformance] [Conformance]
+• [SLOW TEST:9.219 seconds]
+[sig-storage] ConfigMap
+/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:33
+  should be consumable from pods in volume with mappings as non-root [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 ------------------------------
-SS
+SSS
 ------------------------------
-[sig-network] Networking Granular Checks: Pods 
-  should function for node-pod communication: udp [NodeConformance] [Conformance]
+[k8s.io] InitContainer [NodeConformance] 
+  should not start app containers if init containers fail on a RestartAlways pod [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-[BeforeEach] [sig-network] Networking
+[BeforeEach] [k8s.io] InitContainer [NodeConformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
 STEP: Creating a kubernetes client
-May 13 20:16:51.620: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423
-STEP: Building a namespace api object, basename pod-network-test
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-pod-network-test-kjfjj
+Jun 18 13:18:34.041: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
+STEP: Building a namespace api object, basename init-container
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-init-container-cxpw5
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should function for node-pod communication: udp [NodeConformance] [Conformance]
+[BeforeEach] [k8s.io] InitContainer [NodeConformance]
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/init_container.go:43
+[It] should not start app containers if init containers fail on a RestartAlways pod [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-STEP: Performing setup for networking test in namespace e2e-tests-pod-network-test-kjfjj
-STEP: creating a selector
-STEP: Creating the service pods in kubernetes
-May 13 20:16:51.890: INFO: Waiting up to 10m0s for all (but 0) nodes to be schedulable
-STEP: Creating test pods
-May 13 20:17:14.065: INFO: ExecWithOptions {Command:[/bin/sh -c echo 'hostName' | nc -w 1 -u 172.30.227.80 8081 | grep -v '^\s*$'] Namespace:e2e-tests-pod-network-test-kjfjj PodName:host-test-container-pod ContainerName:hostexec Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
-May 13 20:17:14.065: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423
-May 13 20:17:15.275: INFO: Found all expected endpoints: [netserver-0]
-May 13 20:17:15.283: INFO: ExecWithOptions {Command:[/bin/sh -c echo 'hostName' | nc -w 1 -u 172.30.19.125 8081 | grep -v '^\s*$'] Namespace:e2e-tests-pod-network-test-kjfjj PodName:host-test-container-pod ContainerName:hostexec Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
-May 13 20:17:15.283: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423
-May 13 20:17:16.477: INFO: Found all expected endpoints: [netserver-1]
-May 13 20:17:16.528: INFO: ExecWithOptions {Command:[/bin/sh -c echo 'hostName' | nc -w 1 -u 172.30.63.20 8081 | grep -v '^\s*$'] Namespace:e2e-tests-pod-network-test-kjfjj PodName:host-test-container-pod ContainerName:hostexec Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
-May 13 20:17:16.528: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423
-May 13 20:17:17.687: INFO: Found all expected endpoints: [netserver-2]
-[AfterEach] [sig-network] Networking
+STEP: creating the pod
+Jun 18 13:18:35.599: INFO: PodSpec: initContainers in spec.initContainers
+Jun 18 13:19:19.552: INFO: init container has failed twice: &v1.Pod{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"pod-init-93d333f4-91cb-11e9-bce2-ae54e022189f", GenerateName:"", Namespace:"e2e-tests-init-container-cxpw5", SelfLink:"/api/v1/namespaces/e2e-tests-init-container-cxpw5/pods/pod-init-93d333f4-91cb-11e9-bce2-ae54e022189f", UID:"93d5c14c-91cb-11e9-bf44-fa6f350b29f0", ResourceVersion:"112969", Generation:0, CreationTimestamp:v1.Time{Time:time.Time{wall:0x0, ext:63696460715, loc:(*time.Location)(0x7b33b80)}}, DeletionTimestamp:(*v1.Time)(nil), DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"name":"foo", "time":"599161851"}, Annotations:map[string]string{"kubernetes.io/psp":"e2e-test-privileged-psp"}, OwnerReferences:[]v1.OwnerReference(nil), Initializers:(*v1.Initializers)(nil), Finalizers:[]string(nil), ClusterName:""}, Spec:v1.PodSpec{Volumes:[]v1.Volume{v1.Volume{Name:"default-token-sh6rs", VolumeSource:v1.VolumeSource{HostPath:(*v1.HostPathVolumeSource)(nil), EmptyDir:(*v1.EmptyDirVolumeSource)(nil), GCEPersistentDisk:(*v1.GCEPersistentDiskVolumeSource)(nil), AWSElasticBlockStore:(*v1.AWSElasticBlockStoreVolumeSource)(nil), GitRepo:(*v1.GitRepoVolumeSource)(nil), Secret:(*v1.SecretVolumeSource)(0xc001908d80), NFS:(*v1.NFSVolumeSource)(nil), ISCSI:(*v1.ISCSIVolumeSource)(nil), Glusterfs:(*v1.GlusterfsVolumeSource)(nil), PersistentVolumeClaim:(*v1.PersistentVolumeClaimVolumeSource)(nil), RBD:(*v1.RBDVolumeSource)(nil), FlexVolume:(*v1.FlexVolumeSource)(nil), Cinder:(*v1.CinderVolumeSource)(nil), CephFS:(*v1.CephFSVolumeSource)(nil), Flocker:(*v1.FlockerVolumeSource)(nil), DownwardAPI:(*v1.DownwardAPIVolumeSource)(nil), FC:(*v1.FCVolumeSource)(nil), AzureFile:(*v1.AzureFileVolumeSource)(nil), ConfigMap:(*v1.ConfigMapVolumeSource)(nil), VsphereVolume:(*v1.VsphereVirtualDiskVolumeSource)(nil), Quobyte:(*v1.QuobyteVolumeSource)(nil), AzureDisk:(*v1.AzureDiskVolumeSource)(nil), PhotonPersistentDisk:(*v1.PhotonPersistentDiskVolumeSource)(nil), Projected:(*v1.ProjectedVolumeSource)(nil), PortworxVolume:(*v1.PortworxVolumeSource)(nil), ScaleIO:(*v1.ScaleIOVolumeSource)(nil), StorageOS:(*v1.StorageOSVolumeSource)(nil)}}}, InitContainers:[]v1.Container{v1.Container{Name:"init1", Image:"docker.io/library/busybox:1.29", Command:[]string{"/bin/false"}, Args:[]string(nil), WorkingDir:"", Ports:[]v1.ContainerPort(nil), EnvFrom:[]v1.EnvFromSource(nil), Env:[]v1.EnvVar(nil), Resources:v1.ResourceRequirements{Limits:v1.ResourceList(nil), Requests:v1.ResourceList(nil)}, VolumeMounts:[]v1.VolumeMount{v1.VolumeMount{Name:"default-token-sh6rs", ReadOnly:true, MountPath:"/var/run/secrets/kubernetes.io/serviceaccount", SubPath:"", MountPropagation:(*v1.MountPropagationMode)(nil)}}, VolumeDevices:[]v1.VolumeDevice(nil), LivenessProbe:(*v1.Probe)(nil), ReadinessProbe:(*v1.Probe)(nil), Lifecycle:(*v1.Lifecycle)(nil), TerminationMessagePath:"/dev/termination-log", TerminationMessagePolicy:"File", ImagePullPolicy:"IfNotPresent", SecurityContext:(*v1.SecurityContext)(nil), Stdin:false, StdinOnce:false, TTY:false}, v1.Container{Name:"init2", Image:"docker.io/library/busybox:1.29", Command:[]string{"/bin/true"}, Args:[]string(nil), WorkingDir:"", Ports:[]v1.ContainerPort(nil), EnvFrom:[]v1.EnvFromSource(nil), Env:[]v1.EnvVar(nil), Resources:v1.ResourceRequirements{Limits:v1.ResourceList(nil), Requests:v1.ResourceList(nil)}, VolumeMounts:[]v1.VolumeMount{v1.VolumeMount{Name:"default-token-sh6rs", ReadOnly:true, MountPath:"/var/run/secrets/kubernetes.io/serviceaccount", SubPath:"", MountPropagation:(*v1.MountPropagationMode)(nil)}}, VolumeDevices:[]v1.VolumeDevice(nil), LivenessProbe:(*v1.Probe)(nil), ReadinessProbe:(*v1.Probe)(nil), Lifecycle:(*v1.Lifecycle)(nil), TerminationMessagePath:"/dev/termination-log", TerminationMessagePolicy:"File", ImagePullPolicy:"IfNotPresent", SecurityContext:(*v1.SecurityContext)(nil), Stdin:false, StdinOnce:false, TTY:false}}, Containers:[]v1.Container{v1.Container{Name:"run1", Image:"k8s.gcr.io/pause:3.1", Command:[]string(nil), Args:[]string(nil), WorkingDir:"", Ports:[]v1.ContainerPort(nil), EnvFrom:[]v1.EnvFromSource(nil), Env:[]v1.EnvVar(nil), Resources:v1.ResourceRequirements{Limits:v1.ResourceList{"cpu":resource.Quantity{i:resource.int64Amount{value:100, scale:-3}, d:resource.infDecAmount{Dec:(*inf.Dec)(nil)}, s:"100m", Format:"DecimalSI"}, "memory":resource.Quantity{i:resource.int64Amount{value:52428800, scale:0}, d:resource.infDecAmount{Dec:(*inf.Dec)(nil)}, s:"52428800", Format:"DecimalSI"}}, Requests:v1.ResourceList{"cpu":resource.Quantity{i:resource.int64Amount{value:100, scale:-3}, d:resource.infDecAmount{Dec:(*inf.Dec)(nil)}, s:"100m", Format:"DecimalSI"}, "memory":resource.Quantity{i:resource.int64Amount{value:52428800, scale:0}, d:resource.infDecAmount{Dec:(*inf.Dec)(nil)}, s:"52428800", Format:"DecimalSI"}}}, VolumeMounts:[]v1.VolumeMount{v1.VolumeMount{Name:"default-token-sh6rs", ReadOnly:true, MountPath:"/var/run/secrets/kubernetes.io/serviceaccount", SubPath:"", MountPropagation:(*v1.MountPropagationMode)(nil)}}, VolumeDevices:[]v1.VolumeDevice(nil), LivenessProbe:(*v1.Probe)(nil), ReadinessProbe:(*v1.Probe)(nil), Lifecycle:(*v1.Lifecycle)(nil), TerminationMessagePath:"/dev/termination-log", TerminationMessagePolicy:"File", ImagePullPolicy:"IfNotPresent", SecurityContext:(*v1.SecurityContext)(nil), Stdin:false, StdinOnce:false, TTY:false}}, RestartPolicy:"Always", TerminationGracePeriodSeconds:(*int64)(0xc0006aa698), ActiveDeadlineSeconds:(*int64)(nil), DNSPolicy:"ClusterFirst", NodeSelector:map[string]string(nil), ServiceAccountName:"default", DeprecatedServiceAccount:"default", AutomountServiceAccountToken:(*bool)(nil), NodeName:"10.72.74.144", HostNetwork:false, HostPID:false, HostIPC:false, ShareProcessNamespace:(*bool)(nil), SecurityContext:(*v1.PodSecurityContext)(0xc0026e7ec0), ImagePullSecrets:[]v1.LocalObjectReference(nil), Hostname:"", Subdomain:"", Affinity:(*v1.Affinity)(nil), SchedulerName:"default-scheduler", Tolerations:[]v1.Toleration{v1.Toleration{Key:"node.kubernetes.io/not-ready", Operator:"Exists", Value:"", Effect:"NoExecute", TolerationSeconds:(*int64)(0xc0006aa730)}, v1.Toleration{Key:"node.kubernetes.io/unreachable", Operator:"Exists", Value:"", Effect:"NoExecute", TolerationSeconds:(*int64)(0xc0006aa750)}}, HostAliases:[]v1.HostAlias(nil), PriorityClassName:"", Priority:(*int32)(0xc0006aa758), DNSConfig:(*v1.PodDNSConfig)(nil), ReadinessGates:[]v1.PodReadinessGate(nil), RuntimeClassName:(*string)(nil), EnableServiceLinks:(*bool)(0xc0006aa75c)}, Status:v1.PodStatus{Phase:"Pending", Conditions:[]v1.PodCondition{v1.PodCondition{Type:"Initialized", Status:"False", LastProbeTime:v1.Time{Time:time.Time{wall:0x0, ext:0, loc:(*time.Location)(nil)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63696460716, loc:(*time.Location)(0x7b33b80)}}, Reason:"ContainersNotInitialized", Message:"containers with incomplete status: [init1 init2]"}, v1.PodCondition{Type:"Ready", Status:"False", LastProbeTime:v1.Time{Time:time.Time{wall:0x0, ext:0, loc:(*time.Location)(nil)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63696460716, loc:(*time.Location)(0x7b33b80)}}, Reason:"ContainersNotReady", Message:"containers with unready status: [run1]"}, v1.PodCondition{Type:"ContainersReady", Status:"False", LastProbeTime:v1.Time{Time:time.Time{wall:0x0, ext:0, loc:(*time.Location)(nil)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63696460716, loc:(*time.Location)(0x7b33b80)}}, Reason:"ContainersNotReady", Message:"containers with unready status: [run1]"}, v1.PodCondition{Type:"PodScheduled", Status:"True", LastProbeTime:v1.Time{Time:time.Time{wall:0x0, ext:0, loc:(*time.Location)(nil)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63696460716, loc:(*time.Location)(0x7b33b80)}}, Reason:"", Message:""}}, Message:"", Reason:"", NominatedNodeName:"", HostIP:"10.72.74.144", PodIP:"172.30.114.48", StartTime:(*v1.Time)(0xc001a75520), InitContainerStatuses:[]v1.ContainerStatus{v1.ContainerStatus{Name:"init1", State:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(nil), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(0xc00157dd50)}, LastTerminationState:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(nil), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(0xc00157ddc0)}, Ready:false, RestartCount:3, Image:"docker.io/library/busybox:1.29", ImageID:"docker.io/library/busybox@sha256:8ccbac733d19c0dd4d70b4f0c1e12245b5fa3ad24758a11035ee505c629c0796", ContainerID:"containerd://505dd430c8eade065b775afed7cb772dad0b44d7ec300df02845ccd576088ed8"}, v1.ContainerStatus{Name:"init2", State:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(0xc001a75560), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(nil)}, LastTerminationState:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(nil), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(nil)}, Ready:false, RestartCount:0, Image:"docker.io/library/busybox:1.29", ImageID:"", ContainerID:""}}, ContainerStatuses:[]v1.ContainerStatus{v1.ContainerStatus{Name:"run1", State:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(0xc001a75540), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(nil)}, LastTerminationState:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(nil), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(nil)}, Ready:false, RestartCount:0, Image:"k8s.gcr.io/pause:3.1", ImageID:"", ContainerID:""}}, QOSClass:"Guaranteed"}}
+[AfterEach] [k8s.io] InitContainer [NodeConformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
-May 13 20:17:17.687: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "e2e-tests-pod-network-test-kjfjj" for this suite.
-May 13 20:17:35.725: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-May 13 20:17:35.849: INFO: namespace: e2e-tests-pod-network-test-kjfjj, resource: bindings, ignored listing per whitelist
-May 13 20:17:36.049: INFO: namespace e2e-tests-pod-network-test-kjfjj deletion completed in 18.352794377s
+Jun 18 13:19:19.552: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-tests-init-container-cxpw5" for this suite.
+Jun 18 13:19:43.712: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 13:19:44.003: INFO: namespace: e2e-tests-init-container-cxpw5, resource: bindings, ignored listing per whitelist
+Jun 18 13:19:44.301: INFO: namespace e2e-tests-init-container-cxpw5 deletion completed in 24.703338115s
 
-• [SLOW TEST:44.428 seconds]
-[sig-network] Networking
-/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:25
-  Granular Checks: Pods
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:28
-    should function for node-pod communication: udp [NodeConformance] [Conformance]
-    /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
+• [SLOW TEST:70.260 seconds]
+[k8s.io] InitContainer [NodeConformance]
+/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694
+  should not start app containers if init containers fail on a RestartAlways pod [Conformance]
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 ------------------------------
-S
+SSSSSS
 ------------------------------
-[sig-network] Services 
-  should serve multiport endpoints from pods  [Conformance]
+[k8s.io] Probing container 
+  should be restarted with a /healthz http liveness probe [NodeConformance] [Conformance]
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
+[BeforeEach] [k8s.io] Probing container
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
+STEP: Creating a kubernetes client
+Jun 18 13:19:44.302: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
+STEP: Building a namespace api object, basename container-probe
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-container-probe-vpq7l
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [k8s.io] Probing container
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/container_probe.go:48
+[It] should be restarted with a /healthz http liveness probe [NodeConformance] [Conformance]
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
+STEP: Creating pod liveness-http in namespace e2e-tests-container-probe-vpq7l
+Jun 18 13:19:46.812: INFO: Started pod liveness-http in namespace e2e-tests-container-probe-vpq7l
+STEP: checking the pod's current state and verifying that restartCount is present
+Jun 18 13:19:46.827: INFO: Initial restart count of pod liveness-http is 0
+Jun 18 13:20:08.044: INFO: Restart count of pod e2e-tests-container-probe-vpq7l/liveness-http is now 1 (21.216726207s elapsed)
+STEP: deleting the pod
+[AfterEach] [k8s.io] Probing container
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
+Jun 18 13:20:08.086: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-tests-container-probe-vpq7l" for this suite.
+Jun 18 13:20:14.176: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 13:20:14.422: INFO: namespace: e2e-tests-container-probe-vpq7l, resource: bindings, ignored listing per whitelist
+Jun 18 13:20:14.741: INFO: namespace e2e-tests-container-probe-vpq7l deletion completed in 6.632215565s
+
+• [SLOW TEST:30.439 seconds]
+[k8s.io] Probing container
+/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694
+  should be restarted with a /healthz http liveness probe [NodeConformance] [Conformance]
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
+------------------------------
+SSSSSSSSSSSSSSSSSSSS
+------------------------------
+[sig-storage] Projected configMap 
+  should be consumable from pods in volume with mappings and Item mode set [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-[BeforeEach] [sig-network] Services
+[BeforeEach] [sig-storage] Projected configMap
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
 STEP: Creating a kubernetes client
-May 13 20:17:36.049: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423
-STEP: Building a namespace api object, basename services
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-services-khddf
+Jun 18 13:20:14.741: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
+STEP: Building a namespace api object, basename projected
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-projected-269kc
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-network] Services
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:85
-[It] should serve multiport endpoints from pods  [Conformance]
+[It] should be consumable from pods in volume with mappings and Item mode set [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-STEP: creating service multi-endpoint-test in namespace e2e-tests-services-khddf
-STEP: waiting up to 3m0s for service multi-endpoint-test in namespace e2e-tests-services-khddf to expose endpoints map[]
-May 13 20:17:36.367: INFO: Get endpoints failed (5.133695ms elapsed, ignoring for 5s): endpoints "multi-endpoint-test" not found
-May 13 20:17:37.373: INFO: successfully validated that service multi-endpoint-test in namespace e2e-tests-services-khddf exposes endpoints map[] (1.01171746s elapsed)
-STEP: Creating pod pod1 in namespace e2e-tests-services-khddf
-STEP: waiting up to 3m0s for service multi-endpoint-test in namespace e2e-tests-services-khddf to expose endpoints map[pod1:[100]]
-May 13 20:17:39.432: INFO: successfully validated that service multi-endpoint-test in namespace e2e-tests-services-khddf exposes endpoints map[pod1:[100]] (2.041965415s elapsed)
-STEP: Creating pod pod2 in namespace e2e-tests-services-khddf
-STEP: waiting up to 3m0s for service multi-endpoint-test in namespace e2e-tests-services-khddf to expose endpoints map[pod1:[100] pod2:[101]]
-May 13 20:17:41.639: INFO: successfully validated that service multi-endpoint-test in namespace e2e-tests-services-khddf exposes endpoints map[pod1:[100] pod2:[101]] (2.119316868s elapsed)
-STEP: Deleting pod pod1 in namespace e2e-tests-services-khddf
-STEP: waiting up to 3m0s for service multi-endpoint-test in namespace e2e-tests-services-khddf to expose endpoints map[pod2:[101]]
-May 13 20:17:42.678: INFO: successfully validated that service multi-endpoint-test in namespace e2e-tests-services-khddf exposes endpoints map[pod2:[101]] (1.02579511s elapsed)
-STEP: Deleting pod pod2 in namespace e2e-tests-services-khddf
-STEP: waiting up to 3m0s for service multi-endpoint-test in namespace e2e-tests-services-khddf to expose endpoints map[]
-May 13 20:17:43.703: INFO: successfully validated that service multi-endpoint-test in namespace e2e-tests-services-khddf exposes endpoints map[] (1.011492866s elapsed)
-[AfterEach] [sig-network] Services
+STEP: Creating configMap with name projected-configmap-test-volume-map-cf2f7a3a-91cb-11e9-bce2-ae54e022189f
+STEP: Creating a pod to test consume configMaps
+Jun 18 13:20:15.231: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-cf31b0fa-91cb-11e9-bce2-ae54e022189f" in namespace "e2e-tests-projected-269kc" to be "success or failure"
+Jun 18 13:20:15.245: INFO: Pod "pod-projected-configmaps-cf31b0fa-91cb-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 14.039366ms
+Jun 18 13:20:17.260: INFO: Pod "pod-projected-configmaps-cf31b0fa-91cb-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 2.029047569s
+Jun 18 13:20:19.275: INFO: Pod "pod-projected-configmaps-cf31b0fa-91cb-11e9-bce2-ae54e022189f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.04394336s
+STEP: Saw pod success
+Jun 18 13:20:19.275: INFO: Pod "pod-projected-configmaps-cf31b0fa-91cb-11e9-bce2-ae54e022189f" satisfied condition "success or failure"
+Jun 18 13:20:19.290: INFO: Trying to get logs from node 10.72.74.143 pod pod-projected-configmaps-cf31b0fa-91cb-11e9-bce2-ae54e022189f container projected-configmap-volume-test: 
+STEP: delete the pod
+Jun 18 13:20:19.365: INFO: Waiting for pod pod-projected-configmaps-cf31b0fa-91cb-11e9-bce2-ae54e022189f to disappear
+Jun 18 13:20:19.383: INFO: Pod pod-projected-configmaps-cf31b0fa-91cb-11e9-bce2-ae54e022189f no longer exists
+[AfterEach] [sig-storage] Projected configMap
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
-May 13 20:17:43.742: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "e2e-tests-services-khddf" for this suite.
-May 13 20:17:49.794: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-May 13 20:17:49.987: INFO: namespace: e2e-tests-services-khddf, resource: bindings, ignored listing per whitelist
-May 13 20:17:50.042: INFO: namespace e2e-tests-services-khddf deletion completed in 6.276098868s
-[AfterEach] [sig-network] Services
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:90
+Jun 18 13:20:19.383: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-tests-projected-269kc" for this suite.
+Jun 18 13:20:25.479: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 13:20:25.990: INFO: namespace: e2e-tests-projected-269kc, resource: bindings, ignored listing per whitelist
+Jun 18 13:20:26.130: INFO: namespace e2e-tests-projected-269kc deletion completed in 6.724724058s
 
-• [SLOW TEST:13.993 seconds]
-[sig-network] Services
-/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:22
-  should serve multiport endpoints from pods  [Conformance]
+• [SLOW TEST:11.389 seconds]
+[sig-storage] Projected configMap
+/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_configmap.go:34
+  should be consumable from pods in volume with mappings and Item mode set [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 ------------------------------
-SSSSSSSSSS
-------------------------------
 [sig-apps] Deployment 
-  deployment should delete old replica sets [Conformance]
+  RecreateDeployment should delete old pods and create new ones [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 [BeforeEach] [sig-apps] Deployment
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
 STEP: Creating a kubernetes client
-May 13 20:17:50.044: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423
+Jun 18 13:20:26.131: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
 STEP: Building a namespace api object, basename deployment
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-deployment-h4qgj
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-deployment-2lrqs
 STEP: Waiting for a default service account to be provisioned in namespace
 [BeforeEach] [sig-apps] Deployment
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:65
-[It] deployment should delete old replica sets [Conformance]
+[It] RecreateDeployment should delete old pods and create new ones [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-May 13 20:17:50.430: INFO: Pod name cleanup-pod: Found 1 pods out of 1
-STEP: ensuring each pod is running
-May 13 20:17:52.447: INFO: Creating deployment test-cleanup-deployment
-STEP: Waiting for deployment test-cleanup-deployment history to be cleaned up
+Jun 18 13:20:26.578: INFO: Creating deployment "test-recreate-deployment"
+Jun 18 13:20:26.597: INFO: Waiting deployment "test-recreate-deployment" to be updated to revision 1
+Jun 18 13:20:26.624: INFO: new replicaset for deployment "test-recreate-deployment" is yet to be created
+Jun 18 13:20:28.655: INFO: Waiting deployment "test-recreate-deployment" to complete
+Jun 18 13:20:28.669: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63696460826, loc:(*time.Location)(0x7b33b80)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63696460826, loc:(*time.Location)(0x7b33b80)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63696460826, loc:(*time.Location)(0x7b33b80)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63696460826, loc:(*time.Location)(0x7b33b80)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-recreate-deployment-5dfdcc846d\" is progressing."}}, CollisionCount:(*int32)(nil)}
+Jun 18 13:20:30.683: INFO: Triggering a new rollout for deployment "test-recreate-deployment"
+Jun 18 13:20:30.714: INFO: Updating deployment test-recreate-deployment
+Jun 18 13:20:30.714: INFO: Watching deployment "test-recreate-deployment" to verify that new pods will not run with olds pods
 [AfterEach] [sig-apps] Deployment
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:59
-May 13 20:17:52.549: INFO: Deployment "test-cleanup-deployment":
-&Deployment{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-cleanup-deployment,GenerateName:,Namespace:e2e-tests-deployment-h4qgj,SelfLink:/apis/apps/v1/namespaces/e2e-tests-deployment-h4qgj/deployments/test-cleanup-deployment,UID:2fa6a224-75bc-11e9-b786-da20024d205c,ResourceVersion:55171,Generation:1,CreationTimestamp:2019-05-13 20:17:52 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: cleanup-pod,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Spec:DeploymentSpec{Replicas:*1,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: cleanup-pod,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: cleanup-pod,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[],Containers:[{redis gcr.io/kubernetes-e2e-test-images/redis:1.0 [] []  [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,},},Strategy:DeploymentStrategy{Type:RollingUpdate,RollingUpdate:&RollingUpdateDeployment{MaxUnavailable:25%!,(MISSING)MaxSurge:25%!,(MISSING)},},MinReadySeconds:0,RevisionHistoryLimit:*0,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:0,Replicas:0,UpdatedReplicas:0,AvailableReplicas:0,UnavailableReplicas:0,Conditions:[],ReadyReplicas:0,CollisionCount:nil,},}
-
-May 13 20:17:52.556: INFO: New ReplicaSet of Deployment "test-cleanup-deployment" is nil.
+Jun 18 13:20:30.883: INFO: Deployment "test-recreate-deployment":
+&Deployment{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-recreate-deployment,GenerateName:,Namespace:e2e-tests-deployment-2lrqs,SelfLink:/apis/apps/v1/namespaces/e2e-tests-deployment-2lrqs/deployments/test-recreate-deployment,UID:d5fa9605-91cb-11e9-bf44-fa6f350b29f0,ResourceVersion:113234,Generation:2,CreationTimestamp:2019-06-18 13:20:26 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod-3,},Annotations:map[string]string{deployment.kubernetes.io/revision: 2,},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Spec:DeploymentSpec{Replicas:*1,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod-3,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod-3,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] []  [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,},},Strategy:DeploymentStrategy{Type:Recreate,RollingUpdate:nil,},MinReadySeconds:0,RevisionHistoryLimit:*10,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:2,Replicas:1,UpdatedReplicas:1,AvailableReplicas:0,UnavailableReplicas:1,Conditions:[{Available False 2019-06-18 13:20:30 +0000 UTC 2019-06-18 13:20:30 +0000 UTC MinimumReplicasUnavailable Deployment does not have minimum availability.} {Progressing True 2019-06-18 13:20:30 +0000 UTC 2019-06-18 13:20:26 +0000 UTC ReplicaSetUpdated ReplicaSet "test-recreate-deployment-697fbf54bf" is progressing.}],ReadyReplicas:0,CollisionCount:nil,},}
+
+Jun 18 13:20:30.900: INFO: New ReplicaSet "test-recreate-deployment-697fbf54bf" of Deployment "test-recreate-deployment":
+&ReplicaSet{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-recreate-deployment-697fbf54bf,GenerateName:,Namespace:e2e-tests-deployment-2lrqs,SelfLink:/apis/apps/v1/namespaces/e2e-tests-deployment-2lrqs/replicasets/test-recreate-deployment-697fbf54bf,UID:d87c3336-91cb-11e9-bf44-fa6f350b29f0,ResourceVersion:113232,Generation:1,CreationTimestamp:2019-06-18 13:20:30 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod-3,pod-template-hash: 697fbf54bf,},Annotations:map[string]string{deployment.kubernetes.io/desired-replicas: 1,deployment.kubernetes.io/max-replicas: 1,deployment.kubernetes.io/revision: 2,},OwnerReferences:[{apps/v1 Deployment test-recreate-deployment d5fa9605-91cb-11e9-bf44-fa6f350b29f0 0xc00241f7a7 0xc00241f7a8}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:ReplicaSetSpec{Replicas:*1,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod-3,pod-template-hash: 697fbf54bf,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod-3,pod-template-hash: 697fbf54bf,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] []  [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,},},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:1,FullyLabeledReplicas:1,ObservedGeneration:1,ReadyReplicas:0,AvailableReplicas:0,Conditions:[],},}
+Jun 18 13:20:30.900: INFO: All old ReplicaSets of Deployment "test-recreate-deployment":
+Jun 18 13:20:30.900: INFO: &ReplicaSet{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-recreate-deployment-5dfdcc846d,GenerateName:,Namespace:e2e-tests-deployment-2lrqs,SelfLink:/apis/apps/v1/namespaces/e2e-tests-deployment-2lrqs/replicasets/test-recreate-deployment-5dfdcc846d,UID:d5ff7d37-91cb-11e9-bf44-fa6f350b29f0,ResourceVersion:113222,Generation:2,CreationTimestamp:2019-06-18 13:20:26 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod-3,pod-template-hash: 5dfdcc846d,},Annotations:map[string]string{deployment.kubernetes.io/desired-replicas: 1,deployment.kubernetes.io/max-replicas: 1,deployment.kubernetes.io/revision: 1,},OwnerReferences:[{apps/v1 Deployment test-recreate-deployment d5fa9605-91cb-11e9-bf44-fa6f350b29f0 0xc00241f6e7 0xc00241f6e8}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:ReplicaSetSpec{Replicas:*0,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod-3,pod-template-hash: 5dfdcc846d,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod-3,pod-template-hash: 5dfdcc846d,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[],Containers:[{redis gcr.io/kubernetes-e2e-test-images/redis:1.0 [] []  [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,},},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:0,FullyLabeledReplicas:0,ObservedGeneration:2,ReadyReplicas:0,AvailableReplicas:0,Conditions:[],},}
+Jun 18 13:20:30.916: INFO: Pod "test-recreate-deployment-697fbf54bf-q8zx6" is not available:
+&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-recreate-deployment-697fbf54bf-q8zx6,GenerateName:test-recreate-deployment-697fbf54bf-,Namespace:e2e-tests-deployment-2lrqs,SelfLink:/api/v1/namespaces/e2e-tests-deployment-2lrqs/pods/test-recreate-deployment-697fbf54bf-q8zx6,UID:d87dd047-91cb-11e9-bf44-fa6f350b29f0,ResourceVersion:113233,Generation:0,CreationTimestamp:2019-06-18 13:20:30 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod-3,pod-template-hash: 697fbf54bf,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet test-recreate-deployment-697fbf54bf d87c3336-91cb-11e9-bf44-fa6f350b29f0 0xc0023b03e7 0xc0023b03e8}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-lhx4s {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-lhx4s,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] []  [] [] [] {map[] map[]} [{default-token-lhx4s true /var/run/secrets/kubernetes.io/serviceaccount  }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.72.74.149,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists  NoExecute 0xc0023b0bb0} {node.kubernetes.io/unreachable Exists  NoExecute 0xc0023b0bd0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 13:20:30 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 13:20:30 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 13:20:30 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 13:20:30 +0000 UTC  }],Message:,Reason:,HostIP:10.72.74.149,PodIP:,StartTime:2019-06-18 13:20:30 +0000 UTC,ContainerStatuses:[{nginx {ContainerStateWaiting{Reason:ContainerCreating,Message:,} nil nil} {nil nil nil} false 0 docker.io/library/nginx:1.14-alpine  }],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},}
 [AfterEach] [sig-apps] Deployment
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
-May 13 20:17:52.563: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "e2e-tests-deployment-h4qgj" for this suite.
-May 13 20:17:58.600: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-May 13 20:17:58.741: INFO: namespace: e2e-tests-deployment-h4qgj, resource: bindings, ignored listing per whitelist
-May 13 20:17:58.885: INFO: namespace e2e-tests-deployment-h4qgj deletion completed in 6.313801321s
+Jun 18 13:20:30.916: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-tests-deployment-2lrqs" for this suite.
+Jun 18 13:20:38.994: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 13:20:39.134: INFO: namespace: e2e-tests-deployment-2lrqs, resource: bindings, ignored listing per whitelist
+Jun 18 13:20:39.570: INFO: namespace e2e-tests-deployment-2lrqs deletion completed in 8.62998369s
 
-• [SLOW TEST:8.841 seconds]
+• [SLOW TEST:13.439 seconds]
 [sig-apps] Deployment
 /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22
-  deployment should delete old replica sets [Conformance]
+  RecreateDeployment should delete old pods and create new ones [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 ------------------------------
-S
+SSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-scheduling] SchedulerPredicates [Serial] 
-  validates that NodeSelector is respected if not matching  [Conformance]
+[sig-node] ConfigMap 
+  should be consumable via the environment [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-[BeforeEach] [sig-scheduling] SchedulerPredicates [Serial]
+[BeforeEach] [sig-node] ConfigMap
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
 STEP: Creating a kubernetes client
-May 13 20:17:58.885: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423
-STEP: Building a namespace api object, basename sched-pred
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-sched-pred-4xfpc
+Jun 18 13:20:39.571: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
+STEP: Building a namespace api object, basename configmap
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-configmap-9gsfn
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-scheduling] SchedulerPredicates [Serial]
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:79
-May 13 20:17:59.145: INFO: Waiting up to 1m0s for all (but 0) nodes to be ready
-May 13 20:17:59.226: INFO: Waiting for terminating namespaces to be deleted...
-May 13 20:17:59.234: INFO: 
-Logging pods the kubelet thinks is on node 10.170.219.151 before test
-May 13 20:17:59.270: INFO: coredns-58d696879-rrkth from kube-system started at 2019-05-13 16:27:35 +0000 UTC (1 container statuses recorded)
-May 13 20:17:59.270: INFO: 	Container coredns ready: true, restart count 0
-May 13 20:17:59.270: INFO: ibm-file-plugin-7f6d8979bd-p5smn from kube-system started at 2019-05-13 16:27:35 +0000 UTC (1 container statuses recorded)
-May 13 20:17:59.270: INFO: 	Container ibm-file-plugin-container ready: true, restart count 0
-May 13 20:17:59.270: INFO: coredns-58d696879-zmj2g from kube-system started at 2019-05-13 16:27:49 +0000 UTC (1 container statuses recorded)
-May 13 20:17:59.270: INFO: 	Container coredns ready: true, restart count 0
-May 13 20:17:59.270: INFO: calico-node-d9d6s from kube-system started at 2019-05-13 16:27:05 +0000 UTC (1 container statuses recorded)
-May 13 20:17:59.270: INFO: 	Container calico-node ready: true, restart count 0
-May 13 20:17:59.270: INFO: vpn-774cf5c6d4-6p6cr from kube-system started at 2019-05-13 16:27:35 +0000 UTC (1 container statuses recorded)
-May 13 20:17:59.270: INFO: 	Container vpn ready: true, restart count 0
-May 13 20:17:59.270: INFO: ibm-keepalived-watcher-8wshf from kube-system started at 2019-05-13 16:27:05 +0000 UTC (1 container statuses recorded)
-May 13 20:17:59.270: INFO: 	Container keepalived-watcher ready: true, restart count 0
-May 13 20:17:59.270: INFO: kubernetes-dashboard-7996b848f4-mzbwj from kube-system started at 2019-05-13 16:27:35 +0000 UTC (1 container statuses recorded)
-May 13 20:17:59.270: INFO: 	Container kubernetes-dashboard ready: true, restart count 0
-May 13 20:17:59.270: INFO: ibm-kube-fluentd-6fw9c from kube-system started at 2019-05-13 16:32:50 +0000 UTC (1 container statuses recorded)
-May 13 20:17:59.270: INFO: 	Container fluentd ready: true, restart count 0
-May 13 20:17:59.270: INFO: ibm-storage-watcher-845946d5b5-knxcm from kube-system started at 2019-05-13 16:27:35 +0000 UTC (1 container statuses recorded)
-May 13 20:17:59.270: INFO: 	Container ibm-storage-watcher-container ready: true, restart count 0
-May 13 20:17:59.270: INFO: coredns-autoscaler-64f9c5b4df-hzq57 from kube-system started at 2019-05-13 16:27:35 +0000 UTC (1 container statuses recorded)
-May 13 20:17:59.270: INFO: 	Container autoscaler ready: true, restart count 0
-May 13 20:17:59.270: INFO: sonobuoy-systemd-logs-daemon-set-8e5ce9bdf4a94ece-nlv42 from heptio-sonobuoy started at 2019-05-13 18:40:58 +0000 UTC (2 container statuses recorded)
-May 13 20:17:59.270: INFO: 	Container sonobuoy-worker ready: true, restart count 1
-May 13 20:17:59.270: INFO: 	Container systemd-logs ready: true, restart count 1
-May 13 20:17:59.270: INFO: ibm-master-proxy-static-10.170.219.151 from kube-system started at  (0 container statuses recorded)
-May 13 20:17:59.270: INFO: calico-kube-controllers-7dd978d898-r8w74 from kube-system started at 2019-05-13 16:27:35 +0000 UTC (1 container statuses recorded)
-May 13 20:17:59.270: INFO: 	Container calico-kube-controllers ready: true, restart count 0
-May 13 20:17:59.270: INFO: public-cr0ef1a5f1e33c44fe9c21650356d908ee-alb1-8f59cf484-9v5pm from kube-system started at 2019-05-13 16:35:03 +0000 UTC (4 container statuses recorded)
-May 13 20:17:59.270: INFO: 	Container ingress-auth-1 ready: true, restart count 0
-May 13 20:17:59.270: INFO: 	Container ingress-auth-2 ready: true, restart count 0
-May 13 20:17:59.270: INFO: 	Container ingress-auth-3 ready: true, restart count 0
-May 13 20:17:59.270: INFO: 	Container nginx-ingress ready: true, restart count 0
-May 13 20:17:59.270: INFO: 
-Logging pods the kubelet thinks is on node 10.170.219.153 before test
-May 13 20:17:59.299: INFO: sonobuoy-systemd-logs-daemon-set-8e5ce9bdf4a94ece-tgp2l from heptio-sonobuoy started at 2019-05-13 18:40:58 +0000 UTC (2 container statuses recorded)
-May 13 20:17:59.299: INFO: 	Container sonobuoy-worker ready: true, restart count 1
-May 13 20:17:59.299: INFO: 	Container systemd-logs ready: true, restart count 1
-May 13 20:17:59.299: INFO: metrics-server-58dd96c6bd-sggps from kube-system started at 2019-05-13 16:28:04 +0000 UTC (2 container statuses recorded)
-May 13 20:17:59.299: INFO: 	Container metrics-server ready: true, restart count 0
-May 13 20:17:59.299: INFO: 	Container metrics-server-nanny ready: true, restart count 0
-May 13 20:17:59.299: INFO: test-k8s-e2e-pvg-master-verification from default started at 2019-05-13 18:40:52 +0000 UTC (1 container statuses recorded)
-May 13 20:17:59.299: INFO: 	Container test-k8s-e2e-pvg-master-verification ready: true, restart count 0
-May 13 20:17:59.299: INFO: ibm-master-proxy-static-10.170.219.153 from kube-system started at  (0 container statuses recorded)
-May 13 20:17:59.299: INFO: ibm-keepalived-watcher-cwtfg from kube-system started at 2019-05-13 16:27:29 +0000 UTC (1 container statuses recorded)
-May 13 20:17:59.299: INFO: 	Container keepalived-watcher ready: true, restart count 0
-May 13 20:17:59.299: INFO: ibm-kube-fluentd-cdmr5 from kube-system started at 2019-05-13 16:32:50 +0000 UTC (1 container statuses recorded)
-May 13 20:17:59.299: INFO: 	Container fluentd ready: true, restart count 0
-May 13 20:17:59.299: INFO: sonobuoy from heptio-sonobuoy started at 2019-05-13 18:40:57 +0000 UTC (1 container statuses recorded)
-May 13 20:17:59.299: INFO: 	Container kube-sonobuoy ready: true, restart count 0
-May 13 20:17:59.299: INFO: calico-node-phw8q from kube-system started at 2019-05-13 16:27:29 +0000 UTC (1 container statuses recorded)
-May 13 20:17:59.299: INFO: 	Container calico-node ready: true, restart count 0
-May 13 20:17:59.299: INFO: ibm-cloud-provider-ip-169-45-218-130-66c489dcb-z58z4 from ibm-system started at 2019-05-13 16:30:08 +0000 UTC (1 container statuses recorded)
-May 13 20:17:59.299: INFO: 	Container ibm-cloud-provider-ip-169-45-218-130 ready: true, restart count 0
-May 13 20:17:59.299: INFO: 
-Logging pods the kubelet thinks is on node 10.170.219.177 before test
-May 13 20:17:59.319: INFO: ibm-cloud-provider-ip-169-45-218-130-66c489dcb-kd74s from ibm-system started at 2019-05-13 16:30:08 +0000 UTC (1 container statuses recorded)
-May 13 20:17:59.319: INFO: 	Container ibm-cloud-provider-ip-169-45-218-130 ready: true, restart count 0
-May 13 20:17:59.319: INFO: public-cr0ef1a5f1e33c44fe9c21650356d908ee-alb1-8f59cf484-7r5vq from kube-system started at 2019-05-13 16:35:03 +0000 UTC (4 container statuses recorded)
-May 13 20:17:59.319: INFO: 	Container ingress-auth-1 ready: true, restart count 0
-May 13 20:17:59.319: INFO: 	Container ingress-auth-2 ready: true, restart count 0
-May 13 20:17:59.319: INFO: 	Container ingress-auth-3 ready: true, restart count 0
-May 13 20:17:59.319: INFO: 	Container nginx-ingress ready: true, restart count 0
-May 13 20:17:59.319: INFO: ibm-kube-fluentd-zzl5d from kube-system started at 2019-05-13 16:32:50 +0000 UTC (1 container statuses recorded)
-May 13 20:17:59.319: INFO: 	Container fluentd ready: true, restart count 0
-May 13 20:17:59.319: INFO: sonobuoy-e2e-job-afa7e83446444f4e from heptio-sonobuoy started at 2019-05-13 18:40:58 +0000 UTC (2 container statuses recorded)
-May 13 20:17:59.319: INFO: 	Container e2e ready: true, restart count 0
-May 13 20:17:59.319: INFO: 	Container sonobuoy-worker ready: true, restart count 0
-May 13 20:17:59.319: INFO: sonobuoy-systemd-logs-daemon-set-8e5ce9bdf4a94ece-kjc69 from heptio-sonobuoy started at 2019-05-13 18:40:58 +0000 UTC (2 container statuses recorded)
-May 13 20:17:59.319: INFO: 	Container sonobuoy-worker ready: true, restart count 1
-May 13 20:17:59.319: INFO: 	Container systemd-logs ready: true, restart count 1
-May 13 20:17:59.319: INFO: ibm-master-proxy-static-10.170.219.177 from kube-system started at  (0 container statuses recorded)
-May 13 20:17:59.319: INFO: ibm-keepalived-watcher-qszvz from kube-system started at 2019-05-13 16:27:34 +0000 UTC (1 container statuses recorded)
-May 13 20:17:59.319: INFO: 	Container keepalived-watcher ready: true, restart count 0
-May 13 20:17:59.319: INFO: calico-node-twx7l from kube-system started at 2019-05-13 16:27:34 +0000 UTC (1 container statuses recorded)
-May 13 20:17:59.319: INFO: 	Container calico-node ready: true, restart count 0
-[It] validates that NodeSelector is respected if not matching  [Conformance]
+[It] should be consumable via the environment [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-STEP: Trying to schedule Pod with nonempty NodeSelector.
-STEP: Considering event: 
-Type = [Warning], Name = [restricted-pod.159e577495dff002], Reason = [FailedScheduling], Message = [0/3 nodes are available: 3 node(s) didn't match node selector.]
-[AfterEach] [sig-scheduling] SchedulerPredicates [Serial]
+STEP: Creating configMap e2e-tests-configmap-9gsfn/configmap-test-ddfe3170-91cb-11e9-bce2-ae54e022189f
+STEP: Creating a pod to test consume configMaps
+Jun 18 13:20:40.077: INFO: Waiting up to 5m0s for pod "pod-configmaps-de00a116-91cb-11e9-bce2-ae54e022189f" in namespace "e2e-tests-configmap-9gsfn" to be "success or failure"
+Jun 18 13:20:40.093: INFO: Pod "pod-configmaps-de00a116-91cb-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 16.195891ms
+Jun 18 13:20:42.116: INFO: Pod "pod-configmaps-de00a116-91cb-11e9-bce2-ae54e022189f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.03960635s
+STEP: Saw pod success
+Jun 18 13:20:42.117: INFO: Pod "pod-configmaps-de00a116-91cb-11e9-bce2-ae54e022189f" satisfied condition "success or failure"
+Jun 18 13:20:42.132: INFO: Trying to get logs from node 10.72.74.143 pod pod-configmaps-de00a116-91cb-11e9-bce2-ae54e022189f container env-test: 
+STEP: delete the pod
+Jun 18 13:20:42.220: INFO: Waiting for pod pod-configmaps-de00a116-91cb-11e9-bce2-ae54e022189f to disappear
+Jun 18 13:20:42.237: INFO: Pod pod-configmaps-de00a116-91cb-11e9-bce2-ae54e022189f no longer exists
+[AfterEach] [sig-node] ConfigMap
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
-May 13 20:18:00.369: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "e2e-tests-sched-pred-4xfpc" for this suite.
-May 13 20:18:06.472: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-May 13 20:18:06.776: INFO: namespace: e2e-tests-sched-pred-4xfpc, resource: bindings, ignored listing per whitelist
-May 13 20:18:06.904: INFO: namespace e2e-tests-sched-pred-4xfpc deletion completed in 6.484047823s
-[AfterEach] [sig-scheduling] SchedulerPredicates [Serial]
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:70
+Jun 18 13:20:42.237: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-tests-configmap-9gsfn" for this suite.
+Jun 18 13:20:48.323: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 13:20:48.591: INFO: namespace: e2e-tests-configmap-9gsfn, resource: bindings, ignored listing per whitelist
+Jun 18 13:20:48.842: INFO: namespace e2e-tests-configmap-9gsfn deletion completed in 6.58538389s
 
-• [SLOW TEST:8.020 seconds]
-[sig-scheduling] SchedulerPredicates [Serial]
-/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/framework.go:22
-  validates that NodeSelector is respected if not matching  [Conformance]
+• [SLOW TEST:9.272 seconds]
+[sig-node] ConfigMap
+/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/configmap.go:31
+  should be consumable via the environment [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 ------------------------------
-SS
+SSS
 ------------------------------
-[sig-apps] ReplicationController 
-  should adopt matching pods on creation [Conformance]
+[sig-storage] Secrets 
+  should be consumable in multiple volumes in a pod [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-[BeforeEach] [sig-apps] ReplicationController
+[BeforeEach] [sig-storage] Secrets
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
 STEP: Creating a kubernetes client
-May 13 20:18:06.905: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423
-STEP: Building a namespace api object, basename replication-controller
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-replication-controller-8vtxs
+Jun 18 13:20:48.843: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
+STEP: Building a namespace api object, basename secrets
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-secrets-5fwtb
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should adopt matching pods on creation [Conformance]
+[It] should be consumable in multiple volumes in a pod [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-STEP: Given a Pod with a 'name' label pod-adoption is created
-STEP: When a replication controller with a matching selector is created
-STEP: Then the orphan pod is adopted
-[AfterEach] [sig-apps] ReplicationController
+STEP: Creating secret with name secret-test-e3808fac-91cb-11e9-bce2-ae54e022189f
+STEP: Creating a pod to test consume secrets
+Jun 18 13:20:49.312: INFO: Waiting up to 5m0s for pod "pod-secrets-e38285da-91cb-11e9-bce2-ae54e022189f" in namespace "e2e-tests-secrets-5fwtb" to be "success or failure"
+Jun 18 13:20:49.326: INFO: Pod "pod-secrets-e38285da-91cb-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 13.972286ms
+Jun 18 13:20:51.342: INFO: Pod "pod-secrets-e38285da-91cb-11e9-bce2-ae54e022189f": Phase="Running", Reason="", readiness=true. Elapsed: 2.029674669s
+Jun 18 13:20:53.357: INFO: Pod "pod-secrets-e38285da-91cb-11e9-bce2-ae54e022189f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.044894481s
+STEP: Saw pod success
+Jun 18 13:20:53.358: INFO: Pod "pod-secrets-e38285da-91cb-11e9-bce2-ae54e022189f" satisfied condition "success or failure"
+Jun 18 13:20:53.372: INFO: Trying to get logs from node 10.72.74.144 pod pod-secrets-e38285da-91cb-11e9-bce2-ae54e022189f container secret-volume-test: 
+STEP: delete the pod
+Jun 18 13:20:53.484: INFO: Waiting for pod pod-secrets-e38285da-91cb-11e9-bce2-ae54e022189f to disappear
+Jun 18 13:20:53.499: INFO: Pod pod-secrets-e38285da-91cb-11e9-bce2-ae54e022189f no longer exists
+[AfterEach] [sig-storage] Secrets
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
-May 13 20:18:10.345: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "e2e-tests-replication-controller-8vtxs" for this suite.
-May 13 20:18:32.381: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-May 13 20:18:32.475: INFO: namespace: e2e-tests-replication-controller-8vtxs, resource: bindings, ignored listing per whitelist
-May 13 20:18:32.678: INFO: namespace e2e-tests-replication-controller-8vtxs deletion completed in 22.324904654s
+Jun 18 13:20:53.499: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-tests-secrets-5fwtb" for this suite.
+Jun 18 13:21:01.378: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 13:21:03.070: INFO: namespace: e2e-tests-secrets-5fwtb, resource: bindings, ignored listing per whitelist
+Jun 18 13:21:03.606: INFO: namespace e2e-tests-secrets-5fwtb deletion completed in 10.078978876s
 
-• [SLOW TEST:25.773 seconds]
-[sig-apps] ReplicationController
-/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22
-  should adopt matching pods on creation [Conformance]
+• [SLOW TEST:14.763 seconds]
+[sig-storage] Secrets
+/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/secrets_volume.go:34
+  should be consumable in multiple volumes in a pod [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 ------------------------------
 S
 ------------------------------
-[sig-storage] Downward API volume 
-  should set mode on item file [NodeConformance] [Conformance]
+[k8s.io] Docker Containers 
+  should use the image defaults if command and args are blank [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-[BeforeEach] [sig-storage] Downward API volume
+[BeforeEach] [k8s.io] Docker Containers
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
 STEP: Creating a kubernetes client
-May 13 20:18:32.678: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423
-STEP: Building a namespace api object, basename downward-api
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-downward-api-82g8b
+Jun 18 13:21:03.607: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
+STEP: Building a namespace api object, basename containers
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-containers-fqbcd
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-storage] Downward API volume
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:39
-[It] should set mode on item file [NodeConformance] [Conformance]
+[It] should use the image defaults if command and args are blank [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-STEP: Creating a pod to test downward API volume plugin
-May 13 20:18:33.038: INFO: Waiting up to 5m0s for pod "downwardapi-volume-47c9edf5-75bc-11e9-8f67-2632f168be36" in namespace "e2e-tests-downward-api-82g8b" to be "success or failure"
-May 13 20:18:33.046: INFO: Pod "downwardapi-volume-47c9edf5-75bc-11e9-8f67-2632f168be36": Phase="Pending", Reason="", readiness=false. Elapsed: 7.626196ms
-May 13 20:18:35.054: INFO: Pod "downwardapi-volume-47c9edf5-75bc-11e9-8f67-2632f168be36": Phase="Pending", Reason="", readiness=false. Elapsed: 2.015991812s
-May 13 20:18:37.063: INFO: Pod "downwardapi-volume-47c9edf5-75bc-11e9-8f67-2632f168be36": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.025137572s
+STEP: Creating a pod to test use defaults
+Jun 18 13:21:04.101: INFO: Waiting up to 5m0s for pod "client-containers-ec5301db-91cb-11e9-bce2-ae54e022189f" in namespace "e2e-tests-containers-fqbcd" to be "success or failure"
+Jun 18 13:21:04.116: INFO: Pod "client-containers-ec5301db-91cb-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 14.685696ms
+Jun 18 13:21:06.131: INFO: Pod "client-containers-ec5301db-91cb-11e9-bce2-ae54e022189f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.02947169s
 STEP: Saw pod success
-May 13 20:18:37.063: INFO: Pod "downwardapi-volume-47c9edf5-75bc-11e9-8f67-2632f168be36" satisfied condition "success or failure"
-May 13 20:18:37.071: INFO: Trying to get logs from node 10.170.219.153 pod downwardapi-volume-47c9edf5-75bc-11e9-8f67-2632f168be36 container client-container: 
+Jun 18 13:21:06.131: INFO: Pod "client-containers-ec5301db-91cb-11e9-bce2-ae54e022189f" satisfied condition "success or failure"
+Jun 18 13:21:06.146: INFO: Trying to get logs from node 10.72.74.144 pod client-containers-ec5301db-91cb-11e9-bce2-ae54e022189f container test-container: 
 STEP: delete the pod
-May 13 20:18:37.115: INFO: Waiting for pod downwardapi-volume-47c9edf5-75bc-11e9-8f67-2632f168be36 to disappear
-May 13 20:18:37.122: INFO: Pod downwardapi-volume-47c9edf5-75bc-11e9-8f67-2632f168be36 no longer exists
-[AfterEach] [sig-storage] Downward API volume
+Jun 18 13:21:06.222: INFO: Waiting for pod client-containers-ec5301db-91cb-11e9-bce2-ae54e022189f to disappear
+Jun 18 13:21:06.236: INFO: Pod client-containers-ec5301db-91cb-11e9-bce2-ae54e022189f no longer exists
+[AfterEach] [k8s.io] Docker Containers
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
-May 13 20:18:37.122: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "e2e-tests-downward-api-82g8b" for this suite.
-May 13 20:18:43.160: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-May 13 20:18:43.184: INFO: namespace: e2e-tests-downward-api-82g8b, resource: bindings, ignored listing per whitelist
-May 13 20:18:43.453: INFO: namespace e2e-tests-downward-api-82g8b deletion completed in 6.323095867s
+Jun 18 13:21:06.237: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-tests-containers-fqbcd" for this suite.
+Jun 18 13:21:14.316: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 13:21:14.530: INFO: namespace: e2e-tests-containers-fqbcd, resource: bindings, ignored listing per whitelist
+Jun 18 13:21:14.961: INFO: namespace e2e-tests-containers-fqbcd deletion completed in 8.696143974s
 
-• [SLOW TEST:10.775 seconds]
-[sig-storage] Downward API volume
-/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:34
-  should set mode on item file [NodeConformance] [Conformance]
+• [SLOW TEST:11.355 seconds]
+[k8s.io] Docker Containers
+/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694
+  should use the image defaults if command and args are blank [NodeConformance] [Conformance]
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
+------------------------------
+SSSSSSSS
+------------------------------
+[sig-apps] Daemon set [Serial] 
+  should rollback without unnecessary restarts [Conformance]
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
+[BeforeEach] [sig-apps] Daemon set [Serial]
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
+STEP: Creating a kubernetes client
+Jun 18 13:21:14.962: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
+STEP: Building a namespace api object, basename daemonsets
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-daemonsets-hmxxb
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [sig-apps] Daemon set [Serial]
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:102
+[It] should rollback without unnecessary restarts [Conformance]
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
+Jun 18 13:21:15.591: INFO: Requires at least 2 nodes (not -1)
+[AfterEach] [sig-apps] Daemon set [Serial]
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:68
+Jun 18 13:21:15.615: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"selfLink":"/apis/apps/v1/namespaces/e2e-tests-daemonsets-hmxxb/daemonsets","resourceVersion":"113465"},"items":null}
+
+Jun 18 13:21:15.629: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"selfLink":"/api/v1/namespaces/e2e-tests-daemonsets-hmxxb/pods","resourceVersion":"113465"},"items":null}
+
+[AfterEach] [sig-apps] Daemon set [Serial]
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
+Jun 18 13:21:15.687: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-tests-daemonsets-hmxxb" for this suite.
+Jun 18 13:21:21.772: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 13:21:22.128: INFO: namespace: e2e-tests-daemonsets-hmxxb, resource: bindings, ignored listing per whitelist
+Jun 18 13:21:22.293: INFO: namespace e2e-tests-daemonsets-hmxxb deletion completed in 6.587947695s
+
+S [SKIPPING] [7.331 seconds]
+[sig-apps] Daemon set [Serial]
+/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22
+  should rollback without unnecessary restarts [Conformance] [It]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
+
+  Jun 18 13:21:15.591: Requires at least 2 nodes (not -1)
+
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/util.go:292
 ------------------------------
 S
 ------------------------------
+[k8s.io] KubeletManagedEtcHosts 
+  should test kubelet managed /etc/hosts file [NodeConformance] [Conformance]
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
+[BeforeEach] [k8s.io] KubeletManagedEtcHosts
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
+STEP: Creating a kubernetes client
+Jun 18 13:21:22.295: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
+STEP: Building a namespace api object, basename e2e-kubelet-etc-hosts
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-e2e-kubelet-etc-hosts-rgcft
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should test kubelet managed /etc/hosts file [NodeConformance] [Conformance]
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
+STEP: Setting up the test
+STEP: Creating hostNetwork=false pod
+STEP: Creating hostNetwork=true pod
+STEP: Running the test
+STEP: Verifying /etc/hosts of container is kubelet-managed for pod with hostNetwork=false
+Jun 18 13:21:28.919: INFO: ExecWithOptions {Command:[cat /etc/hosts] Namespace:e2e-tests-e2e-kubelet-etc-hosts-rgcft PodName:test-pod ContainerName:busybox-1 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
+Jun 18 13:21:28.919: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
+Jun 18 13:21:29.159: INFO: Exec stderr: ""
+Jun 18 13:21:29.159: INFO: ExecWithOptions {Command:[cat /etc/hosts-original] Namespace:e2e-tests-e2e-kubelet-etc-hosts-rgcft PodName:test-pod ContainerName:busybox-1 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
+Jun 18 13:21:29.159: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
+Jun 18 13:21:29.415: INFO: Exec stderr: ""
+Jun 18 13:21:29.415: INFO: ExecWithOptions {Command:[cat /etc/hosts] Namespace:e2e-tests-e2e-kubelet-etc-hosts-rgcft PodName:test-pod ContainerName:busybox-2 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
+Jun 18 13:21:29.415: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
+Jun 18 13:21:29.684: INFO: Exec stderr: ""
+Jun 18 13:21:29.684: INFO: ExecWithOptions {Command:[cat /etc/hosts-original] Namespace:e2e-tests-e2e-kubelet-etc-hosts-rgcft PodName:test-pod ContainerName:busybox-2 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
+Jun 18 13:21:29.684: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
+Jun 18 13:21:30.524: INFO: Exec stderr: ""
+STEP: Verifying /etc/hosts of container is not kubelet-managed since container specifies /etc/hosts mount
+Jun 18 13:21:30.524: INFO: ExecWithOptions {Command:[cat /etc/hosts] Namespace:e2e-tests-e2e-kubelet-etc-hosts-rgcft PodName:test-pod ContainerName:busybox-3 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
+Jun 18 13:21:30.524: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
+Jun 18 13:21:30.750: INFO: Exec stderr: ""
+Jun 18 13:21:30.750: INFO: ExecWithOptions {Command:[cat /etc/hosts-original] Namespace:e2e-tests-e2e-kubelet-etc-hosts-rgcft PodName:test-pod ContainerName:busybox-3 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
+Jun 18 13:21:30.750: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
+Jun 18 13:21:31.002: INFO: Exec stderr: ""
+STEP: Verifying /etc/hosts content of container is not kubelet-managed for pod with hostNetwork=true
+Jun 18 13:21:31.002: INFO: ExecWithOptions {Command:[cat /etc/hosts] Namespace:e2e-tests-e2e-kubelet-etc-hosts-rgcft PodName:test-host-network-pod ContainerName:busybox-1 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
+Jun 18 13:21:31.003: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
+Jun 18 13:21:31.393: INFO: Exec stderr: ""
+Jun 18 13:21:31.394: INFO: ExecWithOptions {Command:[cat /etc/hosts-original] Namespace:e2e-tests-e2e-kubelet-etc-hosts-rgcft PodName:test-host-network-pod ContainerName:busybox-1 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
+Jun 18 13:21:31.394: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
+Jun 18 13:21:31.767: INFO: Exec stderr: ""
+Jun 18 13:21:31.767: INFO: ExecWithOptions {Command:[cat /etc/hosts] Namespace:e2e-tests-e2e-kubelet-etc-hosts-rgcft PodName:test-host-network-pod ContainerName:busybox-2 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
+Jun 18 13:21:31.767: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
+Jun 18 13:21:32.001: INFO: Exec stderr: ""
+Jun 18 13:21:32.001: INFO: ExecWithOptions {Command:[cat /etc/hosts-original] Namespace:e2e-tests-e2e-kubelet-etc-hosts-rgcft PodName:test-host-network-pod ContainerName:busybox-2 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
+Jun 18 13:21:32.001: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
+Jun 18 13:21:32.413: INFO: Exec stderr: ""
+[AfterEach] [k8s.io] KubeletManagedEtcHosts
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
+Jun 18 13:21:32.413: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-tests-e2e-kubelet-etc-hosts-rgcft" for this suite.
+Jun 18 13:22:24.506: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 13:22:24.899: INFO: namespace: e2e-tests-e2e-kubelet-etc-hosts-rgcft, resource: bindings, ignored listing per whitelist
+Jun 18 13:22:25.063: INFO: namespace e2e-tests-e2e-kubelet-etc-hosts-rgcft deletion completed in 52.609665418s
+
+• [SLOW TEST:62.769 seconds]
+[k8s.io] KubeletManagedEtcHosts
+/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694
+  should test kubelet managed /etc/hosts file [NodeConformance] [Conformance]
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
+------------------------------
+SSSSSSSSSSS
+------------------------------
 [sig-storage] Projected downwardAPI 
-  should provide container's memory limit [NodeConformance] [Conformance]
+  should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 [BeforeEach] [sig-storage] Projected downwardAPI
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
 STEP: Creating a kubernetes client
-May 13 20:18:43.453: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423
+Jun 18 13:22:25.064: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
 STEP: Building a namespace api object, basename projected
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-projected-jq2l8
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-projected-f8792
 STEP: Waiting for a default service account to be provisioned in namespace
 [BeforeEach] [sig-storage] Projected downwardAPI
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:39
-[It] should provide container's memory limit [NodeConformance] [Conformance]
+[It] should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 STEP: Creating a pod to test downward API volume plugin
-May 13 20:18:43.838: INFO: Waiting up to 5m0s for pod "downwardapi-volume-4e398d27-75bc-11e9-8f67-2632f168be36" in namespace "e2e-tests-projected-jq2l8" to be "success or failure"
-May 13 20:18:43.847: INFO: Pod "downwardapi-volume-4e398d27-75bc-11e9-8f67-2632f168be36": Phase="Pending", Reason="", readiness=false. Elapsed: 8.696323ms
-May 13 20:18:46.304: INFO: Pod "downwardapi-volume-4e398d27-75bc-11e9-8f67-2632f168be36": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.4652717s
+Jun 18 13:22:25.594: INFO: Waiting up to 5m0s for pod "downwardapi-volume-1ce5516b-91cc-11e9-bce2-ae54e022189f" in namespace "e2e-tests-projected-f8792" to be "success or failure"
+Jun 18 13:22:25.609: INFO: Pod "downwardapi-volume-1ce5516b-91cc-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 15.092294ms
+Jun 18 13:22:27.625: INFO: Pod "downwardapi-volume-1ce5516b-91cc-11e9-bce2-ae54e022189f": Phase="Running", Reason="", readiness=true. Elapsed: 2.031110444s
+Jun 18 13:22:29.659: INFO: Pod "downwardapi-volume-1ce5516b-91cc-11e9-bce2-ae54e022189f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.064607827s
 STEP: Saw pod success
-May 13 20:18:46.304: INFO: Pod "downwardapi-volume-4e398d27-75bc-11e9-8f67-2632f168be36" satisfied condition "success or failure"
-May 13 20:18:46.312: INFO: Trying to get logs from node 10.170.219.153 pod downwardapi-volume-4e398d27-75bc-11e9-8f67-2632f168be36 container client-container: 
+Jun 18 13:22:29.659: INFO: Pod "downwardapi-volume-1ce5516b-91cc-11e9-bce2-ae54e022189f" satisfied condition "success or failure"
+Jun 18 13:22:29.672: INFO: Trying to get logs from node 10.72.74.149 pod downwardapi-volume-1ce5516b-91cc-11e9-bce2-ae54e022189f container client-container: 
 STEP: delete the pod
-May 13 20:18:46.353: INFO: Waiting for pod downwardapi-volume-4e398d27-75bc-11e9-8f67-2632f168be36 to disappear
-May 13 20:18:46.360: INFO: Pod downwardapi-volume-4e398d27-75bc-11e9-8f67-2632f168be36 no longer exists
+Jun 18 13:22:29.784: INFO: Waiting for pod downwardapi-volume-1ce5516b-91cc-11e9-bce2-ae54e022189f to disappear
+Jun 18 13:22:29.802: INFO: Pod downwardapi-volume-1ce5516b-91cc-11e9-bce2-ae54e022189f no longer exists
 [AfterEach] [sig-storage] Projected downwardAPI
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
-May 13 20:18:46.360: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "e2e-tests-projected-jq2l8" for this suite.
-May 13 20:18:52.408: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-May 13 20:18:52.454: INFO: namespace: e2e-tests-projected-jq2l8, resource: bindings, ignored listing per whitelist
-May 13 20:18:52.673: INFO: namespace e2e-tests-projected-jq2l8 deletion completed in 6.305204466s
+Jun 18 13:22:29.802: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-tests-projected-f8792" for this suite.
+Jun 18 13:22:37.883: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 13:22:38.113: INFO: namespace: e2e-tests-projected-f8792, resource: bindings, ignored listing per whitelist
+Jun 18 13:22:38.374: INFO: namespace e2e-tests-projected-f8792 deletion completed in 8.54847721s
 
-• [SLOW TEST:9.220 seconds]
+• [SLOW TEST:13.311 seconds]
 [sig-storage] Projected downwardAPI
 /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:33
-  should provide container's memory limit [NodeConformance] [Conformance]
+  should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 ------------------------------
-S
+SSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-cli] Kubectl client [k8s.io] Kubectl run --rm job 
-  should create a job from an image, then delete the job  [Conformance]
+[k8s.io] Docker Containers 
+  should be able to override the image's default arguments (docker cmd) [NodeConformance] [Conformance]
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
+[BeforeEach] [k8s.io] Docker Containers
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
+STEP: Creating a kubernetes client
+Jun 18 13:22:38.375: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
+STEP: Building a namespace api object, basename containers
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-containers-rvn8n
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should be able to override the image's default arguments (docker cmd) [NodeConformance] [Conformance]
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
+STEP: Creating a pod to test override arguments
+Jun 18 13:22:38.851: INFO: Waiting up to 5m0s for pod "client-containers-24cc6bd0-91cc-11e9-bce2-ae54e022189f" in namespace "e2e-tests-containers-rvn8n" to be "success or failure"
+Jun 18 13:22:38.873: INFO: Pod "client-containers-24cc6bd0-91cc-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 21.868066ms
+Jun 18 13:22:40.910: INFO: Pod "client-containers-24cc6bd0-91cc-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 2.058221964s
+Jun 18 13:22:42.924: INFO: Pod "client-containers-24cc6bd0-91cc-11e9-bce2-ae54e022189f": Phase="Pending", Reason="", readiness=false. Elapsed: 4.073081425s
+Jun 18 13:22:44.939: INFO: Pod "client-containers-24cc6bd0-91cc-11e9-bce2-ae54e022189f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.087919552s
+STEP: Saw pod success
+Jun 18 13:22:44.939: INFO: Pod "client-containers-24cc6bd0-91cc-11e9-bce2-ae54e022189f" satisfied condition "success or failure"
+Jun 18 13:22:44.954: INFO: Trying to get logs from node 10.72.74.143 pod client-containers-24cc6bd0-91cc-11e9-bce2-ae54e022189f container test-container: 
+STEP: delete the pod
+Jun 18 13:22:45.024: INFO: Waiting for pod client-containers-24cc6bd0-91cc-11e9-bce2-ae54e022189f to disappear
+Jun 18 13:22:45.038: INFO: Pod client-containers-24cc6bd0-91cc-11e9-bce2-ae54e022189f no longer exists
+[AfterEach] [k8s.io] Docker Containers
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
+Jun 18 13:22:45.039: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-tests-containers-rvn8n" for this suite.
+Jun 18 13:22:51.145: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 13:22:51.520: INFO: namespace: e2e-tests-containers-rvn8n, resource: bindings, ignored listing per whitelist
+Jun 18 13:22:51.827: INFO: namespace e2e-tests-containers-rvn8n deletion completed in 6.764418146s
+
+• [SLOW TEST:13.452 seconds]
+[k8s.io] Docker Containers
+/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694
+  should be able to override the image's default arguments (docker cmd) [NodeConformance] [Conformance]
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
+------------------------------
+[sig-cli] Kubectl client [k8s.io] Kubectl run rc 
+  should create an rc from an image  [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 [BeforeEach] [sig-cli] Kubectl client
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
 STEP: Creating a kubernetes client
-May 13 20:18:52.673: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423
+Jun 18 13:22:51.827: INFO: >>> kubeConfig: /tmp/kubeconfig-953583206
 STEP: Building a namespace api object, basename kubectl
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-kubectl-t8x9k
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-kubectl-rmjwh
 STEP: Waiting for a default service account to be provisioned in namespace
 [BeforeEach] [sig-cli] Kubectl client
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:243
-[It] should create a job from an image, then delete the job  [Conformance]
+[BeforeEach] [k8s.io] Kubectl run rc
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1298
+[It] should create an rc from an image  [Conformance]
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-STEP: executing a command with run --rm and attach with stdin
-May 13 20:18:52.944: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-555138423 --namespace=e2e-tests-kubectl-t8x9k run e2e-test-rm-busybox-job --image=docker.io/library/busybox:1.29 --rm=true --generator=job/v1 --restart=OnFailure --attach=true --stdin -- sh -c cat && echo 'stdin closed''
-May 13 20:18:54.919: INFO: stderr: "kubectl run --generator=job/v1 is DEPRECATED and will be removed in a future version. Use kubectl run --generator=run-pod/v1 or kubectl create instead.\nIf you don't see a command prompt, try pressing enter.\n"
-May 13 20:18:54.919: INFO: stdout: "abcd1234stdin closed\njob.batch \"e2e-test-rm-busybox-job\" deleted\n"
-STEP: verifying the job e2e-test-rm-busybox-job was deleted
+STEP: running the image docker.io/library/nginx:1.14-alpine
+Jun 18 13:22:52.389: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 run e2e-test-nginx-rc --image=docker.io/library/nginx:1.14-alpine --generator=run/v1 --namespace=e2e-tests-kubectl-rmjwh'
+Jun 18 13:22:52.781: INFO: stderr: "kubectl run --generator=run/v1 is DEPRECATED and will be removed in a future version. Use kubectl run --generator=run-pod/v1 or kubectl create instead.\n"
+Jun 18 13:22:52.781: INFO: stdout: "replicationcontroller/e2e-test-nginx-rc created\n"
+STEP: verifying the rc e2e-test-nginx-rc was created
+STEP: verifying the pod controlled by rc e2e-test-nginx-rc was created
+STEP: confirm that you can get logs from an rc
+Jun 18 13:22:54.823: INFO: Waiting up to 5m0s for 1 pods to be running and ready: [e2e-test-nginx-rc-rx8x7]
+Jun 18 13:22:54.823: INFO: Waiting up to 5m0s for pod "e2e-test-nginx-rc-rx8x7" in namespace "e2e-tests-kubectl-rmjwh" to be "running and ready"
+Jun 18 13:22:54.899: INFO: Pod "e2e-test-nginx-rc-rx8x7": Phase="Pending", Reason="", readiness=false. Elapsed: 75.941894ms
+Jun 18 13:22:56.914: INFO: Pod "e2e-test-nginx-rc-rx8x7": Phase="Running", Reason="", readiness=true. Elapsed: 2.091159932s
+Jun 18 13:22:56.914: INFO: Pod "e2e-test-nginx-rc-rx8x7" satisfied condition "running and ready"
+Jun 18 13:22:56.914: INFO: Wanted all 1 pods to be running and ready. Result: true. Pods: [e2e-test-nginx-rc-rx8x7]
+Jun 18 13:22:56.914: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 logs rc/e2e-test-nginx-rc --namespace=e2e-tests-kubectl-rmjwh'
+Jun 18 13:22:57.119: INFO: stderr: ""
+Jun 18 13:22:57.119: INFO: stdout: ""
+[AfterEach] [k8s.io] Kubectl run rc
+  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1303
+Jun 18 13:22:57.119: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-953583206 delete rc e2e-test-nginx-rc --namespace=e2e-tests-kubectl-rmjwh'
+Jun 18 13:22:57.288: INFO: stderr: ""
+Jun 18 13:22:57.288: INFO: stdout: "replicationcontroller \"e2e-test-nginx-rc\" deleted\n"
 [AfterEach] [sig-cli] Kubectl client
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
-May 13 20:18:56.929: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "e2e-tests-kubectl-t8x9k" for this suite.
-May 13 20:19:02.977: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-May 13 20:19:03.163: INFO: namespace: e2e-tests-kubectl-t8x9k, resource: bindings, ignored listing per whitelist
-May 13 20:19:03.266: INFO: namespace e2e-tests-kubectl-t8x9k deletion completed in 6.329119229s
+Jun 18 13:22:57.288: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-tests-kubectl-rmjwh" for this suite.
+Jun 18 13:23:03.359: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 13:23:03.448: INFO: namespace: e2e-tests-kubectl-rmjwh, resource: bindings, ignored listing per whitelist
+Jun 18 13:23:03.868: INFO: namespace e2e-tests-kubectl-rmjwh deletion completed in 6.561414796s
 
-• [SLOW TEST:10.593 seconds]
+• [SLOW TEST:12.042 seconds]
 [sig-cli] Kubectl client
 /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:22
-  [k8s.io] Kubectl run --rm job
+  [k8s.io] Kubectl run rc
   /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694
-    should create a job from an image, then delete the job  [Conformance]
+    should create an rc from an image  [Conformance]
     /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
 ------------------------------
-[sig-scheduling] SchedulerPredicates [Serial] 
-  validates resource limits of pods that are allowed to run  [Conformance]
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-[BeforeEach] [sig-scheduling] SchedulerPredicates [Serial]
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
-STEP: Creating a kubernetes client
-May 13 20:19:03.267: INFO: >>> kubeConfig: /tmp/kubeconfig-555138423
-STEP: Building a namespace api object, basename sched-pred
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-sched-pred-5t98x
-STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-scheduling] SchedulerPredicates [Serial]
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:79
-May 13 20:19:03.543: INFO: Waiting up to 1m0s for all (but 0) nodes to be ready
-May 13 20:19:03.556: INFO: Waiting for terminating namespaces to be deleted...
-May 13 20:19:03.564: INFO: 
-Logging pods the kubelet thinks is on node 10.170.219.151 before test
-May 13 20:19:03.589: INFO: calico-kube-controllers-7dd978d898-r8w74 from kube-system started at 2019-05-13 16:27:35 +0000 UTC (1 container statuses recorded)
-May 13 20:19:03.589: INFO: 	Container calico-kube-controllers ready: true, restart count 0
-May 13 20:19:03.589: INFO: public-cr0ef1a5f1e33c44fe9c21650356d908ee-alb1-8f59cf484-9v5pm from kube-system started at 2019-05-13 16:35:03 +0000 UTC (4 container statuses recorded)
-May 13 20:19:03.589: INFO: 	Container ingress-auth-1 ready: true, restart count 0
-May 13 20:19:03.589: INFO: 	Container ingress-auth-2 ready: true, restart count 0
-May 13 20:19:03.589: INFO: 	Container ingress-auth-3 ready: true, restart count 0
-May 13 20:19:03.589: INFO: 	Container nginx-ingress ready: true, restart count 0
-May 13 20:19:03.589: INFO: ibm-master-proxy-static-10.170.219.151 from kube-system started at  (0 container statuses recorded)
-May 13 20:19:03.589: INFO: vpn-774cf5c6d4-6p6cr from kube-system started at 2019-05-13 16:27:35 +0000 UTC (1 container statuses recorded)
-May 13 20:19:03.589: INFO: 	Container vpn ready: true, restart count 0
-May 13 20:19:03.589: INFO: ibm-keepalived-watcher-8wshf from kube-system started at 2019-05-13 16:27:05 +0000 UTC (1 container statuses recorded)
-May 13 20:19:03.589: INFO: 	Container keepalived-watcher ready: true, restart count 0
-May 13 20:19:03.589: INFO: coredns-58d696879-rrkth from kube-system started at 2019-05-13 16:27:35 +0000 UTC (1 container statuses recorded)
-May 13 20:19:03.589: INFO: 	Container coredns ready: true, restart count 0
-May 13 20:19:03.589: INFO: ibm-file-plugin-7f6d8979bd-p5smn from kube-system started at 2019-05-13 16:27:35 +0000 UTC (1 container statuses recorded)
-May 13 20:19:03.589: INFO: 	Container ibm-file-plugin-container ready: true, restart count 0
-May 13 20:19:03.589: INFO: coredns-58d696879-zmj2g from kube-system started at 2019-05-13 16:27:49 +0000 UTC (1 container statuses recorded)
-May 13 20:19:03.589: INFO: 	Container coredns ready: true, restart count 0
-May 13 20:19:03.589: INFO: calico-node-d9d6s from kube-system started at 2019-05-13 16:27:05 +0000 UTC (1 container statuses recorded)
-May 13 20:19:03.589: INFO: 	Container calico-node ready: true, restart count 0
-May 13 20:19:03.589: INFO: ibm-kube-fluentd-6fw9c from kube-system started at 2019-05-13 16:32:50 +0000 UTC (1 container statuses recorded)
-May 13 20:19:03.589: INFO: 	Container fluentd ready: true, restart count 0
-May 13 20:19:03.589: INFO: kubernetes-dashboard-7996b848f4-mzbwj from kube-system started at 2019-05-13 16:27:35 +0000 UTC (1 container statuses recorded)
-May 13 20:19:03.589: INFO: 	Container kubernetes-dashboard ready: true, restart count 0
-May 13 20:19:03.589: INFO: coredns-autoscaler-64f9c5b4df-hzq57 from kube-system started at 2019-05-13 16:27:35 +0000 UTC (1 container statuses recorded)
-May 13 20:19:03.589: INFO: 	Container autoscaler ready: true, restart count 0
-May 13 20:19:03.589: INFO: sonobuoy-systemd-logs-daemon-set-8e5ce9bdf4a94ece-nlv42 from heptio-sonobuoy started at 2019-05-13 18:40:58 +0000 UTC (2 container statuses recorded)
-May 13 20:19:03.589: INFO: 	Container sonobuoy-worker ready: true, restart count 1
-May 13 20:19:03.589: INFO: 	Container systemd-logs ready: true, restart count 1
-May 13 20:19:03.589: INFO: ibm-storage-watcher-845946d5b5-knxcm from kube-system started at 2019-05-13 16:27:35 +0000 UTC (1 container statuses recorded)
-May 13 20:19:03.589: INFO: 	Container ibm-storage-watcher-container ready: true, restart count 0
-May 13 20:19:03.589: INFO: 
-Logging pods the kubelet thinks is on node 10.170.219.153 before test
-May 13 20:19:03.610: INFO: metrics-server-58dd96c6bd-sggps from kube-system started at 2019-05-13 16:28:04 +0000 UTC (2 container statuses recorded)
-May 13 20:19:03.610: INFO: 	Container metrics-server ready: true, restart count 0
-May 13 20:19:03.610: INFO: 	Container metrics-server-nanny ready: true, restart count 0
-May 13 20:19:03.610: INFO: test-k8s-e2e-pvg-master-verification from default started at 2019-05-13 18:40:52 +0000 UTC (1 container statuses recorded)
-May 13 20:19:03.610: INFO: 	Container test-k8s-e2e-pvg-master-verification ready: true, restart count 0
-May 13 20:19:03.610: INFO: sonobuoy-systemd-logs-daemon-set-8e5ce9bdf4a94ece-tgp2l from heptio-sonobuoy started at 2019-05-13 18:40:58 +0000 UTC (2 container statuses recorded)
-May 13 20:19:03.610: INFO: 	Container sonobuoy-worker ready: true, restart count 1
-May 13 20:19:03.610: INFO: 	Container systemd-logs ready: true, restart count 1
-May 13 20:19:03.610: INFO: ibm-master-proxy-static-10.170.219.153 from kube-system started at  (0 container statuses recorded)
-May 13 20:19:03.610: INFO: ibm-keepalived-watcher-cwtfg from kube-system started at 2019-05-13 16:27:29 +0000 UTC (1 container statuses recorded)
-May 13 20:19:03.610: INFO: 	Container keepalived-watcher ready: true, restart count 0
-May 13 20:19:03.610: INFO: calico-node-phw8q from kube-system started at 2019-05-13 16:27:29 +0000 UTC (1 container statuses recorded)
-May 13 20:19:03.610: INFO: 	Container calico-node ready: true, restart count 0
-May 13 20:19:03.610: INFO: ibm-cloud-provider-ip-169-45-218-130-66c489dcb-z58z4 from ibm-system started at 2019-05-13 16:30:08 +0000 UTC (1 container statuses recorded)
-May 13 20:19:03.610: INFO: 	Container ibm-cloud-provider-ip-169-45-218-130 ready: true, restart count 0
-May 13 20:19:03.610: INFO: ibm-kube-fluentd-cdmr5 from kube-system started at 2019-05-13 16:32:50 +0000 UTC (1 container statuses recorded)
-May 13 20:19:03.610: INFO: 	Container fluentd ready: true, restart count 0
-May 13 20:19:03.610: INFO: sonobuoy from heptio-sonobuoy started at 2019-05-13 18:40:57 +0000 UTC (1 container statuses recorded)
-May 13 20:19:03.610: INFO: 	Container kube-sonobuoy ready: true, restart count 0
-May 13 20:19:03.610: INFO: 
-Logging pods the kubelet thinks is on node 10.170.219.177 before test
-May 13 20:19:03.628: INFO: sonobuoy-systemd-logs-daemon-set-8e5ce9bdf4a94ece-kjc69 from heptio-sonobuoy started at 2019-05-13 18:40:58 +0000 UTC (2 container statuses recorded)
-May 13 20:19:03.629: INFO: 	Container sonobuoy-worker ready: true, restart count 1
-May 13 20:19:03.629: INFO: 	Container systemd-logs ready: true, restart count 1
-May 13 20:19:03.629: INFO: ibm-master-proxy-static-10.170.219.177 from kube-system started at  (0 container statuses recorded)
-May 13 20:19:03.629: INFO: ibm-keepalived-watcher-qszvz from kube-system started at 2019-05-13 16:27:34 +0000 UTC (1 container statuses recorded)
-May 13 20:19:03.629: INFO: 	Container keepalived-watcher ready: true, restart count 0
-May 13 20:19:03.629: INFO: calico-node-twx7l from kube-system started at 2019-05-13 16:27:34 +0000 UTC (1 container statuses recorded)
-May 13 20:19:03.629: INFO: 	Container calico-node ready: true, restart count 0
-May 13 20:19:03.629: INFO: ibm-cloud-provider-ip-169-45-218-130-66c489dcb-kd74s from ibm-system started at 2019-05-13 16:30:08 +0000 UTC (1 container statuses recorded)
-May 13 20:19:03.629: INFO: 	Container ibm-cloud-provider-ip-169-45-218-130 ready: true, restart count 0
-May 13 20:19:03.629: INFO: public-cr0ef1a5f1e33c44fe9c21650356d908ee-alb1-8f59cf484-7r5vq from kube-system started at 2019-05-13 16:35:03 +0000 UTC (4 container statuses recorded)
-May 13 20:19:03.629: INFO: 	Container ingress-auth-1 ready: true, restart count 0
-May 13 20:19:03.629: INFO: 	Container ingress-auth-2 ready: true, restart count 0
-May 13 20:19:03.629: INFO: 	Container ingress-auth-3 ready: true, restart count 0
-May 13 20:19:03.629: INFO: 	Container nginx-ingress ready: true, restart count 0
-May 13 20:19:03.629: INFO: ibm-kube-fluentd-zzl5d from kube-system started at 2019-05-13 16:32:50 +0000 UTC (1 container statuses recorded)
-May 13 20:19:03.629: INFO: 	Container fluentd ready: true, restart count 0
-May 13 20:19:03.629: INFO: sonobuoy-e2e-job-afa7e83446444f4e from heptio-sonobuoy started at 2019-05-13 18:40:58 +0000 UTC (2 container statuses recorded)
-May 13 20:19:03.629: INFO: 	Container e2e ready: true, restart count 0
-May 13 20:19:03.629: INFO: 	Container sonobuoy-worker ready: true, restart count 0
-[It] validates resource limits of pods that are allowed to run  [Conformance]
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-STEP: verifying the node has the label node 10.170.219.151
-STEP: verifying the node has the label node 10.170.219.153
-STEP: verifying the node has the label node 10.170.219.177
-May 13 20:19:03.706: INFO: Pod test-k8s-e2e-pvg-master-verification requesting resource cpu=0m on Node 10.170.219.153
-May 13 20:19:03.706: INFO: Pod sonobuoy requesting resource cpu=0m on Node 10.170.219.153
-May 13 20:19:03.706: INFO: Pod sonobuoy-e2e-job-afa7e83446444f4e requesting resource cpu=0m on Node 10.170.219.177
-May 13 20:19:03.706: INFO: Pod sonobuoy-systemd-logs-daemon-set-8e5ce9bdf4a94ece-kjc69 requesting resource cpu=0m on Node 10.170.219.177
-May 13 20:19:03.706: INFO: Pod sonobuoy-systemd-logs-daemon-set-8e5ce9bdf4a94ece-nlv42 requesting resource cpu=0m on Node 10.170.219.151
-May 13 20:19:03.706: INFO: Pod sonobuoy-systemd-logs-daemon-set-8e5ce9bdf4a94ece-tgp2l requesting resource cpu=0m on Node 10.170.219.153
-May 13 20:19:03.706: INFO: Pod ibm-cloud-provider-ip-169-45-218-130-66c489dcb-kd74s requesting resource cpu=5m on Node 10.170.219.177
-May 13 20:19:03.706: INFO: Pod ibm-cloud-provider-ip-169-45-218-130-66c489dcb-z58z4 requesting resource cpu=5m on Node 10.170.219.153
-May 13 20:19:03.706: INFO: Pod calico-kube-controllers-7dd978d898-r8w74 requesting resource cpu=10m on Node 10.170.219.151
-May 13 20:19:03.706: INFO: Pod calico-node-d9d6s requesting resource cpu=250m on Node 10.170.219.151
-May 13 20:19:03.706: INFO: Pod calico-node-phw8q requesting resource cpu=250m on Node 10.170.219.153
-May 13 20:19:03.706: INFO: Pod calico-node-twx7l requesting resource cpu=250m on Node 10.170.219.177
-May 13 20:19:03.706: INFO: Pod coredns-58d696879-rrkth requesting resource cpu=100m on Node 10.170.219.151
-May 13 20:19:03.706: INFO: Pod coredns-58d696879-zmj2g requesting resource cpu=100m on Node 10.170.219.151
-May 13 20:19:03.706: INFO: Pod coredns-autoscaler-64f9c5b4df-hzq57 requesting resource cpu=20m on Node 10.170.219.151
-May 13 20:19:03.706: INFO: Pod ibm-file-plugin-7f6d8979bd-p5smn requesting resource cpu=50m on Node 10.170.219.151
-May 13 20:19:03.706: INFO: Pod ibm-keepalived-watcher-8wshf requesting resource cpu=5m on Node 10.170.219.151
-May 13 20:19:03.706: INFO: Pod ibm-keepalived-watcher-cwtfg requesting resource cpu=5m on Node 10.170.219.153
-May 13 20:19:03.706: INFO: Pod ibm-keepalived-watcher-qszvz requesting resource cpu=5m on Node 10.170.219.177
-May 13 20:19:03.706: INFO: Pod ibm-kube-fluentd-6fw9c requesting resource cpu=25m on Node 10.170.219.151
-May 13 20:19:03.706: INFO: Pod ibm-kube-fluentd-cdmr5 requesting resource cpu=25m on Node 10.170.219.153
-May 13 20:19:03.707: INFO: Pod ibm-kube-fluentd-zzl5d requesting resource cpu=25m on Node 10.170.219.177
-May 13 20:19:03.707: INFO: Pod ibm-master-proxy-static-10.170.219.151 requesting resource cpu=25m on Node 10.170.219.151
-May 13 20:19:03.707: INFO: Pod ibm-master-proxy-static-10.170.219.153 requesting resource cpu=25m on Node 10.170.219.153
-May 13 20:19:03.707: INFO: Pod ibm-master-proxy-static-10.170.219.177 requesting resource cpu=25m on Node 10.170.219.177
-May 13 20:19:03.707: INFO: Pod ibm-storage-watcher-845946d5b5-knxcm requesting resource cpu=50m on Node 10.170.219.151
-May 13 20:19:03.707: INFO: Pod kubernetes-dashboard-7996b848f4-mzbwj requesting resource cpu=50m on Node 10.170.219.151
-May 13 20:19:03.707: INFO: Pod metrics-server-58dd96c6bd-sggps requesting resource cpu=53m on Node 10.170.219.153
-May 13 20:19:03.707: INFO: Pod public-cr0ef1a5f1e33c44fe9c21650356d908ee-alb1-8f59cf484-7r5vq requesting resource cpu=0m on Node 10.170.219.177
-May 13 20:19:03.707: INFO: Pod public-cr0ef1a5f1e33c44fe9c21650356d908ee-alb1-8f59cf484-9v5pm requesting resource cpu=0m on Node 10.170.219.151
-May 13 20:19:03.707: INFO: Pod vpn-774cf5c6d4-6p6cr requesting resource cpu=5m on Node 10.170.219.151
-STEP: Starting Pods to consume most of the cluster CPU.
-STEP: Creating another pod that requires unavailable amount of CPU.
-STEP: Considering event: 
-Type = [Normal], Name = [filler-pod-5a14051f-75bc-11e9-8f67-2632f168be36.159e578392ca77bb], Reason = [Scheduled], Message = [Successfully assigned e2e-tests-sched-pred-5t98x/filler-pod-5a14051f-75bc-11e9-8f67-2632f168be36 to 10.170.219.153]
-STEP: Considering event: 
-Type = [Normal], Name = [filler-pod-5a14051f-75bc-11e9-8f67-2632f168be36.159e5783ced9aafa], Reason = [Pulled], Message = [Container image "k8s.gcr.io/pause:3.1" already present on machine]
-STEP: Considering event: 
-Type = [Normal], Name = [filler-pod-5a14051f-75bc-11e9-8f67-2632f168be36.159e5783d1803979], Reason = [Created], Message = [Created container]
-STEP: Considering event: 
-Type = [Normal], Name = [filler-pod-5a14051f-75bc-11e9-8f67-2632f168be36.159e5783d9b6c056], Reason = [Started], Message = [Started container]
-STEP: Considering event: 
-Type = [Normal], Name = [filler-pod-5a16ac72-75bc-11e9-8f67-2632f168be36.159e578393433592], Reason = [Scheduled], Message = [Successfully assigned e2e-tests-sched-pred-5t98x/filler-pod-5a16ac72-75bc-11e9-8f67-2632f168be36 to 10.170.219.177]
-STEP: Considering event: 
-Type = [Normal], Name = [filler-pod-5a16ac72-75bc-11e9-8f67-2632f168be36.159e5783cdadfbd7], Reason = [Pulled], Message = [Container image "k8s.gcr.io/pause:3.1" already present on machine]
-STEP: Considering event: 
-Type = [Normal], Name = [filler-pod-5a16ac72-75bc-11e9-8f67-2632f168be36.159e5783d0fd8abf], Reason = [Created], Message = [Created container]
-STEP: Considering event: 
-Type = [Normal], Name = [filler-pod-5a16ac72-75bc-11e9-8f67-2632f168be36.159e5783d9ae8f43], Reason = [Started], Message = [Started container]
-STEP: Considering event: 
-Type = [Normal], Name = [filler-pod-5a17ffab-75bc-11e9-8f67-2632f168be36.159e578393cd4319], Reason = [Scheduled], Message = [Successfully assigned e2e-tests-sched-pred-5t98x/filler-pod-5a17ffab-75bc-11e9-8f67-2632f168be36 to 10.170.219.151]
-STEP: Considering event: 
-Type = [Normal], Name = [filler-pod-5a17ffab-75bc-11e9-8f67-2632f168be36.159e5783cda0aca4], Reason = [Pulled], Message = [Container image "k8s.gcr.io/pause:3.1" already present on machine]
-STEP: Considering event: 
-Type = [Normal], Name = [filler-pod-5a17ffab-75bc-11e9-8f67-2632f168be36.159e5783d0fc9b99], Reason = [Created], Message = [Created container]
-STEP: Considering event: 
-Type = [Normal], Name = [filler-pod-5a17ffab-75bc-11e9-8f67-2632f168be36.159e5783d9a373f2], Reason = [Started], Message = [Started container]
-STEP: Considering event: 
-Type = [Warning], Name = [additional-pod.159e578485602a3d], Reason = [FailedScheduling], Message = [0/3 nodes are available: 3 Insufficient cpu.]
-STEP: removing the label node off the node 10.170.219.151
-STEP: verifying the node doesn't have the label node
-STEP: removing the label node off the node 10.170.219.153
-STEP: verifying the node doesn't have the label node
-STEP: removing the label node off the node 10.170.219.177
-STEP: verifying the node doesn't have the label node
-[AfterEach] [sig-scheduling] SchedulerPredicates [Serial]
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
-May 13 20:19:09.264: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "e2e-tests-sched-pred-5t98x" for this suite.
-May 13 20:19:15.302: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-May 13 20:19:15.542: INFO: namespace: e2e-tests-sched-pred-5t98x, resource: bindings, ignored listing per whitelist
-May 13 20:19:15.575: INFO: namespace e2e-tests-sched-pred-5t98x deletion completed in 6.303793619s
-[AfterEach] [sig-scheduling] SchedulerPredicates [Serial]
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:70
-
-• [SLOW TEST:12.309 seconds]
-[sig-scheduling] SchedulerPredicates [Serial]
-/workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/framework.go:22
-  validates resource limits of pods that are allowed to run  [Conformance]
-  /workspace/anago-v1.13.0-rc.2.1+ddf47ac13c1a94/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
-------------------------------
-SSSSSMay 13 20:19:15.575: INFO: Running AfterSuite actions on all nodes
-May 13 20:19:15.576: INFO: Running AfterSuite actions on node 1
-May 13 20:19:15.576: INFO: Skipping dumping logs from cluster
+SSSSSSSSSJun 18 13:23:03.869: INFO: Running AfterSuite actions on all nodes
+Jun 18 13:23:03.869: INFO: Running AfterSuite actions on node 1
+Jun 18 13:23:03.869: INFO: Skipping dumping logs from cluster
 
-Ran 200 of 1946 Specs in 5884.208 seconds
+Ran 200 of 1946 Specs in 6512.437 seconds
 SUCCESS! -- 200 Passed | 0 Failed | 0 Pending | 1746 Skipped PASS
 
-Ginkgo ran 1 suite in 1h38m4.94310774s
+Ginkgo ran 1 suite in 1h48m33.384556665s
 Test Suite Passed
diff --git a/v1.13/iks/junit_01.xml b/v1.13/iks/junit_01.xml
index 35a3945d49..c14e1231ad 100644
--- a/v1.13/iks/junit_01.xml
+++ b/v1.13/iks/junit_01.xml
@@ -1,5441 +1,5441 @@
 
-  
-      
-      
-          
-      
-      
+  
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
-      
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
+      
           
       
-      
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
-      
+      
           
       
-      
+      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
       
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
-      
+      
+      
           
       
-      
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
-      
-      
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
+      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
-      
+      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
       
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+          
+      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
+      
           
       
-      
+      
+      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
       
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
       
           
       
-      
-          
-      
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
+      
           
       
-      
+      
           
       
-      
-      
+      
+      
           
       
-      
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
-      
+      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
+      
+      
       
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
-      
+      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
+      
           
       
-      
+      
+          
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
+      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
       
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
+      
+      
           
       
-      
-      
+      
           
       
-      
-      
+      
           
       
-      
-      
+      
           
       
-      
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
   
\ No newline at end of file
diff --git a/v1.13/iks/version.txt b/v1.13/iks/version.txt
index 264cfdffe7..dd36347a0b 100644
--- a/v1.13/iks/version.txt
+++ b/v1.13/iks/version.txt
@@ -1,2 +1,2 @@
-Client Version: version.Info{Major:"1", Minor:"13", GitVersion:"v1.13.6", GitCommit:"abdda3f9fefa29172298a2e42f5102e777a8ec25", GitTreeState:"clean", BuildDate:"2019-05-08T13:53:53Z", GoVersion:"go1.11.5", Compiler:"gc", Platform:"linux/amd64"}
-Server Version: version.Info{Major:"1", Minor:"13", GitVersion:"v1.13.6+IKS", GitCommit:"ac5f7341d5d0ce8ea8f206ba5b030dc9e9d4cc97", GitTreeState:"clean", BuildDate:"2019-05-09T13:26:51Z", GoVersion:"go1.11.5", Compiler:"gc", Platform:"linux/amd64"}
+Client Version: version.Info{Major:"1", Minor:"13", GitVersion:"v1.13.7", GitCommit:"4683545293d792934a7a7e12f2cc47d20b2dd01b", GitTreeState:"clean", BuildDate:"2019-06-06T01:46:52Z", GoVersion:"go1.11.5", Compiler:"gc", Platform:"linux/amd64"}
+Server Version: version.Info{Major:"1", Minor:"13", GitVersion:"v1.13.7+IKS", GitCommit:"675df39b011fd4f4f54aa131d903bab685cde6b8", GitTreeState:"clean", BuildDate:"2019-06-10T19:51:36Z", GoVersion:"go1.11.5", Compiler:"gc", Platform:"linux/amd64"}
diff --git a/v1.14/iks/PRODUCT.yaml b/v1.14/iks/PRODUCT.yaml
index 5ea1dd4cd9..b12a7402d9 100644
--- a/v1.14/iks/PRODUCT.yaml
+++ b/v1.14/iks/PRODUCT.yaml
@@ -1,6 +1,6 @@
 vendor: IBM
 name: IBM Cloud Kubernetes Service
-version: 1.14.2
+version: 1.14.3
 website_url: https://www.ibm.com/cloud/container-service
 documentation_url: https://cloud.ibm.com/docs/containers?topic=containers-container_index#container_index
 product_logo_url: https://raw.githubusercontent.com/ibm-containers/logo/master/iks.png
diff --git a/v1.14/iks/e2e.log b/v1.14/iks/e2e.log
index 287420ee1d..2b42584b0a 100644
--- a/v1.14/iks/e2e.log
+++ b/v1.14/iks/e2e.log
@@ -1,3486 +1,3040 @@
-I0604 18:27:58.365731      18 test_context.go:405] Using a temporary kubeconfig file from in-cluster config : /tmp/kubeconfig-229234504
-I0604 18:27:58.365802      18 e2e.go:240] Starting e2e run "79e879d1-86f6-11e9-8d1b-467ee19922ac" on Ginkgo node 1
+I0618 11:34:35.438599      20 test_context.go:405] Using a temporary kubeconfig file from in-cluster config : /tmp/kubeconfig-426929150
+I0618 11:34:35.438768      20 e2e.go:240] Starting e2e run "0bbd93b1-91bd-11e9-a25d-8608290c688a" on Ginkgo node 1
 Running Suite: Kubernetes e2e suite
 ===================================
-Random Seed: 1559672877 - Will randomize all specs
+Random Seed: 1560857674 - Will randomize all specs
 Will run 204 of 3585 specs
 
-Jun  4 18:27:58.492: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-Jun  4 18:27:58.494: INFO: Waiting up to 30m0s for all (but 0) nodes to be schedulable
-Jun  4 18:27:58.545: INFO: Waiting up to 10m0s for all pods (need at least 0) in namespace 'kube-system' to be running and ready
-Jun  4 18:27:58.615: INFO: 23 / 23 pods in namespace 'kube-system' are running and ready (0 seconds elapsed)
-Jun  4 18:27:58.615: INFO: expected 11 pod replicas in namespace 'kube-system', 11 are Running and Ready.
-Jun  4 18:27:58.615: INFO: Waiting up to 5m0s for all daemonsets in namespace 'kube-system' to start
-Jun  4 18:27:58.636: INFO: 3 / 3 pods ready in namespace 'kube-system' in daemonset 'calico-node' (0 seconds elapsed)
-Jun  4 18:27:58.636: INFO: 3 / 3 pods ready in namespace 'kube-system' in daemonset 'ibm-keepalived-watcher' (0 seconds elapsed)
-Jun  4 18:27:58.636: INFO: 3 / 3 pods ready in namespace 'kube-system' in daemonset 'ibm-kube-fluentd' (0 seconds elapsed)
-Jun  4 18:27:58.636: INFO: 0 / 0 pods ready in namespace 'kube-system' in daemonset 'ibm-master-proxy' (0 seconds elapsed)
-Jun  4 18:27:58.636: INFO: 0 / 0 pods ready in namespace 'kube-system' in daemonset 'nvidia-driver-installer' (0 seconds elapsed)
-Jun  4 18:27:58.636: INFO: 0 / 0 pods ready in namespace 'kube-system' in daemonset 'nvidia-gpu-device-plugin' (0 seconds elapsed)
-Jun  4 18:27:58.636: INFO: e2e test version: v1.14.2
-Jun  4 18:27:58.639: INFO: kube-apiserver version: v1.14.2+IKS
-SSSS
+Jun 18 11:34:35.636: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+Jun 18 11:34:35.639: INFO: Waiting up to 30m0s for all (but 0) nodes to be schedulable
+Jun 18 11:34:35.690: INFO: Waiting up to 10m0s for all pods (need at least 0) in namespace 'kube-system' to be running and ready
+Jun 18 11:34:35.782: INFO: 23 / 23 pods in namespace 'kube-system' are running and ready (0 seconds elapsed)
+Jun 18 11:34:35.782: INFO: expected 11 pod replicas in namespace 'kube-system', 11 are Running and Ready.
+Jun 18 11:34:35.782: INFO: Waiting up to 5m0s for all daemonsets in namespace 'kube-system' to start
+Jun 18 11:34:35.815: INFO: 3 / 3 pods ready in namespace 'kube-system' in daemonset 'calico-node' (0 seconds elapsed)
+Jun 18 11:34:35.815: INFO: 3 / 3 pods ready in namespace 'kube-system' in daemonset 'ibm-keepalived-watcher' (0 seconds elapsed)
+Jun 18 11:34:35.815: INFO: 3 / 3 pods ready in namespace 'kube-system' in daemonset 'ibm-kube-fluentd' (0 seconds elapsed)
+Jun 18 11:34:35.815: INFO: 0 / 0 pods ready in namespace 'kube-system' in daemonset 'ibm-master-proxy' (0 seconds elapsed)
+Jun 18 11:34:35.815: INFO: 0 / 0 pods ready in namespace 'kube-system' in daemonset 'nvidia-driver-installer' (0 seconds elapsed)
+Jun 18 11:34:35.815: INFO: 0 / 0 pods ready in namespace 'kube-system' in daemonset 'nvidia-gpu-device-plugin' (0 seconds elapsed)
+Jun 18 11:34:35.815: INFO: e2e test version: v1.14.3
+Jun 18 11:34:35.819: INFO: kube-apiserver version: v1.14.3+IKS
+SSSSS
 ------------------------------
 [sig-storage] Projected downwardAPI 
-  should update annotations on modification [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+  should provide container's memory limit [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 [BeforeEach] [sig-storage] Projected downwardAPI
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 18:27:58.639: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
+Jun 18 11:34:35.819: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
 STEP: Building a namespace api object, basename projected
-Jun  4 18:27:58.735: INFO: Found PodSecurityPolicies; assuming PodSecurityPolicy is enabled.
-Jun  4 18:27:58.773: INFO: Found ClusterRoles; assuming RBAC is enabled.
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-2054
+Jun 18 11:34:35.907: INFO: Found PodSecurityPolicies; assuming PodSecurityPolicy is enabled.
+Jun 18 11:34:35.965: INFO: Found ClusterRoles; assuming RBAC is enabled.
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-398
 STEP: Waiting for a default service account to be provisioned in namespace
 [BeforeEach] [sig-storage] Projected downwardAPI
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:39
-[It] should update annotations on modification [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: Creating the pod
-Jun  4 18:28:05.655: INFO: Successfully updated pod "annotationupdate7aa47f31-86f6-11e9-8d1b-467ee19922ac"
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:39
+[It] should provide container's memory limit [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating a pod to test downward API volume plugin
+Jun 18 11:34:36.121: INFO: Waiting up to 5m0s for pod "downwardapi-volume-0cca5f58-91bd-11e9-a25d-8608290c688a" in namespace "projected-398" to be "success or failure"
+Jun 18 11:34:36.134: INFO: Pod "downwardapi-volume-0cca5f58-91bd-11e9-a25d-8608290c688a": Phase="Pending", Reason="", readiness=false. Elapsed: 13.319065ms
+Jun 18 11:34:38.149: INFO: Pod "downwardapi-volume-0cca5f58-91bd-11e9-a25d-8608290c688a": Phase="Pending", Reason="", readiness=false. Elapsed: 2.028002546s
+Jun 18 11:34:40.165: INFO: Pod "downwardapi-volume-0cca5f58-91bd-11e9-a25d-8608290c688a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.044296317s
+STEP: Saw pod success
+Jun 18 11:34:40.165: INFO: Pod "downwardapi-volume-0cca5f58-91bd-11e9-a25d-8608290c688a" satisfied condition "success or failure"
+Jun 18 11:34:40.182: INFO: Trying to get logs from node 10.72.74.184 pod downwardapi-volume-0cca5f58-91bd-11e9-a25d-8608290c688a container client-container: 
+STEP: delete the pod
+Jun 18 11:34:40.307: INFO: Waiting for pod downwardapi-volume-0cca5f58-91bd-11e9-a25d-8608290c688a to disappear
+Jun 18 11:34:40.322: INFO: Pod downwardapi-volume-0cca5f58-91bd-11e9-a25d-8608290c688a no longer exists
 [AfterEach] [sig-storage] Projected downwardAPI
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 18:28:07.707: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "projected-2054" for this suite.
-Jun  4 18:28:31.803: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 18:28:32.195: INFO: namespace projected-2054 deletion completed in 24.439557311s
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 11:34:40.322: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "projected-398" for this suite.
+Jun 18 11:34:46.391: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 11:34:47.488: INFO: namespace projected-398 deletion completed in 7.14981165s
 
-• [SLOW TEST:33.556 seconds]
+• [SLOW TEST:11.669 seconds]
 [sig-storage] Projected downwardAPI
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:33
-  should update annotations on modification [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-------------------------------
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
-------------------------------
-[sig-storage] Secrets 
-  optional updates should be reflected in volume [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [sig-storage] Secrets
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
-STEP: Creating a kubernetes client
-Jun  4 18:28:32.195: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-STEP: Building a namespace api object, basename secrets
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in secrets-4057
-STEP: Waiting for a default service account to be provisioned in namespace
-[It] optional updates should be reflected in volume [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: Creating secret with name s-test-opt-del-8e9d09b5-86f6-11e9-8d1b-467ee19922ac
-STEP: Creating secret with name s-test-opt-upd-8e9d0a03-86f6-11e9-8d1b-467ee19922ac
-STEP: Creating the pod
-STEP: Deleting secret s-test-opt-del-8e9d09b5-86f6-11e9-8d1b-467ee19922ac
-STEP: Updating secret s-test-opt-upd-8e9d0a03-86f6-11e9-8d1b-467ee19922ac
-STEP: Creating secret with name s-test-opt-create-8e9d0a1e-86f6-11e9-8d1b-467ee19922ac
-STEP: waiting to observe update in volume
-[AfterEach] [sig-storage] Secrets
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 18:28:36.789: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "secrets-4057" for this suite.
-Jun  4 18:29:00.865: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 18:29:01.363: INFO: namespace secrets-4057 deletion completed in 24.557598997s
-
-• [SLOW TEST:29.167 seconds]
-[sig-storage] Secrets
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/secrets_volume.go:33
-  optional updates should be reflected in volume [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:33
+  should provide container's memory limit [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+SSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-cli] Kubectl client [k8s.io] Kubectl replace 
-  should update a single-container pod's image  [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[sig-cli] Kubectl client [k8s.io] Update Demo 
+  should do a rolling update of a replication controller  [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 [BeforeEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 18:29:01.363: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
+Jun 18 11:34:47.488: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
 STEP: Building a namespace api object, basename kubectl
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-3671
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-3673
 STEP: Waiting for a default service account to be provisioned in namespace
 [BeforeEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:213
-[BeforeEach] [k8s.io] Kubectl replace
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1619
-[It] should update a single-container pod's image  [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: running the image docker.io/library/nginx:1.14-alpine
-Jun  4 18:29:01.583: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 run e2e-test-nginx-pod --generator=run-pod/v1 --image=docker.io/library/nginx:1.14-alpine --labels=run=e2e-test-nginx-pod --namespace=kubectl-3671'
-Jun  4 18:29:01.770: INFO: stderr: ""
-Jun  4 18:29:01.770: INFO: stdout: "pod/e2e-test-nginx-pod created\n"
-STEP: verifying the pod e2e-test-nginx-pod is running
-STEP: verifying the pod e2e-test-nginx-pod was created
-Jun  4 18:29:11.821: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 get pod e2e-test-nginx-pod --namespace=kubectl-3671 -o json'
-Jun  4 18:29:11.971: INFO: stderr: ""
-Jun  4 18:29:11.971: INFO: stdout: "{\n    \"apiVersion\": \"v1\",\n    \"kind\": \"Pod\",\n    \"metadata\": {\n        \"annotations\": {\n            \"kubernetes.io/psp\": \"e2e-test-privileged-psp\"\n        },\n        \"creationTimestamp\": \"2019-06-04T18:29:01Z\",\n        \"labels\": {\n            \"run\": \"e2e-test-nginx-pod\"\n        },\n        \"name\": \"e2e-test-nginx-pod\",\n        \"namespace\": \"kubectl-3671\",\n        \"resourceVersion\": \"6996\",\n        \"selfLink\": \"/api/v1/namespaces/kubectl-3671/pods/e2e-test-nginx-pod\",\n        \"uid\": \"a015bb38-86f6-11e9-9198-6211642778e8\"\n    },\n    \"spec\": {\n        \"containers\": [\n            {\n                \"image\": \"docker.io/library/nginx:1.14-alpine\",\n                \"imagePullPolicy\": \"IfNotPresent\",\n                \"name\": \"e2e-test-nginx-pod\",\n                \"resources\": {},\n                \"terminationMessagePath\": \"/dev/termination-log\",\n                \"terminationMessagePolicy\": \"File\",\n                \"volumeMounts\": [\n                    {\n                        \"mountPath\": \"/var/run/secrets/kubernetes.io/serviceaccount\",\n                        \"name\": \"default-token-99mmh\",\n                        \"readOnly\": true\n                    }\n                ]\n            }\n        ],\n        \"dnsPolicy\": \"ClusterFirst\",\n        \"enableServiceLinks\": true,\n        \"nodeName\": \"10.212.23.164\",\n        \"priority\": 0,\n        \"restartPolicy\": \"Always\",\n        \"schedulerName\": \"default-scheduler\",\n        \"securityContext\": {},\n        \"serviceAccount\": \"default\",\n        \"serviceAccountName\": \"default\",\n        \"terminationGracePeriodSeconds\": 30,\n        \"tolerations\": [\n            {\n                \"effect\": \"NoExecute\",\n                \"key\": \"node.kubernetes.io/not-ready\",\n                \"operator\": \"Exists\",\n                \"tolerationSeconds\": 300\n            },\n            {\n                \"effect\": \"NoExecute\",\n                \"key\": \"node.kubernetes.io/unreachable\",\n                \"operator\": \"Exists\",\n                \"tolerationSeconds\": 300\n            }\n        ],\n        \"volumes\": [\n            {\n                \"name\": \"default-token-99mmh\",\n                \"secret\": {\n                    \"defaultMode\": 420,\n                    \"secretName\": \"default-token-99mmh\"\n                }\n            }\n        ]\n    },\n    \"status\": {\n        \"conditions\": [\n            {\n                \"lastProbeTime\": null,\n                \"lastTransitionTime\": \"2019-06-04T18:29:01Z\",\n                \"status\": \"True\",\n                \"type\": \"Initialized\"\n            },\n            {\n                \"lastProbeTime\": null,\n                \"lastTransitionTime\": \"2019-06-04T18:29:07Z\",\n                \"status\": \"True\",\n                \"type\": \"Ready\"\n            },\n            {\n                \"lastProbeTime\": null,\n                \"lastTransitionTime\": \"2019-06-04T18:29:07Z\",\n                \"status\": \"True\",\n                \"type\": \"ContainersReady\"\n            },\n            {\n                \"lastProbeTime\": null,\n                \"lastTransitionTime\": \"2019-06-04T18:29:01Z\",\n                \"status\": \"True\",\n                \"type\": \"PodScheduled\"\n            }\n        ],\n        \"containerStatuses\": [\n            {\n                \"containerID\": \"containerd://8761f37de40da4b8402c6a545074cbe9cb9fc19ba72f6edb4a2595d5699cd48c\",\n                \"image\": \"docker.io/library/nginx:1.14-alpine\",\n                \"imageID\": \"docker.io/library/nginx@sha256:485b610fefec7ff6c463ced9623314a04ed67e3945b9c08d7e53a47f6d108dc7\",\n                \"lastState\": {},\n                \"name\": \"e2e-test-nginx-pod\",\n                \"ready\": true,\n                \"restartCount\": 0,\n                \"state\": {\n                    \"running\": {\n                        \"startedAt\": \"2019-06-04T18:29:07Z\"\n                    }\n                }\n            }\n        ],\n        \"hostIP\": \"10.212.23.164\",\n        \"phase\": \"Running\",\n        \"podIP\": \"172.30.254.134\",\n        \"qosClass\": \"BestEffort\",\n        \"startTime\": \"2019-06-04T18:29:01Z\"\n    }\n}\n"
-STEP: replace the image in the pod
-Jun  4 18:29:11.971: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 replace -f - --namespace=kubectl-3671'
-Jun  4 18:29:12.222: INFO: stderr: ""
-Jun  4 18:29:12.222: INFO: stdout: "pod/e2e-test-nginx-pod replaced\n"
-STEP: verifying the pod e2e-test-nginx-pod has the right image docker.io/library/busybox:1.29
-[AfterEach] [k8s.io] Kubectl replace
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1624
-Jun  4 18:29:12.237: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 delete pods e2e-test-nginx-pod --namespace=kubectl-3671'
-Jun  4 18:29:18.057: INFO: stderr: ""
-Jun  4 18:29:18.057: INFO: stdout: "pod \"e2e-test-nginx-pod\" deleted\n"
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:213
+[BeforeEach] [k8s.io] Update Demo
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:265
+[It] should do a rolling update of a replication controller  [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: creating the initial replication controller
+Jun 18 11:34:47.683: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 create -f - --namespace=kubectl-3673'
+Jun 18 11:34:48.063: INFO: stderr: ""
+Jun 18 11:34:48.063: INFO: stdout: "replicationcontroller/update-demo-nautilus created\n"
+STEP: waiting for all containers in name=update-demo pods to come up.
+Jun 18 11:34:48.063: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-3673'
+Jun 18 11:34:48.205: INFO: stderr: ""
+Jun 18 11:34:48.205: INFO: stdout: "update-demo-nautilus-cg8g7 update-demo-nautilus-cgkm5 "
+Jun 18 11:34:48.205: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 get pods update-demo-nautilus-cg8g7 -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-3673'
+Jun 18 11:34:48.340: INFO: stderr: ""
+Jun 18 11:34:48.340: INFO: stdout: ""
+Jun 18 11:34:48.340: INFO: update-demo-nautilus-cg8g7 is created but not running
+Jun 18 11:34:53.340: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-3673'
+Jun 18 11:34:53.476: INFO: stderr: ""
+Jun 18 11:34:53.476: INFO: stdout: "update-demo-nautilus-cg8g7 update-demo-nautilus-cgkm5 "
+Jun 18 11:34:53.476: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 get pods update-demo-nautilus-cg8g7 -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-3673'
+Jun 18 11:34:53.622: INFO: stderr: ""
+Jun 18 11:34:53.622: INFO: stdout: "true"
+Jun 18 11:34:53.622: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 get pods update-demo-nautilus-cg8g7 -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-3673'
+Jun 18 11:34:53.740: INFO: stderr: ""
+Jun 18 11:34:53.740: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0"
+Jun 18 11:34:53.740: INFO: validating pod update-demo-nautilus-cg8g7
+Jun 18 11:34:53.769: INFO: got data: {
+  "image": "nautilus.jpg"
+}
+
+Jun 18 11:34:53.769: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg .
+Jun 18 11:34:53.769: INFO: update-demo-nautilus-cg8g7 is verified up and running
+Jun 18 11:34:53.769: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 get pods update-demo-nautilus-cgkm5 -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-3673'
+Jun 18 11:34:53.913: INFO: stderr: ""
+Jun 18 11:34:53.913: INFO: stdout: "true"
+Jun 18 11:34:53.913: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 get pods update-demo-nautilus-cgkm5 -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-3673'
+Jun 18 11:34:54.068: INFO: stderr: ""
+Jun 18 11:34:54.068: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0"
+Jun 18 11:34:54.068: INFO: validating pod update-demo-nautilus-cgkm5
+Jun 18 11:34:54.095: INFO: got data: {
+  "image": "nautilus.jpg"
+}
+
+Jun 18 11:34:54.095: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg .
+Jun 18 11:34:54.095: INFO: update-demo-nautilus-cgkm5 is verified up and running
+STEP: rolling-update to new replication controller
+Jun 18 11:34:54.097: INFO: scanned /root for discovery docs: 
+Jun 18 11:34:54.097: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 rolling-update update-demo-nautilus --update-period=1s -f - --namespace=kubectl-3673'
+Jun 18 11:35:17.297: INFO: stderr: "Command \"rolling-update\" is deprecated, use \"rollout\" instead\n"
+Jun 18 11:35:17.297: INFO: stdout: "Created update-demo-kitten\nScaling up update-demo-kitten from 0 to 2, scaling down update-demo-nautilus from 2 to 0 (keep 2 pods available, don't exceed 3 pods)\nScaling update-demo-kitten up to 1\nScaling update-demo-nautilus down to 1\nScaling update-demo-kitten up to 2\nScaling update-demo-nautilus down to 0\nUpdate succeeded. Deleting old controller: update-demo-nautilus\nRenaming update-demo-kitten to update-demo-nautilus\nreplicationcontroller/update-demo-nautilus rolling updated\n"
+STEP: waiting for all containers in name=update-demo pods to come up.
+Jun 18 11:35:17.297: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-3673'
+Jun 18 11:35:17.432: INFO: stderr: ""
+Jun 18 11:35:17.432: INFO: stdout: "update-demo-kitten-fgck9 update-demo-kitten-rmjwz "
+Jun 18 11:35:17.432: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 get pods update-demo-kitten-fgck9 -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-3673'
+Jun 18 11:35:17.551: INFO: stderr: ""
+Jun 18 11:35:17.551: INFO: stdout: "true"
+Jun 18 11:35:17.551: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 get pods update-demo-kitten-fgck9 -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-3673'
+Jun 18 11:35:17.678: INFO: stderr: ""
+Jun 18 11:35:17.679: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/kitten:1.0"
+Jun 18 11:35:17.679: INFO: validating pod update-demo-kitten-fgck9
+Jun 18 11:35:17.705: INFO: got data: {
+  "image": "kitten.jpg"
+}
+
+Jun 18 11:35:17.705: INFO: Unmarshalled json jpg/img => {kitten.jpg} , expecting kitten.jpg .
+Jun 18 11:35:17.705: INFO: update-demo-kitten-fgck9 is verified up and running
+Jun 18 11:35:17.705: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 get pods update-demo-kitten-rmjwz -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-3673'
+Jun 18 11:35:17.832: INFO: stderr: ""
+Jun 18 11:35:17.832: INFO: stdout: "true"
+Jun 18 11:35:17.832: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 get pods update-demo-kitten-rmjwz -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-3673'
+Jun 18 11:35:17.973: INFO: stderr: ""
+Jun 18 11:35:17.973: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/kitten:1.0"
+Jun 18 11:35:17.973: INFO: validating pod update-demo-kitten-rmjwz
+Jun 18 11:35:17.998: INFO: got data: {
+  "image": "kitten.jpg"
+}
+
+Jun 18 11:35:17.998: INFO: Unmarshalled json jpg/img => {kitten.jpg} , expecting kitten.jpg .
+Jun 18 11:35:17.998: INFO: update-demo-kitten-rmjwz is verified up and running
 [AfterEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 18:29:18.057: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "kubectl-3671" for this suite.
-Jun  4 18:29:26.118: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 18:29:26.483: INFO: namespace kubectl-3671 deletion completed in 8.40628725s
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 11:35:17.998: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "kubectl-3673" for this suite.
+Jun 18 11:35:42.043: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 11:35:42.485: INFO: namespace kubectl-3673 deletion completed in 24.473254522s
 
-• [SLOW TEST:25.120 seconds]
+• [SLOW TEST:54.997 seconds]
 [sig-cli] Kubectl client
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23
-  [k8s.io] Kubectl replace
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
-    should update a single-container pod's image  [Conformance]
-    /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23
+  [k8s.io] Update Demo
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
+    should do a rolling update of a replication controller  [Conformance]
+    /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-SSSS
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-network] DNS 
-  should provide /etc/hosts entries for the cluster [LinuxOnly] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [sig-network] DNS
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+[k8s.io] InitContainer [NodeConformance] 
+  should not start app containers and fail the pod if init containers fail on a RestartNever pod [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [k8s.io] InitContainer [NodeConformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 18:29:26.483: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-STEP: Building a namespace api object, basename dns
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in dns-5084
+Jun 18 11:35:42.486: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename init-container
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in init-container-1434
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should provide /etc/hosts entries for the cluster [LinuxOnly] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: Running these commands on wheezy: for i in `seq 1 600`; do test -n "$$(getent hosts dns-querier-1.dns-test-service.dns-5084.svc.cluster.local)" && echo OK > /results/wheezy_hosts@dns-querier-1.dns-test-service.dns-5084.svc.cluster.local;test -n "$$(getent hosts dns-querier-1)" && echo OK > /results/wheezy_hosts@dns-querier-1;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-5084.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@PodARecord;sleep 1; done
-
-STEP: Running these commands on jessie: for i in `seq 1 600`; do test -n "$$(getent hosts dns-querier-1.dns-test-service.dns-5084.svc.cluster.local)" && echo OK > /results/jessie_hosts@dns-querier-1.dns-test-service.dns-5084.svc.cluster.local;test -n "$$(getent hosts dns-querier-1)" && echo OK > /results/jessie_hosts@dns-querier-1;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-5084.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_tcp@PodARecord;sleep 1; done
-
-STEP: creating a pod to probe /etc/hosts
-STEP: submitting the pod to kubernetes
-STEP: retrieving the pod
-STEP: looking for the results for each expected name from probers
-Jun  4 18:29:40.935: INFO: DNS probes using dns-5084/dns-test-aef85da2-86f6-11e9-8d1b-467ee19922ac succeeded
-
-STEP: deleting the pod
-[AfterEach] [sig-network] DNS
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 18:29:40.973: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "dns-5084" for this suite.
-Jun  4 18:29:47.040: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 18:29:47.406: INFO: namespace dns-5084 deletion completed in 6.417133576s
+[BeforeEach] [k8s.io] InitContainer [NodeConformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/init_container.go:43
+[It] should not start app containers and fail the pod if init containers fail on a RestartNever pod [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: creating the pod
+Jun 18 11:35:42.698: INFO: PodSpec: initContainers in spec.initContainers
+[AfterEach] [k8s.io] InitContainer [NodeConformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 11:35:48.040: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "init-container-1434" for this suite.
+Jun 18 11:35:54.082: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 11:35:54.599: INFO: namespace init-container-1434 deletion completed in 6.543928159s
 
-• [SLOW TEST:20.923 seconds]
-[sig-network] DNS
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:22
-  should provide /etc/hosts entries for the cluster [LinuxOnly] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+• [SLOW TEST:12.113 seconds]
+[k8s.io] InitContainer [NodeConformance]
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
+  should not start app containers and fail the pod if init containers fail on a RestartNever pod [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-SSSS
+SSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
 [sig-network] Proxy version v1 
-  should proxy logs on node with explicit kubelet port using proxy subresource  [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+  should proxy through a service and a pod  [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 [BeforeEach] version v1
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 18:29:47.407: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
+Jun 18 11:35:54.599: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
 STEP: Building a namespace api object, basename proxy
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in proxy-5472
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in proxy-2583
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should proxy logs on node with explicit kubelet port using proxy subresource  [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-Jun  4 18:29:47.663: INFO: (0) /api/v1/nodes/10.212.23.161:10250/proxy/logs/: 
-alb/
-alternatives.log
-apt/... (200; 21.710734ms)
-Jun  4 18:29:47.680: INFO: (1) /api/v1/nodes/10.212.23.161:10250/proxy/logs/: 
-alb/
-alternatives.log
-apt/... (200; 16.537823ms)
-Jun  4 18:29:47.697: INFO: (2) /api/v1/nodes/10.212.23.161:10250/proxy/logs/: 
-alb/
-alternatives.log
-apt/... (200; 17.616981ms)
-Jun  4 18:29:47.713: INFO: (3) /api/v1/nodes/10.212.23.161:10250/proxy/logs/: 
-alb/
-alternatives.log
-apt/... (200; 15.735465ms)
-Jun  4 18:29:47.729: INFO: (4) /api/v1/nodes/10.212.23.161:10250/proxy/logs/: 
-alb/
-alternatives.log
-apt/... (200; 15.735948ms)
-Jun  4 18:29:47.755: INFO: (5) /api/v1/nodes/10.212.23.161:10250/proxy/logs/: 
-alb/
-alternatives.log
-apt/... (200; 26.254102ms)
-Jun  4 18:29:47.771: INFO: (6) /api/v1/nodes/10.212.23.161:10250/proxy/logs/: 
-alb/
-alternatives.log
-apt/... (200; 16.127811ms)
-Jun  4 18:29:47.788: INFO: (7) /api/v1/nodes/10.212.23.161:10250/proxy/logs/: 
-alb/
-alternatives.log
-apt/... (200; 16.149405ms)
-Jun  4 18:29:47.805: INFO: (8) /api/v1/nodes/10.212.23.161:10250/proxy/logs/: 
-alb/
-alternatives.log
-apt/... (200; 17.131211ms)
-Jun  4 18:29:47.822: INFO: (9) /api/v1/nodes/10.212.23.161:10250/proxy/logs/: 
-alb/
-alternatives.log
-apt/... (200; 16.994644ms)
-Jun  4 18:29:47.839: INFO: (10) /api/v1/nodes/10.212.23.161:10250/proxy/logs/: 
-alb/
-alternatives.log
-apt/... (200; 16.567176ms)
-Jun  4 18:29:47.856: INFO: (11) /api/v1/nodes/10.212.23.161:10250/proxy/logs/: 
-alb/
-alternatives.log
-apt/... (200; 16.766874ms)
-Jun  4 18:29:47.871: INFO: (12) /api/v1/nodes/10.212.23.161:10250/proxy/logs/: 
-alb/
-alternatives.log
-apt/... (200; 15.859679ms)
-Jun  4 18:29:47.897: INFO: (13) /api/v1/nodes/10.212.23.161:10250/proxy/logs/: 
-alb/
-alternatives.log
-apt/... (200; 25.964677ms)
-Jun  4 18:29:47.914: INFO: (14) /api/v1/nodes/10.212.23.161:10250/proxy/logs/: 
-alb/
-alternatives.log
-apt/... (200; 16.005144ms)
-Jun  4 18:29:47.930: INFO: (15) /api/v1/nodes/10.212.23.161:10250/proxy/logs/: 
-alb/
-alternatives.log
-apt/... (200; 16.259596ms)
-Jun  4 18:29:47.948: INFO: (16) /api/v1/nodes/10.212.23.161:10250/proxy/logs/: 
-alb/
-alternatives.log
-apt/... (200; 18.346542ms)
-Jun  4 18:29:47.965: INFO: (17) /api/v1/nodes/10.212.23.161:10250/proxy/logs/: 
-alb/
-alternatives.log
-apt/... (200; 16.408487ms)
-Jun  4 18:29:47.981: INFO: (18) /api/v1/nodes/10.212.23.161:10250/proxy/logs/: 
-alb/
-alternatives.log
-apt/... (200; 16.472914ms)
-Jun  4 18:29:48.005: INFO: (19) /api/v1/nodes/10.212.23.161:10250/proxy/logs/: 
-alb/
-alternatives.log
-apt/... (200; 23.37191ms)
+[It] should proxy through a service and a pod  [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: starting an echo server on multiple ports
+STEP: creating replication controller proxy-service-fsmzd in namespace proxy-2583
+I0618 11:35:54.840796      20 runners.go:184] Created replication controller with name: proxy-service-fsmzd, namespace: proxy-2583, replica count: 1
+I0618 11:35:55.891300      20 runners.go:184] proxy-service-fsmzd Pods: 1 out of 1 created, 0 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady 
+I0618 11:35:56.891612      20 runners.go:184] proxy-service-fsmzd Pods: 1 out of 1 created, 0 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady 
+I0618 11:35:57.891881      20 runners.go:184] proxy-service-fsmzd Pods: 1 out of 1 created, 0 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 1 runningButNotReady 
+I0618 11:35:58.892174      20 runners.go:184] proxy-service-fsmzd Pods: 1 out of 1 created, 0 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 1 runningButNotReady 
+I0618 11:35:59.893004      20 runners.go:184] proxy-service-fsmzd Pods: 1 out of 1 created, 0 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 1 runningButNotReady 
+I0618 11:36:00.893222      20 runners.go:184] proxy-service-fsmzd Pods: 1 out of 1 created, 0 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 1 runningButNotReady 
+I0618 11:36:01.893409      20 runners.go:184] proxy-service-fsmzd Pods: 1 out of 1 created, 0 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 1 runningButNotReady 
+I0618 11:36:02.893621      20 runners.go:184] proxy-service-fsmzd Pods: 1 out of 1 created, 0 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 1 runningButNotReady 
+I0618 11:36:03.893825      20 runners.go:184] proxy-service-fsmzd Pods: 1 out of 1 created, 0 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 1 runningButNotReady 
+I0618 11:36:04.894062      20 runners.go:184] proxy-service-fsmzd Pods: 1 out of 1 created, 1 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady 
+Jun 18 11:36:04.901: INFO: setup took 10.106861558s, starting test cases
+STEP: running 16 cases, 20 attempts per case, 320 total attempts
+Jun 18 11:36:04.927: INFO: (0) /api/v1/namespaces/proxy-2583/pods/http:proxy-service-fsmzd-2gj5q:160/proxy/: foo (200; 25.289269ms)
+Jun 18 11:36:04.933: INFO: (0) /api/v1/namespaces/proxy-2583/services/proxy-service-fsmzd:portname1/proxy/: foo (200; 31.37217ms)
+Jun 18 11:36:04.933: INFO: (0) /api/v1/namespaces/proxy-2583/pods/http:proxy-service-fsmzd-2gj5q:162/proxy/: bar (200; 31.393267ms)
+Jun 18 11:36:04.933: INFO: (0) /api/v1/namespaces/proxy-2583/pods/proxy-service-fsmzd-2gj5q:162/proxy/: bar (200; 31.920493ms)
+Jun 18 11:36:04.934: INFO: (0) /api/v1/namespaces/proxy-2583/services/http:proxy-service-fsmzd:portname1/proxy/: foo (200; 32.364546ms)
+Jun 18 11:36:04.934: INFO: (0) /api/v1/namespaces/proxy-2583/pods/proxy-service-fsmzd-2gj5q/proxy/: test (200; 32.042214ms)
+Jun 18 11:36:04.935: INFO: (0) /api/v1/namespaces/proxy-2583/services/http:proxy-service-fsmzd:portname2/proxy/: bar (200; 32.969129ms)
+Jun 18 11:36:04.935: INFO: (0) /api/v1/namespaces/proxy-2583/pods/http:proxy-service-fsmzd-2gj5q:1080/proxy/: ... (200; 33.036567ms)
+Jun 18 11:36:04.935: INFO: (0) /api/v1/namespaces/proxy-2583/services/proxy-service-fsmzd:portname2/proxy/: bar (200; 33.48674ms)
+Jun 18 11:36:04.936: INFO: (0) /api/v1/namespaces/proxy-2583/pods/proxy-service-fsmzd-2gj5q:1080/proxy/: test<... (200; 34.109524ms)
+Jun 18 11:36:04.940: INFO: (0) /api/v1/namespaces/proxy-2583/pods/proxy-service-fsmzd-2gj5q:160/proxy/: foo (200; 38.438362ms)
+Jun 18 11:36:04.949: INFO: (0) /api/v1/namespaces/proxy-2583/pods/https:proxy-service-fsmzd-2gj5q:443/proxy/: test<... (200; 18.69481ms)
+Jun 18 11:36:04.981: INFO: (1) /api/v1/namespaces/proxy-2583/pods/proxy-service-fsmzd-2gj5q/proxy/: test (200; 24.009046ms)
+Jun 18 11:36:04.981: INFO: (1) /api/v1/namespaces/proxy-2583/services/http:proxy-service-fsmzd:portname1/proxy/: foo (200; 24.429482ms)
+Jun 18 11:36:04.984: INFO: (1) /api/v1/namespaces/proxy-2583/pods/http:proxy-service-fsmzd-2gj5q:162/proxy/: bar (200; 26.949649ms)
+Jun 18 11:36:04.985: INFO: (1) /api/v1/namespaces/proxy-2583/pods/https:proxy-service-fsmzd-2gj5q:462/proxy/: tls qux (200; 27.302717ms)
+Jun 18 11:36:04.985: INFO: (1) /api/v1/namespaces/proxy-2583/pods/https:proxy-service-fsmzd-2gj5q:460/proxy/: tls baz (200; 27.62692ms)
+Jun 18 11:36:04.985: INFO: (1) /api/v1/namespaces/proxy-2583/pods/proxy-service-fsmzd-2gj5q:162/proxy/: bar (200; 27.859053ms)
+Jun 18 11:36:04.985: INFO: (1) /api/v1/namespaces/proxy-2583/pods/https:proxy-service-fsmzd-2gj5q:443/proxy/: ... (200; 28.294249ms)
+Jun 18 11:36:04.987: INFO: (1) /api/v1/namespaces/proxy-2583/services/proxy-service-fsmzd:portname1/proxy/: foo (200; 30.366524ms)
+Jun 18 11:36:04.987: INFO: (1) /api/v1/namespaces/proxy-2583/pods/http:proxy-service-fsmzd-2gj5q:160/proxy/: foo (200; 30.444009ms)
+Jun 18 11:36:04.988: INFO: (1) /api/v1/namespaces/proxy-2583/services/https:proxy-service-fsmzd:tlsportname2/proxy/: tls qux (200; 30.664967ms)
+Jun 18 11:36:04.988: INFO: (1) /api/v1/namespaces/proxy-2583/services/https:proxy-service-fsmzd:tlsportname1/proxy/: tls baz (200; 30.676581ms)
+Jun 18 11:36:04.990: INFO: (1) /api/v1/namespaces/proxy-2583/services/http:proxy-service-fsmzd:portname2/proxy/: bar (200; 33.20094ms)
+Jun 18 11:36:04.991: INFO: (1) /api/v1/namespaces/proxy-2583/services/proxy-service-fsmzd:portname2/proxy/: bar (200; 33.575506ms)
+Jun 18 11:36:05.010: INFO: (2) /api/v1/namespaces/proxy-2583/pods/proxy-service-fsmzd-2gj5q/proxy/: test (200; 18.555032ms)
+Jun 18 11:36:05.016: INFO: (2) /api/v1/namespaces/proxy-2583/pods/proxy-service-fsmzd-2gj5q:162/proxy/: bar (200; 24.419212ms)
+Jun 18 11:36:05.016: INFO: (2) /api/v1/namespaces/proxy-2583/pods/proxy-service-fsmzd-2gj5q:160/proxy/: foo (200; 24.86153ms)
+Jun 18 11:36:05.016: INFO: (2) /api/v1/namespaces/proxy-2583/pods/http:proxy-service-fsmzd-2gj5q:162/proxy/: bar (200; 24.814556ms)
+Jun 18 11:36:05.016: INFO: (2) /api/v1/namespaces/proxy-2583/pods/https:proxy-service-fsmzd-2gj5q:460/proxy/: tls baz (200; 24.849168ms)
+Jun 18 11:36:05.016: INFO: (2) /api/v1/namespaces/proxy-2583/pods/https:proxy-service-fsmzd-2gj5q:443/proxy/: ... (200; 25.000687ms)
+Jun 18 11:36:05.017: INFO: (2) /api/v1/namespaces/proxy-2583/services/https:proxy-service-fsmzd:tlsportname1/proxy/: tls baz (200; 25.138153ms)
+Jun 18 11:36:05.017: INFO: (2) /api/v1/namespaces/proxy-2583/pods/proxy-service-fsmzd-2gj5q:1080/proxy/: test<... (200; 25.292362ms)
+Jun 18 11:36:05.023: INFO: (2) /api/v1/namespaces/proxy-2583/services/http:proxy-service-fsmzd:portname1/proxy/: foo (200; 31.385099ms)
+Jun 18 11:36:05.023: INFO: (2) /api/v1/namespaces/proxy-2583/services/https:proxy-service-fsmzd:tlsportname2/proxy/: tls qux (200; 31.311688ms)
+Jun 18 11:36:05.023: INFO: (2) /api/v1/namespaces/proxy-2583/services/proxy-service-fsmzd:portname2/proxy/: bar (200; 31.578543ms)
+Jun 18 11:36:05.023: INFO: (2) /api/v1/namespaces/proxy-2583/services/proxy-service-fsmzd:portname1/proxy/: foo (200; 31.625648ms)
+Jun 18 11:36:05.023: INFO: (2) /api/v1/namespaces/proxy-2583/services/http:proxy-service-fsmzd:portname2/proxy/: bar (200; 31.62785ms)
+Jun 18 11:36:05.043: INFO: (3) /api/v1/namespaces/proxy-2583/pods/http:proxy-service-fsmzd-2gj5q:162/proxy/: bar (200; 19.433825ms)
+Jun 18 11:36:05.047: INFO: (3) /api/v1/namespaces/proxy-2583/pods/proxy-service-fsmzd-2gj5q:160/proxy/: foo (200; 23.375308ms)
+Jun 18 11:36:05.047: INFO: (3) /api/v1/namespaces/proxy-2583/services/https:proxy-service-fsmzd:tlsportname1/proxy/: tls baz (200; 23.697049ms)
+Jun 18 11:36:05.048: INFO: (3) /api/v1/namespaces/proxy-2583/pods/https:proxy-service-fsmzd-2gj5q:462/proxy/: tls qux (200; 24.348247ms)
+Jun 18 11:36:05.048: INFO: (3) /api/v1/namespaces/proxy-2583/pods/https:proxy-service-fsmzd-2gj5q:443/proxy/: test (200; 24.196469ms)
+Jun 18 11:36:05.050: INFO: (3) /api/v1/namespaces/proxy-2583/pods/proxy-service-fsmzd-2gj5q:162/proxy/: bar (200; 26.846212ms)
+Jun 18 11:36:05.050: INFO: (3) /api/v1/namespaces/proxy-2583/pods/http:proxy-service-fsmzd-2gj5q:1080/proxy/: ... (200; 26.598822ms)
+Jun 18 11:36:05.050: INFO: (3) /api/v1/namespaces/proxy-2583/pods/proxy-service-fsmzd-2gj5q:1080/proxy/: test<... (200; 26.649013ms)
+Jun 18 11:36:05.050: INFO: (3) /api/v1/namespaces/proxy-2583/pods/http:proxy-service-fsmzd-2gj5q:160/proxy/: foo (200; 27.058318ms)
+Jun 18 11:36:05.050: INFO: (3) /api/v1/namespaces/proxy-2583/pods/https:proxy-service-fsmzd-2gj5q:460/proxy/: tls baz (200; 27.119487ms)
+Jun 18 11:36:05.051: INFO: (3) /api/v1/namespaces/proxy-2583/services/proxy-service-fsmzd:portname2/proxy/: bar (200; 27.884453ms)
+Jun 18 11:36:05.052: INFO: (3) /api/v1/namespaces/proxy-2583/services/https:proxy-service-fsmzd:tlsportname2/proxy/: tls qux (200; 28.845582ms)
+Jun 18 11:36:05.052: INFO: (3) /api/v1/namespaces/proxy-2583/services/http:proxy-service-fsmzd:portname1/proxy/: foo (200; 28.735561ms)
+Jun 18 11:36:05.056: INFO: (3) /api/v1/namespaces/proxy-2583/services/http:proxy-service-fsmzd:portname2/proxy/: bar (200; 32.231831ms)
+Jun 18 11:36:05.056: INFO: (3) /api/v1/namespaces/proxy-2583/services/proxy-service-fsmzd:portname1/proxy/: foo (200; 32.655023ms)
+Jun 18 11:36:05.074: INFO: (4) /api/v1/namespaces/proxy-2583/pods/https:proxy-service-fsmzd-2gj5q:443/proxy/: test<... (200; 24.92497ms)
+Jun 18 11:36:05.081: INFO: (4) /api/v1/namespaces/proxy-2583/pods/https:proxy-service-fsmzd-2gj5q:462/proxy/: tls qux (200; 24.85932ms)
+Jun 18 11:36:05.082: INFO: (4) /api/v1/namespaces/proxy-2583/pods/http:proxy-service-fsmzd-2gj5q:1080/proxy/: ... (200; 25.09894ms)
+Jun 18 11:36:05.082: INFO: (4) /api/v1/namespaces/proxy-2583/services/http:proxy-service-fsmzd:portname2/proxy/: bar (200; 24.90527ms)
+Jun 18 11:36:05.082: INFO: (4) /api/v1/namespaces/proxy-2583/pods/proxy-service-fsmzd-2gj5q/proxy/: test (200; 24.759573ms)
+Jun 18 11:36:05.082: INFO: (4) /api/v1/namespaces/proxy-2583/pods/http:proxy-service-fsmzd-2gj5q:162/proxy/: bar (200; 25.397954ms)
+Jun 18 11:36:05.086: INFO: (4) /api/v1/namespaces/proxy-2583/services/proxy-service-fsmzd:portname1/proxy/: foo (200; 29.731245ms)
+Jun 18 11:36:05.086: INFO: (4) /api/v1/namespaces/proxy-2583/services/https:proxy-service-fsmzd:tlsportname2/proxy/: tls qux (200; 29.600554ms)
+Jun 18 11:36:05.087: INFO: (4) /api/v1/namespaces/proxy-2583/services/https:proxy-service-fsmzd:tlsportname1/proxy/: tls baz (200; 29.882956ms)
+Jun 18 11:36:05.087: INFO: (4) /api/v1/namespaces/proxy-2583/services/http:proxy-service-fsmzd:portname1/proxy/: foo (200; 29.815448ms)
+Jun 18 11:36:05.087: INFO: (4) /api/v1/namespaces/proxy-2583/pods/proxy-service-fsmzd-2gj5q:162/proxy/: bar (200; 30.786214ms)
+Jun 18 11:36:05.087: INFO: (4) /api/v1/namespaces/proxy-2583/services/proxy-service-fsmzd:portname2/proxy/: bar (200; 30.590019ms)
+Jun 18 11:36:05.105: INFO: (5) /api/v1/namespaces/proxy-2583/pods/proxy-service-fsmzd-2gj5q/proxy/: test (200; 18.355832ms)
+Jun 18 11:36:05.110: INFO: (5) /api/v1/namespaces/proxy-2583/services/http:proxy-service-fsmzd:portname1/proxy/: foo (200; 23.066438ms)
+Jun 18 11:36:05.110: INFO: (5) /api/v1/namespaces/proxy-2583/services/proxy-service-fsmzd:portname1/proxy/: foo (200; 23.030631ms)
+Jun 18 11:36:05.110: INFO: (5) /api/v1/namespaces/proxy-2583/services/http:proxy-service-fsmzd:portname2/proxy/: bar (200; 23.124334ms)
+Jun 18 11:36:05.112: INFO: (5) /api/v1/namespaces/proxy-2583/pods/proxy-service-fsmzd-2gj5q:1080/proxy/: test<... (200; 24.529325ms)
+Jun 18 11:36:05.112: INFO: (5) /api/v1/namespaces/proxy-2583/pods/https:proxy-service-fsmzd-2gj5q:460/proxy/: tls baz (200; 24.562385ms)
+Jun 18 11:36:05.112: INFO: (5) /api/v1/namespaces/proxy-2583/pods/https:proxy-service-fsmzd-2gj5q:462/proxy/: tls qux (200; 24.494027ms)
+Jun 18 11:36:05.112: INFO: (5) /api/v1/namespaces/proxy-2583/pods/http:proxy-service-fsmzd-2gj5q:1080/proxy/: ... (200; 24.579559ms)
+Jun 18 11:36:05.112: INFO: (5) /api/v1/namespaces/proxy-2583/pods/proxy-service-fsmzd-2gj5q:162/proxy/: bar (200; 24.801177ms)
+Jun 18 11:36:05.113: INFO: (5) /api/v1/namespaces/proxy-2583/pods/https:proxy-service-fsmzd-2gj5q:443/proxy/: ... (200; 23.589482ms)
+Jun 18 11:36:05.141: INFO: (6) /api/v1/namespaces/proxy-2583/pods/https:proxy-service-fsmzd-2gj5q:443/proxy/: test (200; 23.952582ms)
+Jun 18 11:36:05.141: INFO: (6) /api/v1/namespaces/proxy-2583/pods/proxy-service-fsmzd-2gj5q:160/proxy/: foo (200; 23.730695ms)
+Jun 18 11:36:05.143: INFO: (6) /api/v1/namespaces/proxy-2583/pods/proxy-service-fsmzd-2gj5q:1080/proxy/: test<... (200; 24.998993ms)
+Jun 18 11:36:05.143: INFO: (6) /api/v1/namespaces/proxy-2583/pods/https:proxy-service-fsmzd-2gj5q:460/proxy/: tls baz (200; 25.750247ms)
+Jun 18 11:36:05.143: INFO: (6) /api/v1/namespaces/proxy-2583/pods/http:proxy-service-fsmzd-2gj5q:162/proxy/: bar (200; 25.569336ms)
+Jun 18 11:36:05.147: INFO: (6) /api/v1/namespaces/proxy-2583/services/proxy-service-fsmzd:portname2/proxy/: bar (200; 29.932518ms)
+Jun 18 11:36:05.148: INFO: (6) /api/v1/namespaces/proxy-2583/services/proxy-service-fsmzd:portname1/proxy/: foo (200; 30.115297ms)
+Jun 18 11:36:05.148: INFO: (6) /api/v1/namespaces/proxy-2583/services/https:proxy-service-fsmzd:tlsportname1/proxy/: tls baz (200; 30.294092ms)
+Jun 18 11:36:05.148: INFO: (6) /api/v1/namespaces/proxy-2583/pods/http:proxy-service-fsmzd-2gj5q:160/proxy/: foo (200; 31.00938ms)
+Jun 18 11:36:05.148: INFO: (6) /api/v1/namespaces/proxy-2583/services/https:proxy-service-fsmzd:tlsportname2/proxy/: tls qux (200; 30.395224ms)
+Jun 18 11:36:05.148: INFO: (6) /api/v1/namespaces/proxy-2583/services/http:proxy-service-fsmzd:portname2/proxy/: bar (200; 30.575013ms)
+Jun 18 11:36:05.169: INFO: (7) /api/v1/namespaces/proxy-2583/pods/https:proxy-service-fsmzd-2gj5q:460/proxy/: tls baz (200; 20.096721ms)
+Jun 18 11:36:05.169: INFO: (7) /api/v1/namespaces/proxy-2583/pods/https:proxy-service-fsmzd-2gj5q:443/proxy/: test (200; 27.454646ms)
+Jun 18 11:36:05.178: INFO: (7) /api/v1/namespaces/proxy-2583/pods/https:proxy-service-fsmzd-2gj5q:462/proxy/: tls qux (200; 28.772818ms)
+Jun 18 11:36:05.178: INFO: (7) /api/v1/namespaces/proxy-2583/pods/http:proxy-service-fsmzd-2gj5q:160/proxy/: foo (200; 27.904064ms)
+Jun 18 11:36:05.178: INFO: (7) /api/v1/namespaces/proxy-2583/pods/proxy-service-fsmzd-2gj5q:1080/proxy/: test<... (200; 28.419379ms)
+Jun 18 11:36:05.178: INFO: (7) /api/v1/namespaces/proxy-2583/services/proxy-service-fsmzd:portname1/proxy/: foo (200; 28.522101ms)
+Jun 18 11:36:05.178: INFO: (7) /api/v1/namespaces/proxy-2583/pods/http:proxy-service-fsmzd-2gj5q:1080/proxy/: ... (200; 29.17717ms)
+Jun 18 11:36:05.179: INFO: (7) /api/v1/namespaces/proxy-2583/services/https:proxy-service-fsmzd:tlsportname1/proxy/: tls baz (200; 28.698594ms)
+Jun 18 11:36:05.179: INFO: (7) /api/v1/namespaces/proxy-2583/services/https:proxy-service-fsmzd:tlsportname2/proxy/: tls qux (200; 29.744259ms)
+Jun 18 11:36:05.183: INFO: (7) /api/v1/namespaces/proxy-2583/pods/proxy-service-fsmzd-2gj5q:160/proxy/: foo (200; 33.642807ms)
+Jun 18 11:36:05.183: INFO: (7) /api/v1/namespaces/proxy-2583/services/proxy-service-fsmzd:portname2/proxy/: bar (200; 34.634496ms)
+Jun 18 11:36:05.184: INFO: (7) /api/v1/namespaces/proxy-2583/services/http:proxy-service-fsmzd:portname2/proxy/: bar (200; 34.619604ms)
+Jun 18 11:36:05.184: INFO: (7) /api/v1/namespaces/proxy-2583/services/http:proxy-service-fsmzd:portname1/proxy/: foo (200; 33.9007ms)
+Jun 18 11:36:05.203: INFO: (8) /api/v1/namespaces/proxy-2583/pods/http:proxy-service-fsmzd-2gj5q:162/proxy/: bar (200; 19.353751ms)
+Jun 18 11:36:05.208: INFO: (8) /api/v1/namespaces/proxy-2583/pods/https:proxy-service-fsmzd-2gj5q:443/proxy/: test (200; 24.970546ms)
+Jun 18 11:36:05.209: INFO: (8) /api/v1/namespaces/proxy-2583/pods/proxy-service-fsmzd-2gj5q:1080/proxy/: test<... (200; 25.060835ms)
+Jun 18 11:36:05.210: INFO: (8) /api/v1/namespaces/proxy-2583/pods/http:proxy-service-fsmzd-2gj5q:1080/proxy/: ... (200; 25.871305ms)
+Jun 18 11:36:05.210: INFO: (8) /api/v1/namespaces/proxy-2583/pods/http:proxy-service-fsmzd-2gj5q:160/proxy/: foo (200; 26.093702ms)
+Jun 18 11:36:05.210: INFO: (8) /api/v1/namespaces/proxy-2583/pods/proxy-service-fsmzd-2gj5q:162/proxy/: bar (200; 25.893687ms)
+Jun 18 11:36:05.210: INFO: (8) /api/v1/namespaces/proxy-2583/pods/https:proxy-service-fsmzd-2gj5q:460/proxy/: tls baz (200; 26.129628ms)
+Jun 18 11:36:05.215: INFO: (8) /api/v1/namespaces/proxy-2583/services/http:proxy-service-fsmzd:portname1/proxy/: foo (200; 30.656725ms)
+Jun 18 11:36:05.215: INFO: (8) /api/v1/namespaces/proxy-2583/services/proxy-service-fsmzd:portname2/proxy/: bar (200; 31.211039ms)
+Jun 18 11:36:05.215: INFO: (8) /api/v1/namespaces/proxy-2583/services/https:proxy-service-fsmzd:tlsportname2/proxy/: tls qux (200; 31.181052ms)
+Jun 18 11:36:05.215: INFO: (8) /api/v1/namespaces/proxy-2583/services/https:proxy-service-fsmzd:tlsportname1/proxy/: tls baz (200; 31.158807ms)
+Jun 18 11:36:05.216: INFO: (8) /api/v1/namespaces/proxy-2583/services/proxy-service-fsmzd:portname1/proxy/: foo (200; 31.519473ms)
+Jun 18 11:36:05.239: INFO: (9) /api/v1/namespaces/proxy-2583/pods/https:proxy-service-fsmzd-2gj5q:443/proxy/: test (200; 40.563426ms)
+Jun 18 11:36:05.257: INFO: (9) /api/v1/namespaces/proxy-2583/pods/https:proxy-service-fsmzd-2gj5q:460/proxy/: tls baz (200; 41.194791ms)
+Jun 18 11:36:05.257: INFO: (9) /api/v1/namespaces/proxy-2583/pods/http:proxy-service-fsmzd-2gj5q:1080/proxy/: ... (200; 41.103289ms)
+Jun 18 11:36:05.257: INFO: (9) /api/v1/namespaces/proxy-2583/pods/https:proxy-service-fsmzd-2gj5q:462/proxy/: tls qux (200; 41.209571ms)
+Jun 18 11:36:05.257: INFO: (9) /api/v1/namespaces/proxy-2583/pods/proxy-service-fsmzd-2gj5q:160/proxy/: foo (200; 41.175742ms)
+Jun 18 11:36:05.257: INFO: (9) /api/v1/namespaces/proxy-2583/services/proxy-service-fsmzd:portname2/proxy/: bar (200; 41.523903ms)
+Jun 18 11:36:05.257: INFO: (9) /api/v1/namespaces/proxy-2583/pods/http:proxy-service-fsmzd-2gj5q:162/proxy/: bar (200; 41.293094ms)
+Jun 18 11:36:05.257: INFO: (9) /api/v1/namespaces/proxy-2583/pods/http:proxy-service-fsmzd-2gj5q:160/proxy/: foo (200; 41.437437ms)
+Jun 18 11:36:05.265: INFO: (9) /api/v1/namespaces/proxy-2583/pods/proxy-service-fsmzd-2gj5q:1080/proxy/: test<... (200; 49.157913ms)
+Jun 18 11:36:05.265: INFO: (9) /api/v1/namespaces/proxy-2583/services/https:proxy-service-fsmzd:tlsportname2/proxy/: tls qux (200; 49.793855ms)
+Jun 18 11:36:05.266: INFO: (9) /api/v1/namespaces/proxy-2583/services/https:proxy-service-fsmzd:tlsportname1/proxy/: tls baz (200; 49.957174ms)
+Jun 18 11:36:05.271: INFO: (9) /api/v1/namespaces/proxy-2583/services/http:proxy-service-fsmzd:portname2/proxy/: bar (200; 54.783052ms)
+Jun 18 11:36:05.271: INFO: (9) /api/v1/namespaces/proxy-2583/services/proxy-service-fsmzd:portname1/proxy/: foo (200; 54.802014ms)
+Jun 18 11:36:05.271: INFO: (9) /api/v1/namespaces/proxy-2583/services/http:proxy-service-fsmzd:portname1/proxy/: foo (200; 54.714414ms)
+Jun 18 11:36:05.271: INFO: (9) /api/v1/namespaces/proxy-2583/pods/proxy-service-fsmzd-2gj5q:162/proxy/: bar (200; 54.971536ms)
+Jun 18 11:36:05.291: INFO: (10) /api/v1/namespaces/proxy-2583/pods/https:proxy-service-fsmzd-2gj5q:460/proxy/: tls baz (200; 19.147013ms)
+Jun 18 11:36:05.296: INFO: (10) /api/v1/namespaces/proxy-2583/pods/http:proxy-service-fsmzd-2gj5q:160/proxy/: foo (200; 24.654901ms)
+Jun 18 11:36:05.296: INFO: (10) /api/v1/namespaces/proxy-2583/pods/proxy-service-fsmzd-2gj5q:160/proxy/: foo (200; 24.69773ms)
+Jun 18 11:36:05.296: INFO: (10) /api/v1/namespaces/proxy-2583/pods/proxy-service-fsmzd-2gj5q:1080/proxy/: test<... (200; 24.766443ms)
+Jun 18 11:36:05.297: INFO: (10) /api/v1/namespaces/proxy-2583/pods/https:proxy-service-fsmzd-2gj5q:443/proxy/: ... (200; 25.314822ms)
+Jun 18 11:36:05.297: INFO: (10) /api/v1/namespaces/proxy-2583/pods/proxy-service-fsmzd-2gj5q:162/proxy/: bar (200; 25.675937ms)
+Jun 18 11:36:05.297: INFO: (10) /api/v1/namespaces/proxy-2583/services/https:proxy-service-fsmzd:tlsportname1/proxy/: tls baz (200; 25.818194ms)
+Jun 18 11:36:05.297: INFO: (10) /api/v1/namespaces/proxy-2583/pods/https:proxy-service-fsmzd-2gj5q:462/proxy/: tls qux (200; 25.310615ms)
+Jun 18 11:36:05.297: INFO: (10) /api/v1/namespaces/proxy-2583/pods/proxy-service-fsmzd-2gj5q/proxy/: test (200; 25.531568ms)
+Jun 18 11:36:05.297: INFO: (10) /api/v1/namespaces/proxy-2583/pods/http:proxy-service-fsmzd-2gj5q:162/proxy/: bar (200; 25.521314ms)
+Jun 18 11:36:05.302: INFO: (10) /api/v1/namespaces/proxy-2583/services/https:proxy-service-fsmzd:tlsportname2/proxy/: tls qux (200; 30.534015ms)
+Jun 18 11:36:05.302: INFO: (10) /api/v1/namespaces/proxy-2583/services/http:proxy-service-fsmzd:portname1/proxy/: foo (200; 30.442049ms)
+Jun 18 11:36:05.302: INFO: (10) /api/v1/namespaces/proxy-2583/services/proxy-service-fsmzd:portname1/proxy/: foo (200; 30.522851ms)
+Jun 18 11:36:05.302: INFO: (10) /api/v1/namespaces/proxy-2583/services/proxy-service-fsmzd:portname2/proxy/: bar (200; 31.022935ms)
+Jun 18 11:36:05.302: INFO: (10) /api/v1/namespaces/proxy-2583/services/http:proxy-service-fsmzd:portname2/proxy/: bar (200; 30.871252ms)
+Jun 18 11:36:05.320: INFO: (11) /api/v1/namespaces/proxy-2583/pods/http:proxy-service-fsmzd-2gj5q:1080/proxy/: ... (200; 17.477186ms)
+Jun 18 11:36:05.328: INFO: (11) /api/v1/namespaces/proxy-2583/pods/https:proxy-service-fsmzd-2gj5q:460/proxy/: tls baz (200; 25.311536ms)
+Jun 18 11:36:05.328: INFO: (11) /api/v1/namespaces/proxy-2583/services/http:proxy-service-fsmzd:portname2/proxy/: bar (200; 25.379139ms)
+Jun 18 11:36:05.328: INFO: (11) /api/v1/namespaces/proxy-2583/pods/proxy-service-fsmzd-2gj5q:1080/proxy/: test<... (200; 25.016965ms)
+Jun 18 11:36:05.328: INFO: (11) /api/v1/namespaces/proxy-2583/pods/proxy-service-fsmzd-2gj5q:162/proxy/: bar (200; 25.638655ms)
+Jun 18 11:36:05.328: INFO: (11) /api/v1/namespaces/proxy-2583/pods/proxy-service-fsmzd-2gj5q/proxy/: test (200; 25.302016ms)
+Jun 18 11:36:05.328: INFO: (11) /api/v1/namespaces/proxy-2583/pods/https:proxy-service-fsmzd-2gj5q:462/proxy/: tls qux (200; 25.690453ms)
+Jun 18 11:36:05.328: INFO: (11) /api/v1/namespaces/proxy-2583/pods/http:proxy-service-fsmzd-2gj5q:160/proxy/: foo (200; 25.532591ms)
+Jun 18 11:36:05.328: INFO: (11) /api/v1/namespaces/proxy-2583/pods/proxy-service-fsmzd-2gj5q:160/proxy/: foo (200; 25.418401ms)
+Jun 18 11:36:05.328: INFO: (11) /api/v1/namespaces/proxy-2583/pods/https:proxy-service-fsmzd-2gj5q:443/proxy/: test<... (200; 25.078974ms)
+Jun 18 11:36:05.359: INFO: (12) /api/v1/namespaces/proxy-2583/pods/proxy-service-fsmzd-2gj5q/proxy/: test (200; 24.675461ms)
+Jun 18 11:36:05.359: INFO: (12) /api/v1/namespaces/proxy-2583/pods/proxy-service-fsmzd-2gj5q:160/proxy/: foo (200; 25.03328ms)
+Jun 18 11:36:05.359: INFO: (12) /api/v1/namespaces/proxy-2583/pods/http:proxy-service-fsmzd-2gj5q:160/proxy/: foo (200; 25.178684ms)
+Jun 18 11:36:05.360: INFO: (12) /api/v1/namespaces/proxy-2583/pods/http:proxy-service-fsmzd-2gj5q:1080/proxy/: ... (200; 24.593524ms)
+Jun 18 11:36:05.364: INFO: (12) /api/v1/namespaces/proxy-2583/services/proxy-service-fsmzd:portname2/proxy/: bar (200; 30.12243ms)
+Jun 18 11:36:05.364: INFO: (12) /api/v1/namespaces/proxy-2583/services/https:proxy-service-fsmzd:tlsportname1/proxy/: tls baz (200; 29.568889ms)
+Jun 18 11:36:05.364: INFO: (12) /api/v1/namespaces/proxy-2583/services/http:proxy-service-fsmzd:portname2/proxy/: bar (200; 30.18753ms)
+Jun 18 11:36:05.364: INFO: (12) /api/v1/namespaces/proxy-2583/services/http:proxy-service-fsmzd:portname1/proxy/: foo (200; 29.432669ms)
+Jun 18 11:36:05.364: INFO: (12) /api/v1/namespaces/proxy-2583/services/proxy-service-fsmzd:portname1/proxy/: foo (200; 29.941926ms)
+Jun 18 11:36:05.383: INFO: (13) /api/v1/namespaces/proxy-2583/pods/proxy-service-fsmzd-2gj5q:162/proxy/: bar (200; 18.002773ms)
+Jun 18 11:36:05.383: INFO: (13) /api/v1/namespaces/proxy-2583/pods/http:proxy-service-fsmzd-2gj5q:160/proxy/: foo (200; 17.732258ms)
+Jun 18 11:36:05.389: INFO: (13) /api/v1/namespaces/proxy-2583/pods/proxy-service-fsmzd-2gj5q/proxy/: test (200; 23.45242ms)
+Jun 18 11:36:05.389: INFO: (13) /api/v1/namespaces/proxy-2583/pods/proxy-service-fsmzd-2gj5q:1080/proxy/: test<... (200; 23.895185ms)
+Jun 18 11:36:05.389: INFO: (13) /api/v1/namespaces/proxy-2583/pods/https:proxy-service-fsmzd-2gj5q:460/proxy/: tls baz (200; 23.335494ms)
+Jun 18 11:36:05.392: INFO: (13) /api/v1/namespaces/proxy-2583/pods/http:proxy-service-fsmzd-2gj5q:1080/proxy/: ... (200; 26.490768ms)
+Jun 18 11:36:05.392: INFO: (13) /api/v1/namespaces/proxy-2583/pods/proxy-service-fsmzd-2gj5q:160/proxy/: foo (200; 27.164878ms)
+Jun 18 11:36:05.392: INFO: (13) /api/v1/namespaces/proxy-2583/services/proxy-service-fsmzd:portname2/proxy/: bar (200; 27.689811ms)
+Jun 18 11:36:05.392: INFO: (13) /api/v1/namespaces/proxy-2583/pods/http:proxy-service-fsmzd-2gj5q:162/proxy/: bar (200; 26.79869ms)
+Jun 18 11:36:05.392: INFO: (13) /api/v1/namespaces/proxy-2583/pods/https:proxy-service-fsmzd-2gj5q:462/proxy/: tls qux (200; 26.431244ms)
+Jun 18 11:36:05.392: INFO: (13) /api/v1/namespaces/proxy-2583/pods/https:proxy-service-fsmzd-2gj5q:443/proxy/: test (200; 24.404056ms)
+Jun 18 11:36:05.424: INFO: (14) /api/v1/namespaces/proxy-2583/pods/https:proxy-service-fsmzd-2gj5q:460/proxy/: tls baz (200; 24.879153ms)
+Jun 18 11:36:05.424: INFO: (14) /api/v1/namespaces/proxy-2583/pods/http:proxy-service-fsmzd-2gj5q:160/proxy/: foo (200; 25.095609ms)
+Jun 18 11:36:05.424: INFO: (14) /api/v1/namespaces/proxy-2583/pods/http:proxy-service-fsmzd-2gj5q:1080/proxy/: ... (200; 25.326551ms)
+Jun 18 11:36:05.424: INFO: (14) /api/v1/namespaces/proxy-2583/pods/proxy-service-fsmzd-2gj5q:160/proxy/: foo (200; 25.348801ms)
+Jun 18 11:36:05.424: INFO: (14) /api/v1/namespaces/proxy-2583/pods/https:proxy-service-fsmzd-2gj5q:443/proxy/: test<... (200; 25.905911ms)
+Jun 18 11:36:05.425: INFO: (14) /api/v1/namespaces/proxy-2583/services/http:proxy-service-fsmzd:portname2/proxy/: bar (200; 26.122224ms)
+Jun 18 11:36:05.431: INFO: (14) /api/v1/namespaces/proxy-2583/pods/http:proxy-service-fsmzd-2gj5q:162/proxy/: bar (200; 31.625661ms)
+Jun 18 11:36:05.431: INFO: (14) /api/v1/namespaces/proxy-2583/services/https:proxy-service-fsmzd:tlsportname1/proxy/: tls baz (200; 31.71215ms)
+Jun 18 11:36:05.431: INFO: (14) /api/v1/namespaces/proxy-2583/services/proxy-service-fsmzd:portname2/proxy/: bar (200; 31.793248ms)
+Jun 18 11:36:05.431: INFO: (14) /api/v1/namespaces/proxy-2583/services/https:proxy-service-fsmzd:tlsportname2/proxy/: tls qux (200; 31.910412ms)
+Jun 18 11:36:05.431: INFO: (14) /api/v1/namespaces/proxy-2583/services/http:proxy-service-fsmzd:portname1/proxy/: foo (200; 32.106189ms)
+Jun 18 11:36:05.431: INFO: (14) /api/v1/namespaces/proxy-2583/services/proxy-service-fsmzd:portname1/proxy/: foo (200; 32.386209ms)
+Jun 18 11:36:05.451: INFO: (15) /api/v1/namespaces/proxy-2583/pods/http:proxy-service-fsmzd-2gj5q:160/proxy/: foo (200; 20.101724ms)
+Jun 18 11:36:05.451: INFO: (15) /api/v1/namespaces/proxy-2583/pods/proxy-service-fsmzd-2gj5q/proxy/: test (200; 19.970641ms)
+Jun 18 11:36:05.451: INFO: (15) /api/v1/namespaces/proxy-2583/pods/http:proxy-service-fsmzd-2gj5q:162/proxy/: bar (200; 20.335921ms)
+Jun 18 11:36:05.458: INFO: (15) /api/v1/namespaces/proxy-2583/pods/proxy-service-fsmzd-2gj5q:160/proxy/: foo (200; 26.365009ms)
+Jun 18 11:36:05.458: INFO: (15) /api/v1/namespaces/proxy-2583/services/https:proxy-service-fsmzd:tlsportname1/proxy/: tls baz (200; 26.227668ms)
+Jun 18 11:36:05.458: INFO: (15) /api/v1/namespaces/proxy-2583/pods/proxy-service-fsmzd-2gj5q:1080/proxy/: test<... (200; 26.442854ms)
+Jun 18 11:36:05.458: INFO: (15) /api/v1/namespaces/proxy-2583/pods/https:proxy-service-fsmzd-2gj5q:443/proxy/: ... (200; 26.964169ms)
+Jun 18 11:36:05.459: INFO: (15) /api/v1/namespaces/proxy-2583/pods/proxy-service-fsmzd-2gj5q:162/proxy/: bar (200; 27.005178ms)
+Jun 18 11:36:05.464: INFO: (15) /api/v1/namespaces/proxy-2583/services/http:proxy-service-fsmzd:portname1/proxy/: foo (200; 32.365731ms)
+Jun 18 11:36:05.464: INFO: (15) /api/v1/namespaces/proxy-2583/services/https:proxy-service-fsmzd:tlsportname2/proxy/: tls qux (200; 32.434565ms)
+Jun 18 11:36:05.464: INFO: (15) /api/v1/namespaces/proxy-2583/services/http:proxy-service-fsmzd:portname2/proxy/: bar (200; 32.481317ms)
+Jun 18 11:36:05.464: INFO: (15) /api/v1/namespaces/proxy-2583/services/proxy-service-fsmzd:portname2/proxy/: bar (200; 32.876685ms)
+Jun 18 11:36:05.484: INFO: (16) /api/v1/namespaces/proxy-2583/pods/https:proxy-service-fsmzd-2gj5q:462/proxy/: tls qux (200; 19.984993ms)
+Jun 18 11:36:05.489: INFO: (16) /api/v1/namespaces/proxy-2583/services/http:proxy-service-fsmzd:portname2/proxy/: bar (200; 24.508624ms)
+Jun 18 11:36:05.489: INFO: (16) /api/v1/namespaces/proxy-2583/pods/proxy-service-fsmzd-2gj5q:1080/proxy/: test<... (200; 24.398025ms)
+Jun 18 11:36:05.489: INFO: (16) /api/v1/namespaces/proxy-2583/pods/proxy-service-fsmzd-2gj5q/proxy/: test (200; 23.910647ms)
+Jun 18 11:36:05.490: INFO: (16) /api/v1/namespaces/proxy-2583/pods/https:proxy-service-fsmzd-2gj5q:460/proxy/: tls baz (200; 23.886306ms)
+Jun 18 11:36:05.490: INFO: (16) /api/v1/namespaces/proxy-2583/pods/proxy-service-fsmzd-2gj5q:160/proxy/: foo (200; 25.015006ms)
+Jun 18 11:36:05.490: INFO: (16) /api/v1/namespaces/proxy-2583/pods/http:proxy-service-fsmzd-2gj5q:160/proxy/: foo (200; 24.798791ms)
+Jun 18 11:36:05.490: INFO: (16) /api/v1/namespaces/proxy-2583/pods/http:proxy-service-fsmzd-2gj5q:162/proxy/: bar (200; 24.44571ms)
+Jun 18 11:36:05.490: INFO: (16) /api/v1/namespaces/proxy-2583/pods/http:proxy-service-fsmzd-2gj5q:1080/proxy/: ... (200; 24.158826ms)
+Jun 18 11:36:05.490: INFO: (16) /api/v1/namespaces/proxy-2583/pods/https:proxy-service-fsmzd-2gj5q:443/proxy/: test (200; 24.675855ms)
+Jun 18 11:36:05.522: INFO: (17) /api/v1/namespaces/proxy-2583/pods/http:proxy-service-fsmzd-2gj5q:1080/proxy/: ... (200; 25.10003ms)
+Jun 18 11:36:05.522: INFO: (17) /api/v1/namespaces/proxy-2583/services/http:proxy-service-fsmzd:portname1/proxy/: foo (200; 25.668527ms)
+Jun 18 11:36:05.522: INFO: (17) /api/v1/namespaces/proxy-2583/pods/http:proxy-service-fsmzd-2gj5q:160/proxy/: foo (200; 25.516587ms)
+Jun 18 11:36:05.523: INFO: (17) /api/v1/namespaces/proxy-2583/pods/https:proxy-service-fsmzd-2gj5q:460/proxy/: tls baz (200; 26.054198ms)
+Jun 18 11:36:05.523: INFO: (17) /api/v1/namespaces/proxy-2583/pods/proxy-service-fsmzd-2gj5q:1080/proxy/: test<... (200; 25.929247ms)
+Jun 18 11:36:05.523: INFO: (17) /api/v1/namespaces/proxy-2583/pods/https:proxy-service-fsmzd-2gj5q:462/proxy/: tls qux (200; 26.058618ms)
+Jun 18 11:36:05.522: INFO: (17) /api/v1/namespaces/proxy-2583/pods/http:proxy-service-fsmzd-2gj5q:162/proxy/: bar (200; 25.668984ms)
+Jun 18 11:36:05.527: INFO: (17) /api/v1/namespaces/proxy-2583/services/proxy-service-fsmzd:portname2/proxy/: bar (200; 30.610361ms)
+Jun 18 11:36:05.527: INFO: (17) /api/v1/namespaces/proxy-2583/services/https:proxy-service-fsmzd:tlsportname2/proxy/: tls qux (200; 30.796138ms)
+Jun 18 11:36:05.528: INFO: (17) /api/v1/namespaces/proxy-2583/services/https:proxy-service-fsmzd:tlsportname1/proxy/: tls baz (200; 30.722595ms)
+Jun 18 11:36:05.528: INFO: (17) /api/v1/namespaces/proxy-2583/services/proxy-service-fsmzd:portname1/proxy/: foo (200; 31.047157ms)
+Jun 18 11:36:05.528: INFO: (17) /api/v1/namespaces/proxy-2583/pods/proxy-service-fsmzd-2gj5q:160/proxy/: foo (200; 31.000118ms)
+Jun 18 11:36:05.528: INFO: (17) /api/v1/namespaces/proxy-2583/services/http:proxy-service-fsmzd:portname2/proxy/: bar (200; 31.238328ms)
+Jun 18 11:36:05.547: INFO: (18) /api/v1/namespaces/proxy-2583/pods/https:proxy-service-fsmzd-2gj5q:443/proxy/: test (200; 25.188348ms)
+Jun 18 11:36:05.553: INFO: (18) /api/v1/namespaces/proxy-2583/pods/proxy-service-fsmzd-2gj5q:162/proxy/: bar (200; 25.317638ms)
+Jun 18 11:36:05.553: INFO: (18) /api/v1/namespaces/proxy-2583/pods/proxy-service-fsmzd-2gj5q:160/proxy/: foo (200; 25.325308ms)
+Jun 18 11:36:05.553: INFO: (18) /api/v1/namespaces/proxy-2583/pods/http:proxy-service-fsmzd-2gj5q:162/proxy/: bar (200; 25.288123ms)
+Jun 18 11:36:05.553: INFO: (18) /api/v1/namespaces/proxy-2583/pods/proxy-service-fsmzd-2gj5q:1080/proxy/: test<... (200; 25.409898ms)
+Jun 18 11:36:05.554: INFO: (18) /api/v1/namespaces/proxy-2583/pods/https:proxy-service-fsmzd-2gj5q:460/proxy/: tls baz (200; 25.768256ms)
+Jun 18 11:36:05.554: INFO: (18) /api/v1/namespaces/proxy-2583/pods/http:proxy-service-fsmzd-2gj5q:1080/proxy/: ... (200; 25.491882ms)
+Jun 18 11:36:05.559: INFO: (18) /api/v1/namespaces/proxy-2583/services/https:proxy-service-fsmzd:tlsportname1/proxy/: tls baz (200; 30.876042ms)
+Jun 18 11:36:05.559: INFO: (18) /api/v1/namespaces/proxy-2583/services/https:proxy-service-fsmzd:tlsportname2/proxy/: tls qux (200; 31.342329ms)
+Jun 18 11:36:05.559: INFO: (18) /api/v1/namespaces/proxy-2583/services/http:proxy-service-fsmzd:portname2/proxy/: bar (200; 30.804557ms)
+Jun 18 11:36:05.559: INFO: (18) /api/v1/namespaces/proxy-2583/services/proxy-service-fsmzd:portname2/proxy/: bar (200; 30.862334ms)
+Jun 18 11:36:05.559: INFO: (18) /api/v1/namespaces/proxy-2583/services/proxy-service-fsmzd:portname1/proxy/: foo (200; 31.077639ms)
+Jun 18 11:36:05.559: INFO: (18) /api/v1/namespaces/proxy-2583/services/http:proxy-service-fsmzd:portname1/proxy/: foo (200; 30.954765ms)
+Jun 18 11:36:05.577: INFO: (19) /api/v1/namespaces/proxy-2583/pods/proxy-service-fsmzd-2gj5q:162/proxy/: bar (200; 17.661033ms)
+Jun 18 11:36:05.587: INFO: (19) /api/v1/namespaces/proxy-2583/pods/http:proxy-service-fsmzd-2gj5q:162/proxy/: bar (200; 27.941581ms)
+Jun 18 11:36:05.587: INFO: (19) /api/v1/namespaces/proxy-2583/pods/http:proxy-service-fsmzd-2gj5q:160/proxy/: foo (200; 27.636143ms)
+Jun 18 11:36:05.588: INFO: (19) /api/v1/namespaces/proxy-2583/pods/http:proxy-service-fsmzd-2gj5q:1080/proxy/: ... (200; 27.972489ms)
+Jun 18 11:36:05.588: INFO: (19) /api/v1/namespaces/proxy-2583/pods/https:proxy-service-fsmzd-2gj5q:443/proxy/: test<... (200; 28.042368ms)
+Jun 18 11:36:05.588: INFO: (19) /api/v1/namespaces/proxy-2583/pods/proxy-service-fsmzd-2gj5q/proxy/: test (200; 28.725286ms)
+Jun 18 11:36:05.588: INFO: (19) /api/v1/namespaces/proxy-2583/pods/https:proxy-service-fsmzd-2gj5q:462/proxy/: tls qux (200; 28.976997ms)
+Jun 18 11:36:05.589: INFO: (19) /api/v1/namespaces/proxy-2583/pods/https:proxy-service-fsmzd-2gj5q:460/proxy/: tls baz (200; 28.998107ms)
+Jun 18 11:36:05.597: INFO: (19) /api/v1/namespaces/proxy-2583/services/proxy-service-fsmzd:portname2/proxy/: bar (200; 37.396726ms)
+Jun 18 11:36:05.597: INFO: (19) /api/v1/namespaces/proxy-2583/services/http:proxy-service-fsmzd:portname2/proxy/: bar (200; 37.638003ms)
+Jun 18 11:36:05.597: INFO: (19) /api/v1/namespaces/proxy-2583/services/http:proxy-service-fsmzd:portname1/proxy/: foo (200; 37.671738ms)
+Jun 18 11:36:05.598: INFO: (19) /api/v1/namespaces/proxy-2583/services/https:proxy-service-fsmzd:tlsportname2/proxy/: tls qux (200; 38.049789ms)
+Jun 18 11:36:05.598: INFO: (19) /api/v1/namespaces/proxy-2583/services/proxy-service-fsmzd:portname1/proxy/: foo (200; 38.33986ms)
+Jun 18 11:36:05.598: INFO: (19) /api/v1/namespaces/proxy-2583/services/https:proxy-service-fsmzd:tlsportname1/proxy/: tls baz (200; 38.166098ms)
+STEP: deleting ReplicationController proxy-service-fsmzd in namespace proxy-2583, will wait for the garbage collector to delete the pods
+Jun 18 11:36:05.701: INFO: Deleting ReplicationController proxy-service-fsmzd took: 40.893306ms
+Jun 18 11:36:05.902: INFO: Terminating ReplicationController proxy-service-fsmzd pods took: 200.253121ms
 [AfterEach] version v1
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 18:29:48.005: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "proxy-5472" for this suite.
-Jun  4 18:29:54.062: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 18:29:54.412: INFO: namespace proxy-5472 deletion completed in 6.394846502s
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 11:36:07.602: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "proxy-2583" for this suite.
+Jun 18 11:36:15.643: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 11:36:16.054: INFO: namespace proxy-2583 deletion completed in 8.437603153s
 
-• [SLOW TEST:7.005 seconds]
+• [SLOW TEST:21.454 seconds]
 [sig-network] Proxy
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:22
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:22
   version v1
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/proxy.go:56
-    should proxy logs on node with explicit kubelet port using proxy subresource  [Conformance]
-    /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/proxy.go:56
+    should proxy through a service and a pod  [Conformance]
+    /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-SSSSSSSSSSSSSSSSSS
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-network] Services 
-  should serve multiport endpoints from pods  [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [sig-network] Services
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+[sig-node] ConfigMap 
+  should fail to create ConfigMap with empty key [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-node] ConfigMap
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 18:29:54.412: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-STEP: Building a namespace api object, basename services
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in services-8045
+Jun 18 11:36:16.056: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename configmap
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in configmap-323
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-network] Services
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:86
-[It] should serve multiport endpoints from pods  [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: creating service multi-endpoint-test in namespace services-8045
-STEP: waiting up to 3m0s for service multi-endpoint-test in namespace services-8045 to expose endpoints map[]
-Jun  4 18:29:54.672: INFO: successfully validated that service multi-endpoint-test in namespace services-8045 exposes endpoints map[] (11.223599ms elapsed)
-STEP: Creating pod pod1 in namespace services-8045
-STEP: waiting up to 3m0s for service multi-endpoint-test in namespace services-8045 to expose endpoints map[pod1:[100]]
-Jun  4 18:29:58.992: INFO: successfully validated that service multi-endpoint-test in namespace services-8045 exposes endpoints map[pod1:[100]] (4.302540919s elapsed)
-STEP: Creating pod pod2 in namespace services-8045
-STEP: waiting up to 3m0s for service multi-endpoint-test in namespace services-8045 to expose endpoints map[pod1:[100] pod2:[101]]
-Jun  4 18:30:03.254: INFO: successfully validated that service multi-endpoint-test in namespace services-8045 exposes endpoints map[pod1:[100] pod2:[101]] (4.246812496s elapsed)
-STEP: Deleting pod pod1 in namespace services-8045
-STEP: waiting up to 3m0s for service multi-endpoint-test in namespace services-8045 to expose endpoints map[pod2:[101]]
-Jun  4 18:30:03.302: INFO: successfully validated that service multi-endpoint-test in namespace services-8045 exposes endpoints map[pod2:[101]] (24.381248ms elapsed)
-STEP: Deleting pod pod2 in namespace services-8045
-STEP: waiting up to 3m0s for service multi-endpoint-test in namespace services-8045 to expose endpoints map[]
-Jun  4 18:30:03.332: INFO: successfully validated that service multi-endpoint-test in namespace services-8045 exposes endpoints map[] (10.372225ms elapsed)
-[AfterEach] [sig-network] Services
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 18:30:03.471: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "services-8045" for this suite.
-Jun  4 18:30:27.550: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 18:30:27.912: INFO: namespace services-8045 deletion completed in 24.414718461s
-[AfterEach] [sig-network] Services
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:91
+[It] should fail to create ConfigMap with empty key [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating configMap that has name configmap-test-emptyKey-487e1719-91bd-11e9-a25d-8608290c688a
+[AfterEach] [sig-node] ConfigMap
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 11:36:16.261: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "configmap-323" for this suite.
+Jun 18 11:36:22.301: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 11:36:22.688: INFO: namespace configmap-323 deletion completed in 6.412567431s
 
-• [SLOW TEST:33.500 seconds]
-[sig-network] Services
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:22
-  should serve multiport endpoints from pods  [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-------------------------------
-SSSSSSSS
+• [SLOW TEST:6.632 seconds]
+[sig-node] ConfigMap
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/configmap.go:32
+  should fail to create ConfigMap with empty key [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-[k8s.io] Probing container 
-  should be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [k8s.io] Probing container
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+[sig-storage] Downward API volume 
+  should provide container's memory limit [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-storage] Downward API volume
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 18:30:27.912: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-STEP: Building a namespace api object, basename container-probe
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in container-probe-2148
+Jun 18 11:36:22.689: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename downward-api
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in downward-api-102
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [k8s.io] Probing container
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/container_probe.go:51
-[It] should be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: Creating pod liveness-exec in namespace container-probe-2148
-Jun  4 18:30:34.188: INFO: Started pod liveness-exec in namespace container-probe-2148
-STEP: checking the pod's current state and verifying that restartCount is present
-Jun  4 18:30:34.201: INFO: Initial restart count of pod liveness-exec is 0
-Jun  4 18:31:26.635: INFO: Restart count of pod container-probe-2148/liveness-exec is now 1 (52.434281344s elapsed)
-STEP: deleting the pod
-[AfterEach] [k8s.io] Probing container
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 18:31:26.672: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "container-probe-2148" for this suite.
-Jun  4 18:31:32.802: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 18:31:33.311: INFO: namespace container-probe-2148 deletion completed in 6.625085979s
+[BeforeEach] [sig-storage] Downward API volume
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:39
+[It] should provide container's memory limit [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating a pod to test downward API volume plugin
+Jun 18 11:36:22.907: INFO: Waiting up to 5m0s for pod "downwardapi-volume-4c70e794-91bd-11e9-a25d-8608290c688a" in namespace "downward-api-102" to be "success or failure"
+Jun 18 11:36:22.920: INFO: Pod "downwardapi-volume-4c70e794-91bd-11e9-a25d-8608290c688a": Phase="Pending", Reason="", readiness=false. Elapsed: 12.918418ms
+Jun 18 11:36:25.465: INFO: Pod "downwardapi-volume-4c70e794-91bd-11e9-a25d-8608290c688a": Phase="Running", Reason="", readiness=true. Elapsed: 2.557723329s
+Jun 18 11:36:27.479: INFO: Pod "downwardapi-volume-4c70e794-91bd-11e9-a25d-8608290c688a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.571646235s
+STEP: Saw pod success
+Jun 18 11:36:27.479: INFO: Pod "downwardapi-volume-4c70e794-91bd-11e9-a25d-8608290c688a" satisfied condition "success or failure"
+Jun 18 11:36:27.492: INFO: Trying to get logs from node 10.72.74.138 pod downwardapi-volume-4c70e794-91bd-11e9-a25d-8608290c688a container client-container: 
+STEP: delete the pod
+Jun 18 11:36:27.608: INFO: Waiting for pod downwardapi-volume-4c70e794-91bd-11e9-a25d-8608290c688a to disappear
+Jun 18 11:36:27.627: INFO: Pod downwardapi-volume-4c70e794-91bd-11e9-a25d-8608290c688a no longer exists
+[AfterEach] [sig-storage] Downward API volume
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 11:36:27.627: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "downward-api-102" for this suite.
+Jun 18 11:36:33.665: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 11:36:34.080: INFO: namespace downward-api-102 deletion completed in 6.439165422s
 
-• [SLOW TEST:65.399 seconds]
-[k8s.io] Probing container
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
-  should be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+• [SLOW TEST:11.392 seconds]
+[sig-storage] Downward API volume
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:34
+  should provide container's memory limit [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+SSSSSSSSS
 ------------------------------
-[sig-network] DNS 
-  should provide DNS for the cluster  [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [sig-network] DNS
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+[k8s.io] Variable Expansion 
+  should allow composing env vars into new env vars [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [k8s.io] Variable Expansion
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 18:31:33.312: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-STEP: Building a namespace api object, basename dns
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in dns-4097
+Jun 18 11:36:34.080: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename var-expansion
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in var-expansion-6590
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should provide DNS for the cluster  [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: Running these commands on wheezy: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search kubernetes.default.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_udp@kubernetes.default.svc.cluster.local;check="$$(dig +tcp +noall +answer +search kubernetes.default.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@kubernetes.default.svc.cluster.local;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-4097.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@PodARecord;sleep 1; done
-
-STEP: Running these commands on jessie: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search kubernetes.default.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_udp@kubernetes.default.svc.cluster.local;check="$$(dig +tcp +noall +answer +search kubernetes.default.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_tcp@kubernetes.default.svc.cluster.local;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-4097.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_tcp@PodARecord;sleep 1; done
-
-STEP: creating a pod to probe DNS
-STEP: submitting the pod to kubernetes
-STEP: retrieving the pod
-STEP: looking for the results for each expected name from probers
-Jun  4 18:31:35.729: INFO: DNS probes using dns-4097/dns-test-fa908a54-86f6-11e9-8d1b-467ee19922ac succeeded
-
-STEP: deleting the pod
-[AfterEach] [sig-network] DNS
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 18:31:35.767: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "dns-4097" for this suite.
-Jun  4 18:31:41.828: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 18:31:42.178: INFO: namespace dns-4097 deletion completed in 6.39082238s
+[It] should allow composing env vars into new env vars [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating a pod to test env composition
+Jun 18 11:36:34.316: INFO: Waiting up to 5m0s for pod "var-expansion-533dbd02-91bd-11e9-a25d-8608290c688a" in namespace "var-expansion-6590" to be "success or failure"
+Jun 18 11:36:34.328: INFO: Pod "var-expansion-533dbd02-91bd-11e9-a25d-8608290c688a": Phase="Pending", Reason="", readiness=false. Elapsed: 12.446469ms
+Jun 18 11:36:36.343: INFO: Pod "var-expansion-533dbd02-91bd-11e9-a25d-8608290c688a": Phase="Pending", Reason="", readiness=false. Elapsed: 2.027553315s
+Jun 18 11:36:38.357: INFO: Pod "var-expansion-533dbd02-91bd-11e9-a25d-8608290c688a": Phase="Running", Reason="", readiness=true. Elapsed: 4.041563661s
+Jun 18 11:36:40.371: INFO: Pod "var-expansion-533dbd02-91bd-11e9-a25d-8608290c688a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.055108342s
+STEP: Saw pod success
+Jun 18 11:36:40.371: INFO: Pod "var-expansion-533dbd02-91bd-11e9-a25d-8608290c688a" satisfied condition "success or failure"
+Jun 18 11:36:40.387: INFO: Trying to get logs from node 10.72.74.184 pod var-expansion-533dbd02-91bd-11e9-a25d-8608290c688a container dapi-container: 
+STEP: delete the pod
+Jun 18 11:36:40.475: INFO: Waiting for pod var-expansion-533dbd02-91bd-11e9-a25d-8608290c688a to disappear
+Jun 18 11:36:40.487: INFO: Pod var-expansion-533dbd02-91bd-11e9-a25d-8608290c688a no longer exists
+[AfterEach] [k8s.io] Variable Expansion
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 11:36:40.487: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "var-expansion-6590" for this suite.
+Jun 18 11:36:48.527: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 11:36:48.926: INFO: namespace var-expansion-6590 deletion completed in 8.42541647s
 
-• [SLOW TEST:8.866 seconds]
-[sig-network] DNS
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:22
-  should provide DNS for the cluster  [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+• [SLOW TEST:14.846 seconds]
+[k8s.io] Variable Expansion
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
+  should allow composing env vars into new env vars [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-SSSSSSS
+SSSSSSSSS
 ------------------------------
-[sig-cli] Kubectl client [k8s.io] Proxy server 
-  should support --unix-socket=/path  [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[sig-cli] Kubectl client [k8s.io] Kubectl api-versions 
+  should check if v1 is in available api versions  [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 [BeforeEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 18:31:42.178: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
+Jun 18 11:36:48.927: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
 STEP: Building a namespace api object, basename kubectl
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-3722
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-7484
 STEP: Waiting for a default service account to be provisioned in namespace
 [BeforeEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:213
-[It] should support --unix-socket=/path  [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: Starting the proxy
-Jun  4 18:31:42.465: INFO: Asynchronously running '/usr/local/bin/kubectl kubectl --kubeconfig=/tmp/kubeconfig-229234504 proxy --unix-socket=/tmp/kubectl-proxy-unix430166023/test'
-STEP: retrieving proxy /api/ output
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:213
+[It] should check if v1 is in available api versions  [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: validating api versions
+Jun 18 11:36:49.124: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 api-versions'
+Jun 18 11:36:49.270: INFO: stderr: ""
+Jun 18 11:36:49.270: INFO: stdout: "admissionregistration.k8s.io/v1beta1\napiextensions.k8s.io/v1beta1\napiregistration.k8s.io/v1\napiregistration.k8s.io/v1beta1\napps/v1\napps/v1beta1\napps/v1beta2\nauthentication.k8s.io/v1\nauthentication.k8s.io/v1beta1\nauthorization.k8s.io/v1\nauthorization.k8s.io/v1beta1\nautoscaling/v1\nautoscaling/v2beta1\nautoscaling/v2beta2\nbatch/v1\nbatch/v1beta1\nbatch/v2alpha1\ncertificates.k8s.io/v1beta1\ncoordination.k8s.io/v1\ncoordination.k8s.io/v1beta1\nevents.k8s.io/v1beta1\nextensions/v1beta1\nmetrics.k8s.io/v1beta1\nnetworking.k8s.io/v1\nnetworking.k8s.io/v1beta1\npolicy/v1beta1\nrbac.authorization.k8s.io/v1\nrbac.authorization.k8s.io/v1beta1\nscheduling.k8s.io/v1\nscheduling.k8s.io/v1beta1\nstorage.k8s.io/v1\nstorage.k8s.io/v1beta1\nv1\n"
 [AfterEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 18:31:42.509: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "kubectl-3722" for this suite.
-Jun  4 18:31:48.567: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 18:31:48.985: INFO: namespace kubectl-3722 deletion completed in 6.462581596s
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 11:36:49.270: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "kubectl-7484" for this suite.
+Jun 18 11:36:55.308: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 11:36:55.685: INFO: namespace kubectl-7484 deletion completed in 6.402370154s
 
-• [SLOW TEST:6.807 seconds]
+• [SLOW TEST:6.758 seconds]
 [sig-cli] Kubectl client
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23
-  [k8s.io] Proxy server
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
-    should support --unix-socket=/path  [Conformance]
-    /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23
+  [k8s.io] Kubectl api-versions
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
+    should check if v1 is in available api versions  [Conformance]
+    /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-[k8s.io] Container Lifecycle Hook when create a pod with lifecycle hook 
-  should execute prestop exec hook properly [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [k8s.io] Container Lifecycle Hook
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+SSSSSSS
+------------------------------
+[sig-storage] Secrets 
+  should be consumable in multiple volumes in a pod [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-storage] Secrets
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 18:31:48.985: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-STEP: Building a namespace api object, basename container-lifecycle-hook
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in container-lifecycle-hook-2429
+Jun 18 11:36:55.685: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename secrets
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in secrets-7507
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] when create a pod with lifecycle hook
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:61
-STEP: create the container to handle the HTTPGet hook request.
-[It] should execute prestop exec hook properly [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: create the pod with lifecycle hook
-STEP: delete the pod with lifecycle hook
-Jun  4 18:32:01.555: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear
-Jun  4 18:32:01.566: INFO: Pod pod-with-prestop-exec-hook still exists
-Jun  4 18:32:03.566: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear
-Jun  4 18:32:03.579: INFO: Pod pod-with-prestop-exec-hook still exists
-Jun  4 18:32:05.566: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear
-Jun  4 18:32:05.578: INFO: Pod pod-with-prestop-exec-hook still exists
-Jun  4 18:32:07.566: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear
-Jun  4 18:32:07.579: INFO: Pod pod-with-prestop-exec-hook still exists
-Jun  4 18:32:09.566: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear
-Jun  4 18:32:09.578: INFO: Pod pod-with-prestop-exec-hook still exists
-Jun  4 18:32:11.566: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear
-Jun  4 18:32:11.585: INFO: Pod pod-with-prestop-exec-hook still exists
-Jun  4 18:32:13.566: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear
-Jun  4 18:32:13.579: INFO: Pod pod-with-prestop-exec-hook still exists
-Jun  4 18:32:15.566: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear
-Jun  4 18:32:15.580: INFO: Pod pod-with-prestop-exec-hook still exists
-Jun  4 18:32:17.566: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear
-Jun  4 18:32:17.581: INFO: Pod pod-with-prestop-exec-hook still exists
-Jun  4 18:32:19.566: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear
-Jun  4 18:32:19.579: INFO: Pod pod-with-prestop-exec-hook no longer exists
-STEP: check prestop hook
-[AfterEach] [k8s.io] Container Lifecycle Hook
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 18:32:19.636: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "container-lifecycle-hook-2429" for this suite.
-Jun  4 18:32:43.696: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 18:32:44.095: INFO: namespace container-lifecycle-hook-2429 deletion completed in 24.443982548s
+[It] should be consumable in multiple volumes in a pod [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating secret with name secret-test-601bd85a-91bd-11e9-a25d-8608290c688a
+STEP: Creating a pod to test consume secrets
+Jun 18 11:36:55.917: INFO: Waiting up to 5m0s for pod "pod-secrets-601d8b10-91bd-11e9-a25d-8608290c688a" in namespace "secrets-7507" to be "success or failure"
+Jun 18 11:36:55.936: INFO: Pod "pod-secrets-601d8b10-91bd-11e9-a25d-8608290c688a": Phase="Pending", Reason="", readiness=false. Elapsed: 18.649134ms
+Jun 18 11:36:57.950: INFO: Pod "pod-secrets-601d8b10-91bd-11e9-a25d-8608290c688a": Phase="Running", Reason="", readiness=true. Elapsed: 2.032529244s
+Jun 18 11:36:59.963: INFO: Pod "pod-secrets-601d8b10-91bd-11e9-a25d-8608290c688a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.045920456s
+STEP: Saw pod success
+Jun 18 11:36:59.963: INFO: Pod "pod-secrets-601d8b10-91bd-11e9-a25d-8608290c688a" satisfied condition "success or failure"
+Jun 18 11:36:59.976: INFO: Trying to get logs from node 10.72.74.138 pod pod-secrets-601d8b10-91bd-11e9-a25d-8608290c688a container secret-volume-test: 
+STEP: delete the pod
+Jun 18 11:37:00.054: INFO: Waiting for pod pod-secrets-601d8b10-91bd-11e9-a25d-8608290c688a to disappear
+Jun 18 11:37:00.068: INFO: Pod pod-secrets-601d8b10-91bd-11e9-a25d-8608290c688a no longer exists
+[AfterEach] [sig-storage] Secrets
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 11:37:00.068: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "secrets-7507" for this suite.
+Jun 18 11:37:06.117: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 11:37:06.550: INFO: namespace secrets-7507 deletion completed in 6.460555591s
 
-• [SLOW TEST:55.110 seconds]
-[k8s.io] Container Lifecycle Hook
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
-  when create a pod with lifecycle hook
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:40
-    should execute prestop exec hook properly [NodeConformance] [Conformance]
-    /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+• [SLOW TEST:10.865 seconds]
+[sig-storage] Secrets
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/secrets_volume.go:33
+  should be consumable in multiple volumes in a pod [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-SS
+SSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[k8s.io] Container Lifecycle Hook when create a pod with lifecycle hook 
-  should execute poststart http hook properly [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [k8s.io] Container Lifecycle Hook
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+[sig-storage] EmptyDir volumes 
+  should support (non-root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-storage] EmptyDir volumes
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 18:32:44.095: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-STEP: Building a namespace api object, basename container-lifecycle-hook
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in container-lifecycle-hook-6776
+Jun 18 11:37:06.550: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename emptydir
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in emptydir-9145
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] when create a pod with lifecycle hook
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:61
-STEP: create the container to handle the HTTPGet hook request.
-[It] should execute poststart http hook properly [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: create the pod with lifecycle hook
-STEP: check poststart hook
-STEP: delete the pod with lifecycle hook
-Jun  4 18:32:54.555: INFO: Waiting for pod pod-with-poststart-http-hook to disappear
-Jun  4 18:32:54.574: INFO: Pod pod-with-poststart-http-hook still exists
-Jun  4 18:32:56.575: INFO: Waiting for pod pod-with-poststart-http-hook to disappear
-Jun  4 18:32:56.587: INFO: Pod pod-with-poststart-http-hook still exists
-Jun  4 18:32:58.575: INFO: Waiting for pod pod-with-poststart-http-hook to disappear
-Jun  4 18:32:58.655: INFO: Pod pod-with-poststart-http-hook no longer exists
-[AfterEach] [k8s.io] Container Lifecycle Hook
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 18:32:58.655: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "container-lifecycle-hook-6776" for this suite.
-Jun  4 18:33:22.717: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 18:33:23.051: INFO: namespace container-lifecycle-hook-6776 deletion completed in 24.378677153s
+[It] should support (non-root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating a pod to test emptydir 0777 on tmpfs
+Jun 18 11:37:06.772: INFO: Waiting up to 5m0s for pod "pod-66960615-91bd-11e9-a25d-8608290c688a" in namespace "emptydir-9145" to be "success or failure"
+Jun 18 11:37:06.784: INFO: Pod "pod-66960615-91bd-11e9-a25d-8608290c688a": Phase="Pending", Reason="", readiness=false. Elapsed: 12.458935ms
+Jun 18 11:37:08.798: INFO: Pod "pod-66960615-91bd-11e9-a25d-8608290c688a": Phase="Pending", Reason="", readiness=false. Elapsed: 2.026083851s
+Jun 18 11:37:10.812: INFO: Pod "pod-66960615-91bd-11e9-a25d-8608290c688a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.040032722s
+STEP: Saw pod success
+Jun 18 11:37:10.812: INFO: Pod "pod-66960615-91bd-11e9-a25d-8608290c688a" satisfied condition "success or failure"
+Jun 18 11:37:10.826: INFO: Trying to get logs from node 10.72.74.138 pod pod-66960615-91bd-11e9-a25d-8608290c688a container test-container: 
+STEP: delete the pod
+Jun 18 11:37:11.133: INFO: Waiting for pod pod-66960615-91bd-11e9-a25d-8608290c688a to disappear
+Jun 18 11:37:11.160: INFO: Pod pod-66960615-91bd-11e9-a25d-8608290c688a no longer exists
+[AfterEach] [sig-storage] EmptyDir volumes
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 11:37:11.160: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "emptydir-9145" for this suite.
+Jun 18 11:37:17.202: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 11:37:18.816: INFO: namespace emptydir-9145 deletion completed in 7.642085037s
 
-• [SLOW TEST:38.955 seconds]
-[k8s.io] Container Lifecycle Hook
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
-  when create a pod with lifecycle hook
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:40
-    should execute poststart http hook properly [NodeConformance] [Conformance]
-    /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+• [SLOW TEST:12.266 seconds]
+[sig-storage] EmptyDir volumes
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:41
+  should support (non-root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-SS
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-cli] Kubectl client [k8s.io] Kubectl cluster-info 
-  should check if Kubernetes master services is included in cluster-info  [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+[sig-storage] EmptyDir volumes 
+  should support (non-root,0666,default) [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-storage] EmptyDir volumes
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 18:33:23.051: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-STEP: Building a namespace api object, basename kubectl
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-7442
+Jun 18 11:37:18.817: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename emptydir
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in emptydir-821
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:213
-[It] should check if Kubernetes master services is included in cluster-info  [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: validating cluster-info
-Jun  4 18:33:23.268: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 cluster-info'
-Jun  4 18:33:23.355: INFO: stderr: ""
-Jun  4 18:33:23.355: INFO: stdout: "\x1b[0;32mKubernetes master\x1b[0m is running at \x1b[0;33mhttps://172.21.0.1:443\x1b[0m\n\x1b[0;32mCoreDNS\x1b[0m is running at \x1b[0;33mhttps://172.21.0.1:443/api/v1/namespaces/kube-system/services/kube-dns:dns/proxy\x1b[0m\n\x1b[0;32mkubernetes-dashboard\x1b[0m is running at \x1b[0;33mhttps://172.21.0.1:443/api/v1/namespaces/kube-system/services/https:kubernetes-dashboard:/proxy\x1b[0m\n\x1b[0;32mMetrics-server\x1b[0m is running at \x1b[0;33mhttps://172.21.0.1:443/api/v1/namespaces/kube-system/services/https:metrics-server:/proxy\x1b[0m\n\nTo further debug and diagnose cluster problems, use 'kubectl cluster-info dump'.\n"
-[AfterEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 18:33:23.355: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "kubectl-7442" for this suite.
-Jun  4 18:33:29.433: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 18:33:29.757: INFO: namespace kubectl-7442 deletion completed in 6.388138047s
+[It] should support (non-root,0666,default) [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating a pod to test emptydir 0666 on node default medium
+Jun 18 11:37:19.054: INFO: Waiting up to 5m0s for pod "pod-6de6dc45-91bd-11e9-a25d-8608290c688a" in namespace "emptydir-821" to be "success or failure"
+Jun 18 11:37:19.074: INFO: Pod "pod-6de6dc45-91bd-11e9-a25d-8608290c688a": Phase="Pending", Reason="", readiness=false. Elapsed: 20.013835ms
+Jun 18 11:37:21.088: INFO: Pod "pod-6de6dc45-91bd-11e9-a25d-8608290c688a": Phase="Running", Reason="", readiness=true. Elapsed: 2.033882863s
+Jun 18 11:37:23.756: INFO: Pod "pod-6de6dc45-91bd-11e9-a25d-8608290c688a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.702186856s
+STEP: Saw pod success
+Jun 18 11:37:23.757: INFO: Pod "pod-6de6dc45-91bd-11e9-a25d-8608290c688a" satisfied condition "success or failure"
+Jun 18 11:37:23.769: INFO: Trying to get logs from node 10.72.74.138 pod pod-6de6dc45-91bd-11e9-a25d-8608290c688a container test-container: 
+STEP: delete the pod
+Jun 18 11:37:23.846: INFO: Waiting for pod pod-6de6dc45-91bd-11e9-a25d-8608290c688a to disappear
+Jun 18 11:37:23.859: INFO: Pod pod-6de6dc45-91bd-11e9-a25d-8608290c688a no longer exists
+[AfterEach] [sig-storage] EmptyDir volumes
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 11:37:23.859: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "emptydir-821" for this suite.
+Jun 18 11:37:31.898: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 11:37:32.292: INFO: namespace emptydir-821 deletion completed in 8.420149419s
 
-• [SLOW TEST:6.706 seconds]
-[sig-cli] Kubectl client
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23
-  [k8s.io] Kubectl cluster-info
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
-    should check if Kubernetes master services is included in cluster-info  [Conformance]
-    /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+• [SLOW TEST:13.476 seconds]
+[sig-storage] EmptyDir volumes
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:41
+  should support (non-root,0666,default) [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-SSSSS
+SSSSSSSSS
 ------------------------------
-[sig-cli] Kubectl client [k8s.io] Update Demo 
-  should create and stop a replication controller  [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+[sig-storage] Projected secret 
+  should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-storage] Projected secret
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 18:33:29.758: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-STEP: Building a namespace api object, basename kubectl
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-3468
+Jun 18 11:37:32.293: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename projected
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-6180
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:213
-[BeforeEach] [k8s.io] Update Demo
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:265
-[It] should create and stop a replication controller  [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: creating a replication controller
-Jun  4 18:33:30.064: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 create -f - --namespace=kubectl-3468'
-Jun  4 18:33:30.237: INFO: stderr: ""
-Jun  4 18:33:30.237: INFO: stdout: "replicationcontroller/update-demo-nautilus created\n"
-STEP: waiting for all containers in name=update-demo pods to come up.
-Jun  4 18:33:30.237: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-3468'
-Jun  4 18:33:30.327: INFO: stderr: ""
-Jun  4 18:33:30.327: INFO: stdout: "update-demo-nautilus-xmrvg update-demo-nautilus-zmmdt "
-Jun  4 18:33:30.327: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 get pods update-demo-nautilus-xmrvg -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-3468'
-Jun  4 18:33:30.412: INFO: stderr: ""
-Jun  4 18:33:30.412: INFO: stdout: ""
-Jun  4 18:33:30.412: INFO: update-demo-nautilus-xmrvg is created but not running
-Jun  4 18:33:35.413: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-3468'
-Jun  4 18:33:35.590: INFO: stderr: ""
-Jun  4 18:33:35.590: INFO: stdout: "update-demo-nautilus-xmrvg update-demo-nautilus-zmmdt "
-Jun  4 18:33:35.590: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 get pods update-demo-nautilus-xmrvg -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-3468'
-Jun  4 18:33:35.682: INFO: stderr: ""
-Jun  4 18:33:35.682: INFO: stdout: "true"
-Jun  4 18:33:35.682: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 get pods update-demo-nautilus-xmrvg -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-3468'
-Jun  4 18:33:35.771: INFO: stderr: ""
-Jun  4 18:33:35.771: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0"
-Jun  4 18:33:35.771: INFO: validating pod update-demo-nautilus-xmrvg
-Jun  4 18:33:35.791: INFO: got data: {
-  "image": "nautilus.jpg"
-}
-
-Jun  4 18:33:35.791: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg .
-Jun  4 18:33:35.791: INFO: update-demo-nautilus-xmrvg is verified up and running
-Jun  4 18:33:35.791: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 get pods update-demo-nautilus-zmmdt -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-3468'
-Jun  4 18:33:35.886: INFO: stderr: ""
-Jun  4 18:33:35.886: INFO: stdout: "true"
-Jun  4 18:33:35.886: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 get pods update-demo-nautilus-zmmdt -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-3468'
-Jun  4 18:33:35.988: INFO: stderr: ""
-Jun  4 18:33:35.988: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0"
-Jun  4 18:33:35.988: INFO: validating pod update-demo-nautilus-zmmdt
-Jun  4 18:33:36.008: INFO: got data: {
-  "image": "nautilus.jpg"
-}
-
-Jun  4 18:33:36.008: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg .
-Jun  4 18:33:36.008: INFO: update-demo-nautilus-zmmdt is verified up and running
-STEP: using delete to clean up resources
-Jun  4 18:33:36.008: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 delete --grace-period=0 --force -f - --namespace=kubectl-3468'
-Jun  4 18:33:36.124: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n"
-Jun  4 18:33:36.124: INFO: stdout: "replicationcontroller \"update-demo-nautilus\" force deleted\n"
-Jun  4 18:33:36.124: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 get rc,svc -l name=update-demo --no-headers --namespace=kubectl-3468'
-Jun  4 18:33:36.223: INFO: stderr: "No resources found.\n"
-Jun  4 18:33:36.223: INFO: stdout: ""
-Jun  4 18:33:36.223: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 get pods -l name=update-demo --namespace=kubectl-3468 -o go-template={{ range .items }}{{ if not .metadata.deletionTimestamp }}{{ .metadata.name }}{{ "\n" }}{{ end }}{{ end }}'
-Jun  4 18:33:36.319: INFO: stderr: ""
-Jun  4 18:33:36.319: INFO: stdout: ""
-[AfterEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 18:33:36.320: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "kubectl-3468" for this suite.
-Jun  4 18:33:42.379: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 18:33:42.920: INFO: namespace kubectl-3468 deletion completed in 6.584389713s
+[It] should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating projection with secret that has name projected-secret-test-75eda075-91bd-11e9-a25d-8608290c688a
+STEP: Creating a pod to test consume secrets
+Jun 18 11:37:32.525: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-75ef73cf-91bd-11e9-a25d-8608290c688a" in namespace "projected-6180" to be "success or failure"
+Jun 18 11:37:32.538: INFO: Pod "pod-projected-secrets-75ef73cf-91bd-11e9-a25d-8608290c688a": Phase="Pending", Reason="", readiness=false. Elapsed: 12.692251ms
+Jun 18 11:37:34.553: INFO: Pod "pod-projected-secrets-75ef73cf-91bd-11e9-a25d-8608290c688a": Phase="Pending", Reason="", readiness=false. Elapsed: 2.027982601s
+Jun 18 11:37:36.567: INFO: Pod "pod-projected-secrets-75ef73cf-91bd-11e9-a25d-8608290c688a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.041837211s
+STEP: Saw pod success
+Jun 18 11:37:36.567: INFO: Pod "pod-projected-secrets-75ef73cf-91bd-11e9-a25d-8608290c688a" satisfied condition "success or failure"
+Jun 18 11:37:36.587: INFO: Trying to get logs from node 10.72.74.184 pod pod-projected-secrets-75ef73cf-91bd-11e9-a25d-8608290c688a container projected-secret-volume-test: 
+STEP: delete the pod
+Jun 18 11:37:36.671: INFO: Waiting for pod pod-projected-secrets-75ef73cf-91bd-11e9-a25d-8608290c688a to disappear
+Jun 18 11:37:36.693: INFO: Pod pod-projected-secrets-75ef73cf-91bd-11e9-a25d-8608290c688a no longer exists
+[AfterEach] [sig-storage] Projected secret
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 11:37:36.693: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "projected-6180" for this suite.
+Jun 18 11:37:42.733: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 11:37:43.180: INFO: namespace projected-6180 deletion completed in 6.472250725s
 
-• [SLOW TEST:13.163 seconds]
-[sig-cli] Kubectl client
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23
-  [k8s.io] Update Demo
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
-    should create and stop a replication controller  [Conformance]
-    /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+• [SLOW TEST:10.888 seconds]
+[sig-storage] Projected secret
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_secret.go:33
+  should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-SSSSSSSSSSSSSSSSSSSS
+SSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-auth] ServiceAccounts 
-  should allow opting out of API token automount  [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [sig-auth] ServiceAccounts
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+[sig-api-machinery] Garbage collector 
+  should orphan RS created by deployment when deleteOptions.PropagationPolicy is Orphan [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-api-machinery] Garbage collector
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 18:33:42.921: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-STEP: Building a namespace api object, basename svcaccounts
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in svcaccounts-1412
+Jun 18 11:37:43.183: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename gc
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in gc-1887
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should allow opting out of API token automount  [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: getting the auto-created API token
-Jun  4 18:33:43.777: INFO: created pod pod-service-account-defaultsa
-Jun  4 18:33:43.777: INFO: pod pod-service-account-defaultsa service account token volume mount: true
-Jun  4 18:33:43.793: INFO: created pod pod-service-account-mountsa
-Jun  4 18:33:43.794: INFO: pod pod-service-account-mountsa service account token volume mount: true
-Jun  4 18:33:43.807: INFO: created pod pod-service-account-nomountsa
-Jun  4 18:33:43.807: INFO: pod pod-service-account-nomountsa service account token volume mount: false
-Jun  4 18:33:43.829: INFO: created pod pod-service-account-defaultsa-mountspec
-Jun  4 18:33:43.829: INFO: pod pod-service-account-defaultsa-mountspec service account token volume mount: true
-Jun  4 18:33:43.842: INFO: created pod pod-service-account-mountsa-mountspec
-Jun  4 18:33:43.842: INFO: pod pod-service-account-mountsa-mountspec service account token volume mount: true
-Jun  4 18:33:43.857: INFO: created pod pod-service-account-nomountsa-mountspec
-Jun  4 18:33:43.857: INFO: pod pod-service-account-nomountsa-mountspec service account token volume mount: true
-Jun  4 18:33:43.874: INFO: created pod pod-service-account-defaultsa-nomountspec
-Jun  4 18:33:43.874: INFO: pod pod-service-account-defaultsa-nomountspec service account token volume mount: false
-Jun  4 18:33:43.888: INFO: created pod pod-service-account-mountsa-nomountspec
-Jun  4 18:33:43.888: INFO: pod pod-service-account-mountsa-nomountspec service account token volume mount: false
-Jun  4 18:33:43.900: INFO: created pod pod-service-account-nomountsa-nomountspec
-Jun  4 18:33:43.900: INFO: pod pod-service-account-nomountsa-nomountspec service account token volume mount: false
-[AfterEach] [sig-auth] ServiceAccounts
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 18:33:43.901: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "svcaccounts-1412" for this suite.
-Jun  4 18:34:07.966: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 18:34:08.389: INFO: namespace svcaccounts-1412 deletion completed in 24.472849525s
+[It] should orphan RS created by deployment when deleteOptions.PropagationPolicy is Orphan [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: create the deployment
+STEP: Wait for the Deployment to create new ReplicaSet
+STEP: delete the deployment
+STEP: wait for 30 seconds to see if the garbage collector mistakenly deletes the rs
+STEP: Gathering metrics
+W0618 11:38:13.499864      20 metrics_grabber.go:79] Master node is not registered. Grabbing metrics from Scheduler, ControllerManager and ClusterAutoscaler is disabled.
+Jun 18 11:38:13.499: INFO: For apiserver_request_total:
+For apiserver_request_latencies_summary:
+For apiserver_init_events_total:
+For garbage_collector_attempt_to_delete_queue_latency:
+For garbage_collector_attempt_to_delete_work_duration:
+For garbage_collector_attempt_to_orphan_queue_latency:
+For garbage_collector_attempt_to_orphan_work_duration:
+For garbage_collector_dirty_processing_latency_microseconds:
+For garbage_collector_event_processing_latency_microseconds:
+For garbage_collector_graph_changes_queue_latency:
+For garbage_collector_graph_changes_work_duration:
+For garbage_collector_orphan_processing_latency_microseconds:
+For namespace_queue_latency:
+For namespace_queue_latency_sum:
+For namespace_queue_latency_count:
+For namespace_retries:
+For namespace_work_duration:
+For namespace_work_duration_sum:
+For namespace_work_duration_count:
+For function_duration_seconds:
+For errors_total:
+For evicted_pods_total:
 
-• [SLOW TEST:25.468 seconds]
-[sig-auth] ServiceAccounts
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/auth/framework.go:22
-  should allow opting out of API token automount  [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-------------------------------
-SSSSSSS
-------------------------------
-[sig-network] Services 
-  should provide secure master service  [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [sig-network] Services
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
-STEP: Creating a kubernetes client
-Jun  4 18:34:08.390: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-STEP: Building a namespace api object, basename services
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in services-9542
-STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-network] Services
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:86
-[It] should provide secure master service  [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[AfterEach] [sig-network] Services
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 18:34:08.623: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "services-9542" for this suite.
-Jun  4 18:34:14.677: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 18:34:15.017: INFO: namespace services-9542 deletion completed in 6.381174566s
-[AfterEach] [sig-network] Services
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:91
+[AfterEach] [sig-api-machinery] Garbage collector
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 11:38:13.499: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "gc-1887" for this suite.
+Jun 18 11:38:21.547: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 11:38:22.708: INFO: namespace gc-1887 deletion completed in 9.194548874s
 
-• [SLOW TEST:6.627 seconds]
-[sig-network] Services
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:22
-  should provide secure master service  [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+• [SLOW TEST:39.525 seconds]
+[sig-api-machinery] Garbage collector
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:22
+  should orphan RS created by deployment when deleteOptions.PropagationPolicy is Orphan [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-SSSSSSSSS
+SSSS
 ------------------------------
-[k8s.io] Docker Containers 
-  should be able to override the image's default command and arguments [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [k8s.io] Docker Containers
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+[k8s.io] Variable Expansion 
+  should allow substituting values in a container's args [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [k8s.io] Variable Expansion
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 18:34:15.017: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-STEP: Building a namespace api object, basename containers
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in containers-8790
+Jun 18 11:38:22.708: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename var-expansion
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in var-expansion-7416
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should be able to override the image's default command and arguments [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: Creating a pod to test override all
-Jun  4 18:34:15.254: INFO: Waiting up to 5m0s for pod "client-containers-5af17882-86f7-11e9-8d1b-467ee19922ac" in namespace "containers-8790" to be "success or failure"
-Jun  4 18:34:15.267: INFO: Pod "client-containers-5af17882-86f7-11e9-8d1b-467ee19922ac": Phase="Pending", Reason="", readiness=false. Elapsed: 13.204716ms
-Jun  4 18:34:17.280: INFO: Pod "client-containers-5af17882-86f7-11e9-8d1b-467ee19922ac": Phase="Pending", Reason="", readiness=false. Elapsed: 2.02646302s
-Jun  4 18:34:19.293: INFO: Pod "client-containers-5af17882-86f7-11e9-8d1b-467ee19922ac": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.038837344s
+[It] should allow substituting values in a container's args [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating a pod to test substitution in container's args
+Jun 18 11:38:22.937: INFO: Waiting up to 5m0s for pod "var-expansion-93fbf7d7-91bd-11e9-a25d-8608290c688a" in namespace "var-expansion-7416" to be "success or failure"
+Jun 18 11:38:22.967: INFO: Pod "var-expansion-93fbf7d7-91bd-11e9-a25d-8608290c688a": Phase="Pending", Reason="", readiness=false. Elapsed: 30.032934ms
+Jun 18 11:38:24.980: INFO: Pod "var-expansion-93fbf7d7-91bd-11e9-a25d-8608290c688a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.043407344s
 STEP: Saw pod success
-Jun  4 18:34:19.293: INFO: Pod "client-containers-5af17882-86f7-11e9-8d1b-467ee19922ac" satisfied condition "success or failure"
-Jun  4 18:34:19.304: INFO: Trying to get logs from node 10.212.23.164 pod client-containers-5af17882-86f7-11e9-8d1b-467ee19922ac container test-container: 
+Jun 18 11:38:24.981: INFO: Pod "var-expansion-93fbf7d7-91bd-11e9-a25d-8608290c688a" satisfied condition "success or failure"
+Jun 18 11:38:24.994: INFO: Trying to get logs from node 10.72.74.138 pod var-expansion-93fbf7d7-91bd-11e9-a25d-8608290c688a container dapi-container: 
 STEP: delete the pod
-Jun  4 18:34:19.364: INFO: Waiting for pod client-containers-5af17882-86f7-11e9-8d1b-467ee19922ac to disappear
-Jun  4 18:34:19.378: INFO: Pod client-containers-5af17882-86f7-11e9-8d1b-467ee19922ac no longer exists
-[AfterEach] [k8s.io] Docker Containers
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 18:34:19.379: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "containers-8790" for this suite.
-Jun  4 18:34:25.441: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 18:34:25.984: INFO: namespace containers-8790 deletion completed in 6.589806496s
+Jun 18 11:38:25.085: INFO: Waiting for pod var-expansion-93fbf7d7-91bd-11e9-a25d-8608290c688a to disappear
+Jun 18 11:38:25.097: INFO: Pod var-expansion-93fbf7d7-91bd-11e9-a25d-8608290c688a no longer exists
+[AfterEach] [k8s.io] Variable Expansion
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 11:38:25.098: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "var-expansion-7416" for this suite.
+Jun 18 11:38:31.147: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 11:38:31.525: INFO: namespace var-expansion-7416 deletion completed in 6.413028648s
 
-• [SLOW TEST:10.967 seconds]
-[k8s.io] Docker Containers
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
-  should be able to override the image's default command and arguments [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+• [SLOW TEST:8.816 seconds]
+[k8s.io] Variable Expansion
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
+  should allow substituting values in a container's args [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-SSSSSSSSSSSSS
+SSS
 ------------------------------
-[sig-storage] Secrets 
-  should be able to mount in a volume regardless of a different secret existing with same name in different namespace [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [sig-storage] Secrets
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+[sig-storage] Projected combined 
+  should project all components that make up the projection API [Projection][NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-storage] Projected combined
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 18:34:25.986: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-STEP: Building a namespace api object, basename secrets
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in secrets-9040
+Jun 18 11:38:31.525: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename projected
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-7718
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should be able to mount in a volume regardless of a different secret existing with same name in different namespace [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in secret-namespace-732
-STEP: Creating secret with name secret-test-617ae6f6-86f7-11e9-8d1b-467ee19922ac
-STEP: Creating a pod to test consume secrets
-Jun  4 18:34:26.455: INFO: Waiting up to 5m0s for pod "pod-secrets-619e596a-86f7-11e9-8d1b-467ee19922ac" in namespace "secrets-9040" to be "success or failure"
-Jun  4 18:34:26.467: INFO: Pod "pod-secrets-619e596a-86f7-11e9-8d1b-467ee19922ac": Phase="Pending", Reason="", readiness=false. Elapsed: 11.693559ms
-Jun  4 18:34:28.480: INFO: Pod "pod-secrets-619e596a-86f7-11e9-8d1b-467ee19922ac": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.024985209s
+[It] should project all components that make up the projection API [Projection][NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating configMap with name configmap-projected-all-test-volume-993bbe40-91bd-11e9-a25d-8608290c688a
+STEP: Creating secret with name secret-projected-all-test-volume-993bbc04-91bd-11e9-a25d-8608290c688a
+STEP: Creating a pod to test Check all projections for projected volume plugin
+Jun 18 11:38:31.770: INFO: Waiting up to 5m0s for pod "projected-volume-993bbbb5-91bd-11e9-a25d-8608290c688a" in namespace "projected-7718" to be "success or failure"
+Jun 18 11:38:31.789: INFO: Pod "projected-volume-993bbbb5-91bd-11e9-a25d-8608290c688a": Phase="Pending", Reason="", readiness=false. Elapsed: 18.335307ms
+Jun 18 11:38:33.803: INFO: Pod "projected-volume-993bbbb5-91bd-11e9-a25d-8608290c688a": Phase="Running", Reason="", readiness=true. Elapsed: 2.032773308s
+Jun 18 11:38:35.817: INFO: Pod "projected-volume-993bbbb5-91bd-11e9-a25d-8608290c688a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.046914489s
 STEP: Saw pod success
-Jun  4 18:34:28.480: INFO: Pod "pod-secrets-619e596a-86f7-11e9-8d1b-467ee19922ac" satisfied condition "success or failure"
-Jun  4 18:34:28.492: INFO: Trying to get logs from node 10.212.23.189 pod pod-secrets-619e596a-86f7-11e9-8d1b-467ee19922ac container secret-volume-test: 
+Jun 18 11:38:35.817: INFO: Pod "projected-volume-993bbbb5-91bd-11e9-a25d-8608290c688a" satisfied condition "success or failure"
+Jun 18 11:38:35.830: INFO: Trying to get logs from node 10.72.74.184 pod projected-volume-993bbbb5-91bd-11e9-a25d-8608290c688a container projected-all-volume-test: 
 STEP: delete the pod
-Jun  4 18:34:28.595: INFO: Waiting for pod pod-secrets-619e596a-86f7-11e9-8d1b-467ee19922ac to disappear
-Jun  4 18:34:28.606: INFO: Pod pod-secrets-619e596a-86f7-11e9-8d1b-467ee19922ac no longer exists
-[AfterEach] [sig-storage] Secrets
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 18:34:28.606: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "secrets-9040" for this suite.
-Jun  4 18:34:34.701: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 18:34:35.061: INFO: namespace secrets-9040 deletion completed in 6.442127146s
-STEP: Destroying namespace "secret-namespace-732" for this suite.
-Jun  4 18:34:41.106: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 18:34:41.443: INFO: namespace secret-namespace-732 deletion completed in 6.381028236s
-
-• [SLOW TEST:15.457 seconds]
-[sig-storage] Secrets
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/secrets_volume.go:33
-  should be able to mount in a volume regardless of a different secret existing with same name in different namespace [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+Jun 18 11:38:35.907: INFO: Waiting for pod projected-volume-993bbbb5-91bd-11e9-a25d-8608290c688a to disappear
+Jun 18 11:38:35.919: INFO: Pod projected-volume-993bbbb5-91bd-11e9-a25d-8608290c688a no longer exists
+[AfterEach] [sig-storage] Projected combined
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 11:38:35.919: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "projected-7718" for this suite.
+Jun 18 11:38:41.957: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 11:38:42.351: INFO: namespace projected-7718 deletion completed in 6.418169805s
+
+• [SLOW TEST:10.826 seconds]
+[sig-storage] Projected combined
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_combined.go:31
+  should project all components that make up the projection API [Projection][NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-SS
+SSSS
 ------------------------------
-[sig-storage] EmptyDir volumes 
-  should support (non-root,0644,default) [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [sig-storage] EmptyDir volumes
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+[sig-cli] Kubectl client [k8s.io] Kubectl cluster-info 
+  should check if Kubernetes master services is included in cluster-info  [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-cli] Kubectl client
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 18:34:41.443: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-STEP: Building a namespace api object, basename emptydir
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in emptydir-1881
+Jun 18 11:38:42.353: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename kubectl
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-6322
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should support (non-root,0644,default) [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: Creating a pod to test emptydir 0644 on node default medium
-Jun  4 18:34:41.805: INFO: Waiting up to 5m0s for pod "pod-6ac52a35-86f7-11e9-8d1b-467ee19922ac" in namespace "emptydir-1881" to be "success or failure"
-Jun  4 18:34:41.817: INFO: Pod "pod-6ac52a35-86f7-11e9-8d1b-467ee19922ac": Phase="Pending", Reason="", readiness=false. Elapsed: 11.719527ms
-Jun  4 18:34:43.834: INFO: Pod "pod-6ac52a35-86f7-11e9-8d1b-467ee19922ac": Phase="Pending", Reason="", readiness=false. Elapsed: 2.029142931s
-Jun  4 18:34:45.847: INFO: Pod "pod-6ac52a35-86f7-11e9-8d1b-467ee19922ac": Phase="Pending", Reason="", readiness=false. Elapsed: 4.042384871s
-Jun  4 18:34:47.860: INFO: Pod "pod-6ac52a35-86f7-11e9-8d1b-467ee19922ac": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.05536927s
-STEP: Saw pod success
-Jun  4 18:34:47.860: INFO: Pod "pod-6ac52a35-86f7-11e9-8d1b-467ee19922ac" satisfied condition "success or failure"
-Jun  4 18:34:47.871: INFO: Trying to get logs from node 10.212.23.161 pod pod-6ac52a35-86f7-11e9-8d1b-467ee19922ac container test-container: 
-STEP: delete the pod
-Jun  4 18:34:47.935: INFO: Waiting for pod pod-6ac52a35-86f7-11e9-8d1b-467ee19922ac to disappear
-Jun  4 18:34:47.946: INFO: Pod pod-6ac52a35-86f7-11e9-8d1b-467ee19922ac no longer exists
-[AfterEach] [sig-storage] EmptyDir volumes
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 18:34:47.946: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "emptydir-1881" for this suite.
-Jun  4 18:34:54.100: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 18:34:54.496: INFO: namespace emptydir-1881 deletion completed in 6.536233025s
+[BeforeEach] [sig-cli] Kubectl client
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:213
+[It] should check if Kubernetes master services is included in cluster-info  [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: validating cluster-info
+Jun 18 11:38:42.553: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 cluster-info'
+Jun 18 11:38:42.671: INFO: stderr: ""
+Jun 18 11:38:42.672: INFO: stdout: "\x1b[0;32mKubernetes master\x1b[0m is running at \x1b[0;33mhttps://172.21.0.1:443\x1b[0m\n\x1b[0;32mCoreDNS\x1b[0m is running at \x1b[0;33mhttps://172.21.0.1:443/api/v1/namespaces/kube-system/services/kube-dns:dns/proxy\x1b[0m\n\x1b[0;32mkubernetes-dashboard\x1b[0m is running at \x1b[0;33mhttps://172.21.0.1:443/api/v1/namespaces/kube-system/services/https:kubernetes-dashboard:/proxy\x1b[0m\n\x1b[0;32mMetrics-server\x1b[0m is running at \x1b[0;33mhttps://172.21.0.1:443/api/v1/namespaces/kube-system/services/https:metrics-server:/proxy\x1b[0m\n\nTo further debug and diagnose cluster problems, use 'kubectl cluster-info dump'.\n"
+[AfterEach] [sig-cli] Kubectl client
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 11:38:42.672: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "kubectl-6322" for this suite.
+Jun 18 11:38:48.710: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 11:38:49.117: INFO: namespace kubectl-6322 deletion completed in 6.432635287s
 
-• [SLOW TEST:13.053 seconds]
-[sig-storage] EmptyDir volumes
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:41
-  should support (non-root,0644,default) [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+• [SLOW TEST:6.764 seconds]
+[sig-cli] Kubectl client
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23
+  [k8s.io] Kubectl cluster-info
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
+    should check if Kubernetes master services is included in cluster-info  [Conformance]
+    /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-SSSSSSSSSS
+SSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[k8s.io] Pods 
-  should get a host IP [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [k8s.io] Pods
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+[sig-cli] Kubectl client [k8s.io] Kubectl run pod 
+  should create a pod from an image when restart is Never  [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-cli] Kubectl client
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 18:34:54.497: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-STEP: Building a namespace api object, basename pods
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in pods-3131
+Jun 18 11:38:49.117: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename kubectl
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-735
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [k8s.io] Pods
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:135
-[It] should get a host IP [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: creating pod
-Jun  4 18:34:56.792: INFO: Pod pod-hostip-727a6b79-86f7-11e9-8d1b-467ee19922ac has hostIP: 10.212.23.164
-[AfterEach] [k8s.io] Pods
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 18:34:56.792: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "pods-3131" for this suite.
-Jun  4 18:35:20.852: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 18:35:21.321: INFO: namespace pods-3131 deletion completed in 24.513322311s
+[BeforeEach] [sig-cli] Kubectl client
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:213
+[BeforeEach] [k8s.io] Kubectl run pod
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1583
+[It] should create a pod from an image when restart is Never  [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: running the image docker.io/library/nginx:1.14-alpine
+Jun 18 11:38:49.319: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 run e2e-test-nginx-pod --restart=Never --generator=run-pod/v1 --image=docker.io/library/nginx:1.14-alpine --namespace=kubectl-735'
+Jun 18 11:38:49.448: INFO: stderr: ""
+Jun 18 11:38:49.448: INFO: stdout: "pod/e2e-test-nginx-pod created\n"
+STEP: verifying the pod e2e-test-nginx-pod was created
+[AfterEach] [k8s.io] Kubectl run pod
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1588
+Jun 18 11:38:49.468: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 delete pods e2e-test-nginx-pod --namespace=kubectl-735'
+Jun 18 11:38:53.884: INFO: stderr: ""
+Jun 18 11:38:53.884: INFO: stdout: "pod \"e2e-test-nginx-pod\" deleted\n"
+[AfterEach] [sig-cli] Kubectl client
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 11:38:53.889: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "kubectl-735" for this suite.
+Jun 18 11:38:59.936: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 11:39:00.358: INFO: namespace kubectl-735 deletion completed in 6.449208564s
 
-• [SLOW TEST:26.824 seconds]
-[k8s.io] Pods
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
-  should get a host IP [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-------------------------------
-SS
+• [SLOW TEST:11.242 seconds]
+[sig-cli] Kubectl client
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23
+  [k8s.io] Kubectl run pod
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
+    should create a pod from an image when restart is Never  [Conformance]
+    /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-[sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] 
-  should perform rolling updates and roll backs of template modifications [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [sig-apps] StatefulSet
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+[sig-storage] Projected configMap 
+  updates should be reflected in volume [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-storage] Projected configMap
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 18:35:21.322: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-STEP: Building a namespace api object, basename statefulset
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in statefulset-6176
+Jun 18 11:39:00.359: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename projected
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-9530
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-apps] StatefulSet
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:59
-[BeforeEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:74
-STEP: Creating service test in namespace statefulset-6176
-[It] should perform rolling updates and roll backs of template modifications [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: Creating a new StatefulSet
-Jun  4 18:35:21.592: INFO: Found 0 stateful pods, waiting for 3
-Jun  4 18:35:31.605: INFO: Waiting for pod ss2-0 to enter Running - Ready=true, currently Running - Ready=true
-Jun  4 18:35:31.605: INFO: Waiting for pod ss2-1 to enter Running - Ready=true, currently Running - Ready=true
-Jun  4 18:35:31.605: INFO: Waiting for pod ss2-2 to enter Running - Ready=true, currently Pending - Ready=false
-Jun  4 18:35:41.604: INFO: Waiting for pod ss2-0 to enter Running - Ready=true, currently Running - Ready=true
-Jun  4 18:35:41.604: INFO: Waiting for pod ss2-1 to enter Running - Ready=true, currently Running - Ready=true
-Jun  4 18:35:41.604: INFO: Waiting for pod ss2-2 to enter Running - Ready=true, currently Running - Ready=true
-Jun  4 18:35:41.641: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 exec --namespace=statefulset-6176 ss2-1 -- /bin/sh -x -c mv -v /usr/share/nginx/html/index.html /tmp/ || true'
-Jun  4 18:35:41.923: INFO: stderr: "+ mv -v /usr/share/nginx/html/index.html /tmp/\n"
-Jun  4 18:35:41.923: INFO: stdout: "'/usr/share/nginx/html/index.html' -> '/tmp/index.html'\n"
-Jun  4 18:35:41.923: INFO: stdout of mv -v /usr/share/nginx/html/index.html /tmp/ || true on ss2-1: '/usr/share/nginx/html/index.html' -> '/tmp/index.html'
-
-STEP: Updating StatefulSet template: update image from docker.io/library/nginx:1.14-alpine to docker.io/library/nginx:1.15-alpine
-Jun  4 18:35:51.993: INFO: Updating stateful set ss2
-STEP: Creating a new revision
-STEP: Updating Pods in reverse ordinal order
-Jun  4 18:36:02.369: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 exec --namespace=statefulset-6176 ss2-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
-Jun  4 18:36:02.654: INFO: stderr: "+ mv -v /tmp/index.html /usr/share/nginx/html/\n"
-Jun  4 18:36:02.654: INFO: stdout: "'/tmp/index.html' -> '/usr/share/nginx/html/index.html'\n"
-Jun  4 18:36:02.654: INFO: stdout of mv -v /tmp/index.html /usr/share/nginx/html/ || true on ss2-1: '/tmp/index.html' -> '/usr/share/nginx/html/index.html'
-
-Jun  4 18:36:12.731: INFO: Waiting for StatefulSet statefulset-6176/ss2 to complete update
-Jun  4 18:36:12.731: INFO: Waiting for Pod statefulset-6176/ss2-0 to have revision ss2-c79899b9 update revision ss2-787997d666
-Jun  4 18:36:12.731: INFO: Waiting for Pod statefulset-6176/ss2-1 to have revision ss2-c79899b9 update revision ss2-787997d666
-Jun  4 18:36:22.753: INFO: Waiting for StatefulSet statefulset-6176/ss2 to complete update
-Jun  4 18:36:22.753: INFO: Waiting for Pod statefulset-6176/ss2-0 to have revision ss2-c79899b9 update revision ss2-787997d666
-Jun  4 18:36:22.753: INFO: Waiting for Pod statefulset-6176/ss2-1 to have revision ss2-c79899b9 update revision ss2-787997d666
-Jun  4 18:36:32.753: INFO: Waiting for StatefulSet statefulset-6176/ss2 to complete update
-Jun  4 18:36:32.753: INFO: Waiting for Pod statefulset-6176/ss2-0 to have revision ss2-c79899b9 update revision ss2-787997d666
-Jun  4 18:36:42.752: INFO: Waiting for StatefulSet statefulset-6176/ss2 to complete update
-Jun  4 18:36:42.752: INFO: Waiting for Pod statefulset-6176/ss2-0 to have revision ss2-c79899b9 update revision ss2-787997d666
-STEP: Rolling back to a previous revision
-Jun  4 18:36:52.765: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 exec --namespace=statefulset-6176 ss2-1 -- /bin/sh -x -c mv -v /usr/share/nginx/html/index.html /tmp/ || true'
-Jun  4 18:36:53.049: INFO: stderr: "+ mv -v /usr/share/nginx/html/index.html /tmp/\n"
-Jun  4 18:36:53.049: INFO: stdout: "'/usr/share/nginx/html/index.html' -> '/tmp/index.html'\n"
-Jun  4 18:36:53.049: INFO: stdout of mv -v /usr/share/nginx/html/index.html /tmp/ || true on ss2-1: '/usr/share/nginx/html/index.html' -> '/tmp/index.html'
-
-Jun  4 18:37:03.122: INFO: Updating stateful set ss2
-STEP: Rolling back update in reverse ordinal order
-Jun  4 18:37:13.255: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 exec --namespace=statefulset-6176 ss2-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
-Jun  4 18:37:13.516: INFO: stderr: "+ mv -v /tmp/index.html /usr/share/nginx/html/\n"
-Jun  4 18:37:13.517: INFO: stdout: "'/tmp/index.html' -> '/usr/share/nginx/html/index.html'\n"
-Jun  4 18:37:13.517: INFO: stdout of mv -v /tmp/index.html /usr/share/nginx/html/ || true on ss2-1: '/tmp/index.html' -> '/usr/share/nginx/html/index.html'
-
-Jun  4 18:37:13.558: INFO: Waiting for StatefulSet statefulset-6176/ss2 to complete update
-Jun  4 18:37:13.558: INFO: Waiting for Pod statefulset-6176/ss2-0 to have revision ss2-787997d666 update revision ss2-c79899b9
-Jun  4 18:37:13.558: INFO: Waiting for Pod statefulset-6176/ss2-1 to have revision ss2-787997d666 update revision ss2-c79899b9
-Jun  4 18:37:13.558: INFO: Waiting for Pod statefulset-6176/ss2-2 to have revision ss2-787997d666 update revision ss2-c79899b9
-Jun  4 18:37:23.667: INFO: Waiting for StatefulSet statefulset-6176/ss2 to complete update
-Jun  4 18:37:23.667: INFO: Waiting for Pod statefulset-6176/ss2-0 to have revision ss2-787997d666 update revision ss2-c79899b9
-Jun  4 18:37:23.667: INFO: Waiting for Pod statefulset-6176/ss2-1 to have revision ss2-787997d666 update revision ss2-c79899b9
-Jun  4 18:37:33.581: INFO: Waiting for StatefulSet statefulset-6176/ss2 to complete update
-Jun  4 18:37:33.581: INFO: Waiting for Pod statefulset-6176/ss2-0 to have revision ss2-787997d666 update revision ss2-c79899b9
-[AfterEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:85
-Jun  4 18:37:43.586: INFO: Deleting all statefulset in ns statefulset-6176
-Jun  4 18:37:43.594: INFO: Scaling statefulset ss2 to 0
-Jun  4 18:38:13.703: INFO: Waiting for statefulset status.replicas updated to 0
-Jun  4 18:38:13.711: INFO: Deleting statefulset ss2
-[AfterEach] [sig-apps] StatefulSet
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 18:38:13.760: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "statefulset-6176" for this suite.
-Jun  4 18:38:21.816: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 18:38:22.322: INFO: namespace statefulset-6176 deletion completed in 8.546717674s
+[It] updates should be reflected in volume [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating projection with configMap that has name projected-configmap-test-upd-aa6e0ce6-91bd-11e9-a25d-8608290c688a
+STEP: Creating the pod
+STEP: Updating configmap projected-configmap-test-upd-aa6e0ce6-91bd-11e9-a25d-8608290c688a
+STEP: waiting to observe update in volume
+[AfterEach] [sig-storage] Projected configMap
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 11:39:04.774: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "projected-9530" for this suite.
+Jun 18 11:39:28.829: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 11:39:29.256: INFO: namespace projected-9530 deletion completed in 24.465420546s
 
-• [SLOW TEST:181.000 seconds]
-[sig-apps] StatefulSet
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22
-  [k8s.io] Basic StatefulSet functionality [StatefulSetBasic]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
-    should perform rolling updates and roll backs of template modifications [Conformance]
-    /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+• [SLOW TEST:28.897 seconds]
+[sig-storage] Projected configMap
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_configmap.go:33
+  updates should be reflected in volume [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-SSSSSSS
+SSSSSSSSS
 ------------------------------
-[sig-apps] ReplicationController 
-  should adopt matching pods on creation [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [sig-apps] ReplicationController
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+[sig-storage] Downward API volume 
+  should provide podname only [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-storage] Downward API volume
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 18:38:22.322: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-STEP: Building a namespace api object, basename replication-controller
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in replication-controller-3681
+Jun 18 11:39:29.257: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename downward-api
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in downward-api-2977
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should adopt matching pods on creation [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: Given a Pod with a 'name' label pod-adoption is created
-STEP: When a replication controller with a matching selector is created
-STEP: Then the orphan pod is adopted
-[AfterEach] [sig-apps] ReplicationController
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 18:38:25.694: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "replication-controller-3681" for this suite.
-Jun  4 18:38:49.754: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 18:38:50.199: INFO: namespace replication-controller-3681 deletion completed in 24.487654414s
+[BeforeEach] [sig-storage] Downward API volume
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:39
+[It] should provide podname only [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating a pod to test downward API volume plugin
+Jun 18 11:39:29.479: INFO: Waiting up to 5m0s for pod "downwardapi-volume-bba56a95-91bd-11e9-a25d-8608290c688a" in namespace "downward-api-2977" to be "success or failure"
+Jun 18 11:39:29.493: INFO: Pod "downwardapi-volume-bba56a95-91bd-11e9-a25d-8608290c688a": Phase="Pending", Reason="", readiness=false. Elapsed: 13.796555ms
+Jun 18 11:39:31.507: INFO: Pod "downwardapi-volume-bba56a95-91bd-11e9-a25d-8608290c688a": Phase="Running", Reason="", readiness=true. Elapsed: 2.028014497s
+Jun 18 11:39:33.524: INFO: Pod "downwardapi-volume-bba56a95-91bd-11e9-a25d-8608290c688a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.044787853s
+STEP: Saw pod success
+Jun 18 11:39:33.524: INFO: Pod "downwardapi-volume-bba56a95-91bd-11e9-a25d-8608290c688a" satisfied condition "success or failure"
+Jun 18 11:39:33.539: INFO: Trying to get logs from node 10.72.74.138 pod downwardapi-volume-bba56a95-91bd-11e9-a25d-8608290c688a container client-container: 
+STEP: delete the pod
+Jun 18 11:39:33.620: INFO: Waiting for pod downwardapi-volume-bba56a95-91bd-11e9-a25d-8608290c688a to disappear
+Jun 18 11:39:33.632: INFO: Pod downwardapi-volume-bba56a95-91bd-11e9-a25d-8608290c688a no longer exists
+[AfterEach] [sig-storage] Downward API volume
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 11:39:33.632: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "downward-api-2977" for this suite.
+Jun 18 11:39:39.672: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 11:39:40.078: INFO: namespace downward-api-2977 deletion completed in 6.43173082s
 
-• [SLOW TEST:27.877 seconds]
-[sig-apps] ReplicationController
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22
-  should adopt matching pods on creation [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+• [SLOW TEST:10.820 seconds]
+[sig-storage] Downward API volume
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:34
+  should provide podname only [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-SSSSSSSSSSSSSSS
+S
 ------------------------------
-[sig-apps] Daemon set [Serial] 
-  should retry creating failed daemon pods [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [sig-apps] Daemon set [Serial]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+[sig-api-machinery] Garbage collector 
+  should orphan pods created by rc if delete options say so [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-api-machinery] Garbage collector
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 18:38:50.199: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-STEP: Building a namespace api object, basename daemonsets
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in daemonsets-4739
+Jun 18 11:39:40.078: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename gc
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in gc-9034
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-apps] Daemon set [Serial]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:102
-[It] should retry creating failed daemon pods [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: Creating a simple DaemonSet "daemon-set"
-STEP: Check that daemon pods launch on every node of the cluster.
-Jun  4 18:38:50.541: INFO: Number of nodes with available pods: 0
-Jun  4 18:38:50.541: INFO: Node 10.212.23.161 is running more than one daemon pod
-Jun  4 18:38:51.668: INFO: Number of nodes with available pods: 0
-Jun  4 18:38:51.668: INFO: Node 10.212.23.161 is running more than one daemon pod
-Jun  4 18:38:52.569: INFO: Number of nodes with available pods: 2
-Jun  4 18:38:52.569: INFO: Node 10.212.23.164 is running more than one daemon pod
-Jun  4 18:38:53.568: INFO: Number of nodes with available pods: 3
-Jun  4 18:38:53.568: INFO: Number of running nodes: 3, number of available pods: 3
-STEP: Set a daemon pod's phase to 'Failed', check that the daemon pod is revived.
-Jun  4 18:38:53.629: INFO: Number of nodes with available pods: 2
-Jun  4 18:38:53.629: INFO: Node 10.212.23.189 is running more than one daemon pod
-Jun  4 18:38:54.855: INFO: Number of nodes with available pods: 2
-Jun  4 18:38:54.855: INFO: Node 10.212.23.189 is running more than one daemon pod
-Jun  4 18:38:55.657: INFO: Number of nodes with available pods: 2
-Jun  4 18:38:55.657: INFO: Node 10.212.23.189 is running more than one daemon pod
-Jun  4 18:38:56.657: INFO: Number of nodes with available pods: 3
-Jun  4 18:38:56.657: INFO: Number of running nodes: 3, number of available pods: 3
-STEP: Wait for the failed daemon pod to be completely deleted.
-[AfterEach] [sig-apps] Daemon set [Serial]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:68
-STEP: Deleting DaemonSet "daemon-set"
-STEP: deleting DaemonSet.extensions daemon-set in namespace daemonsets-4739, will wait for the garbage collector to delete the pods
-Jun  4 18:38:56.765: INFO: Deleting DaemonSet.extensions daemon-set took: 27.950538ms
-Jun  4 18:38:56.865: INFO: Terminating DaemonSet.extensions daemon-set pods took: 100.238451ms
-Jun  4 18:39:06.979: INFO: Number of nodes with available pods: 0
-Jun  4 18:39:06.979: INFO: Number of running nodes: 0, number of available pods: 0
-Jun  4 18:39:06.992: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"selfLink":"/apis/apps/v1/namespaces/daemonsets-4739/daemonsets","resourceVersion":"9405"},"items":null}
-
-Jun  4 18:39:07.055: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"selfLink":"/api/v1/namespaces/daemonsets-4739/pods","resourceVersion":"9405"},"items":null}
+[It] should orphan pods created by rc if delete options say so [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: create the rc
+STEP: delete the rc
+STEP: wait for the rc to be deleted
+STEP: wait for 30 seconds to see if the garbage collector mistakenly deletes the pods
+STEP: Gathering metrics
+W0618 11:40:20.427713      20 metrics_grabber.go:79] Master node is not registered. Grabbing metrics from Scheduler, ControllerManager and ClusterAutoscaler is disabled.
+Jun 18 11:40:20.427: INFO: For apiserver_request_total:
+For apiserver_request_latencies_summary:
+For apiserver_init_events_total:
+For garbage_collector_attempt_to_delete_queue_latency:
+For garbage_collector_attempt_to_delete_work_duration:
+For garbage_collector_attempt_to_orphan_queue_latency:
+For garbage_collector_attempt_to_orphan_work_duration:
+For garbage_collector_dirty_processing_latency_microseconds:
+For garbage_collector_event_processing_latency_microseconds:
+For garbage_collector_graph_changes_queue_latency:
+For garbage_collector_graph_changes_work_duration:
+For garbage_collector_orphan_processing_latency_microseconds:
+For namespace_queue_latency:
+For namespace_queue_latency_sum:
+For namespace_queue_latency_count:
+For namespace_retries:
+For namespace_work_duration:
+For namespace_work_duration_sum:
+For namespace_work_duration_count:
+For function_duration_seconds:
+For errors_total:
+For evicted_pods_total:
 
-[AfterEach] [sig-apps] Daemon set [Serial]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 18:39:07.109: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "daemonsets-4739" for this suite.
-Jun  4 18:39:15.164: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 18:39:15.580: INFO: namespace daemonsets-4739 deletion completed in 8.456652313s
+[AfterEach] [sig-api-machinery] Garbage collector
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 11:40:20.427: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "gc-9034" for this suite.
+Jun 18 11:40:28.470: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 11:40:28.941: INFO: namespace gc-9034 deletion completed in 8.501930589s
 
-• [SLOW TEST:25.380 seconds]
-[sig-apps] Daemon set [Serial]
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22
-  should retry creating failed daemon pods [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+• [SLOW TEST:48.862 seconds]
+[sig-api-machinery] Garbage collector
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:22
+  should orphan pods created by rc if delete options say so [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-SSSSSSSSS
+SSSSSSSSSSS
 ------------------------------
 [sig-storage] Downward API volume 
-  should provide container's memory request [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+  should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 [BeforeEach] [sig-storage] Downward API volume
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 18:39:15.580: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
+Jun 18 11:40:28.941: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
 STEP: Building a namespace api object, basename downward-api
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in downward-api-7049
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in downward-api-8833
 STEP: Waiting for a default service account to be provisioned in namespace
 [BeforeEach] [sig-storage] Downward API volume
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:39
-[It] should provide container's memory request [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:39
+[It] should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 STEP: Creating a pod to test downward API volume plugin
-Jun  4 18:39:15.817: INFO: Waiting up to 5m0s for pod "downwardapi-volume-0e17df66-86f8-11e9-8d1b-467ee19922ac" in namespace "downward-api-7049" to be "success or failure"
-Jun  4 18:39:15.827: INFO: Pod "downwardapi-volume-0e17df66-86f8-11e9-8d1b-467ee19922ac": Phase="Pending", Reason="", readiness=false. Elapsed: 10.35411ms
-Jun  4 18:39:17.840: INFO: Pod "downwardapi-volume-0e17df66-86f8-11e9-8d1b-467ee19922ac": Phase="Pending", Reason="", readiness=false. Elapsed: 2.023428996s
-Jun  4 18:39:19.854: INFO: Pod "downwardapi-volume-0e17df66-86f8-11e9-8d1b-467ee19922ac": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.036989784s
+Jun 18 11:40:29.176: INFO: Waiting up to 5m0s for pod "downwardapi-volume-df38fb3c-91bd-11e9-a25d-8608290c688a" in namespace "downward-api-8833" to be "success or failure"
+Jun 18 11:40:29.200: INFO: Pod "downwardapi-volume-df38fb3c-91bd-11e9-a25d-8608290c688a": Phase="Pending", Reason="", readiness=false. Elapsed: 23.778656ms
+Jun 18 11:40:31.215: INFO: Pod "downwardapi-volume-df38fb3c-91bd-11e9-a25d-8608290c688a": Phase="Running", Reason="", readiness=true. Elapsed: 2.038370907s
+Jun 18 11:40:33.230: INFO: Pod "downwardapi-volume-df38fb3c-91bd-11e9-a25d-8608290c688a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.053527745s
 STEP: Saw pod success
-Jun  4 18:39:19.854: INFO: Pod "downwardapi-volume-0e17df66-86f8-11e9-8d1b-467ee19922ac" satisfied condition "success or failure"
-Jun  4 18:39:19.865: INFO: Trying to get logs from node 10.212.23.164 pod downwardapi-volume-0e17df66-86f8-11e9-8d1b-467ee19922ac container client-container: 
+Jun 18 11:40:33.230: INFO: Pod "downwardapi-volume-df38fb3c-91bd-11e9-a25d-8608290c688a" satisfied condition "success or failure"
+Jun 18 11:40:33.246: INFO: Trying to get logs from node 10.72.74.184 pod downwardapi-volume-df38fb3c-91bd-11e9-a25d-8608290c688a container client-container: 
 STEP: delete the pod
-Jun  4 18:39:19.941: INFO: Waiting for pod downwardapi-volume-0e17df66-86f8-11e9-8d1b-467ee19922ac to disappear
-Jun  4 18:39:19.952: INFO: Pod downwardapi-volume-0e17df66-86f8-11e9-8d1b-467ee19922ac no longer exists
+Jun 18 11:40:33.334: INFO: Waiting for pod downwardapi-volume-df38fb3c-91bd-11e9-a25d-8608290c688a to disappear
+Jun 18 11:40:33.349: INFO: Pod downwardapi-volume-df38fb3c-91bd-11e9-a25d-8608290c688a no longer exists
 [AfterEach] [sig-storage] Downward API volume
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 18:39:19.952: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "downward-api-7049" for this suite.
-Jun  4 18:39:26.010: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 18:39:26.366: INFO: namespace downward-api-7049 deletion completed in 6.400404764s
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 11:40:33.349: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "downward-api-8833" for this suite.
+Jun 18 11:40:39.396: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 11:40:39.783: INFO: namespace downward-api-8833 deletion completed in 6.413895931s
 
-• [SLOW TEST:10.787 seconds]
+• [SLOW TEST:10.843 seconds]
 [sig-storage] Downward API volume
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:34
-  should provide container's memory request [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:34
+  should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-SSS
+S
 ------------------------------
 [sig-apps] Deployment 
   RollingUpdateDeployment should delete old pods and create new ones [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 [BeforeEach] [sig-apps] Deployment
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 18:39:26.366: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
+Jun 18 11:40:39.783: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
 STEP: Building a namespace api object, basename deployment
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in deployment-8999
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in deployment-1932
 STEP: Waiting for a default service account to be provisioned in namespace
 [BeforeEach] [sig-apps] Deployment
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:65
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:65
 [It] RollingUpdateDeployment should delete old pods and create new ones [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-Jun  4 18:39:26.582: INFO: Creating replica set "test-rolling-update-controller" (going to be adopted)
-Jun  4 18:39:26.608: INFO: Pod name sample-pod: Found 0 pods out of 1
-Jun  4 18:39:31.623: INFO: Pod name sample-pod: Found 1 pods out of 1
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+Jun 18 11:40:40.020: INFO: Creating replica set "test-rolling-update-controller" (going to be adopted)
+Jun 18 11:40:40.056: INFO: Pod name sample-pod: Found 0 pods out of 1
+Jun 18 11:40:45.070: INFO: Pod name sample-pod: Found 1 pods out of 1
 STEP: ensuring each pod is running
-Jun  4 18:39:31.623: INFO: Creating deployment "test-rolling-update-deployment"
-Jun  4 18:39:31.633: INFO: Ensuring deployment "test-rolling-update-deployment" gets the next revision from the one the adopted replica set "test-rolling-update-controller" has
-Jun  4 18:39:31.650: INFO: new replicaset for deployment "test-rolling-update-deployment" is yet to be created
-Jun  4 18:39:33.668: INFO: Ensuring status for deployment "test-rolling-update-deployment" is the expected
-Jun  4 18:39:33.681: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:2, UpdatedReplicas:1, ReadyReplicas:1, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63695270371, loc:(*time.Location)(0x8a140e0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63695270371, loc:(*time.Location)(0x8a140e0)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63695270371, loc:(*time.Location)(0x8a140e0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63695270371, loc:(*time.Location)(0x8a140e0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rolling-update-deployment-67599b4d9\" is progressing."}}, CollisionCount:(*int32)(nil)}
-Jun  4 18:39:35.690: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:2, UpdatedReplicas:1, ReadyReplicas:1, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63695270371, loc:(*time.Location)(0x8a140e0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63695270371, loc:(*time.Location)(0x8a140e0)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63695270371, loc:(*time.Location)(0x8a140e0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63695270371, loc:(*time.Location)(0x8a140e0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rolling-update-deployment-67599b4d9\" is progressing."}}, CollisionCount:(*int32)(nil)}
-Jun  4 18:39:37.692: INFO: Ensuring deployment "test-rolling-update-deployment" has one old replica set (the one it adopted)
+Jun 18 11:40:45.070: INFO: Creating deployment "test-rolling-update-deployment"
+Jun 18 11:40:45.079: INFO: Ensuring deployment "test-rolling-update-deployment" gets the next revision from the one the adopted replica set "test-rolling-update-controller" has
+Jun 18 11:40:45.101: INFO: new replicaset for deployment "test-rolling-update-deployment" is yet to be created
+Jun 18 11:40:47.120: INFO: Ensuring status for deployment "test-rolling-update-deployment" is the expected
+Jun 18 11:40:47.127: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:2, UpdatedReplicas:1, ReadyReplicas:1, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63696454845, loc:(*time.Location)(0x8a1a0e0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63696454845, loc:(*time.Location)(0x8a1a0e0)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63696454845, loc:(*time.Location)(0x8a1a0e0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63696454845, loc:(*time.Location)(0x8a1a0e0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rolling-update-deployment-67599b4d9\" is progressing."}}, CollisionCount:(*int32)(nil)}
+Jun 18 11:40:49.135: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:2, UpdatedReplicas:1, ReadyReplicas:1, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63696454845, loc:(*time.Location)(0x8a1a0e0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63696454845, loc:(*time.Location)(0x8a1a0e0)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63696454845, loc:(*time.Location)(0x8a1a0e0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63696454845, loc:(*time.Location)(0x8a1a0e0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rolling-update-deployment-67599b4d9\" is progressing."}}, CollisionCount:(*int32)(nil)}
+Jun 18 11:40:51.135: INFO: Ensuring deployment "test-rolling-update-deployment" has one old replica set (the one it adopted)
 [AfterEach] [sig-apps] Deployment
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:59
-Jun  4 18:39:37.715: INFO: Deployment "test-rolling-update-deployment":
-&Deployment{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-rolling-update-deployment,GenerateName:,Namespace:deployment-8999,SelfLink:/apis/apps/v1/namespaces/deployment-8999/deployments/test-rolling-update-deployment,UID:17873bf1-86f8-11e9-8318-1e5386706511,ResourceVersion:9579,Generation:1,CreationTimestamp:2019-06-04 18:39:31 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod,},Annotations:map[string]string{deployment.kubernetes.io/revision: 3546343826724305833,},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:DeploymentSpec{Replicas:*1,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[],Containers:[{redis gcr.io/kubernetes-e2e-test-images/redis:1.0 [] []  [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,},},Strategy:DeploymentStrategy{Type:RollingUpdate,RollingUpdate:&RollingUpdateDeployment{MaxUnavailable:25%!,(MISSING)MaxSurge:25%!,(MISSING)},},MinReadySeconds:0,RevisionHistoryLimit:*10,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:1,Replicas:1,UpdatedReplicas:1,AvailableReplicas:1,UnavailableReplicas:0,Conditions:[{Available True 2019-06-04 18:39:31 +0000 UTC 2019-06-04 18:39:31 +0000 UTC MinimumReplicasAvailable Deployment has minimum availability.} {Progressing True 2019-06-04 18:39:36 +0000 UTC 2019-06-04 18:39:31 +0000 UTC NewReplicaSetAvailable ReplicaSet "test-rolling-update-deployment-67599b4d9" has successfully progressed.}],ReadyReplicas:1,CollisionCount:nil,},}
-
-Jun  4 18:39:37.724: INFO: New ReplicaSet "test-rolling-update-deployment-67599b4d9" of Deployment "test-rolling-update-deployment":
-&ReplicaSet{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-rolling-update-deployment-67599b4d9,GenerateName:,Namespace:deployment-8999,SelfLink:/apis/apps/v1/namespaces/deployment-8999/replicasets/test-rolling-update-deployment-67599b4d9,UID:178b9953-86f8-11e9-9509-923735b172b1,ResourceVersion:9568,Generation:1,CreationTimestamp:2019-06-04 18:39:31 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod,pod-template-hash: 67599b4d9,},Annotations:map[string]string{deployment.kubernetes.io/desired-replicas: 1,deployment.kubernetes.io/max-replicas: 2,deployment.kubernetes.io/revision: 3546343826724305833,},OwnerReferences:[{apps/v1 Deployment test-rolling-update-deployment 17873bf1-86f8-11e9-8318-1e5386706511 0xc00243def0 0xc00243def1}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:ReplicaSetSpec{Replicas:*1,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod,pod-template-hash: 67599b4d9,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod,pod-template-hash: 67599b4d9,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[],Containers:[{redis gcr.io/kubernetes-e2e-test-images/redis:1.0 [] []  [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,},},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:1,FullyLabeledReplicas:1,ObservedGeneration:1,ReadyReplicas:1,AvailableReplicas:1,Conditions:[],},}
-Jun  4 18:39:37.724: INFO: All old ReplicaSets of Deployment "test-rolling-update-deployment":
-Jun  4 18:39:37.724: INFO: &ReplicaSet{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-rolling-update-controller,GenerateName:,Namespace:deployment-8999,SelfLink:/apis/apps/v1/namespaces/deployment-8999/replicasets/test-rolling-update-controller,UID:14860c78-86f8-11e9-8318-1e5386706511,ResourceVersion:9578,Generation:2,CreationTimestamp:2019-06-04 18:39:26 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod,pod: nginx,},Annotations:map[string]string{deployment.kubernetes.io/desired-replicas: 1,deployment.kubernetes.io/max-replicas: 2,deployment.kubernetes.io/revision: 3546343826724305832,},OwnerReferences:[{apps/v1 Deployment test-rolling-update-deployment 17873bf1-86f8-11e9-8318-1e5386706511 0xc00243de1f 0xc00243de30}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:ReplicaSetSpec{Replicas:*0,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod,pod: nginx,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod,pod: nginx,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] []  [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,},},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:0,FullyLabeledReplicas:0,ObservedGeneration:2,ReadyReplicas:0,AvailableReplicas:0,Conditions:[],},}
-Jun  4 18:39:37.736: INFO: Pod "test-rolling-update-deployment-67599b4d9-nrrdr" is available:
-&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-rolling-update-deployment-67599b4d9-nrrdr,GenerateName:test-rolling-update-deployment-67599b4d9-,Namespace:deployment-8999,SelfLink:/api/v1/namespaces/deployment-8999/pods/test-rolling-update-deployment-67599b4d9-nrrdr,UID:178d1871-86f8-11e9-9509-923735b172b1,ResourceVersion:9567,Generation:0,CreationTimestamp:2019-06-04 18:39:31 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod,pod-template-hash: 67599b4d9,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet test-rolling-update-deployment-67599b4d9 178b9953-86f8-11e9-9509-923735b172b1 0xc00266c760 0xc00266c761}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-fb6qs {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-fb6qs,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{redis gcr.io/kubernetes-e2e-test-images/redis:1.0 [] []  [] [] [] {map[] map[]} [{default-token-fb6qs true /var/run/secrets/kubernetes.io/serviceaccount   }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.212.23.189,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists  NoExecute 0xc00266c7d0} {node.kubernetes.io/unreachable Exists  NoExecute 0xc00266c7f0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Running,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 18:39:31 +0000 UTC  } {Ready True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 18:39:35 +0000 UTC  } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 18:39:35 +0000 UTC  } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 18:39:31 +0000 UTC  }],Message:,Reason:,HostIP:10.212.23.189,PodIP:172.30.193.20,StartTime:2019-06-04 18:39:31 +0000 UTC,ContainerStatuses:[{redis {nil ContainerStateRunning{StartedAt:2019-06-04 18:39:35 +0000 UTC,} nil} {nil nil nil} true 0 gcr.io/kubernetes-e2e-test-images/redis:1.0 gcr.io/kubernetes-e2e-test-images/redis@sha256:af4748d1655c08dc54d4be5182135395db9ce87aba2d4699b26b14ae197c5830 containerd://11775b6d8df2fb1b912c791977fc79bdedd4ee193bd10181406c460fdbdb81cb}],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},}
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:59
+Jun 18 11:40:51.167: INFO: Deployment "test-rolling-update-deployment":
+&Deployment{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-rolling-update-deployment,GenerateName:,Namespace:deployment-1932,SelfLink:/apis/apps/v1/namespaces/deployment-1932/deployments/test-rolling-update-deployment,UID:e8b822a3-91bd-11e9-a08a-ee7a14707756,ResourceVersion:87018,Generation:1,CreationTimestamp:2019-06-18 11:40:45 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod,},Annotations:map[string]string{deployment.kubernetes.io/revision: 3546343826724305833,},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:DeploymentSpec{Replicas:*1,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[],Containers:[{redis gcr.io/kubernetes-e2e-test-images/redis:1.0 [] []  [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,},},Strategy:DeploymentStrategy{Type:RollingUpdate,RollingUpdate:&RollingUpdateDeployment{MaxUnavailable:25%!,(MISSING)MaxSurge:25%!,(MISSING)},},MinReadySeconds:0,RevisionHistoryLimit:*10,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:1,Replicas:1,UpdatedReplicas:1,AvailableReplicas:1,UnavailableReplicas:0,Conditions:[{Available True 2019-06-18 11:40:45 +0000 UTC 2019-06-18 11:40:45 +0000 UTC MinimumReplicasAvailable Deployment has minimum availability.} {Progressing True 2019-06-18 11:40:49 +0000 UTC 2019-06-18 11:40:45 +0000 UTC NewReplicaSetAvailable ReplicaSet "test-rolling-update-deployment-67599b4d9" has successfully progressed.}],ReadyReplicas:1,CollisionCount:nil,},}
+
+Jun 18 11:40:51.180: INFO: New ReplicaSet "test-rolling-update-deployment-67599b4d9" of Deployment "test-rolling-update-deployment":
+&ReplicaSet{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-rolling-update-deployment-67599b4d9,GenerateName:,Namespace:deployment-1932,SelfLink:/apis/apps/v1/namespaces/deployment-1932/replicasets/test-rolling-update-deployment-67599b4d9,UID:e8bbccd6-91bd-11e9-a08a-ee7a14707756,ResourceVersion:87008,Generation:1,CreationTimestamp:2019-06-18 11:40:45 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod,pod-template-hash: 67599b4d9,},Annotations:map[string]string{deployment.kubernetes.io/desired-replicas: 1,deployment.kubernetes.io/max-replicas: 2,deployment.kubernetes.io/revision: 3546343826724305833,},OwnerReferences:[{apps/v1 Deployment test-rolling-update-deployment e8b822a3-91bd-11e9-a08a-ee7a14707756 0xc002cdbba0 0xc002cdbba1}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:ReplicaSetSpec{Replicas:*1,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod,pod-template-hash: 67599b4d9,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod,pod-template-hash: 67599b4d9,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[],Containers:[{redis gcr.io/kubernetes-e2e-test-images/redis:1.0 [] []  [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,},},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:1,FullyLabeledReplicas:1,ObservedGeneration:1,ReadyReplicas:1,AvailableReplicas:1,Conditions:[],},}
+Jun 18 11:40:51.180: INFO: All old ReplicaSets of Deployment "test-rolling-update-deployment":
+Jun 18 11:40:51.181: INFO: &ReplicaSet{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-rolling-update-controller,GenerateName:,Namespace:deployment-1932,SelfLink:/apis/apps/v1/namespaces/deployment-1932/replicasets/test-rolling-update-controller,UID:e5b58576-91bd-11e9-a08a-ee7a14707756,ResourceVersion:87017,Generation:2,CreationTimestamp:2019-06-18 11:40:40 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod,pod: nginx,},Annotations:map[string]string{deployment.kubernetes.io/desired-replicas: 1,deployment.kubernetes.io/max-replicas: 2,deployment.kubernetes.io/revision: 3546343826724305832,},OwnerReferences:[{apps/v1 Deployment test-rolling-update-deployment e8b822a3-91bd-11e9-a08a-ee7a14707756 0xc002cdbad7 0xc002cdbad8}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:ReplicaSetSpec{Replicas:*0,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod,pod: nginx,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod,pod: nginx,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] []  [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,},},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:0,FullyLabeledReplicas:0,ObservedGeneration:2,ReadyReplicas:0,AvailableReplicas:0,Conditions:[],},}
+Jun 18 11:40:51.194: INFO: Pod "test-rolling-update-deployment-67599b4d9-chnt9" is available:
+&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-rolling-update-deployment-67599b4d9-chnt9,GenerateName:test-rolling-update-deployment-67599b4d9-,Namespace:deployment-1932,SelfLink:/api/v1/namespaces/deployment-1932/pods/test-rolling-update-deployment-67599b4d9-chnt9,UID:e8bd6669-91bd-11e9-a08a-ee7a14707756,ResourceVersion:87007,Generation:0,CreationTimestamp:2019-06-18 11:40:45 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod,pod-template-hash: 67599b4d9,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet test-rolling-update-deployment-67599b4d9 e8bbccd6-91bd-11e9-a08a-ee7a14707756 0xc00289c740 0xc00289c741}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-84r62 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-84r62,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{redis gcr.io/kubernetes-e2e-test-images/redis:1.0 [] []  [] [] [] {map[] map[]} [{default-token-84r62 true /var/run/secrets/kubernetes.io/serviceaccount   }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.72.74.184,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists  NoExecute 0xc00289c7b0} {node.kubernetes.io/unreachable Exists  NoExecute 0xc00289c7d0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Running,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:40:45 +0000 UTC  } {Ready True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:40:49 +0000 UTC  } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:40:49 +0000 UTC  } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:40:45 +0000 UTC  }],Message:,Reason:,HostIP:10.72.74.184,PodIP:172.30.142.20,StartTime:2019-06-18 11:40:45 +0000 UTC,ContainerStatuses:[{redis {nil ContainerStateRunning{StartedAt:2019-06-18 11:40:48 +0000 UTC,} nil} {nil nil nil} true 0 gcr.io/kubernetes-e2e-test-images/redis:1.0 gcr.io/kubernetes-e2e-test-images/redis@sha256:af4748d1655c08dc54d4be5182135395db9ce87aba2d4699b26b14ae197c5830 containerd://00c7db4d97d52b43af639d47f73133216b42fbf10cf7ad884b14c00e94016a15}],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},}
 [AfterEach] [sig-apps] Deployment
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 18:39:37.736: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "deployment-8999" for this suite.
-Jun  4 18:39:43.794: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 18:39:44.328: INFO: namespace deployment-8999 deletion completed in 6.577059576s
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 11:40:51.195: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "deployment-1932" for this suite.
+Jun 18 11:40:59.236: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 11:40:59.627: INFO: namespace deployment-1932 deletion completed in 8.41897963s
 
-• [SLOW TEST:17.962 seconds]
+• [SLOW TEST:19.843 seconds]
 [sig-apps] Deployment
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22
   RollingUpdateDeployment should delete old pods and create new ones [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-------------------------------
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
-------------------------------
-[sig-node] ConfigMap 
-  should be consumable via environment variable [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [sig-node] ConfigMap
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
-STEP: Creating a kubernetes client
-Jun  4 18:39:44.329: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-STEP: Building a namespace api object, basename configmap
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in configmap-7739
-STEP: Waiting for a default service account to be provisioned in namespace
-[It] should be consumable via environment variable [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: Creating configMap configmap-7739/configmap-test-1f3c4581-86f8-11e9-8d1b-467ee19922ac
-STEP: Creating a pod to test consume configMaps
-Jun  4 18:39:44.590: INFO: Waiting up to 5m0s for pod "pod-configmaps-1f3e5455-86f8-11e9-8d1b-467ee19922ac" in namespace "configmap-7739" to be "success or failure"
-Jun  4 18:39:44.601: INFO: Pod "pod-configmaps-1f3e5455-86f8-11e9-8d1b-467ee19922ac": Phase="Pending", Reason="", readiness=false. Elapsed: 10.810837ms
-Jun  4 18:39:46.667: INFO: Pod "pod-configmaps-1f3e5455-86f8-11e9-8d1b-467ee19922ac": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.076854501s
-STEP: Saw pod success
-Jun  4 18:39:46.667: INFO: Pod "pod-configmaps-1f3e5455-86f8-11e9-8d1b-467ee19922ac" satisfied condition "success or failure"
-Jun  4 18:39:46.678: INFO: Trying to get logs from node 10.212.23.161 pod pod-configmaps-1f3e5455-86f8-11e9-8d1b-467ee19922ac container env-test: 
-STEP: delete the pod
-Jun  4 18:39:46.747: INFO: Waiting for pod pod-configmaps-1f3e5455-86f8-11e9-8d1b-467ee19922ac to disappear
-Jun  4 18:39:46.758: INFO: Pod pod-configmaps-1f3e5455-86f8-11e9-8d1b-467ee19922ac no longer exists
-[AfterEach] [sig-node] ConfigMap
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 18:39:46.758: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "configmap-7739" for this suite.
-Jun  4 18:39:52.818: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 18:39:53.148: INFO: namespace configmap-7739 deletion completed in 6.375843173s
-
-• [SLOW TEST:8.818 seconds]
-[sig-node] ConfigMap
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/configmap.go:32
-  should be consumable via environment variable [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-------------------------------
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
-------------------------------
-[k8s.io] KubeletManagedEtcHosts 
-  should test kubelet managed /etc/hosts file [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [k8s.io] KubeletManagedEtcHosts
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
-STEP: Creating a kubernetes client
-Jun  4 18:39:53.148: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-STEP: Building a namespace api object, basename e2e-kubelet-etc-hosts
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-kubelet-etc-hosts-4637
-STEP: Waiting for a default service account to be provisioned in namespace
-[It] should test kubelet managed /etc/hosts file [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: Setting up the test
-STEP: Creating hostNetwork=false pod
-STEP: Creating hostNetwork=true pod
-STEP: Running the test
-STEP: Verifying /etc/hosts of container is kubelet-managed for pod with hostNetwork=false
-Jun  4 18:40:01.473: INFO: ExecWithOptions {Command:[cat /etc/hosts] Namespace:e2e-kubelet-etc-hosts-4637 PodName:test-pod ContainerName:busybox-1 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
-Jun  4 18:40:01.473: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-Jun  4 18:40:01.681: INFO: Exec stderr: ""
-Jun  4 18:40:01.681: INFO: ExecWithOptions {Command:[cat /etc/hosts-original] Namespace:e2e-kubelet-etc-hosts-4637 PodName:test-pod ContainerName:busybox-1 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
-Jun  4 18:40:01.681: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-Jun  4 18:40:01.853: INFO: Exec stderr: ""
-Jun  4 18:40:01.853: INFO: ExecWithOptions {Command:[cat /etc/hosts] Namespace:e2e-kubelet-etc-hosts-4637 PodName:test-pod ContainerName:busybox-2 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
-Jun  4 18:40:01.853: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-Jun  4 18:40:02.072: INFO: Exec stderr: ""
-Jun  4 18:40:02.072: INFO: ExecWithOptions {Command:[cat /etc/hosts-original] Namespace:e2e-kubelet-etc-hosts-4637 PodName:test-pod ContainerName:busybox-2 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
-Jun  4 18:40:02.072: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-Jun  4 18:40:02.353: INFO: Exec stderr: ""
-STEP: Verifying /etc/hosts of container is not kubelet-managed since container specifies /etc/hosts mount
-Jun  4 18:40:02.353: INFO: ExecWithOptions {Command:[cat /etc/hosts] Namespace:e2e-kubelet-etc-hosts-4637 PodName:test-pod ContainerName:busybox-3 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
-Jun  4 18:40:02.354: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-Jun  4 18:40:02.539: INFO: Exec stderr: ""
-Jun  4 18:40:02.539: INFO: ExecWithOptions {Command:[cat /etc/hosts-original] Namespace:e2e-kubelet-etc-hosts-4637 PodName:test-pod ContainerName:busybox-3 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
-Jun  4 18:40:02.539: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-Jun  4 18:40:02.733: INFO: Exec stderr: ""
-STEP: Verifying /etc/hosts content of container is not kubelet-managed for pod with hostNetwork=true
-Jun  4 18:40:02.733: INFO: ExecWithOptions {Command:[cat /etc/hosts] Namespace:e2e-kubelet-etc-hosts-4637 PodName:test-host-network-pod ContainerName:busybox-1 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
-Jun  4 18:40:02.733: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-Jun  4 18:40:02.924: INFO: Exec stderr: ""
-Jun  4 18:40:02.924: INFO: ExecWithOptions {Command:[cat /etc/hosts-original] Namespace:e2e-kubelet-etc-hosts-4637 PodName:test-host-network-pod ContainerName:busybox-1 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
-Jun  4 18:40:02.924: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-Jun  4 18:40:03.107: INFO: Exec stderr: ""
-Jun  4 18:40:03.107: INFO: ExecWithOptions {Command:[cat /etc/hosts] Namespace:e2e-kubelet-etc-hosts-4637 PodName:test-host-network-pod ContainerName:busybox-2 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
-Jun  4 18:40:03.107: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-Jun  4 18:40:03.270: INFO: Exec stderr: ""
-Jun  4 18:40:03.270: INFO: ExecWithOptions {Command:[cat /etc/hosts-original] Namespace:e2e-kubelet-etc-hosts-4637 PodName:test-host-network-pod ContainerName:busybox-2 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
-Jun  4 18:40:03.271: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-Jun  4 18:40:03.444: INFO: Exec stderr: ""
-[AfterEach] [k8s.io] KubeletManagedEtcHosts
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 18:40:03.444: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "e2e-kubelet-etc-hosts-4637" for this suite.
-Jun  4 18:40:49.536: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 18:40:49.877: INFO: namespace e2e-kubelet-etc-hosts-4637 deletion completed in 46.392817281s
-
-• [SLOW TEST:56.729 seconds]
-[k8s.io] KubeletManagedEtcHosts
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
-  should test kubelet managed /etc/hosts file [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
 SSSSSSS
 ------------------------------
-[sig-storage] EmptyDir volumes 
-  should support (root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [sig-storage] EmptyDir volumes
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
-STEP: Creating a kubernetes client
-Jun  4 18:40:49.878: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-STEP: Building a namespace api object, basename emptydir
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in emptydir-75
-STEP: Waiting for a default service account to be provisioned in namespace
-[It] should support (root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: Creating a pod to test emptydir 0666 on tmpfs
-Jun  4 18:40:50.223: INFO: Waiting up to 5m0s for pod "pod-465d0de1-86f8-11e9-8d1b-467ee19922ac" in namespace "emptydir-75" to be "success or failure"
-Jun  4 18:40:50.234: INFO: Pod "pod-465d0de1-86f8-11e9-8d1b-467ee19922ac": Phase="Pending", Reason="", readiness=false. Elapsed: 11.301608ms
-Jun  4 18:40:52.247: INFO: Pod "pod-465d0de1-86f8-11e9-8d1b-467ee19922ac": Phase="Running", Reason="", readiness=true. Elapsed: 2.023519511s
-Jun  4 18:40:54.260: INFO: Pod "pod-465d0de1-86f8-11e9-8d1b-467ee19922ac": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.036362253s
-STEP: Saw pod success
-Jun  4 18:40:54.260: INFO: Pod "pod-465d0de1-86f8-11e9-8d1b-467ee19922ac" satisfied condition "success or failure"
-Jun  4 18:40:54.272: INFO: Trying to get logs from node 10.212.23.161 pod pod-465d0de1-86f8-11e9-8d1b-467ee19922ac container test-container: 
-STEP: delete the pod
-Jun  4 18:40:54.333: INFO: Waiting for pod pod-465d0de1-86f8-11e9-8d1b-467ee19922ac to disappear
-Jun  4 18:40:54.344: INFO: Pod pod-465d0de1-86f8-11e9-8d1b-467ee19922ac no longer exists
-[AfterEach] [sig-storage] EmptyDir volumes
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 18:40:54.344: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "emptydir-75" for this suite.
-Jun  4 18:41:00.401: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 18:41:00.746: INFO: namespace emptydir-75 deletion completed in 6.387011418s
-
-• [SLOW TEST:10.868 seconds]
-[sig-storage] EmptyDir volumes
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:41
-  should support (root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-------------------------------
-SSSSSSSS
-------------------------------
-[sig-storage] Projected configMap 
-  should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [sig-storage] Projected configMap
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+[sig-api-machinery] Watchers 
+  should be able to restart watching from the last resource version observed by the previous watch [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-api-machinery] Watchers
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 18:41:00.746: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-STEP: Building a namespace api object, basename projected
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-2700
+Jun 18 11:40:59.627: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename watch
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in watch-2129
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: Creating configMap with name projected-configmap-test-volume-4cc7b80a-86f8-11e9-8d1b-467ee19922ac
-STEP: Creating a pod to test consume configMaps
-Jun  4 18:41:01.000: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-4cc9ca6d-86f8-11e9-8d1b-467ee19922ac" in namespace "projected-2700" to be "success or failure"
-Jun  4 18:41:01.012: INFO: Pod "pod-projected-configmaps-4cc9ca6d-86f8-11e9-8d1b-467ee19922ac": Phase="Pending", Reason="", readiness=false. Elapsed: 12.174975ms
-Jun  4 18:41:03.024: INFO: Pod "pod-projected-configmaps-4cc9ca6d-86f8-11e9-8d1b-467ee19922ac": Phase="Running", Reason="", readiness=true. Elapsed: 2.024353282s
-Jun  4 18:41:05.037: INFO: Pod "pod-projected-configmaps-4cc9ca6d-86f8-11e9-8d1b-467ee19922ac": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.037281909s
-STEP: Saw pod success
-Jun  4 18:41:05.037: INFO: Pod "pod-projected-configmaps-4cc9ca6d-86f8-11e9-8d1b-467ee19922ac" satisfied condition "success or failure"
-Jun  4 18:41:05.048: INFO: Trying to get logs from node 10.212.23.164 pod pod-projected-configmaps-4cc9ca6d-86f8-11e9-8d1b-467ee19922ac container projected-configmap-volume-test: 
-STEP: delete the pod
-Jun  4 18:41:05.107: INFO: Waiting for pod pod-projected-configmaps-4cc9ca6d-86f8-11e9-8d1b-467ee19922ac to disappear
-Jun  4 18:41:05.117: INFO: Pod pod-projected-configmaps-4cc9ca6d-86f8-11e9-8d1b-467ee19922ac no longer exists
-[AfterEach] [sig-storage] Projected configMap
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 18:41:05.118: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "projected-2700" for this suite.
-Jun  4 18:41:11.173: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 18:41:11.595: INFO: namespace projected-2700 deletion completed in 6.463452383s
+[It] should be able to restart watching from the last resource version observed by the previous watch [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: creating a watch on configmaps
+STEP: creating a new configmap
+STEP: modifying the configmap once
+STEP: closing the watch once it receives two notifications
+Jun 18 11:40:59.864: INFO: Got : ADDED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-watch-closed,GenerateName:,Namespace:watch-2129,SelfLink:/api/v1/namespaces/watch-2129/configmaps/e2e-watch-test-watch-closed,UID:f1832d68-91bd-11e9-a08a-ee7a14707756,ResourceVersion:87078,Generation:0,CreationTimestamp:2019-06-18 11:40:59 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: watch-closed-and-restarted,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{},BinaryData:map[string][]byte{},}
+Jun 18 11:40:59.864: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-watch-closed,GenerateName:,Namespace:watch-2129,SelfLink:/api/v1/namespaces/watch-2129/configmaps/e2e-watch-test-watch-closed,UID:f1832d68-91bd-11e9-a08a-ee7a14707756,ResourceVersion:87079,Generation:0,CreationTimestamp:2019-06-18 11:40:59 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: watch-closed-and-restarted,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{mutation: 1,},BinaryData:map[string][]byte{},}
+STEP: modifying the configmap a second time, while the watch is closed
+STEP: creating a new watch on configmaps from the last resource version observed by the first watch
+STEP: deleting the configmap
+STEP: Expecting to observe notifications for all changes to the configmap since the first watch closed
+Jun 18 11:40:59.925: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-watch-closed,GenerateName:,Namespace:watch-2129,SelfLink:/api/v1/namespaces/watch-2129/configmaps/e2e-watch-test-watch-closed,UID:f1832d68-91bd-11e9-a08a-ee7a14707756,ResourceVersion:87080,Generation:0,CreationTimestamp:2019-06-18 11:40:59 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: watch-closed-and-restarted,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},}
+Jun 18 11:40:59.925: INFO: Got : DELETED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-watch-closed,GenerateName:,Namespace:watch-2129,SelfLink:/api/v1/namespaces/watch-2129/configmaps/e2e-watch-test-watch-closed,UID:f1832d68-91bd-11e9-a08a-ee7a14707756,ResourceVersion:87081,Generation:0,CreationTimestamp:2019-06-18 11:40:59 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: watch-closed-and-restarted,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},}
+[AfterEach] [sig-api-machinery] Watchers
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 11:40:59.925: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "watch-2129" for this suite.
+Jun 18 11:41:07.964: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 11:41:08.400: INFO: namespace watch-2129 deletion completed in 8.461913259s
 
-• [SLOW TEST:10.849 seconds]
-[sig-storage] Projected configMap
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_configmap.go:33
-  should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+• [SLOW TEST:8.772 seconds]
+[sig-api-machinery] Watchers
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:22
+  should be able to restart watching from the last resource version observed by the previous watch [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-SSS
+SSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
 [sig-storage] Projected downwardAPI 
-  should provide podname only [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+  should provide container's memory request [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 [BeforeEach] [sig-storage] Projected downwardAPI
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 18:41:11.595: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
+Jun 18 11:41:08.400: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
 STEP: Building a namespace api object, basename projected
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-9198
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-714
 STEP: Waiting for a default service account to be provisioned in namespace
 [BeforeEach] [sig-storage] Projected downwardAPI
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:39
-[It] should provide podname only [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:39
+[It] should provide container's memory request [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 STEP: Creating a pod to test downward API volume plugin
-Jun  4 18:41:11.833: INFO: Waiting up to 5m0s for pod "downwardapi-volume-533e7c0b-86f8-11e9-8d1b-467ee19922ac" in namespace "projected-9198" to be "success or failure"
-Jun  4 18:41:11.849: INFO: Pod "downwardapi-volume-533e7c0b-86f8-11e9-8d1b-467ee19922ac": Phase="Pending", Reason="", readiness=false. Elapsed: 15.261687ms
-Jun  4 18:41:13.862: INFO: Pod "downwardapi-volume-533e7c0b-86f8-11e9-8d1b-467ee19922ac": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.028290563s
+Jun 18 11:41:08.622: INFO: Waiting up to 5m0s for pod "downwardapi-volume-f6bd50f1-91bd-11e9-a25d-8608290c688a" in namespace "projected-714" to be "success or failure"
+Jun 18 11:41:08.635: INFO: Pod "downwardapi-volume-f6bd50f1-91bd-11e9-a25d-8608290c688a": Phase="Pending", Reason="", readiness=false. Elapsed: 13.114603ms
+Jun 18 11:41:10.651: INFO: Pod "downwardapi-volume-f6bd50f1-91bd-11e9-a25d-8608290c688a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.028752651s
 STEP: Saw pod success
-Jun  4 18:41:13.862: INFO: Pod "downwardapi-volume-533e7c0b-86f8-11e9-8d1b-467ee19922ac" satisfied condition "success or failure"
-Jun  4 18:41:13.873: INFO: Trying to get logs from node 10.212.23.189 pod downwardapi-volume-533e7c0b-86f8-11e9-8d1b-467ee19922ac container client-container: 
+Jun 18 11:41:10.651: INFO: Pod "downwardapi-volume-f6bd50f1-91bd-11e9-a25d-8608290c688a" satisfied condition "success or failure"
+Jun 18 11:41:10.666: INFO: Trying to get logs from node 10.72.74.138 pod downwardapi-volume-f6bd50f1-91bd-11e9-a25d-8608290c688a container client-container: 
 STEP: delete the pod
-Jun  4 18:41:13.933: INFO: Waiting for pod downwardapi-volume-533e7c0b-86f8-11e9-8d1b-467ee19922ac to disappear
-Jun  4 18:41:13.944: INFO: Pod downwardapi-volume-533e7c0b-86f8-11e9-8d1b-467ee19922ac no longer exists
+Jun 18 11:41:10.752: INFO: Waiting for pod downwardapi-volume-f6bd50f1-91bd-11e9-a25d-8608290c688a to disappear
+Jun 18 11:41:10.766: INFO: Pod downwardapi-volume-f6bd50f1-91bd-11e9-a25d-8608290c688a no longer exists
 [AfterEach] [sig-storage] Projected downwardAPI
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 18:41:13.944: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "projected-9198" for this suite.
-Jun  4 18:41:20.000: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 18:41:20.495: INFO: namespace projected-9198 deletion completed in 6.537410075s
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 11:41:10.766: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "projected-714" for this suite.
+Jun 18 11:41:16.809: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 11:41:17.224: INFO: namespace projected-714 deletion completed in 6.444414423s
 
-• [SLOW TEST:8.900 seconds]
+• [SLOW TEST:8.823 seconds]
 [sig-storage] Projected downwardAPI
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:33
-  should provide podname only [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:33
+  should provide container's memory request [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-SSSSSSSS
+SSSSSSSSSSSSSSS
 ------------------------------
-[sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] 
-  Burst scaling should run to completion even with unhealthy pods [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [sig-apps] StatefulSet
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+[sig-storage] ConfigMap 
+  binary data should be reflected in volume [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-storage] ConfigMap
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 18:41:20.496: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-STEP: Building a namespace api object, basename statefulset
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in statefulset-8606
+Jun 18 11:41:17.224: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename configmap
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in configmap-1337
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-apps] StatefulSet
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:59
-[BeforeEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:74
-STEP: Creating service test in namespace statefulset-8606
-[It] Burst scaling should run to completion even with unhealthy pods [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: Creating stateful set ss in namespace statefulset-8606
-STEP: Waiting until all stateful set ss replicas will be running in namespace statefulset-8606
-Jun  4 18:41:20.793: INFO: Found 0 stateful pods, waiting for 1
-Jun  4 18:41:30.808: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true
-STEP: Confirming that stateful set scale up will not halt with unhealthy stateful pod
-Jun  4 18:41:30.820: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 exec --namespace=statefulset-8606 ss-0 -- /bin/sh -x -c mv -v /usr/share/nginx/html/index.html /tmp/ || true'
-Jun  4 18:41:31.148: INFO: stderr: "+ mv -v /usr/share/nginx/html/index.html /tmp/\n"
-Jun  4 18:41:31.148: INFO: stdout: "'/usr/share/nginx/html/index.html' -> '/tmp/index.html'\n"
-Jun  4 18:41:31.148: INFO: stdout of mv -v /usr/share/nginx/html/index.html /tmp/ || true on ss-0: '/usr/share/nginx/html/index.html' -> '/tmp/index.html'
-
-Jun  4 18:41:31.161: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=true
-Jun  4 18:41:41.176: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=false
-Jun  4 18:41:41.176: INFO: Waiting for statefulset status.replicas updated to 0
-Jun  4 18:41:41.214: INFO: POD   NODE           PHASE    GRACE  CONDITIONS
-Jun  4 18:41:41.214: INFO: ss-0  10.212.23.161  Running         [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 18:41:20 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-04 18:41:31 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-04 18:41:31 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 18:41:20 +0000 UTC  }]
-Jun  4 18:41:41.214: INFO: 
-Jun  4 18:41:41.214: INFO: StatefulSet ss has not reached scale 3, at 1
-Jun  4 18:41:42.226: INFO: Verifying statefulset ss doesn't scale past 3 for another 8.988756791s
-Jun  4 18:41:43.237: INFO: Verifying statefulset ss doesn't scale past 3 for another 7.977190007s
-Jun  4 18:41:44.251: INFO: Verifying statefulset ss doesn't scale past 3 for another 6.965407702s
-Jun  4 18:41:45.269: INFO: Verifying statefulset ss doesn't scale past 3 for another 5.952245492s
-Jun  4 18:41:46.281: INFO: Verifying statefulset ss doesn't scale past 3 for another 4.934282231s
-Jun  4 18:41:47.292: INFO: Verifying statefulset ss doesn't scale past 3 for another 3.922145303s
-Jun  4 18:41:48.305: INFO: Verifying statefulset ss doesn't scale past 3 for another 2.910746593s
-Jun  4 18:41:49.319: INFO: Verifying statefulset ss doesn't scale past 3 for another 1.89829989s
-Jun  4 18:41:50.331: INFO: Verifying statefulset ss doesn't scale past 3 for another 884.183151ms
-STEP: Scaling up stateful set ss to 3 replicas and waiting until all of them will be running in namespace statefulset-8606
-Jun  4 18:41:51.345: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 exec --namespace=statefulset-8606 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
-Jun  4 18:41:51.659: INFO: stderr: "+ mv -v /tmp/index.html /usr/share/nginx/html/\n"
-Jun  4 18:41:51.659: INFO: stdout: "'/tmp/index.html' -> '/usr/share/nginx/html/index.html'\n"
-Jun  4 18:41:51.659: INFO: stdout of mv -v /tmp/index.html /usr/share/nginx/html/ || true on ss-0: '/tmp/index.html' -> '/usr/share/nginx/html/index.html'
-
-Jun  4 18:41:51.659: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 exec --namespace=statefulset-8606 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
-Jun  4 18:41:51.934: INFO: stderr: "+ mv -v /tmp/index.html /usr/share/nginx/html/\nmv: can't rename '/tmp/index.html': No such file or directory\n+ true\n"
-Jun  4 18:41:51.934: INFO: stdout: "'/tmp/index.html' -> '/usr/share/nginx/html/index.html'\n"
-Jun  4 18:41:51.934: INFO: stdout of mv -v /tmp/index.html /usr/share/nginx/html/ || true on ss-1: '/tmp/index.html' -> '/usr/share/nginx/html/index.html'
-
-Jun  4 18:41:51.934: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 exec --namespace=statefulset-8606 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
-Jun  4 18:41:52.216: INFO: stderr: "+ mv -v /tmp/index.html /usr/share/nginx/html/\nmv: can't rename '/tmp/index.html': No such file or directory\n+ true\n"
-Jun  4 18:41:52.216: INFO: stdout: "'/tmp/index.html' -> '/usr/share/nginx/html/index.html'\n"
-Jun  4 18:41:52.216: INFO: stdout of mv -v /tmp/index.html /usr/share/nginx/html/ || true on ss-2: '/tmp/index.html' -> '/usr/share/nginx/html/index.html'
-
-Jun  4 18:41:52.235: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true
-Jun  4 18:41:52.235: INFO: Waiting for pod ss-1 to enter Running - Ready=true, currently Running - Ready=true
-Jun  4 18:41:52.235: INFO: Waiting for pod ss-2 to enter Running - Ready=true, currently Running - Ready=true
-STEP: Scale down will not halt with unhealthy stateful pod
-Jun  4 18:41:52.256: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 exec --namespace=statefulset-8606 ss-0 -- /bin/sh -x -c mv -v /usr/share/nginx/html/index.html /tmp/ || true'
-Jun  4 18:41:52.531: INFO: stderr: "+ mv -v /usr/share/nginx/html/index.html /tmp/\n"
-Jun  4 18:41:52.532: INFO: stdout: "'/usr/share/nginx/html/index.html' -> '/tmp/index.html'\n"
-Jun  4 18:41:52.532: INFO: stdout of mv -v /usr/share/nginx/html/index.html /tmp/ || true on ss-0: '/usr/share/nginx/html/index.html' -> '/tmp/index.html'
-
-Jun  4 18:41:52.532: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 exec --namespace=statefulset-8606 ss-1 -- /bin/sh -x -c mv -v /usr/share/nginx/html/index.html /tmp/ || true'
-Jun  4 18:41:52.791: INFO: stderr: "+ mv -v /usr/share/nginx/html/index.html /tmp/\n"
-Jun  4 18:41:52.791: INFO: stdout: "'/usr/share/nginx/html/index.html' -> '/tmp/index.html'\n"
-Jun  4 18:41:52.791: INFO: stdout of mv -v /usr/share/nginx/html/index.html /tmp/ || true on ss-1: '/usr/share/nginx/html/index.html' -> '/tmp/index.html'
-
-Jun  4 18:41:52.791: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 exec --namespace=statefulset-8606 ss-2 -- /bin/sh -x -c mv -v /usr/share/nginx/html/index.html /tmp/ || true'
-Jun  4 18:41:53.070: INFO: stderr: "+ mv -v /usr/share/nginx/html/index.html /tmp/\n"
-Jun  4 18:41:53.070: INFO: stdout: "'/usr/share/nginx/html/index.html' -> '/tmp/index.html'\n"
-Jun  4 18:41:53.070: INFO: stdout of mv -v /usr/share/nginx/html/index.html /tmp/ || true on ss-2: '/usr/share/nginx/html/index.html' -> '/tmp/index.html'
-
-Jun  4 18:41:53.070: INFO: Waiting for statefulset status.replicas updated to 0
-Jun  4 18:41:53.079: INFO: Waiting for stateful set status.readyReplicas to become 0, currently 1
-Jun  4 18:42:03.166: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=false
-Jun  4 18:42:03.166: INFO: Waiting for pod ss-1 to enter Running - Ready=false, currently Running - Ready=false
-Jun  4 18:42:03.166: INFO: Waiting for pod ss-2 to enter Running - Ready=false, currently Running - Ready=false
-Jun  4 18:42:03.214: INFO: POD   NODE           PHASE    GRACE  CONDITIONS
-Jun  4 18:42:03.214: INFO: ss-0  10.212.23.161  Running         [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 18:41:20 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-04 18:41:52 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-04 18:41:52 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 18:41:20 +0000 UTC  }]
-Jun  4 18:42:03.214: INFO: ss-1  10.212.23.189  Running  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 18:41:41 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-04 18:41:52 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-04 18:41:52 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 18:41:41 +0000 UTC  }]
-Jun  4 18:42:03.214: INFO: ss-2  10.212.23.164  Running  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 18:41:41 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-04 18:41:54 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-04 18:41:54 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 18:41:41 +0000 UTC  }]
-Jun  4 18:42:03.214: INFO: 
-Jun  4 18:42:03.214: INFO: StatefulSet ss has not reached scale 0, at 3
-Jun  4 18:42:04.227: INFO: POD   NODE           PHASE    GRACE  CONDITIONS
-Jun  4 18:42:04.227: INFO: ss-0  10.212.23.161  Running  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 18:41:20 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-04 18:41:52 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-04 18:41:52 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 18:41:20 +0000 UTC  }]
-Jun  4 18:42:04.227: INFO: ss-1  10.212.23.189  Running  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 18:41:41 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-04 18:41:52 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-04 18:41:52 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 18:41:41 +0000 UTC  }]
-Jun  4 18:42:04.227: INFO: ss-2  10.212.23.164  Running  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 18:41:41 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-04 18:41:54 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-04 18:41:54 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 18:41:41 +0000 UTC  }]
-Jun  4 18:42:04.227: INFO: 
-Jun  4 18:42:04.227: INFO: StatefulSet ss has not reached scale 0, at 3
-Jun  4 18:42:05.244: INFO: POD   NODE           PHASE    GRACE  CONDITIONS
-Jun  4 18:42:05.244: INFO: ss-0  10.212.23.161  Pending  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 18:41:20 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-04 18:41:52 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-04 18:41:52 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 18:41:20 +0000 UTC  }]
-Jun  4 18:42:05.244: INFO: ss-1  10.212.23.189  Running  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 18:41:41 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-04 18:41:52 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-04 18:41:52 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 18:41:41 +0000 UTC  }]
-Jun  4 18:42:05.244: INFO: ss-2  10.212.23.164  Running  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 18:41:41 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-04 18:41:54 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-04 18:41:54 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 18:41:41 +0000 UTC  }]
-Jun  4 18:42:05.244: INFO: 
-Jun  4 18:42:05.244: INFO: StatefulSet ss has not reached scale 0, at 3
-Jun  4 18:42:06.256: INFO: POD   NODE           PHASE    GRACE  CONDITIONS
-Jun  4 18:42:06.256: INFO: ss-1  10.212.23.189  Running  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 18:41:41 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-04 18:41:52 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-04 18:41:52 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 18:41:41 +0000 UTC  }]
-Jun  4 18:42:06.256: INFO: ss-2  10.212.23.164  Running  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 18:41:41 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-04 18:41:54 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-04 18:41:54 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 18:41:41 +0000 UTC  }]
-Jun  4 18:42:06.256: INFO: 
-Jun  4 18:42:06.256: INFO: StatefulSet ss has not reached scale 0, at 2
-Jun  4 18:42:07.270: INFO: POD   NODE           PHASE    GRACE  CONDITIONS
-Jun  4 18:42:07.270: INFO: ss-1  10.212.23.189  Running  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 18:41:41 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-04 18:41:52 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-04 18:41:52 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 18:41:41 +0000 UTC  }]
-Jun  4 18:42:07.270: INFO: ss-2  10.212.23.164  Running  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 18:41:41 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-04 18:41:54 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-04 18:41:54 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 18:41:41 +0000 UTC  }]
-Jun  4 18:42:07.270: INFO: 
-Jun  4 18:42:07.270: INFO: StatefulSet ss has not reached scale 0, at 2
-Jun  4 18:42:08.284: INFO: POD   NODE           PHASE    GRACE  CONDITIONS
-Jun  4 18:42:08.284: INFO: ss-1  10.212.23.189  Running  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 18:41:41 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-04 18:41:52 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-04 18:41:52 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 18:41:41 +0000 UTC  }]
-Jun  4 18:42:08.284: INFO: 
-Jun  4 18:42:08.284: INFO: StatefulSet ss has not reached scale 0, at 1
-Jun  4 18:42:09.297: INFO: POD   NODE           PHASE    GRACE  CONDITIONS
-Jun  4 18:42:09.297: INFO: ss-1  10.212.23.189  Running  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 18:41:41 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-04 18:41:52 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-04 18:41:52 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 18:41:41 +0000 UTC  }]
-Jun  4 18:42:09.297: INFO: 
-Jun  4 18:42:09.297: INFO: StatefulSet ss has not reached scale 0, at 1
-Jun  4 18:42:10.309: INFO: POD   NODE           PHASE    GRACE  CONDITIONS
-Jun  4 18:42:10.309: INFO: ss-1  10.212.23.189  Running  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 18:41:41 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-04 18:41:52 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-04 18:41:52 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 18:41:41 +0000 UTC  }]
-Jun  4 18:42:10.309: INFO: 
-Jun  4 18:42:10.309: INFO: StatefulSet ss has not reached scale 0, at 1
-Jun  4 18:42:11.321: INFO: POD   NODE           PHASE    GRACE  CONDITIONS
-Jun  4 18:42:11.321: INFO: ss-1  10.212.23.189  Running  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 18:41:41 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-04 18:41:52 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-04 18:41:52 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 18:41:41 +0000 UTC  }]
-Jun  4 18:42:11.321: INFO: 
-Jun  4 18:42:11.321: INFO: StatefulSet ss has not reached scale 0, at 1
-Jun  4 18:42:12.334: INFO: POD   NODE           PHASE    GRACE  CONDITIONS
-Jun  4 18:42:12.334: INFO: ss-1  10.212.23.189  Running  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 18:41:41 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-04 18:41:52 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-04 18:41:52 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 18:41:41 +0000 UTC  }]
-Jun  4 18:42:12.334: INFO: 
-Jun  4 18:42:12.334: INFO: StatefulSet ss has not reached scale 0, at 1
-STEP: Scaling down stateful set ss to 0 replicas and waiting until none of pods will run in namespacestatefulset-8606
-Jun  4 18:42:13.377: INFO: Scaling statefulset ss to 0
-Jun  4 18:42:13.405: INFO: Waiting for statefulset status.replicas updated to 0
-[AfterEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:85
-Jun  4 18:42:13.411: INFO: Deleting all statefulset in ns statefulset-8606
-Jun  4 18:42:13.418: INFO: Scaling statefulset ss to 0
-Jun  4 18:42:13.474: INFO: Waiting for statefulset status.replicas updated to 0
-Jun  4 18:42:13.482: INFO: Deleting statefulset ss
-[AfterEach] [sig-apps] StatefulSet
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 18:42:13.517: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "statefulset-8606" for this suite.
-Jun  4 18:42:21.581: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 18:42:21.972: INFO: namespace statefulset-8606 deletion completed in 8.437294827s
+[It] binary data should be reflected in volume [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating configMap with name configmap-test-upd-fc025375-91bd-11e9-a25d-8608290c688a
+STEP: Creating the pod
+STEP: Waiting for pod with text data
+STEP: Waiting for pod with binary data
+[AfterEach] [sig-storage] ConfigMap
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 11:41:21.719: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "configmap-1337" for this suite.
+Jun 18 11:41:46.067: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 11:41:46.477: INFO: namespace configmap-1337 deletion completed in 24.743348394s
 
-• [SLOW TEST:61.477 seconds]
-[sig-apps] StatefulSet
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22
-  [k8s.io] Basic StatefulSet functionality [StatefulSetBasic]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
-    Burst scaling should run to completion even with unhealthy pods [Conformance]
-    /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-------------------------------
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+• [SLOW TEST:29.253 seconds]
+[sig-storage] ConfigMap
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:32
+  binary data should be reflected in volume [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-[sig-storage] Projected secret 
-  should be consumable from pods in volume with mappings and Item Mode set [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [sig-storage] Projected secret
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+[sig-storage] Secrets 
+  should be able to mount in a volume regardless of a different secret existing with same name in different namespace [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-storage] Secrets
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 18:42:21.973: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-STEP: Building a namespace api object, basename projected
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-8979
+Jun 18 11:41:46.477: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename secrets
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in secrets-1576
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should be consumable from pods in volume with mappings and Item Mode set [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: Creating projection with secret that has name projected-secret-test-map-7d311ddc-86f8-11e9-8d1b-467ee19922ac
+[It] should be able to mount in a volume regardless of a different secret existing with same name in different namespace [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in secret-namespace-9496
+STEP: Creating secret with name secret-test-0d6f3696-91be-11e9-a25d-8608290c688a
 STEP: Creating a pod to test consume secrets
-Jun  4 18:42:22.220: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-7d32ab0b-86f8-11e9-8d1b-467ee19922ac" in namespace "projected-8979" to be "success or failure"
-Jun  4 18:42:22.231: INFO: Pod "pod-projected-secrets-7d32ab0b-86f8-11e9-8d1b-467ee19922ac": Phase="Pending", Reason="", readiness=false. Elapsed: 10.946832ms
-Jun  4 18:42:24.245: INFO: Pod "pod-projected-secrets-7d32ab0b-86f8-11e9-8d1b-467ee19922ac": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.024957817s
+Jun 18 11:41:46.905: INFO: Waiting up to 5m0s for pod "pod-secrets-0d8f1ebb-91be-11e9-a25d-8608290c688a" in namespace "secrets-1576" to be "success or failure"
+Jun 18 11:41:46.917: INFO: Pod "pod-secrets-0d8f1ebb-91be-11e9-a25d-8608290c688a": Phase="Pending", Reason="", readiness=false. Elapsed: 12.116535ms
+Jun 18 11:41:48.931: INFO: Pod "pod-secrets-0d8f1ebb-91be-11e9-a25d-8608290c688a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.025622775s
 STEP: Saw pod success
-Jun  4 18:42:24.245: INFO: Pod "pod-projected-secrets-7d32ab0b-86f8-11e9-8d1b-467ee19922ac" satisfied condition "success or failure"
-Jun  4 18:42:24.268: INFO: Trying to get logs from node 10.212.23.161 pod pod-projected-secrets-7d32ab0b-86f8-11e9-8d1b-467ee19922ac container projected-secret-volume-test: 
+Jun 18 11:41:48.931: INFO: Pod "pod-secrets-0d8f1ebb-91be-11e9-a25d-8608290c688a" satisfied condition "success or failure"
+Jun 18 11:41:48.944: INFO: Trying to get logs from node 10.72.74.138 pod pod-secrets-0d8f1ebb-91be-11e9-a25d-8608290c688a container secret-volume-test: 
 STEP: delete the pod
-Jun  4 18:42:24.330: INFO: Waiting for pod pod-projected-secrets-7d32ab0b-86f8-11e9-8d1b-467ee19922ac to disappear
-Jun  4 18:42:24.341: INFO: Pod pod-projected-secrets-7d32ab0b-86f8-11e9-8d1b-467ee19922ac no longer exists
-[AfterEach] [sig-storage] Projected secret
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 18:42:24.341: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "projected-8979" for this suite.
-Jun  4 18:42:30.396: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 18:42:30.775: INFO: namespace projected-8979 deletion completed in 6.420862124s
-
-• [SLOW TEST:8.802 seconds]
-[sig-storage] Projected secret
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_secret.go:33
-  should be consumable from pods in volume with mappings and Item Mode set [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+Jun 18 11:41:49.669: INFO: Waiting for pod pod-secrets-0d8f1ebb-91be-11e9-a25d-8608290c688a to disappear
+Jun 18 11:41:49.681: INFO: Pod pod-secrets-0d8f1ebb-91be-11e9-a25d-8608290c688a no longer exists
+[AfterEach] [sig-storage] Secrets
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 11:41:49.681: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "secrets-1576" for this suite.
+Jun 18 11:41:55.720: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 11:41:56.122: INFO: namespace secrets-1576 deletion completed in 6.427362641s
+STEP: Destroying namespace "secret-namespace-9496" for this suite.
+Jun 18 11:42:02.148: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 11:42:02.544: INFO: namespace secret-namespace-9496 deletion completed in 6.422349253s
+
+• [SLOW TEST:16.067 seconds]
+[sig-storage] Secrets
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/secrets_volume.go:33
+  should be able to mount in a volume regardless of a different secret existing with same name in different namespace [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-SSSSSSSSSSSSSSS
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
 [sig-storage] Downward API volume 
   should provide container's cpu request [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 [BeforeEach] [sig-storage] Downward API volume
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 18:42:30.776: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
+Jun 18 11:42:02.546: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
 STEP: Building a namespace api object, basename downward-api
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in downward-api-5820
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in downward-api-8097
 STEP: Waiting for a default service account to be provisioned in namespace
 [BeforeEach] [sig-storage] Downward API volume
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:39
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:39
 [It] should provide container's cpu request [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 STEP: Creating a pod to test downward API volume plugin
-Jun  4 18:42:31.015: INFO: Waiting up to 5m0s for pod "downwardapi-volume-8270277a-86f8-11e9-8d1b-467ee19922ac" in namespace "downward-api-5820" to be "success or failure"
-Jun  4 18:42:31.029: INFO: Pod "downwardapi-volume-8270277a-86f8-11e9-8d1b-467ee19922ac": Phase="Pending", Reason="", readiness=false. Elapsed: 14.557698ms
-Jun  4 18:42:33.042: INFO: Pod "downwardapi-volume-8270277a-86f8-11e9-8d1b-467ee19922ac": Phase="Running", Reason="", readiness=true. Elapsed: 2.027303583s
-Jun  4 18:42:35.056: INFO: Pod "downwardapi-volume-8270277a-86f8-11e9-8d1b-467ee19922ac": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.041464887s
+Jun 18 11:42:02.769: INFO: Waiting up to 5m0s for pod "downwardapi-volume-17035272-91be-11e9-a25d-8608290c688a" in namespace "downward-api-8097" to be "success or failure"
+Jun 18 11:42:02.784: INFO: Pod "downwardapi-volume-17035272-91be-11e9-a25d-8608290c688a": Phase="Pending", Reason="", readiness=false. Elapsed: 14.109522ms
+Jun 18 11:42:04.798: INFO: Pod "downwardapi-volume-17035272-91be-11e9-a25d-8608290c688a": Phase="Pending", Reason="", readiness=false. Elapsed: 2.028321802s
+Jun 18 11:42:06.811: INFO: Pod "downwardapi-volume-17035272-91be-11e9-a25d-8608290c688a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.04204316s
 STEP: Saw pod success
-Jun  4 18:42:35.056: INFO: Pod "downwardapi-volume-8270277a-86f8-11e9-8d1b-467ee19922ac" satisfied condition "success or failure"
-Jun  4 18:42:35.155: INFO: Trying to get logs from node 10.212.23.164 pod downwardapi-volume-8270277a-86f8-11e9-8d1b-467ee19922ac container client-container: 
+Jun 18 11:42:06.811: INFO: Pod "downwardapi-volume-17035272-91be-11e9-a25d-8608290c688a" satisfied condition "success or failure"
+Jun 18 11:42:06.827: INFO: Trying to get logs from node 10.72.74.184 pod downwardapi-volume-17035272-91be-11e9-a25d-8608290c688a container client-container: 
 STEP: delete the pod
-Jun  4 18:42:35.226: INFO: Waiting for pod downwardapi-volume-8270277a-86f8-11e9-8d1b-467ee19922ac to disappear
-Jun  4 18:42:35.236: INFO: Pod downwardapi-volume-8270277a-86f8-11e9-8d1b-467ee19922ac no longer exists
+Jun 18 11:42:06.909: INFO: Waiting for pod downwardapi-volume-17035272-91be-11e9-a25d-8608290c688a to disappear
+Jun 18 11:42:06.921: INFO: Pod downwardapi-volume-17035272-91be-11e9-a25d-8608290c688a no longer exists
 [AfterEach] [sig-storage] Downward API volume
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 18:42:35.236: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "downward-api-5820" for this suite.
-Jun  4 18:42:41.305: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 18:42:41.697: INFO: namespace downward-api-5820 deletion completed in 6.442433875s
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 11:42:06.921: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "downward-api-8097" for this suite.
+Jun 18 11:42:12.963: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 11:42:13.385: INFO: namespace downward-api-8097 deletion completed in 6.449444012s
 
-• [SLOW TEST:10.921 seconds]
+• [SLOW TEST:10.840 seconds]
 [sig-storage] Downward API volume
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:34
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:34
   should provide container's cpu request [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-SSSSSSSSSSSSSSSSSSSSSSSSSSS
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-node] Downward API 
-  should provide host IP as an env var [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [sig-node] Downward API
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+[k8s.io] Container Lifecycle Hook when create a pod with lifecycle hook 
+  should execute poststart exec hook properly [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [k8s.io] Container Lifecycle Hook
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 18:42:41.698: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-STEP: Building a namespace api object, basename downward-api
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in downward-api-8951
+Jun 18 11:42:13.386: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename container-lifecycle-hook
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in container-lifecycle-hook-9636
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should provide host IP as an env var [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: Creating a pod to test downward api env vars
-Jun  4 18:42:41.938: INFO: Waiting up to 5m0s for pod "downward-api-88f3ab05-86f8-11e9-8d1b-467ee19922ac" in namespace "downward-api-8951" to be "success or failure"
-Jun  4 18:42:41.955: INFO: Pod "downward-api-88f3ab05-86f8-11e9-8d1b-467ee19922ac": Phase="Pending", Reason="", readiness=false. Elapsed: 16.434009ms
-Jun  4 18:42:43.967: INFO: Pod "downward-api-88f3ab05-86f8-11e9-8d1b-467ee19922ac": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.028747678s
-STEP: Saw pod success
-Jun  4 18:42:43.967: INFO: Pod "downward-api-88f3ab05-86f8-11e9-8d1b-467ee19922ac" satisfied condition "success or failure"
-Jun  4 18:42:43.978: INFO: Trying to get logs from node 10.212.23.164 pod downward-api-88f3ab05-86f8-11e9-8d1b-467ee19922ac container dapi-container: 
-STEP: delete the pod
-Jun  4 18:42:44.094: INFO: Waiting for pod downward-api-88f3ab05-86f8-11e9-8d1b-467ee19922ac to disappear
-Jun  4 18:42:44.104: INFO: Pod downward-api-88f3ab05-86f8-11e9-8d1b-467ee19922ac no longer exists
-[AfterEach] [sig-node] Downward API
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 18:42:44.104: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "downward-api-8951" for this suite.
-Jun  4 18:42:50.165: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 18:42:50.604: INFO: namespace downward-api-8951 deletion completed in 6.48346777s
-
-• [SLOW TEST:8.906 seconds]
-[sig-node] Downward API
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downward_api.go:38
-  should provide host IP as an env var [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] when create a pod with lifecycle hook
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:61
+STEP: create the container to handle the HTTPGet hook request.
+[It] should execute poststart exec hook properly [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: create the pod with lifecycle hook
+STEP: check poststart hook
+STEP: delete the pod with lifecycle hook
+Jun 18 11:42:26.513: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear
+Jun 18 11:42:26.525: INFO: Pod pod-with-poststart-exec-hook still exists
+Jun 18 11:42:28.526: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear
+Jun 18 11:42:28.539: INFO: Pod pod-with-poststart-exec-hook still exists
+Jun 18 11:42:30.526: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear
+Jun 18 11:42:30.539: INFO: Pod pod-with-poststart-exec-hook still exists
+Jun 18 11:42:32.526: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear
+Jun 18 11:42:32.540: INFO: Pod pod-with-poststart-exec-hook still exists
+Jun 18 11:42:34.526: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear
+Jun 18 11:42:34.539: INFO: Pod pod-with-poststart-exec-hook still exists
+Jun 18 11:42:36.526: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear
+Jun 18 11:42:36.540: INFO: Pod pod-with-poststart-exec-hook still exists
+Jun 18 11:42:38.526: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear
+Jun 18 11:42:38.539: INFO: Pod pod-with-poststart-exec-hook still exists
+Jun 18 11:42:40.526: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear
+Jun 18 11:42:41.024: INFO: Pod pod-with-poststart-exec-hook still exists
+Jun 18 11:42:42.526: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear
+Jun 18 11:42:42.539: INFO: Pod pod-with-poststart-exec-hook still exists
+Jun 18 11:42:44.526: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear
+Jun 18 11:42:44.540: INFO: Pod pod-with-poststart-exec-hook still exists
+Jun 18 11:42:46.526: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear
+Jun 18 11:42:46.539: INFO: Pod pod-with-poststart-exec-hook still exists
+Jun 18 11:42:48.526: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear
+Jun 18 11:42:48.541: INFO: Pod pod-with-poststart-exec-hook still exists
+Jun 18 11:42:50.526: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear
+Jun 18 11:42:50.539: INFO: Pod pod-with-poststart-exec-hook still exists
+Jun 18 11:42:52.526: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear
+Jun 18 11:42:52.539: INFO: Pod pod-with-poststart-exec-hook no longer exists
+[AfterEach] [k8s.io] Container Lifecycle Hook
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 11:42:52.539: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "container-lifecycle-hook-9636" for this suite.
+Jun 18 11:43:16.594: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 11:43:17.014: INFO: namespace container-lifecycle-hook-9636 deletion completed in 24.446904428s
+
+• [SLOW TEST:63.628 seconds]
+[k8s.io] Container Lifecycle Hook
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
+  when create a pod with lifecycle hook
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:40
+    should execute poststart exec hook properly [NodeConformance] [Conformance]
+    /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-SSSSS
+SSSSSSSSSSSS
 ------------------------------
-[sig-node] ConfigMap 
-  should fail to create ConfigMap with empty key [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [sig-node] ConfigMap
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+[k8s.io] InitContainer [NodeConformance] 
+  should invoke init containers on a RestartAlways pod [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [k8s.io] InitContainer [NodeConformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 18:42:50.604: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-STEP: Building a namespace api object, basename configmap
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in configmap-6141
+Jun 18 11:43:17.015: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename init-container
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in init-container-9930
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should fail to create ConfigMap with empty key [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: Creating configMap that has name configmap-test-emptyKey-8e430c22-86f8-11e9-8d1b-467ee19922ac
-[AfterEach] [sig-node] ConfigMap
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 18:42:50.833: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "configmap-6141" for this suite.
-Jun  4 18:42:56.892: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 18:42:57.281: INFO: namespace configmap-6141 deletion completed in 6.430881042s
+[BeforeEach] [k8s.io] InitContainer [NodeConformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/init_container.go:43
+[It] should invoke init containers on a RestartAlways pod [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: creating the pod
+Jun 18 11:43:17.215: INFO: PodSpec: initContainers in spec.initContainers
+[AfterEach] [k8s.io] InitContainer [NodeConformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 11:43:22.096: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "init-container-9930" for this suite.
+Jun 18 11:43:46.141: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 11:43:46.530: INFO: namespace init-container-9930 deletion completed in 24.420357876s
 
-• [SLOW TEST:6.677 seconds]
-[sig-node] ConfigMap
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/configmap.go:32
-  should fail to create ConfigMap with empty key [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+• [SLOW TEST:29.515 seconds]
+[k8s.io] InitContainer [NodeConformance]
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
+  should invoke init containers on a RestartAlways pod [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-SSSSSSSSSSS
+SSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[k8s.io] [sig-node] Pods Extended [k8s.io] Pods Set QOS Class 
-  should be submitted and removed  [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [k8s.io] [sig-node] Pods Extended
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+[sig-apps] ReplicationController 
+  should adopt matching pods on creation [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-apps] ReplicationController
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 18:42:57.281: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-STEP: Building a namespace api object, basename pods
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in pods-1617
+Jun 18 11:43:46.531: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename replication-controller
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in replication-controller-582
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [k8s.io] Pods Set QOS Class
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/node/pods.go:177
-[It] should be submitted and removed  [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: creating the pod
-STEP: submitting the pod to kubernetes
-STEP: verifying QOS class is set on the pod
-[AfterEach] [k8s.io] [sig-node] Pods Extended
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 18:42:57.528: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "pods-1617" for this suite.
-Jun  4 18:43:21.587: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 18:43:21.977: INFO: namespace pods-1617 deletion completed in 24.431045305s
+[It] should adopt matching pods on creation [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Given a Pod with a 'name' label pod-adoption is created
+STEP: When a replication controller with a matching selector is created
+STEP: Then the orphan pod is adopted
+[AfterEach] [sig-apps] ReplicationController
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 11:43:49.838: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "replication-controller-582" for this suite.
+Jun 18 11:44:11.878: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 11:44:12.276: INFO: namespace replication-controller-582 deletion completed in 22.42491148s
 
-• [SLOW TEST:24.696 seconds]
-[k8s.io] [sig-node] Pods Extended
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
-  [k8s.io] Pods Set QOS Class
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
-    should be submitted and removed  [Conformance]
-    /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+• [SLOW TEST:25.746 seconds]
+[sig-apps] ReplicationController
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22
+  should adopt matching pods on creation [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-storage] ConfigMap 
-  should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [sig-storage] ConfigMap
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+[sig-storage] Downward API volume 
+  should set mode on item file [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-storage] Downward API volume
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 18:43:21.977: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-STEP: Building a namespace api object, basename configmap
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in configmap-3445
+Jun 18 11:44:12.277: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename downward-api
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in downward-api-7443
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: Creating configMap with name configmap-test-volume-map-a0f4f830-86f8-11e9-8d1b-467ee19922ac
-STEP: Creating a pod to test consume configMaps
-Jun  4 18:43:22.228: INFO: Waiting up to 5m0s for pod "pod-configmaps-a0f72453-86f8-11e9-8d1b-467ee19922ac" in namespace "configmap-3445" to be "success or failure"
-Jun  4 18:43:22.240: INFO: Pod "pod-configmaps-a0f72453-86f8-11e9-8d1b-467ee19922ac": Phase="Pending", Reason="", readiness=false. Elapsed: 12.598206ms
-Jun  4 18:43:24.253: INFO: Pod "pod-configmaps-a0f72453-86f8-11e9-8d1b-467ee19922ac": Phase="Running", Reason="", readiness=true. Elapsed: 2.025573875s
-Jun  4 18:43:26.266: INFO: Pod "pod-configmaps-a0f72453-86f8-11e9-8d1b-467ee19922ac": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.038441473s
+[BeforeEach] [sig-storage] Downward API volume
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:39
+[It] should set mode on item file [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating a pod to test downward API volume plugin
+Jun 18 11:44:12.515: INFO: Waiting up to 5m0s for pod "downwardapi-volume-64591943-91be-11e9-a25d-8608290c688a" in namespace "downward-api-7443" to be "success or failure"
+Jun 18 11:44:12.529: INFO: Pod "downwardapi-volume-64591943-91be-11e9-a25d-8608290c688a": Phase="Pending", Reason="", readiness=false. Elapsed: 13.242444ms
+Jun 18 11:44:14.542: INFO: Pod "downwardapi-volume-64591943-91be-11e9-a25d-8608290c688a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.026527582s
 STEP: Saw pod success
-Jun  4 18:43:26.266: INFO: Pod "pod-configmaps-a0f72453-86f8-11e9-8d1b-467ee19922ac" satisfied condition "success or failure"
-Jun  4 18:43:26.278: INFO: Trying to get logs from node 10.212.23.164 pod pod-configmaps-a0f72453-86f8-11e9-8d1b-467ee19922ac container configmap-volume-test: 
+Jun 18 11:44:14.542: INFO: Pod "downwardapi-volume-64591943-91be-11e9-a25d-8608290c688a" satisfied condition "success or failure"
+Jun 18 11:44:14.555: INFO: Trying to get logs from node 10.72.74.138 pod downwardapi-volume-64591943-91be-11e9-a25d-8608290c688a container client-container: 
 STEP: delete the pod
-Jun  4 18:43:26.341: INFO: Waiting for pod pod-configmaps-a0f72453-86f8-11e9-8d1b-467ee19922ac to disappear
-Jun  4 18:43:26.352: INFO: Pod pod-configmaps-a0f72453-86f8-11e9-8d1b-467ee19922ac no longer exists
-[AfterEach] [sig-storage] ConfigMap
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 18:43:26.352: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "configmap-3445" for this suite.
-Jun  4 18:43:32.405: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 18:43:32.739: INFO: namespace configmap-3445 deletion completed in 6.375020359s
+Jun 18 11:44:14.731: INFO: Waiting for pod downwardapi-volume-64591943-91be-11e9-a25d-8608290c688a to disappear
+Jun 18 11:44:14.748: INFO: Pod downwardapi-volume-64591943-91be-11e9-a25d-8608290c688a no longer exists
+[AfterEach] [sig-storage] Downward API volume
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 11:44:14.748: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "downward-api-7443" for this suite.
+Jun 18 11:44:20.790: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 11:44:21.367: INFO: namespace downward-api-7443 deletion completed in 6.605222213s
 
-• [SLOW TEST:10.762 seconds]
-[sig-storage] ConfigMap
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:32
-  should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+• [SLOW TEST:9.090 seconds]
+[sig-storage] Downward API volume
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:34
+  should set mode on item file [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSS
+------------------------------
+[sig-api-machinery] Namespaces [Serial] 
+  should ensure that all pods are removed when a namespace is deleted [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-api-machinery] Namespaces [Serial]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Jun 18 11:44:21.368: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename namespaces
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in namespaces-1477
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should ensure that all pods are removed when a namespace is deleted [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating a test namespace
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in nsdeletetest-9498
+STEP: Waiting for a default service account to be provisioned in namespace
+STEP: Creating a pod in the namespace
+STEP: Waiting for the pod to have running status
+STEP: Deleting the namespace
+STEP: Waiting for the namespace to be removed.
+STEP: Recreating the namespace
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in nsdeletetest-7121
+STEP: Verifying there are no pods in the namespace
+[AfterEach] [sig-api-machinery] Namespaces [Serial]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 11:44:49.050: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "namespaces-1477" for this suite.
+Jun 18 11:44:55.091: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 11:44:55.493: INFO: namespace namespaces-1477 deletion completed in 6.428584943s
+STEP: Destroying namespace "nsdeletetest-9498" for this suite.
+Jun 18 11:44:55.500: INFO: Namespace nsdeletetest-9498 was already deleted
+STEP: Destroying namespace "nsdeletetest-7121" for this suite.
+Jun 18 11:45:03.526: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 11:45:03.921: INFO: namespace nsdeletetest-7121 deletion completed in 8.420518438s
+
+• [SLOW TEST:42.553 seconds]
+[sig-api-machinery] Namespaces [Serial]
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:22
+  should ensure that all pods are removed when a namespace is deleted [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+------------------------------
+[sig-node] Downward API 
+  should provide host IP as an env var [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-node] Downward API
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Jun 18 11:45:03.922: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename downward-api
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in downward-api-5500
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should provide host IP as an env var [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating a pod to test downward api env vars
+Jun 18 11:45:04.147: INFO: Waiting up to 5m0s for pod "downward-api-831fda49-91be-11e9-a25d-8608290c688a" in namespace "downward-api-5500" to be "success or failure"
+Jun 18 11:45:04.169: INFO: Pod "downward-api-831fda49-91be-11e9-a25d-8608290c688a": Phase="Pending", Reason="", readiness=false. Elapsed: 22.122225ms
+Jun 18 11:45:06.183: INFO: Pod "downward-api-831fda49-91be-11e9-a25d-8608290c688a": Phase="Running", Reason="", readiness=true. Elapsed: 2.035607665s
+Jun 18 11:45:08.197: INFO: Pod "downward-api-831fda49-91be-11e9-a25d-8608290c688a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.049567887s
+STEP: Saw pod success
+Jun 18 11:45:08.197: INFO: Pod "downward-api-831fda49-91be-11e9-a25d-8608290c688a" satisfied condition "success or failure"
+Jun 18 11:45:08.210: INFO: Trying to get logs from node 10.72.74.138 pod downward-api-831fda49-91be-11e9-a25d-8608290c688a container dapi-container: 
+STEP: delete the pod
+Jun 18 11:45:08.287: INFO: Waiting for pod downward-api-831fda49-91be-11e9-a25d-8608290c688a to disappear
+Jun 18 11:45:08.300: INFO: Pod downward-api-831fda49-91be-11e9-a25d-8608290c688a no longer exists
+[AfterEach] [sig-node] Downward API
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 11:45:08.300: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "downward-api-5500" for this suite.
+Jun 18 11:45:14.347: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 11:45:14.742: INFO: namespace downward-api-5500 deletion completed in 6.421173895s
+
+• [SLOW TEST:10.820 seconds]
+[sig-node] Downward API
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downward_api.go:38
+  should provide host IP as an env var [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+[k8s.io] Pods 
+  should be submitted and removed [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [k8s.io] Pods
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Jun 18 11:45:14.742: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename pods
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in pods-3023
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [k8s.io] Pods
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:135
+[It] should be submitted and removed [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: creating the pod
+STEP: setting up watch
+STEP: submitting the pod to kubernetes
+Jun 18 11:45:14.958: INFO: observed the pod list
+STEP: verifying the pod is in kubernetes
+STEP: verifying pod creation was observed
+STEP: deleting the pod gracefully
+STEP: verifying the kubelet observed the termination notice
+Jun 18 11:45:24.104: INFO: no pod exists with the name we were looking for, assuming the termination request was observed and completed
+STEP: verifying pod deletion was observed
+[AfterEach] [k8s.io] Pods
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 11:45:24.118: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "pods-3023" for this suite.
+Jun 18 11:45:30.164: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 11:45:30.623: INFO: namespace pods-3023 deletion completed in 6.492208539s
+
+• [SLOW TEST:15.881 seconds]
+[k8s.io] Pods
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
+  should be submitted and removed [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
 [sig-scheduling] SchedulerPredicates [Serial] 
   validates resource limits of pods that are allowed to run  [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 [BeforeEach] [sig-scheduling] SchedulerPredicates [Serial]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 18:43:32.739: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
+Jun 18 11:45:30.623: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
 STEP: Building a namespace api object, basename sched-pred
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in sched-pred-4369
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in sched-pred-6868
 STEP: Waiting for a default service account to be provisioned in namespace
 [BeforeEach] [sig-scheduling] SchedulerPredicates [Serial]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:79
-Jun  4 18:43:32.955: INFO: Waiting up to 1m0s for all (but 0) nodes to be ready
-Jun  4 18:43:32.983: INFO: Waiting for terminating namespaces to be deleted...
-Jun  4 18:43:33.001: INFO: 
-Logging pods the kubelet thinks is on node 10.212.23.161 before test
-Jun  4 18:43:33.033: INFO: public-cra6d87a4394f440e3becc50f237563475-alb1-79855b4ffb-mqcml from kube-system started at 2019-06-04 17:49:42 +0000 UTC (4 container statuses recorded)
-Jun  4 18:43:33.033: INFO: 	Container ingress-auth-1 ready: true, restart count 0
-Jun  4 18:43:33.033: INFO: 	Container ingress-auth-2 ready: true, restart count 0
-Jun  4 18:43:33.033: INFO: 	Container ingress-auth-3 ready: true, restart count 0
-Jun  4 18:43:33.033: INFO: 	Container nginx-ingress ready: true, restart count 0
-Jun  4 18:43:33.033: INFO: sonobuoy-systemd-logs-daemon-set-d63a0136eef44275-l8n68 from heptio-sonobuoy started at 2019-06-04 18:27:33 +0000 UTC (2 container statuses recorded)
-Jun  4 18:43:33.033: INFO: 	Container sonobuoy-worker ready: true, restart count 0
-Jun  4 18:43:33.033: INFO: 	Container systemd-logs ready: true, restart count 0
-Jun  4 18:43:33.033: INFO: vpn-7bbfb5bc49-4mjxm from kube-system started at 2019-06-04 17:44:57 +0000 UTC (1 container statuses recorded)
-Jun  4 18:43:33.033: INFO: 	Container vpn ready: true, restart count 0
-Jun  4 18:43:33.033: INFO: calico-kube-controllers-64ccf5d5d4-k2qnd from kube-system started at 2019-06-04 17:44:57 +0000 UTC (1 container statuses recorded)
-Jun  4 18:43:33.033: INFO: 	Container calico-kube-controllers ready: true, restart count 0
-Jun  4 18:43:33.033: INFO: kubernetes-dashboard-b4f5d97fc-rrmfx from kube-system started at 2019-06-04 17:44:57 +0000 UTC (1 container statuses recorded)
-Jun  4 18:43:33.033: INFO: 	Container kubernetes-dashboard ready: true, restart count 0
-Jun  4 18:43:33.033: INFO: ibm-storage-watcher-f74547b86-jhnt2 from kube-system started at 2019-06-04 17:44:57 +0000 UTC (1 container statuses recorded)
-Jun  4 18:43:33.033: INFO: 	Container ibm-storage-watcher-container ready: true, restart count 0
-Jun  4 18:43:33.033: INFO: ibm-kube-fluentd-cj2l2 from kube-system started at 2019-06-04 17:45:04 +0000 UTC (1 container statuses recorded)
-Jun  4 18:43:33.033: INFO: 	Container fluentd ready: true, restart count 0
-Jun  4 18:43:33.033: INFO: calico-node-wtt96 from kube-system started at 2019-06-04 17:44:38 +0000 UTC (1 container statuses recorded)
-Jun  4 18:43:33.033: INFO: 	Container calico-node ready: true, restart count 0
-Jun  4 18:43:33.033: INFO: coredns-autoscaler-5846b97547-t7mrd from kube-system started at 2019-06-04 17:44:57 +0000 UTC (1 container statuses recorded)
-Jun  4 18:43:33.033: INFO: 	Container autoscaler ready: true, restart count 0
-Jun  4 18:43:33.033: INFO: ibm-master-proxy-static-10.212.23.161 from kube-system started at  (0 container statuses recorded)
-Jun  4 18:43:33.033: INFO: coredns-cc5d484bb-9nhk2 from kube-system started at 2019-06-04 17:44:57 +0000 UTC (1 container statuses recorded)
-Jun  4 18:43:33.033: INFO: 	Container coredns ready: true, restart count 0
-Jun  4 18:43:33.033: INFO: ibm-keepalived-watcher-vl7mx from kube-system started at 2019-06-04 17:44:38 +0000 UTC (1 container statuses recorded)
-Jun  4 18:43:33.033: INFO: 	Container keepalived-watcher ready: true, restart count 0
-Jun  4 18:43:33.033: INFO: ibm-file-plugin-69d4bc9c8f-jwhqt from kube-system started at 2019-06-04 17:44:57 +0000 UTC (1 container statuses recorded)
-Jun  4 18:43:33.033: INFO: 	Container ibm-file-plugin-container ready: true, restart count 0
-Jun  4 18:43:33.033: INFO: 
-Logging pods the kubelet thinks is on node 10.212.23.164 before test
-Jun  4 18:43:33.058: INFO: metrics-server-66cfcd4b8d-xg6wk from kube-system started at 2019-06-04 17:45:26 +0000 UTC (2 container statuses recorded)
-Jun  4 18:43:33.058: INFO: 	Container metrics-server ready: true, restart count 0
-Jun  4 18:43:33.058: INFO: 	Container metrics-server-nanny ready: true, restart count 0
-Jun  4 18:43:33.058: INFO: ibm-master-proxy-static-10.212.23.164 from kube-system started at  (0 container statuses recorded)
-Jun  4 18:43:33.058: INFO: ibm-keepalived-watcher-lfldx from kube-system started at 2019-06-04 17:44:49 +0000 UTC (1 container statuses recorded)
-Jun  4 18:43:33.058: INFO: 	Container keepalived-watcher ready: true, restart count 0
-Jun  4 18:43:33.058: INFO: calico-node-sphpm from kube-system started at 2019-06-04 17:44:49 +0000 UTC (1 container statuses recorded)
-Jun  4 18:43:33.058: INFO: 	Container calico-node ready: true, restart count 0
-Jun  4 18:43:33.058: INFO: sonobuoy-systemd-logs-daemon-set-d63a0136eef44275-b6j5c from heptio-sonobuoy started at 2019-06-04 18:27:33 +0000 UTC (2 container statuses recorded)
-Jun  4 18:43:33.058: INFO: 	Container sonobuoy-worker ready: true, restart count 0
-Jun  4 18:43:33.058: INFO: 	Container systemd-logs ready: true, restart count 0
-Jun  4 18:43:33.058: INFO: ibm-cloud-provider-ip-162-133-73-165-ffb96fbff-7g5nv from ibm-system started at 2019-06-04 17:47:12 +0000 UTC (1 container statuses recorded)
-Jun  4 18:43:33.058: INFO: 	Container ibm-cloud-provider-ip-162-133-73-165 ready: true, restart count 0
-Jun  4 18:43:33.058: INFO: test-k8s-e2e-pvg-master-verification from default started at 2019-06-04 18:27:15 +0000 UTC (1 container statuses recorded)
-Jun  4 18:43:33.058: INFO: 	Container test-k8s-e2e-pvg-master-verification ready: true, restart count 0
-Jun  4 18:43:33.058: INFO: sonobuoy-e2e-job-39546bc4ddc3410d from heptio-sonobuoy started at 2019-06-04 18:27:33 +0000 UTC (2 container statuses recorded)
-Jun  4 18:43:33.058: INFO: 	Container e2e ready: true, restart count 0
-Jun  4 18:43:33.058: INFO: 	Container sonobuoy-worker ready: true, restart count 0
-Jun  4 18:43:33.058: INFO: ibm-kube-fluentd-nr9cm from kube-system started at 2019-06-04 17:45:04 +0000 UTC (1 container statuses recorded)
-Jun  4 18:43:33.058: INFO: 	Container fluentd ready: true, restart count 0
-Jun  4 18:43:33.058: INFO: sonobuoy from heptio-sonobuoy started at 2019-06-04 18:27:24 +0000 UTC (1 container statuses recorded)
-Jun  4 18:43:33.058: INFO: 	Container kube-sonobuoy ready: true, restart count 0
-Jun  4 18:43:33.058: INFO: 
-Logging pods the kubelet thinks is on node 10.212.23.189 before test
-Jun  4 18:43:33.093: INFO: calico-node-v6hns from kube-system started at 2019-06-04 17:45:04 +0000 UTC (1 container statuses recorded)
-Jun  4 18:43:33.093: INFO: 	Container calico-node ready: true, restart count 0
-Jun  4 18:43:33.093: INFO: ibm-keepalived-watcher-dqk2k from kube-system started at 2019-06-04 17:45:04 +0000 UTC (1 container statuses recorded)
-Jun  4 18:43:33.093: INFO: 	Container keepalived-watcher ready: true, restart count 0
-Jun  4 18:43:33.093: INFO: coredns-cc5d484bb-nlgcf from kube-system started at 2019-06-04 17:45:30 +0000 UTC (1 container statuses recorded)
-Jun  4 18:43:33.093: INFO: 	Container coredns ready: true, restart count 0
-Jun  4 18:43:33.093: INFO: ibm-cloud-provider-ip-162-133-73-165-ffb96fbff-jf4l4 from ibm-system started at 2019-06-04 17:47:12 +0000 UTC (1 container statuses recorded)
-Jun  4 18:43:33.093: INFO: 	Container ibm-cloud-provider-ip-162-133-73-165 ready: true, restart count 0
-Jun  4 18:43:33.093: INFO: public-cra6d87a4394f440e3becc50f237563475-alb1-79855b4ffb-89n6x from kube-system started at 2019-06-04 17:49:42 +0000 UTC (4 container statuses recorded)
-Jun  4 18:43:33.093: INFO: 	Container ingress-auth-1 ready: true, restart count 0
-Jun  4 18:43:33.093: INFO: 	Container ingress-auth-2 ready: true, restart count 0
-Jun  4 18:43:33.093: INFO: 	Container ingress-auth-3 ready: true, restart count 0
-Jun  4 18:43:33.093: INFO: 	Container nginx-ingress ready: true, restart count 0
-Jun  4 18:43:33.093: INFO: sonobuoy-systemd-logs-daemon-set-d63a0136eef44275-hnnhz from heptio-sonobuoy started at 2019-06-04 18:27:33 +0000 UTC (2 container statuses recorded)
-Jun  4 18:43:33.093: INFO: 	Container sonobuoy-worker ready: true, restart count 0
-Jun  4 18:43:33.093: INFO: 	Container systemd-logs ready: true, restart count 0
-Jun  4 18:43:33.093: INFO: ibm-master-proxy-static-10.212.23.189 from kube-system started at  (0 container statuses recorded)
-Jun  4 18:43:33.093: INFO: ibm-kube-fluentd-k56t7 from kube-system started at 2019-06-04 17:45:04 +0000 UTC (1 container statuses recorded)
-Jun  4 18:43:33.093: INFO: 	Container fluentd ready: true, restart count 0
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:79
+Jun 18 11:45:30.815: INFO: Waiting up to 1m0s for all (but 0) nodes to be ready
+Jun 18 11:45:30.841: INFO: Waiting for terminating namespaces to be deleted...
+Jun 18 11:45:30.849: INFO: 
+Logging pods the kubelet thinks is on node 10.72.74.138 before test
+Jun 18 11:45:30.888: INFO: ibm-cloud-provider-ip-158-176-120-140-d8f5f45f5-4tksx from ibm-system started at 2019-06-17 21:41:02 +0000 UTC (1 container statuses recorded)
+Jun 18 11:45:30.888: INFO: 	Container ibm-cloud-provider-ip-158-176-120-140 ready: true, restart count 0
+Jun 18 11:45:30.888: INFO: public-crd9ae0c6aaa554167a4f1d2c76b9ad9f3-alb1-fd9fb76d-q646z from kube-system started at 2019-06-17 21:41:07 +0000 UTC (4 container statuses recorded)
+Jun 18 11:45:30.888: INFO: 	Container ingress-auth-1 ready: true, restart count 0
+Jun 18 11:45:30.888: INFO: 	Container ingress-auth-2 ready: true, restart count 0
+Jun 18 11:45:30.888: INFO: 	Container ingress-auth-3 ready: true, restart count 0
+Jun 18 11:45:30.888: INFO: 	Container nginx-ingress ready: true, restart count 0
+Jun 18 11:45:30.888: INFO: ibm-master-proxy-static-10.72.74.138 from kube-system started at  (0 container statuses recorded)
+Jun 18 11:45:30.888: INFO: calico-node-prxhz from kube-system started at 2019-06-17 21:37:25 +0000 UTC (1 container statuses recorded)
+Jun 18 11:45:30.888: INFO: 	Container calico-node ready: true, restart count 0
+Jun 18 11:45:30.888: INFO: ibm-keepalived-watcher-4l788 from kube-system started at 2019-06-17 21:37:25 +0000 UTC (1 container statuses recorded)
+Jun 18 11:45:30.888: INFO: 	Container keepalived-watcher ready: true, restart count 0
+Jun 18 11:45:30.888: INFO: ibm-kube-fluentd-gmp54 from kube-system started at 2019-06-17 21:37:34 +0000 UTC (1 container statuses recorded)
+Jun 18 11:45:30.888: INFO: 	Container fluentd ready: true, restart count 0
+Jun 18 11:45:30.888: INFO: sonobuoy-systemd-logs-daemon-set-376e58a3dd534c11-ccq8b from heptio-sonobuoy started at 2019-06-18 11:33:56 +0000 UTC (2 container statuses recorded)
+Jun 18 11:45:30.888: INFO: 	Container sonobuoy-worker ready: true, restart count 0
+Jun 18 11:45:30.888: INFO: 	Container systemd-logs ready: true, restart count 0
+Jun 18 11:45:30.888: INFO: 
+Logging pods the kubelet thinks is on node 10.72.74.184 before test
+Jun 18 11:45:30.924: INFO: test-k8s-e2e-pvg-master-verification from default started at 2019-06-18 11:33:41 +0000 UTC (1 container statuses recorded)
+Jun 18 11:45:30.924: INFO: 	Container test-k8s-e2e-pvg-master-verification ready: true, restart count 0
+Jun 18 11:45:30.924: INFO: sonobuoy-systemd-logs-daemon-set-376e58a3dd534c11-x9mmr from heptio-sonobuoy started at 2019-06-18 11:33:56 +0000 UTC (2 container statuses recorded)
+Jun 18 11:45:30.924: INFO: 	Container sonobuoy-worker ready: true, restart count 0
+Jun 18 11:45:30.924: INFO: 	Container systemd-logs ready: true, restart count 0
+Jun 18 11:45:30.924: INFO: ibm-master-proxy-static-10.72.74.184 from kube-system started at  (0 container statuses recorded)
+Jun 18 11:45:30.924: INFO: ibm-kube-fluentd-66f2t from kube-system started at 2019-06-17 21:37:34 +0000 UTC (1 container statuses recorded)
+Jun 18 11:45:30.924: INFO: 	Container fluentd ready: true, restart count 0
+Jun 18 11:45:30.924: INFO: ibm-cloud-provider-ip-158-176-120-140-d8f5f45f5-kszmp from ibm-system started at 2019-06-17 21:41:02 +0000 UTC (1 container statuses recorded)
+Jun 18 11:45:30.924: INFO: 	Container ibm-cloud-provider-ip-158-176-120-140 ready: true, restart count 0
+Jun 18 11:45:30.924: INFO: sonobuoy from heptio-sonobuoy started at 2019-06-18 11:33:48 +0000 UTC (1 container statuses recorded)
+Jun 18 11:45:30.924: INFO: 	Container kube-sonobuoy ready: true, restart count 0
+Jun 18 11:45:30.924: INFO: calico-node-2jll6 from kube-system started at 2019-06-17 21:37:03 +0000 UTC (1 container statuses recorded)
+Jun 18 11:45:30.924: INFO: 	Container calico-node ready: true, restart count 0
+Jun 18 11:45:30.924: INFO: ibm-keepalived-watcher-xkcvg from kube-system started at 2019-06-17 21:37:03 +0000 UTC (1 container statuses recorded)
+Jun 18 11:45:30.924: INFO: 	Container keepalived-watcher ready: true, restart count 0
+Jun 18 11:45:30.924: INFO: coredns-78cff85d65-jvwmh from kube-system started at 2019-06-17 21:37:29 +0000 UTC (1 container statuses recorded)
+Jun 18 11:45:30.924: INFO: 	Container coredns ready: true, restart count 0
+Jun 18 11:45:30.924: INFO: metrics-server-5cdb4c5b-67974 from kube-system started at 2019-06-17 21:37:27 +0000 UTC (2 container statuses recorded)
+Jun 18 11:45:30.924: INFO: 	Container metrics-server ready: true, restart count 0
+Jun 18 11:45:30.924: INFO: 	Container metrics-server-nanny ready: true, restart count 0
+Jun 18 11:45:30.924: INFO: 
+Logging pods the kubelet thinks is on node 10.72.74.189 before test
+Jun 18 11:45:31.008: INFO: ibm-keepalived-watcher-klm4c from kube-system started at 2019-06-17 21:36:59 +0000 UTC (1 container statuses recorded)
+Jun 18 11:45:31.008: INFO: 	Container keepalived-watcher ready: true, restart count 0
+Jun 18 11:45:31.008: INFO: vpn-cf6ff59b-gqxpw from kube-system started at 2019-06-17 21:37:09 +0000 UTC (1 container statuses recorded)
+Jun 18 11:45:31.008: INFO: 	Container vpn ready: true, restart count 0
+Jun 18 11:45:31.008: INFO: kubernetes-dashboard-6f5f8c6896-rbvmt from kube-system started at 2019-06-17 21:37:09 +0000 UTC (1 container statuses recorded)
+Jun 18 11:45:31.008: INFO: 	Container kubernetes-dashboard ready: true, restart count 0
+Jun 18 11:45:31.008: INFO: sonobuoy-e2e-job-68f893029cbd431f from heptio-sonobuoy started at 2019-06-18 11:33:56 +0000 UTC (2 container statuses recorded)
+Jun 18 11:45:31.008: INFO: 	Container e2e ready: true, restart count 0
+Jun 18 11:45:31.008: INFO: 	Container sonobuoy-worker ready: true, restart count 0
+Jun 18 11:45:31.008: INFO: ibm-master-proxy-static-10.72.74.189 from kube-system started at  (0 container statuses recorded)
+Jun 18 11:45:31.008: INFO: ibm-storage-watcher-964758dd-rsgxw from kube-system started at 2019-06-17 21:37:09 +0000 UTC (1 container statuses recorded)
+Jun 18 11:45:31.008: INFO: 	Container ibm-storage-watcher-container ready: true, restart count 0
+Jun 18 11:45:31.008: INFO: coredns-78cff85d65-2mm72 from kube-system started at 2019-06-17 21:37:09 +0000 UTC (1 container statuses recorded)
+Jun 18 11:45:31.008: INFO: 	Container coredns ready: true, restart count 0
+Jun 18 11:45:31.008: INFO: calico-kube-controllers-85d6c48f7-ggvwd from kube-system started at 2019-06-17 21:37:09 +0000 UTC (1 container statuses recorded)
+Jun 18 11:45:31.008: INFO: 	Container calico-kube-controllers ready: true, restart count 0
+Jun 18 11:45:31.008: INFO: public-crd9ae0c6aaa554167a4f1d2c76b9ad9f3-alb1-fd9fb76d-sf25n from kube-system started at 2019-06-17 21:41:07 +0000 UTC (4 container statuses recorded)
+Jun 18 11:45:31.008: INFO: 	Container ingress-auth-1 ready: true, restart count 0
+Jun 18 11:45:31.008: INFO: 	Container ingress-auth-2 ready: true, restart count 0
+Jun 18 11:45:31.008: INFO: 	Container ingress-auth-3 ready: true, restart count 0
+Jun 18 11:45:31.008: INFO: 	Container nginx-ingress ready: true, restart count 0
+Jun 18 11:45:31.008: INFO: coredns-autoscaler-6854575d6-p69qn from kube-system started at 2019-06-17 21:37:09 +0000 UTC (1 container statuses recorded)
+Jun 18 11:45:31.008: INFO: 	Container autoscaler ready: true, restart count 0
+Jun 18 11:45:31.008: INFO: ibm-file-plugin-98f5986d7-bbzfx from kube-system started at 2019-06-17 21:37:09 +0000 UTC (1 container statuses recorded)
+Jun 18 11:45:31.008: INFO: 	Container ibm-file-plugin-container ready: true, restart count 0
+Jun 18 11:45:31.008: INFO: ibm-kube-fluentd-qg9mz from kube-system started at 2019-06-17 21:37:34 +0000 UTC (1 container statuses recorded)
+Jun 18 11:45:31.008: INFO: 	Container fluentd ready: true, restart count 0
+Jun 18 11:45:31.008: INFO: sonobuoy-systemd-logs-daemon-set-376e58a3dd534c11-rn8dw from heptio-sonobuoy started at 2019-06-18 11:33:56 +0000 UTC (2 container statuses recorded)
+Jun 18 11:45:31.008: INFO: 	Container sonobuoy-worker ready: true, restart count 0
+Jun 18 11:45:31.008: INFO: 	Container systemd-logs ready: true, restart count 0
+Jun 18 11:45:31.008: INFO: calico-node-htlqc from kube-system started at 2019-06-17 21:36:59 +0000 UTC (1 container statuses recorded)
+Jun 18 11:45:31.008: INFO: 	Container calico-node ready: true, restart count 0
 [It] validates resource limits of pods that are allowed to run  [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: verifying the node has the label node 10.212.23.161
-STEP: verifying the node has the label node 10.212.23.164
-STEP: verifying the node has the label node 10.212.23.189
-Jun  4 18:43:33.228: INFO: Pod test-k8s-e2e-pvg-master-verification requesting resource cpu=0m on Node 10.212.23.164
-Jun  4 18:43:33.229: INFO: Pod sonobuoy requesting resource cpu=0m on Node 10.212.23.164
-Jun  4 18:43:33.229: INFO: Pod sonobuoy-e2e-job-39546bc4ddc3410d requesting resource cpu=0m on Node 10.212.23.164
-Jun  4 18:43:33.229: INFO: Pod sonobuoy-systemd-logs-daemon-set-d63a0136eef44275-b6j5c requesting resource cpu=0m on Node 10.212.23.164
-Jun  4 18:43:33.229: INFO: Pod sonobuoy-systemd-logs-daemon-set-d63a0136eef44275-hnnhz requesting resource cpu=0m on Node 10.212.23.189
-Jun  4 18:43:33.229: INFO: Pod sonobuoy-systemd-logs-daemon-set-d63a0136eef44275-l8n68 requesting resource cpu=0m on Node 10.212.23.161
-Jun  4 18:43:33.229: INFO: Pod ibm-cloud-provider-ip-162-133-73-165-ffb96fbff-7g5nv requesting resource cpu=5m on Node 10.212.23.164
-Jun  4 18:43:33.229: INFO: Pod ibm-cloud-provider-ip-162-133-73-165-ffb96fbff-jf4l4 requesting resource cpu=5m on Node 10.212.23.189
-Jun  4 18:43:33.229: INFO: Pod calico-kube-controllers-64ccf5d5d4-k2qnd requesting resource cpu=10m on Node 10.212.23.161
-Jun  4 18:43:33.229: INFO: Pod calico-node-sphpm requesting resource cpu=250m on Node 10.212.23.164
-Jun  4 18:43:33.229: INFO: Pod calico-node-v6hns requesting resource cpu=250m on Node 10.212.23.189
-Jun  4 18:43:33.229: INFO: Pod calico-node-wtt96 requesting resource cpu=250m on Node 10.212.23.161
-Jun  4 18:43:33.229: INFO: Pod coredns-autoscaler-5846b97547-t7mrd requesting resource cpu=20m on Node 10.212.23.161
-Jun  4 18:43:33.229: INFO: Pod coredns-cc5d484bb-9nhk2 requesting resource cpu=100m on Node 10.212.23.161
-Jun  4 18:43:33.229: INFO: Pod coredns-cc5d484bb-nlgcf requesting resource cpu=100m on Node 10.212.23.189
-Jun  4 18:43:33.229: INFO: Pod ibm-file-plugin-69d4bc9c8f-jwhqt requesting resource cpu=50m on Node 10.212.23.161
-Jun  4 18:43:33.229: INFO: Pod ibm-keepalived-watcher-dqk2k requesting resource cpu=5m on Node 10.212.23.189
-Jun  4 18:43:33.229: INFO: Pod ibm-keepalived-watcher-lfldx requesting resource cpu=5m on Node 10.212.23.164
-Jun  4 18:43:33.229: INFO: Pod ibm-keepalived-watcher-vl7mx requesting resource cpu=5m on Node 10.212.23.161
-Jun  4 18:43:33.229: INFO: Pod ibm-kube-fluentd-cj2l2 requesting resource cpu=25m on Node 10.212.23.161
-Jun  4 18:43:33.229: INFO: Pod ibm-kube-fluentd-k56t7 requesting resource cpu=25m on Node 10.212.23.189
-Jun  4 18:43:33.229: INFO: Pod ibm-kube-fluentd-nr9cm requesting resource cpu=25m on Node 10.212.23.164
-Jun  4 18:43:33.229: INFO: Pod ibm-master-proxy-static-10.212.23.161 requesting resource cpu=25m on Node 10.212.23.161
-Jun  4 18:43:33.229: INFO: Pod ibm-master-proxy-static-10.212.23.164 requesting resource cpu=25m on Node 10.212.23.164
-Jun  4 18:43:33.229: INFO: Pod ibm-master-proxy-static-10.212.23.189 requesting resource cpu=25m on Node 10.212.23.189
-Jun  4 18:43:33.229: INFO: Pod ibm-storage-watcher-f74547b86-jhnt2 requesting resource cpu=50m on Node 10.212.23.161
-Jun  4 18:43:33.229: INFO: Pod kubernetes-dashboard-b4f5d97fc-rrmfx requesting resource cpu=50m on Node 10.212.23.161
-Jun  4 18:43:33.229: INFO: Pod metrics-server-66cfcd4b8d-xg6wk requesting resource cpu=53m on Node 10.212.23.164
-Jun  4 18:43:33.229: INFO: Pod public-cra6d87a4394f440e3becc50f237563475-alb1-79855b4ffb-89n6x requesting resource cpu=0m on Node 10.212.23.189
-Jun  4 18:43:33.229: INFO: Pod public-cra6d87a4394f440e3becc50f237563475-alb1-79855b4ffb-mqcml requesting resource cpu=0m on Node 10.212.23.161
-Jun  4 18:43:33.229: INFO: Pod vpn-7bbfb5bc49-4mjxm requesting resource cpu=5m on Node 10.212.23.161
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: verifying the node has the label node 10.72.74.138
+STEP: verifying the node has the label node 10.72.74.184
+STEP: verifying the node has the label node 10.72.74.189
+Jun 18 11:45:31.152: INFO: Pod test-k8s-e2e-pvg-master-verification requesting resource cpu=0m on Node 10.72.74.184
+Jun 18 11:45:31.152: INFO: Pod sonobuoy requesting resource cpu=0m on Node 10.72.74.184
+Jun 18 11:45:31.152: INFO: Pod sonobuoy-e2e-job-68f893029cbd431f requesting resource cpu=0m on Node 10.72.74.189
+Jun 18 11:45:31.152: INFO: Pod sonobuoy-systemd-logs-daemon-set-376e58a3dd534c11-ccq8b requesting resource cpu=0m on Node 10.72.74.138
+Jun 18 11:45:31.152: INFO: Pod sonobuoy-systemd-logs-daemon-set-376e58a3dd534c11-rn8dw requesting resource cpu=0m on Node 10.72.74.189
+Jun 18 11:45:31.152: INFO: Pod sonobuoy-systemd-logs-daemon-set-376e58a3dd534c11-x9mmr requesting resource cpu=0m on Node 10.72.74.184
+Jun 18 11:45:31.152: INFO: Pod ibm-cloud-provider-ip-158-176-120-140-d8f5f45f5-4tksx requesting resource cpu=5m on Node 10.72.74.138
+Jun 18 11:45:31.152: INFO: Pod ibm-cloud-provider-ip-158-176-120-140-d8f5f45f5-kszmp requesting resource cpu=5m on Node 10.72.74.184
+Jun 18 11:45:31.152: INFO: Pod calico-kube-controllers-85d6c48f7-ggvwd requesting resource cpu=10m on Node 10.72.74.189
+Jun 18 11:45:31.152: INFO: Pod calico-node-2jll6 requesting resource cpu=250m on Node 10.72.74.184
+Jun 18 11:45:31.152: INFO: Pod calico-node-htlqc requesting resource cpu=250m on Node 10.72.74.189
+Jun 18 11:45:31.152: INFO: Pod calico-node-prxhz requesting resource cpu=250m on Node 10.72.74.138
+Jun 18 11:45:31.152: INFO: Pod coredns-78cff85d65-2mm72 requesting resource cpu=100m on Node 10.72.74.189
+Jun 18 11:45:31.152: INFO: Pod coredns-78cff85d65-jvwmh requesting resource cpu=100m on Node 10.72.74.184
+Jun 18 11:45:31.152: INFO: Pod coredns-autoscaler-6854575d6-p69qn requesting resource cpu=20m on Node 10.72.74.189
+Jun 18 11:45:31.152: INFO: Pod ibm-file-plugin-98f5986d7-bbzfx requesting resource cpu=50m on Node 10.72.74.189
+Jun 18 11:45:31.152: INFO: Pod ibm-keepalived-watcher-4l788 requesting resource cpu=5m on Node 10.72.74.138
+Jun 18 11:45:31.152: INFO: Pod ibm-keepalived-watcher-klm4c requesting resource cpu=5m on Node 10.72.74.189
+Jun 18 11:45:31.152: INFO: Pod ibm-keepalived-watcher-xkcvg requesting resource cpu=5m on Node 10.72.74.184
+Jun 18 11:45:31.152: INFO: Pod ibm-kube-fluentd-66f2t requesting resource cpu=25m on Node 10.72.74.184
+Jun 18 11:45:31.152: INFO: Pod ibm-kube-fluentd-gmp54 requesting resource cpu=25m on Node 10.72.74.138
+Jun 18 11:45:31.152: INFO: Pod ibm-kube-fluentd-qg9mz requesting resource cpu=25m on Node 10.72.74.189
+Jun 18 11:45:31.152: INFO: Pod ibm-master-proxy-static-10.72.74.138 requesting resource cpu=25m on Node 10.72.74.138
+Jun 18 11:45:31.152: INFO: Pod ibm-master-proxy-static-10.72.74.184 requesting resource cpu=25m on Node 10.72.74.184
+Jun 18 11:45:31.152: INFO: Pod ibm-master-proxy-static-10.72.74.189 requesting resource cpu=25m on Node 10.72.74.189
+Jun 18 11:45:31.152: INFO: Pod ibm-storage-watcher-964758dd-rsgxw requesting resource cpu=50m on Node 10.72.74.189
+Jun 18 11:45:31.152: INFO: Pod kubernetes-dashboard-6f5f8c6896-rbvmt requesting resource cpu=50m on Node 10.72.74.189
+Jun 18 11:45:31.152: INFO: Pod metrics-server-5cdb4c5b-67974 requesting resource cpu=53m on Node 10.72.74.184
+Jun 18 11:45:31.152: INFO: Pod public-crd9ae0c6aaa554167a4f1d2c76b9ad9f3-alb1-fd9fb76d-q646z requesting resource cpu=0m on Node 10.72.74.138
+Jun 18 11:45:31.152: INFO: Pod public-crd9ae0c6aaa554167a4f1d2c76b9ad9f3-alb1-fd9fb76d-sf25n requesting resource cpu=0m on Node 10.72.74.189
+Jun 18 11:45:31.152: INFO: Pod vpn-cf6ff59b-gqxpw requesting resource cpu=5m on Node 10.72.74.189
 STEP: Starting Pods to consume most of the cluster CPU.
 STEP: Creating another pod that requires unavailable amount of CPU.
 STEP: Considering event: 
-Type = [Normal], Name = [filler-pod-a788f8da-86f8-11e9-8d1b-467ee19922ac.15a51311d48fc264], Reason = [Scheduled], Message = [Successfully assigned sched-pred-4369/filler-pod-a788f8da-86f8-11e9-8d1b-467ee19922ac to 10.212.23.161]
+Type = [Normal], Name = [filler-pod-933c46d3-91be-11e9-a25d-8608290c688a.15a94861e79dc566], Reason = [Scheduled], Message = [Successfully assigned sched-pred-6868/filler-pod-933c46d3-91be-11e9-a25d-8608290c688a to 10.72.74.138]
 STEP: Considering event: 
-Type = [Normal], Name = [filler-pod-a788f8da-86f8-11e9-8d1b-467ee19922ac.15a513121083e07a], Reason = [Pulled], Message = [Container image "k8s.gcr.io/pause:3.1" already present on machine]
+Type = [Normal], Name = [filler-pod-933c46d3-91be-11e9-a25d-8608290c688a.15a9486224001d2a], Reason = [Pulled], Message = [Container image "k8s.gcr.io/pause:3.1" already present on machine]
 STEP: Considering event: 
-Type = [Normal], Name = [filler-pod-a788f8da-86f8-11e9-8d1b-467ee19922ac.15a51312138a42c5], Reason = [Created], Message = [Created container filler-pod-a788f8da-86f8-11e9-8d1b-467ee19922ac]
+Type = [Normal], Name = [filler-pod-933c46d3-91be-11e9-a25d-8608290c688a.15a948622822234e], Reason = [Created], Message = [Created container filler-pod-933c46d3-91be-11e9-a25d-8608290c688a]
 STEP: Considering event: 
-Type = [Normal], Name = [filler-pod-a788f8da-86f8-11e9-8d1b-467ee19922ac.15a513121c17cb52], Reason = [Started], Message = [Started container filler-pod-a788f8da-86f8-11e9-8d1b-467ee19922ac]
+Type = [Normal], Name = [filler-pod-933c46d3-91be-11e9-a25d-8608290c688a.15a9486232f9410f], Reason = [Started], Message = [Started container filler-pod-933c46d3-91be-11e9-a25d-8608290c688a]
 STEP: Considering event: 
-Type = [Normal], Name = [filler-pod-a78c027a-86f8-11e9-8d1b-467ee19922ac.15a51311d56120f6], Reason = [Scheduled], Message = [Successfully assigned sched-pred-4369/filler-pod-a78c027a-86f8-11e9-8d1b-467ee19922ac to 10.212.23.164]
+Type = [Normal], Name = [filler-pod-9340be73-91be-11e9-a25d-8608290c688a.15a94861e8ceb9ee], Reason = [Scheduled], Message = [Successfully assigned sched-pred-6868/filler-pod-9340be73-91be-11e9-a25d-8608290c688a to 10.72.74.184]
 STEP: Considering event: 
-Type = [Normal], Name = [filler-pod-a78c027a-86f8-11e9-8d1b-467ee19922ac.15a5131210d2a3c8], Reason = [Pulled], Message = [Container image "k8s.gcr.io/pause:3.1" already present on machine]
+Type = [Normal], Name = [filler-pod-9340be73-91be-11e9-a25d-8608290c688a.15a948622a94f641], Reason = [Pulled], Message = [Container image "k8s.gcr.io/pause:3.1" already present on machine]
 STEP: Considering event: 
-Type = [Normal], Name = [filler-pod-a78c027a-86f8-11e9-8d1b-467ee19922ac.15a5131213b932a3], Reason = [Created], Message = [Created container filler-pod-a78c027a-86f8-11e9-8d1b-467ee19922ac]
+Type = [Normal], Name = [filler-pod-9340be73-91be-11e9-a25d-8608290c688a.15a948622dc7057f], Reason = [Created], Message = [Created container filler-pod-9340be73-91be-11e9-a25d-8608290c688a]
 STEP: Considering event: 
-Type = [Normal], Name = [filler-pod-a78c027a-86f8-11e9-8d1b-467ee19922ac.15a513121cb2c827], Reason = [Started], Message = [Started container filler-pod-a78c027a-86f8-11e9-8d1b-467ee19922ac]
+Type = [Normal], Name = [filler-pod-9340be73-91be-11e9-a25d-8608290c688a.15a948623831a0ec], Reason = [Started], Message = [Started container filler-pod-9340be73-91be-11e9-a25d-8608290c688a]
 STEP: Considering event: 
-Type = [Normal], Name = [filler-pod-a78e4e7c-86f8-11e9-8d1b-467ee19922ac.15a51311d64d958e], Reason = [Scheduled], Message = [Successfully assigned sched-pred-4369/filler-pod-a78e4e7c-86f8-11e9-8d1b-467ee19922ac to 10.212.23.189]
+Type = [Normal], Name = [filler-pod-9343b223-91be-11e9-a25d-8608290c688a.15a94861e9a80236], Reason = [Scheduled], Message = [Successfully assigned sched-pred-6868/filler-pod-9343b223-91be-11e9-a25d-8608290c688a to 10.72.74.189]
 STEP: Considering event: 
-Type = [Normal], Name = [filler-pod-a78e4e7c-86f8-11e9-8d1b-467ee19922ac.15a51312145d4fad], Reason = [Pulled], Message = [Container image "k8s.gcr.io/pause:3.1" already present on machine]
+Type = [Normal], Name = [filler-pod-9343b223-91be-11e9-a25d-8608290c688a.15a9486223aa45f7], Reason = [Pulling], Message = [Pulling image "k8s.gcr.io/pause:3.1"]
 STEP: Considering event: 
-Type = [Normal], Name = [filler-pod-a78e4e7c-86f8-11e9-8d1b-467ee19922ac.15a513121759e126], Reason = [Created], Message = [Created container filler-pod-a78e4e7c-86f8-11e9-8d1b-467ee19922ac]
+Type = [Normal], Name = [filler-pod-9343b223-91be-11e9-a25d-8608290c688a.15a9486258d322cb], Reason = [Pulled], Message = [Successfully pulled image "k8s.gcr.io/pause:3.1"]
 STEP: Considering event: 
-Type = [Normal], Name = [filler-pod-a78e4e7c-86f8-11e9-8d1b-467ee19922ac.15a5131220a9bc7b], Reason = [Started], Message = [Started container filler-pod-a78e4e7c-86f8-11e9-8d1b-467ee19922ac]
+Type = [Normal], Name = [filler-pod-9343b223-91be-11e9-a25d-8608290c688a.15a948625cf20a7b], Reason = [Created], Message = [Created container filler-pod-9343b223-91be-11e9-a25d-8608290c688a]
 STEP: Considering event: 
-Type = [Warning], Name = [additional-pod.15a51312ca140ee6], Reason = [FailedScheduling], Message = [0/3 nodes are available: 3 Insufficient cpu.]
-STEP: removing the label node off the node 10.212.23.164
+Type = [Normal], Name = [filler-pod-9343b223-91be-11e9-a25d-8608290c688a.15a9486266ea4280], Reason = [Started], Message = [Started container filler-pod-9343b223-91be-11e9-a25d-8608290c688a]
+STEP: Considering event: 
+Type = [Warning], Name = [additional-pod.15a94862dd2b9ed3], Reason = [FailedScheduling], Message = [0/3 nodes are available: 3 Insufficient cpu.]
+STEP: removing the label node off the node 10.72.74.138
 STEP: verifying the node doesn't have the label node
-STEP: removing the label node off the node 10.212.23.189
+STEP: removing the label node off the node 10.72.74.184
 STEP: verifying the node doesn't have the label node
-STEP: removing the label node off the node 10.212.23.161
+STEP: removing the label node off the node 10.72.74.189
 STEP: verifying the node doesn't have the label node
 [AfterEach] [sig-scheduling] SchedulerPredicates [Serial]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 18:43:38.693: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "sched-pred-4369" for this suite.
-Jun  4 18:43:44.745: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 18:43:45.078: INFO: namespace sched-pred-4369 deletion completed in 6.374263719s
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 11:45:36.468: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "sched-pred-6868" for this suite.
+Jun 18 11:45:44.507: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 11:45:44.902: INFO: namespace sched-pred-6868 deletion completed in 8.420880069s
 [AfterEach] [sig-scheduling] SchedulerPredicates [Serial]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:70
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:70
 
-• [SLOW TEST:12.339 seconds]
+• [SLOW TEST:14.279 seconds]
 [sig-scheduling] SchedulerPredicates [Serial]
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/framework.go:22
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/framework.go:22
   validates resource limits of pods that are allowed to run  [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-[sig-storage] Downward API volume 
-  should provide podname only [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [sig-storage] Downward API volume
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+------------------------------
+[sig-storage] ConfigMap 
+  should be consumable from pods in volume with mappings [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-storage] ConfigMap
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 18:43:45.079: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-STEP: Building a namespace api object, basename downward-api
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in downward-api-2981
+Jun 18 11:45:44.907: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename configmap
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in configmap-1253
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-storage] Downward API volume
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:39
-[It] should provide podname only [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: Creating a pod to test downward API volume plugin
-Jun  4 18:43:45.645: INFO: Waiting up to 5m0s for pod "downwardapi-volume-aeba0b31-86f8-11e9-8d1b-467ee19922ac" in namespace "downward-api-2981" to be "success or failure"
-Jun  4 18:43:45.661: INFO: Pod "downwardapi-volume-aeba0b31-86f8-11e9-8d1b-467ee19922ac": Phase="Pending", Reason="", readiness=false. Elapsed: 15.542125ms
-Jun  4 18:43:47.674: INFO: Pod "downwardapi-volume-aeba0b31-86f8-11e9-8d1b-467ee19922ac": Phase="Pending", Reason="", readiness=false. Elapsed: 2.029341159s
-Jun  4 18:43:49.688: INFO: Pod "downwardapi-volume-aeba0b31-86f8-11e9-8d1b-467ee19922ac": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.043288506s
+[It] should be consumable from pods in volume with mappings [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating configMap with name configmap-test-volume-map-9b8dcc8a-91be-11e9-a25d-8608290c688a
+STEP: Creating a pod to test consume configMaps
+Jun 18 11:45:45.149: INFO: Waiting up to 5m0s for pod "pod-configmaps-9b900c7e-91be-11e9-a25d-8608290c688a" in namespace "configmap-1253" to be "success or failure"
+Jun 18 11:45:45.167: INFO: Pod "pod-configmaps-9b900c7e-91be-11e9-a25d-8608290c688a": Phase="Pending", Reason="", readiness=false. Elapsed: 18.382668ms
+Jun 18 11:45:47.182: INFO: Pod "pod-configmaps-9b900c7e-91be-11e9-a25d-8608290c688a": Phase="Running", Reason="", readiness=true. Elapsed: 2.032886109s
+Jun 18 11:45:49.195: INFO: Pod "pod-configmaps-9b900c7e-91be-11e9-a25d-8608290c688a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.046423567s
 STEP: Saw pod success
-Jun  4 18:43:49.688: INFO: Pod "downwardapi-volume-aeba0b31-86f8-11e9-8d1b-467ee19922ac" satisfied condition "success or failure"
-Jun  4 18:43:49.700: INFO: Trying to get logs from node 10.212.23.161 pod downwardapi-volume-aeba0b31-86f8-11e9-8d1b-467ee19922ac container client-container: 
+Jun 18 11:45:49.195: INFO: Pod "pod-configmaps-9b900c7e-91be-11e9-a25d-8608290c688a" satisfied condition "success or failure"
+Jun 18 11:45:49.219: INFO: Trying to get logs from node 10.72.74.138 pod pod-configmaps-9b900c7e-91be-11e9-a25d-8608290c688a container configmap-volume-test: 
 STEP: delete the pod
-Jun  4 18:43:49.761: INFO: Waiting for pod downwardapi-volume-aeba0b31-86f8-11e9-8d1b-467ee19922ac to disappear
-Jun  4 18:43:49.855: INFO: Pod downwardapi-volume-aeba0b31-86f8-11e9-8d1b-467ee19922ac no longer exists
-[AfterEach] [sig-storage] Downward API volume
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 18:43:49.856: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "downward-api-2981" for this suite.
-Jun  4 18:43:55.918: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 18:43:56.526: INFO: namespace downward-api-2981 deletion completed in 6.652982802s
+Jun 18 11:45:49.730: INFO: Waiting for pod pod-configmaps-9b900c7e-91be-11e9-a25d-8608290c688a to disappear
+Jun 18 11:45:49.742: INFO: Pod pod-configmaps-9b900c7e-91be-11e9-a25d-8608290c688a no longer exists
+[AfterEach] [sig-storage] ConfigMap
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 11:45:49.742: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "configmap-1253" for this suite.
+Jun 18 11:45:55.785: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 11:45:56.169: INFO: namespace configmap-1253 deletion completed in 6.411500024s
 
-• [SLOW TEST:11.448 seconds]
-[sig-storage] Downward API volume
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:34
-  should provide podname only [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+• [SLOW TEST:11.262 seconds]
+[sig-storage] ConfigMap
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:32
+  should be consumable from pods in volume with mappings [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-S
+SSSSSSS
 ------------------------------
-[sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] 
-  Should recreate evicted statefulset [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [sig-apps] StatefulSet
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+[sig-storage] Subpath Atomic writer volumes 
+  should support subpaths with configmap pod [LinuxOnly] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-storage] Subpath
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 18:43:56.527: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-STEP: Building a namespace api object, basename statefulset
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in statefulset-8847
+Jun 18 11:45:56.169: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename subpath
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in subpath-8570
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-apps] StatefulSet
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:59
-[BeforeEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:74
-STEP: Creating service test in namespace statefulset-8847
-[It] Should recreate evicted statefulset [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: Looking for a node to schedule stateful set and pod
-STEP: Creating pod with conflicting port in namespace statefulset-8847
-STEP: Creating statefulset with conflicting port in namespace statefulset-8847
-STEP: Waiting until pod test-pod will start running in namespace statefulset-8847
-STEP: Waiting until stateful pod ss-0 will be recreated and deleted at least once in namespace statefulset-8847
-Jun  4 18:44:00.842: INFO: Observed stateful pod in namespace: statefulset-8847, name: ss-0, uid: b7bfaaf1-86f8-11e9-9509-923735b172b1, status phase: Pending. Waiting for statefulset controller to delete.
-Jun  4 18:44:01.012: INFO: Observed stateful pod in namespace: statefulset-8847, name: ss-0, uid: b7bfaaf1-86f8-11e9-9509-923735b172b1, status phase: Failed. Waiting for statefulset controller to delete.
-Jun  4 18:44:01.025: INFO: Observed stateful pod in namespace: statefulset-8847, name: ss-0, uid: b7bfaaf1-86f8-11e9-9509-923735b172b1, status phase: Failed. Waiting for statefulset controller to delete.
-Jun  4 18:44:01.035: INFO: Observed delete event for stateful pod ss-0 in namespace statefulset-8847
-STEP: Removing pod with conflicting port in namespace statefulset-8847
-STEP: Waiting when stateful pod ss-0 will be recreated in namespace statefulset-8847 and will be in running state
-[AfterEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:85
-Jun  4 18:44:05.108: INFO: Deleting all statefulset in ns statefulset-8847
-Jun  4 18:44:05.116: INFO: Scaling statefulset ss to 0
-Jun  4 18:44:15.165: INFO: Waiting for statefulset status.replicas updated to 0
-Jun  4 18:44:15.173: INFO: Deleting statefulset ss
-[AfterEach] [sig-apps] StatefulSet
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 18:44:15.255: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "statefulset-8847" for this suite.
-Jun  4 18:44:23.493: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 18:44:24.038: INFO: namespace statefulset-8847 deletion completed in 8.766327685s
+[BeforeEach] Atomic writer volumes
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:38
+STEP: Setting up data
+[It] should support subpaths with configmap pod [LinuxOnly] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating pod pod-subpath-test-configmap-gtnr
+STEP: Creating a pod to test atomic-volume-subpath
+Jun 18 11:45:56.423: INFO: Waiting up to 5m0s for pod "pod-subpath-test-configmap-gtnr" in namespace "subpath-8570" to be "success or failure"
+Jun 18 11:45:56.435: INFO: Pod "pod-subpath-test-configmap-gtnr": Phase="Pending", Reason="", readiness=false. Elapsed: 11.861926ms
+Jun 18 11:45:58.450: INFO: Pod "pod-subpath-test-configmap-gtnr": Phase="Pending", Reason="", readiness=false. Elapsed: 2.027307221s
+Jun 18 11:46:00.464: INFO: Pod "pod-subpath-test-configmap-gtnr": Phase="Running", Reason="", readiness=true. Elapsed: 4.040907821s
+Jun 18 11:46:02.481: INFO: Pod "pod-subpath-test-configmap-gtnr": Phase="Running", Reason="", readiness=true. Elapsed: 6.057857864s
+Jun 18 11:46:04.496: INFO: Pod "pod-subpath-test-configmap-gtnr": Phase="Running", Reason="", readiness=true. Elapsed: 8.073155257s
+Jun 18 11:46:06.510: INFO: Pod "pod-subpath-test-configmap-gtnr": Phase="Running", Reason="", readiness=true. Elapsed: 10.086723791s
+Jun 18 11:46:08.524: INFO: Pod "pod-subpath-test-configmap-gtnr": Phase="Running", Reason="", readiness=true. Elapsed: 12.101323056s
+Jun 18 11:46:10.539: INFO: Pod "pod-subpath-test-configmap-gtnr": Phase="Running", Reason="", readiness=true. Elapsed: 14.115569719s
+Jun 18 11:46:12.552: INFO: Pod "pod-subpath-test-configmap-gtnr": Phase="Running", Reason="", readiness=true. Elapsed: 16.129489137s
+Jun 18 11:46:14.566: INFO: Pod "pod-subpath-test-configmap-gtnr": Phase="Running", Reason="", readiness=true. Elapsed: 18.14349946s
+Jun 18 11:46:16.580: INFO: Pod "pod-subpath-test-configmap-gtnr": Phase="Running", Reason="", readiness=true. Elapsed: 20.156712973s
+Jun 18 11:46:18.593: INFO: Pod "pod-subpath-test-configmap-gtnr": Phase="Succeeded", Reason="", readiness=false. Elapsed: 22.170182843s
+STEP: Saw pod success
+Jun 18 11:46:18.593: INFO: Pod "pod-subpath-test-configmap-gtnr" satisfied condition "success or failure"
+Jun 18 11:46:18.606: INFO: Trying to get logs from node 10.72.74.184 pod pod-subpath-test-configmap-gtnr container test-container-subpath-configmap-gtnr: 
+STEP: delete the pod
+Jun 18 11:46:18.684: INFO: Waiting for pod pod-subpath-test-configmap-gtnr to disappear
+Jun 18 11:46:18.696: INFO: Pod pod-subpath-test-configmap-gtnr no longer exists
+STEP: Deleting pod pod-subpath-test-configmap-gtnr
+Jun 18 11:46:18.696: INFO: Deleting pod "pod-subpath-test-configmap-gtnr" in namespace "subpath-8570"
+[AfterEach] [sig-storage] Subpath
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 11:46:18.709: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "subpath-8570" for this suite.
+Jun 18 11:46:24.750: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 11:46:25.960: INFO: namespace subpath-8570 deletion completed in 7.237655867s
 
-• [SLOW TEST:27.511 seconds]
-[sig-apps] StatefulSet
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22
-  [k8s.io] Basic StatefulSet functionality [StatefulSetBasic]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
-    Should recreate evicted statefulset [Conformance]
-    /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+• [SLOW TEST:29.792 seconds]
+[sig-storage] Subpath
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:22
+  Atomic writer volumes
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:34
+    should support subpaths with configmap pod [LinuxOnly] [Conformance]
+    /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-SSSSSSSSSSSSSSSSSSSSS
+SSSSS
 ------------------------------
-[sig-api-machinery] CustomResourceDefinition resources Simple CustomResourceDefinition 
-  creating/deleting custom resource definition objects works  [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [sig-api-machinery] CustomResourceDefinition resources
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+[k8s.io] InitContainer [NodeConformance] 
+  should not start app containers if init containers fail on a RestartAlways pod [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [k8s.io] InitContainer [NodeConformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 18:44:24.040: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-STEP: Building a namespace api object, basename custom-resource-definition
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in custom-resource-definition-9983
+Jun 18 11:46:25.964: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename init-container
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in init-container-949
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] creating/deleting custom resource definition objects works  [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-Jun  4 18:44:24.283: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-[AfterEach] [sig-api-machinery] CustomResourceDefinition resources
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 18:44:25.449: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "custom-resource-definition-9983" for this suite.
-Jun  4 18:44:31.524: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 18:44:32.002: INFO: namespace custom-resource-definition-9983 deletion completed in 6.531038981s
+[BeforeEach] [k8s.io] InitContainer [NodeConformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/init_container.go:43
+[It] should not start app containers if init containers fail on a RestartAlways pod [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: creating the pod
+Jun 18 11:46:26.161: INFO: PodSpec: initContainers in spec.initContainers
+Jun 18 11:47:14.798: INFO: init container has failed twice: &v1.Pod{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"pod-init-b405ed50-91be-11e9-a25d-8608290c688a", GenerateName:"", Namespace:"init-container-949", SelfLink:"/api/v1/namespaces/init-container-949/pods/pod-init-b405ed50-91be-11e9-a25d-8608290c688a", UID:"b408109e-91be-11e9-a08a-ee7a14707756", ResourceVersion:"88440", Generation:0, CreationTimestamp:v1.Time{Time:time.Time{wall:0x0, ext:63696455186, loc:(*time.Location)(0x8a1a0e0)}}, DeletionTimestamp:(*v1.Time)(nil), DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"name":"foo", "time":"161204773"}, Annotations:map[string]string{"kubernetes.io/psp":"e2e-test-privileged-psp"}, OwnerReferences:[]v1.OwnerReference(nil), Initializers:(*v1.Initializers)(nil), Finalizers:[]string(nil), ClusterName:"", ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v1.PodSpec{Volumes:[]v1.Volume{v1.Volume{Name:"default-token-hc6qw", VolumeSource:v1.VolumeSource{HostPath:(*v1.HostPathVolumeSource)(nil), EmptyDir:(*v1.EmptyDirVolumeSource)(nil), GCEPersistentDisk:(*v1.GCEPersistentDiskVolumeSource)(nil), AWSElasticBlockStore:(*v1.AWSElasticBlockStoreVolumeSource)(nil), GitRepo:(*v1.GitRepoVolumeSource)(nil), Secret:(*v1.SecretVolumeSource)(0xc0020c4900), NFS:(*v1.NFSVolumeSource)(nil), ISCSI:(*v1.ISCSIVolumeSource)(nil), Glusterfs:(*v1.GlusterfsVolumeSource)(nil), PersistentVolumeClaim:(*v1.PersistentVolumeClaimVolumeSource)(nil), RBD:(*v1.RBDVolumeSource)(nil), FlexVolume:(*v1.FlexVolumeSource)(nil), Cinder:(*v1.CinderVolumeSource)(nil), CephFS:(*v1.CephFSVolumeSource)(nil), Flocker:(*v1.FlockerVolumeSource)(nil), DownwardAPI:(*v1.DownwardAPIVolumeSource)(nil), FC:(*v1.FCVolumeSource)(nil), AzureFile:(*v1.AzureFileVolumeSource)(nil), ConfigMap:(*v1.ConfigMapVolumeSource)(nil), VsphereVolume:(*v1.VsphereVirtualDiskVolumeSource)(nil), Quobyte:(*v1.QuobyteVolumeSource)(nil), AzureDisk:(*v1.AzureDiskVolumeSource)(nil), PhotonPersistentDisk:(*v1.PhotonPersistentDiskVolumeSource)(nil), Projected:(*v1.ProjectedVolumeSource)(nil), PortworxVolume:(*v1.PortworxVolumeSource)(nil), ScaleIO:(*v1.ScaleIOVolumeSource)(nil), StorageOS:(*v1.StorageOSVolumeSource)(nil), CSI:(*v1.CSIVolumeSource)(nil)}}}, InitContainers:[]v1.Container{v1.Container{Name:"init1", Image:"docker.io/library/busybox:1.29", Command:[]string{"/bin/false"}, Args:[]string(nil), WorkingDir:"", Ports:[]v1.ContainerPort(nil), EnvFrom:[]v1.EnvFromSource(nil), Env:[]v1.EnvVar(nil), Resources:v1.ResourceRequirements{Limits:v1.ResourceList(nil), Requests:v1.ResourceList(nil)}, VolumeMounts:[]v1.VolumeMount{v1.VolumeMount{Name:"default-token-hc6qw", ReadOnly:true, MountPath:"/var/run/secrets/kubernetes.io/serviceaccount", SubPath:"", MountPropagation:(*v1.MountPropagationMode)(nil), SubPathExpr:""}}, VolumeDevices:[]v1.VolumeDevice(nil), LivenessProbe:(*v1.Probe)(nil), ReadinessProbe:(*v1.Probe)(nil), Lifecycle:(*v1.Lifecycle)(nil), TerminationMessagePath:"/dev/termination-log", TerminationMessagePolicy:"File", ImagePullPolicy:"IfNotPresent", SecurityContext:(*v1.SecurityContext)(nil), Stdin:false, StdinOnce:false, TTY:false}, v1.Container{Name:"init2", Image:"docker.io/library/busybox:1.29", Command:[]string{"/bin/true"}, Args:[]string(nil), WorkingDir:"", Ports:[]v1.ContainerPort(nil), EnvFrom:[]v1.EnvFromSource(nil), Env:[]v1.EnvVar(nil), Resources:v1.ResourceRequirements{Limits:v1.ResourceList(nil), Requests:v1.ResourceList(nil)}, VolumeMounts:[]v1.VolumeMount{v1.VolumeMount{Name:"default-token-hc6qw", ReadOnly:true, MountPath:"/var/run/secrets/kubernetes.io/serviceaccount", SubPath:"", MountPropagation:(*v1.MountPropagationMode)(nil), SubPathExpr:""}}, VolumeDevices:[]v1.VolumeDevice(nil), LivenessProbe:(*v1.Probe)(nil), ReadinessProbe:(*v1.Probe)(nil), Lifecycle:(*v1.Lifecycle)(nil), TerminationMessagePath:"/dev/termination-log", TerminationMessagePolicy:"File", ImagePullPolicy:"IfNotPresent", SecurityContext:(*v1.SecurityContext)(nil), Stdin:false, StdinOnce:false, TTY:false}}, Containers:[]v1.Container{v1.Container{Name:"run1", Image:"k8s.gcr.io/pause:3.1", Command:[]string(nil), Args:[]string(nil), WorkingDir:"", Ports:[]v1.ContainerPort(nil), EnvFrom:[]v1.EnvFromSource(nil), Env:[]v1.EnvVar(nil), Resources:v1.ResourceRequirements{Limits:v1.ResourceList{"cpu":resource.Quantity{i:resource.int64Amount{value:100, scale:-3}, d:resource.infDecAmount{Dec:(*inf.Dec)(nil)}, s:"100m", Format:"DecimalSI"}, "memory":resource.Quantity{i:resource.int64Amount{value:52428800, scale:0}, d:resource.infDecAmount{Dec:(*inf.Dec)(nil)}, s:"52428800", Format:"DecimalSI"}}, Requests:v1.ResourceList{"cpu":resource.Quantity{i:resource.int64Amount{value:100, scale:-3}, d:resource.infDecAmount{Dec:(*inf.Dec)(nil)}, s:"100m", Format:"DecimalSI"}, "memory":resource.Quantity{i:resource.int64Amount{value:52428800, scale:0}, d:resource.infDecAmount{Dec:(*inf.Dec)(nil)}, s:"52428800", Format:"DecimalSI"}}}, VolumeMounts:[]v1.VolumeMount{v1.VolumeMount{Name:"default-token-hc6qw", ReadOnly:true, MountPath:"/var/run/secrets/kubernetes.io/serviceaccount", SubPath:"", MountPropagation:(*v1.MountPropagationMode)(nil), SubPathExpr:""}}, VolumeDevices:[]v1.VolumeDevice(nil), LivenessProbe:(*v1.Probe)(nil), ReadinessProbe:(*v1.Probe)(nil), Lifecycle:(*v1.Lifecycle)(nil), TerminationMessagePath:"/dev/termination-log", TerminationMessagePolicy:"File", ImagePullPolicy:"IfNotPresent", SecurityContext:(*v1.SecurityContext)(nil), Stdin:false, StdinOnce:false, TTY:false}}, RestartPolicy:"Always", TerminationGracePeriodSeconds:(*int64)(0xc002943c68), ActiveDeadlineSeconds:(*int64)(nil), DNSPolicy:"ClusterFirst", NodeSelector:map[string]string(nil), ServiceAccountName:"default", DeprecatedServiceAccount:"default", AutomountServiceAccountToken:(*bool)(nil), NodeName:"10.72.74.138", HostNetwork:false, HostPID:false, HostIPC:false, ShareProcessNamespace:(*bool)(nil), SecurityContext:(*v1.PodSecurityContext)(0xc002c25140), ImagePullSecrets:[]v1.LocalObjectReference(nil), Hostname:"", Subdomain:"", Affinity:(*v1.Affinity)(nil), SchedulerName:"default-scheduler", Tolerations:[]v1.Toleration{v1.Toleration{Key:"node.kubernetes.io/not-ready", Operator:"Exists", Value:"", Effect:"NoExecute", TolerationSeconds:(*int64)(0xc002943cf0)}, v1.Toleration{Key:"node.kubernetes.io/unreachable", Operator:"Exists", Value:"", Effect:"NoExecute", TolerationSeconds:(*int64)(0xc002943d10)}}, HostAliases:[]v1.HostAlias(nil), PriorityClassName:"", Priority:(*int32)(0xc002943d18), DNSConfig:(*v1.PodDNSConfig)(nil), ReadinessGates:[]v1.PodReadinessGate(nil), RuntimeClassName:(*string)(nil), EnableServiceLinks:(*bool)(0xc002943d1c)}, Status:v1.PodStatus{Phase:"Pending", Conditions:[]v1.PodCondition{v1.PodCondition{Type:"Initialized", Status:"False", LastProbeTime:v1.Time{Time:time.Time{wall:0x0, ext:0, loc:(*time.Location)(nil)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63696455186, loc:(*time.Location)(0x8a1a0e0)}}, Reason:"ContainersNotInitialized", Message:"containers with incomplete status: [init1 init2]"}, v1.PodCondition{Type:"Ready", Status:"False", LastProbeTime:v1.Time{Time:time.Time{wall:0x0, ext:0, loc:(*time.Location)(nil)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63696455186, loc:(*time.Location)(0x8a1a0e0)}}, Reason:"ContainersNotReady", Message:"containers with unready status: [run1]"}, v1.PodCondition{Type:"ContainersReady", Status:"False", LastProbeTime:v1.Time{Time:time.Time{wall:0x0, ext:0, loc:(*time.Location)(nil)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63696455186, loc:(*time.Location)(0x8a1a0e0)}}, Reason:"ContainersNotReady", Message:"containers with unready status: [run1]"}, v1.PodCondition{Type:"PodScheduled", Status:"True", LastProbeTime:v1.Time{Time:time.Time{wall:0x0, ext:0, loc:(*time.Location)(nil)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63696455186, loc:(*time.Location)(0x8a1a0e0)}}, Reason:"", Message:""}}, Message:"", Reason:"", NominatedNodeName:"", HostIP:"10.72.74.138", PodIP:"172.30.206.153", StartTime:(*v1.Time)(0xc001fe1980), InitContainerStatuses:[]v1.ContainerStatus{v1.ContainerStatus{Name:"init1", State:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(nil), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(0xc000b1f500)}, LastTerminationState:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(nil), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(0xc000b1f5e0)}, Ready:false, RestartCount:3, Image:"docker.io/library/busybox:1.29", ImageID:"docker.io/library/busybox@sha256:8ccbac733d19c0dd4d70b4f0c1e12245b5fa3ad24758a11035ee505c629c0796", ContainerID:"containerd://d4867cab864b6454fd95d49b2346624ba21a1b9a8f43409aa74e59e4f54b2d29"}, v1.ContainerStatus{Name:"init2", State:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(0xc001fe19c0), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(nil)}, LastTerminationState:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(nil), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(nil)}, Ready:false, RestartCount:0, Image:"docker.io/library/busybox:1.29", ImageID:"", ContainerID:""}}, ContainerStatuses:[]v1.ContainerStatus{v1.ContainerStatus{Name:"run1", State:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(0xc001fe19a0), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(nil)}, LastTerminationState:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(nil), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(nil)}, Ready:false, RestartCount:0, Image:"k8s.gcr.io/pause:3.1", ImageID:"", ContainerID:""}}, QOSClass:"Guaranteed"}}
+[AfterEach] [k8s.io] InitContainer [NodeConformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 11:47:14.799: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "init-container-949" for this suite.
+Jun 18 11:47:38.841: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 11:47:39.220: INFO: namespace init-container-949 deletion completed in 24.407791607s
 
-• [SLOW TEST:7.962 seconds]
-[sig-api-machinery] CustomResourceDefinition resources
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:22
-  Simple CustomResourceDefinition
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/custom_resource_definition.go:35
-    creating/deleting custom resource definition objects works  [Conformance]
-    /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+• [SLOW TEST:73.257 seconds]
+[k8s.io] InitContainer [NodeConformance]
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
+  should not start app containers if init containers fail on a RestartAlways pod [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-S
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-storage] EmptyDir wrapper volumes 
-  should not cause race condition when used for configmaps [Serial] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [sig-storage] EmptyDir wrapper volumes
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+[sig-storage] Secrets 
+  should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-storage] Secrets
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 18:44:32.002: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-STEP: Building a namespace api object, basename emptydir-wrapper
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in emptydir-wrapper-6413
+Jun 18 11:47:39.220: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename secrets
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in secrets-3194
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should not cause race condition when used for configmaps [Serial] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: Creating 50 configmaps
-STEP: Creating RC which spawns configmap-volume pods
-Jun  4 18:44:32.885: INFO: Pod name wrapped-volume-race-cb113739-86f8-11e9-8d1b-467ee19922ac: Found 0 pods out of 5
-Jun  4 18:44:37.906: INFO: Pod name wrapped-volume-race-cb113739-86f8-11e9-8d1b-467ee19922ac: Found 5 pods out of 5
-STEP: Ensuring each pod is running
-STEP: deleting ReplicationController wrapped-volume-race-cb113739-86f8-11e9-8d1b-467ee19922ac in namespace emptydir-wrapper-6413, will wait for the garbage collector to delete the pods
-Jun  4 18:44:48.202: INFO: Deleting ReplicationController wrapped-volume-race-cb113739-86f8-11e9-8d1b-467ee19922ac took: 31.405169ms
-Jun  4 18:44:48.302: INFO: Terminating ReplicationController wrapped-volume-race-cb113739-86f8-11e9-8d1b-467ee19922ac pods took: 100.385209ms
-STEP: Creating RC which spawns configmap-volume pods
-Jun  4 18:45:27.160: INFO: Pod name wrapped-volume-race-eb68ac5a-86f8-11e9-8d1b-467ee19922ac: Found 0 pods out of 5
-Jun  4 18:45:32.180: INFO: Pod name wrapped-volume-race-eb68ac5a-86f8-11e9-8d1b-467ee19922ac: Found 5 pods out of 5
-STEP: Ensuring each pod is running
-STEP: deleting ReplicationController wrapped-volume-race-eb68ac5a-86f8-11e9-8d1b-467ee19922ac in namespace emptydir-wrapper-6413, will wait for the garbage collector to delete the pods
-Jun  4 18:45:42.462: INFO: Deleting ReplicationController wrapped-volume-race-eb68ac5a-86f8-11e9-8d1b-467ee19922ac took: 30.027762ms
-Jun  4 18:45:42.562: INFO: Terminating ReplicationController wrapped-volume-race-eb68ac5a-86f8-11e9-8d1b-467ee19922ac pods took: 100.138239ms
-STEP: Creating RC which spawns configmap-volume pods
-Jun  4 18:46:27.091: INFO: Pod name wrapped-volume-race-0f16a5d9-86f9-11e9-8d1b-467ee19922ac: Found 0 pods out of 5
-Jun  4 18:46:32.110: INFO: Pod name wrapped-volume-race-0f16a5d9-86f9-11e9-8d1b-467ee19922ac: Found 5 pods out of 5
-STEP: Ensuring each pod is running
-STEP: deleting ReplicationController wrapped-volume-race-0f16a5d9-86f9-11e9-8d1b-467ee19922ac in namespace emptydir-wrapper-6413, will wait for the garbage collector to delete the pods
-Jun  4 18:46:42.281: INFO: Deleting ReplicationController wrapped-volume-race-0f16a5d9-86f9-11e9-8d1b-467ee19922ac took: 27.894139ms
-Jun  4 18:46:42.482: INFO: Terminating ReplicationController wrapped-volume-race-0f16a5d9-86f9-11e9-8d1b-467ee19922ac pods took: 200.452456ms
-STEP: Cleaning up the configMaps
-[AfterEach] [sig-storage] EmptyDir wrapper volumes
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 18:47:18.902: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "emptydir-wrapper-6413" for this suite.
-Jun  4 18:47:28.971: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 18:47:29.400: INFO: namespace emptydir-wrapper-6413 deletion completed in 10.474177332s
+[It] should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating secret with name secret-test-dfaf7393-91be-11e9-a25d-8608290c688a
+STEP: Creating a pod to test consume secrets
+Jun 18 11:47:39.452: INFO: Waiting up to 5m0s for pod "pod-secrets-dfb14e2f-91be-11e9-a25d-8608290c688a" in namespace "secrets-3194" to be "success or failure"
+Jun 18 11:47:39.472: INFO: Pod "pod-secrets-dfb14e2f-91be-11e9-a25d-8608290c688a": Phase="Pending", Reason="", readiness=false. Elapsed: 19.917016ms
+Jun 18 11:47:41.485: INFO: Pod "pod-secrets-dfb14e2f-91be-11e9-a25d-8608290c688a": Phase="Running", Reason="", readiness=true. Elapsed: 2.033428399s
+Jun 18 11:47:43.499: INFO: Pod "pod-secrets-dfb14e2f-91be-11e9-a25d-8608290c688a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.047255846s
+STEP: Saw pod success
+Jun 18 11:47:43.499: INFO: Pod "pod-secrets-dfb14e2f-91be-11e9-a25d-8608290c688a" satisfied condition "success or failure"
+Jun 18 11:47:43.513: INFO: Trying to get logs from node 10.72.74.184 pod pod-secrets-dfb14e2f-91be-11e9-a25d-8608290c688a container secret-volume-test: 
+STEP: delete the pod
+Jun 18 11:47:43.583: INFO: Waiting for pod pod-secrets-dfb14e2f-91be-11e9-a25d-8608290c688a to disappear
+Jun 18 11:47:43.596: INFO: Pod pod-secrets-dfb14e2f-91be-11e9-a25d-8608290c688a no longer exists
+[AfterEach] [sig-storage] Secrets
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 11:47:43.596: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "secrets-3194" for this suite.
+Jun 18 11:47:49.636: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 11:47:50.050: INFO: namespace secrets-3194 deletion completed in 6.439107783s
 
-• [SLOW TEST:177.398 seconds]
-[sig-storage] EmptyDir wrapper volumes
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:22
-  should not cause race condition when used for configmaps [Serial] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+• [SLOW TEST:10.829 seconds]
+[sig-storage] Secrets
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/secrets_volume.go:33
+  should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-SSSSSSS
+SSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[k8s.io] Container Lifecycle Hook when create a pod with lifecycle hook 
-  should execute prestop http hook properly [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [k8s.io] Container Lifecycle Hook
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+[sig-network] DNS 
+  should provide /etc/hosts entries for the cluster [LinuxOnly] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-network] DNS
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 18:47:29.401: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-STEP: Building a namespace api object, basename container-lifecycle-hook
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in container-lifecycle-hook-2560
+Jun 18 11:47:50.052: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename dns
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in dns-2966
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] when create a pod with lifecycle hook
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:61
-STEP: create the container to handle the HTTPGet hook request.
-[It] should execute prestop http hook properly [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: create the pod with lifecycle hook
-STEP: delete the pod with lifecycle hook
-Jun  4 18:47:33.744: INFO: Waiting for pod pod-with-prestop-http-hook to disappear
-Jun  4 18:47:33.759: INFO: Pod pod-with-prestop-http-hook still exists
-Jun  4 18:47:35.759: INFO: Waiting for pod pod-with-prestop-http-hook to disappear
-Jun  4 18:47:35.772: INFO: Pod pod-with-prestop-http-hook still exists
-Jun  4 18:47:37.759: INFO: Waiting for pod pod-with-prestop-http-hook to disappear
-Jun  4 18:47:37.778: INFO: Pod pod-with-prestop-http-hook still exists
-Jun  4 18:47:39.760: INFO: Waiting for pod pod-with-prestop-http-hook to disappear
-Jun  4 18:47:39.772: INFO: Pod pod-with-prestop-http-hook still exists
-Jun  4 18:47:41.760: INFO: Waiting for pod pod-with-prestop-http-hook to disappear
-Jun  4 18:47:41.776: INFO: Pod pod-with-prestop-http-hook still exists
-Jun  4 18:47:43.759: INFO: Waiting for pod pod-with-prestop-http-hook to disappear
-Jun  4 18:47:43.773: INFO: Pod pod-with-prestop-http-hook still exists
-Jun  4 18:47:45.760: INFO: Waiting for pod pod-with-prestop-http-hook to disappear
-Jun  4 18:47:45.777: INFO: Pod pod-with-prestop-http-hook still exists
-Jun  4 18:47:47.760: INFO: Waiting for pod pod-with-prestop-http-hook to disappear
-Jun  4 18:47:47.773: INFO: Pod pod-with-prestop-http-hook still exists
-Jun  4 18:47:49.759: INFO: Waiting for pod pod-with-prestop-http-hook to disappear
-Jun  4 18:47:49.774: INFO: Pod pod-with-prestop-http-hook no longer exists
-STEP: check prestop hook
-[AfterEach] [k8s.io] Container Lifecycle Hook
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 18:47:49.981: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "container-lifecycle-hook-2560" for this suite.
-Jun  4 18:48:14.099: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 18:48:14.551: INFO: namespace container-lifecycle-hook-2560 deletion completed in 24.552937184s
+[It] should provide /etc/hosts entries for the cluster [LinuxOnly] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Running these commands on wheezy: for i in `seq 1 600`; do test -n "$$(getent hosts dns-querier-1.dns-test-service.dns-2966.svc.cluster.local)" && echo OK > /results/wheezy_hosts@dns-querier-1.dns-test-service.dns-2966.svc.cluster.local;test -n "$$(getent hosts dns-querier-1)" && echo OK > /results/wheezy_hosts@dns-querier-1;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-2966.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@PodARecord;sleep 1; done
 
-• [SLOW TEST:45.150 seconds]
-[k8s.io] Container Lifecycle Hook
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
-  when create a pod with lifecycle hook
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:40
-    should execute prestop http hook properly [NodeConformance] [Conformance]
-    /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Running these commands on jessie: for i in `seq 1 600`; do test -n "$$(getent hosts dns-querier-1.dns-test-service.dns-2966.svc.cluster.local)" && echo OK > /results/jessie_hosts@dns-querier-1.dns-test-service.dns-2966.svc.cluster.local;test -n "$$(getent hosts dns-querier-1)" && echo OK > /results/jessie_hosts@dns-querier-1;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-2966.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_tcp@PodARecord;sleep 1; done
+
+STEP: creating a pod to probe /etc/hosts
+STEP: submitting the pod to kubernetes
+STEP: retrieving the pod
+STEP: looking for the results for each expected name from probers
+Jun 18 11:48:06.500: INFO: DNS probes using dns-2966/dns-test-e626a39c-91be-11e9-a25d-8608290c688a succeeded
+
+STEP: deleting the pod
+[AfterEach] [sig-network] DNS
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 11:48:06.548: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "dns-2966" for this suite.
+Jun 18 11:48:14.589: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 11:48:14.997: INFO: namespace dns-2966 deletion completed in 8.434748041s
+
+• [SLOW TEST:24.945 seconds]
+[sig-network] DNS
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:22
+  should provide /etc/hosts entries for the cluster [LinuxOnly] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-SS
+SSSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[k8s.io] Pods 
-  should allow activeDeadlineSeconds to be updated [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [k8s.io] Pods
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+[sig-storage] Downward API volume 
+  should update annotations on modification [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-storage] Downward API volume
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 18:48:14.551: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-STEP: Building a namespace api object, basename pods
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in pods-5750
+Jun 18 11:48:15.000: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename downward-api
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in downward-api-2550
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [k8s.io] Pods
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:135
-[It] should allow activeDeadlineSeconds to be updated [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: creating the pod
-STEP: submitting the pod to kubernetes
-STEP: verifying the pod is in kubernetes
-STEP: updating the pod
-Jun  4 18:48:17.586: INFO: Successfully updated pod "pod-update-activedeadlineseconds-4f66cf47-86f9-11e9-8d1b-467ee19922ac"
-Jun  4 18:48:17.586: INFO: Waiting up to 5m0s for pod "pod-update-activedeadlineseconds-4f66cf47-86f9-11e9-8d1b-467ee19922ac" in namespace "pods-5750" to be "terminated due to deadline exceeded"
-Jun  4 18:48:17.598: INFO: Pod "pod-update-activedeadlineseconds-4f66cf47-86f9-11e9-8d1b-467ee19922ac": Phase="Running", Reason="", readiness=true. Elapsed: 12.337798ms
-Jun  4 18:48:19.610: INFO: Pod "pod-update-activedeadlineseconds-4f66cf47-86f9-11e9-8d1b-467ee19922ac": Phase="Running", Reason="", readiness=true. Elapsed: 2.024790538s
-Jun  4 18:48:21.629: INFO: Pod "pod-update-activedeadlineseconds-4f66cf47-86f9-11e9-8d1b-467ee19922ac": Phase="Failed", Reason="DeadlineExceeded", readiness=false. Elapsed: 4.043915415s
-Jun  4 18:48:21.630: INFO: Pod "pod-update-activedeadlineseconds-4f66cf47-86f9-11e9-8d1b-467ee19922ac" satisfied condition "terminated due to deadline exceeded"
-[AfterEach] [k8s.io] Pods
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 18:48:21.630: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "pods-5750" for this suite.
-Jun  4 18:48:27.692: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 18:48:28.077: INFO: namespace pods-5750 deletion completed in 6.427498666s
+[BeforeEach] [sig-storage] Downward API volume
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:39
+[It] should update annotations on modification [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating the pod
+Jun 18 11:48:17.856: INFO: Successfully updated pod "annotationupdatef50593dc-91be-11e9-a25d-8608290c688a"
+[AfterEach] [sig-storage] Downward API volume
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 11:48:19.920: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "downward-api-2550" for this suite.
+Jun 18 11:48:43.969: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 11:48:45.077: INFO: namespace downward-api-2550 deletion completed in 25.14305531s
 
-• [SLOW TEST:13.526 seconds]
-[k8s.io] Pods
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
-  should allow activeDeadlineSeconds to be updated [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+• [SLOW TEST:30.077 seconds]
+[sig-storage] Downward API volume
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:34
+  should update annotations on modification [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-SSSSSSSSSSSSSSSSSSSSS
+SSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-apps] ReplicaSet 
-  should serve a basic image on each replica with a public image  [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [sig-apps] ReplicaSet
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+[sig-storage] Projected secret 
+  should be consumable from pods in volume [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-storage] Projected secret
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 18:48:28.078: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-STEP: Building a namespace api object, basename replicaset
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in replicaset-2586
+Jun 18 11:48:45.078: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename projected
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-5602
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should serve a basic image on each replica with a public image  [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-Jun  4 18:48:28.297: INFO: Creating ReplicaSet my-hostname-basic-5768b9a0-86f9-11e9-8d1b-467ee19922ac
-Jun  4 18:48:28.323: INFO: Pod name my-hostname-basic-5768b9a0-86f9-11e9-8d1b-467ee19922ac: Found 0 pods out of 1
-Jun  4 18:48:33.335: INFO: Pod name my-hostname-basic-5768b9a0-86f9-11e9-8d1b-467ee19922ac: Found 1 pods out of 1
-Jun  4 18:48:33.335: INFO: Ensuring a pod for ReplicaSet "my-hostname-basic-5768b9a0-86f9-11e9-8d1b-467ee19922ac" is running
-Jun  4 18:48:33.349: INFO: Pod "my-hostname-basic-5768b9a0-86f9-11e9-8d1b-467ee19922ac-vwsdn" is running (conditions: [{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2019-06-04 18:48:28 +0000 UTC Reason: Message:} {Type:Ready Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2019-06-04 18:48:32 +0000 UTC Reason: Message:} {Type:ContainersReady Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2019-06-04 18:48:32 +0000 UTC Reason: Message:} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2019-06-04 18:48:28 +0000 UTC Reason: Message:}])
-Jun  4 18:48:33.349: INFO: Trying to dial the pod
-Jun  4 18:48:38.396: INFO: Controller my-hostname-basic-5768b9a0-86f9-11e9-8d1b-467ee19922ac: Got expected result from replica 1 [my-hostname-basic-5768b9a0-86f9-11e9-8d1b-467ee19922ac-vwsdn]: "my-hostname-basic-5768b9a0-86f9-11e9-8d1b-467ee19922ac-vwsdn", 1 of 1 required successes so far
-[AfterEach] [sig-apps] ReplicaSet
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 18:48:38.396: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "replicaset-2586" for this suite.
-Jun  4 18:48:44.477: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 18:48:44.841: INFO: namespace replicaset-2586 deletion completed in 6.430447872s
+[It] should be consumable from pods in volume [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating projection with secret that has name projected-secret-test-06f0cdf6-91bf-11e9-a25d-8608290c688a
+STEP: Creating a pod to test consume secrets
+Jun 18 11:48:45.311: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-06f29e52-91bf-11e9-a25d-8608290c688a" in namespace "projected-5602" to be "success or failure"
+Jun 18 11:48:45.324: INFO: Pod "pod-projected-secrets-06f29e52-91bf-11e9-a25d-8608290c688a": Phase="Pending", Reason="", readiness=false. Elapsed: 13.264565ms
+Jun 18 11:48:47.338: INFO: Pod "pod-projected-secrets-06f29e52-91bf-11e9-a25d-8608290c688a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.026696575s
+STEP: Saw pod success
+Jun 18 11:48:47.338: INFO: Pod "pod-projected-secrets-06f29e52-91bf-11e9-a25d-8608290c688a" satisfied condition "success or failure"
+Jun 18 11:48:47.351: INFO: Trying to get logs from node 10.72.74.138 pod pod-projected-secrets-06f29e52-91bf-11e9-a25d-8608290c688a container projected-secret-volume-test: 
+STEP: delete the pod
+Jun 18 11:48:47.432: INFO: Waiting for pod pod-projected-secrets-06f29e52-91bf-11e9-a25d-8608290c688a to disappear
+Jun 18 11:48:47.444: INFO: Pod pod-projected-secrets-06f29e52-91bf-11e9-a25d-8608290c688a no longer exists
+[AfterEach] [sig-storage] Projected secret
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 11:48:47.444: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "projected-5602" for this suite.
+Jun 18 11:48:53.494: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 11:48:53.870: INFO: namespace projected-5602 deletion completed in 6.41192246s
 
-• [SLOW TEST:16.763 seconds]
-[sig-apps] ReplicaSet
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22
-  should serve a basic image on each replica with a public image  [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+• [SLOW TEST:8.792 seconds]
+[sig-storage] Projected secret
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_secret.go:33
+  should be consumable from pods in volume [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-SS
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[k8s.io] [sig-node] PreStop 
-  should call prestop when killing a pod  [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [k8s.io] [sig-node] PreStop
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+[k8s.io] Kubelet when scheduling a read only busybox container 
+  should not write to root filesystem [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [k8s.io] Kubelet
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 18:48:44.841: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-STEP: Building a namespace api object, basename prestop
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in prestop-8668
+Jun 18 11:48:53.874: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename kubelet-test
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubelet-test-8518
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [k8s.io] [sig-node] PreStop
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/node/pre_stop.go:167
-[It] should call prestop when killing a pod  [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: Creating server pod server in namespace prestop-8668
-STEP: Waiting for pods to come up.
-STEP: Creating tester pod tester in namespace prestop-8668
-STEP: Deleting pre-stop pod
-Jun  4 18:49:00.209: INFO: Saw: {
-	"Hostname": "server",
-	"Sent": null,
-	"Received": {
-		"prestop": 1
-	},
-	"Errors": null,
-	"Log": [
-		"default/nettest has 0 endpoints ([]), which is less than 8 as expected. Waiting for all endpoints to come up.",
-		"default/nettest has 0 endpoints ([]), which is less than 8 as expected. Waiting for all endpoints to come up.",
-		"default/nettest has 0 endpoints ([]), which is less than 8 as expected. Waiting for all endpoints to come up."
-	],
-	"StillContactingPeers": true
-}
-STEP: Deleting the server pod
-[AfterEach] [k8s.io] [sig-node] PreStop
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 18:49:00.230: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "prestop-8668" for this suite.
-Jun  4 18:49:40.294: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 18:49:40.714: INFO: namespace prestop-8668 deletion completed in 40.46886833s
+[BeforeEach] [k8s.io] Kubelet
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:37
+[It] should not write to root filesystem [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[AfterEach] [k8s.io] Kubelet
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 11:48:57.254: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "kubelet-test-8518" for this suite.
+Jun 18 11:49:39.292: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 11:49:39.696: INFO: namespace kubelet-test-8518 deletion completed in 42.428403775s
 
-• [SLOW TEST:55.872 seconds]
-[k8s.io] [sig-node] PreStop
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
-  should call prestop when killing a pod  [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+• [SLOW TEST:45.823 seconds]
+[k8s.io] Kubelet
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
+  when scheduling a read only busybox container
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:187
+    should not write to root filesystem [LinuxOnly] [NodeConformance] [Conformance]
+    /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-SSSS
+SSSSSSSSSSSSSSSS
 ------------------------------
-[k8s.io] Container Runtime blackbox test when starting a container that exits 
-  should run with the expected status [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [k8s.io] Container Runtime
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+[sig-storage] Projected downwardAPI 
+  should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-storage] Projected downwardAPI
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 18:49:40.714: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-STEP: Building a namespace api object, basename container-runtime
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in container-runtime-686
+Jun 18 11:49:39.699: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename projected
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-9206
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should run with the expected status [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: Container 'terminate-cmd-rpa': should get the expected 'RestartCount'
-STEP: Container 'terminate-cmd-rpa': should get the expected 'Phase'
-STEP: Container 'terminate-cmd-rpa': should get the expected 'Ready' condition
-STEP: Container 'terminate-cmd-rpa': should get the expected 'State'
-STEP: Container 'terminate-cmd-rpa': should be possible to delete [NodeConformance]
-STEP: Container 'terminate-cmd-rpof': should get the expected 'RestartCount'
-STEP: Container 'terminate-cmd-rpof': should get the expected 'Phase'
-STEP: Container 'terminate-cmd-rpof': should get the expected 'Ready' condition
-STEP: Container 'terminate-cmd-rpof': should get the expected 'State'
-STEP: Container 'terminate-cmd-rpof': should be possible to delete [NodeConformance]
-STEP: Container 'terminate-cmd-rpn': should get the expected 'RestartCount'
-STEP: Container 'terminate-cmd-rpn': should get the expected 'Phase'
-STEP: Container 'terminate-cmd-rpn': should get the expected 'Ready' condition
-STEP: Container 'terminate-cmd-rpn': should get the expected 'State'
-STEP: Container 'terminate-cmd-rpn': should be possible to delete [NodeConformance]
-[AfterEach] [k8s.io] Container Runtime
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 18:50:01.735: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "container-runtime-686" for this suite.
-Jun  4 18:50:07.800: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 18:50:08.231: INFO: namespace container-runtime-686 deletion completed in 6.479595811s
+[BeforeEach] [sig-storage] Projected downwardAPI
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:39
+[It] should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating a pod to test downward API volume plugin
+Jun 18 11:49:39.924: INFO: Waiting up to 5m0s for pod "downwardapi-volume-277f5551-91bf-11e9-a25d-8608290c688a" in namespace "projected-9206" to be "success or failure"
+Jun 18 11:49:39.937: INFO: Pod "downwardapi-volume-277f5551-91bf-11e9-a25d-8608290c688a": Phase="Pending", Reason="", readiness=false. Elapsed: 12.592013ms
+Jun 18 11:49:41.950: INFO: Pod "downwardapi-volume-277f5551-91bf-11e9-a25d-8608290c688a": Phase="Running", Reason="", readiness=true. Elapsed: 2.025921914s
+Jun 18 11:49:43.964: INFO: Pod "downwardapi-volume-277f5551-91bf-11e9-a25d-8608290c688a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.039818761s
+STEP: Saw pod success
+Jun 18 11:49:43.964: INFO: Pod "downwardapi-volume-277f5551-91bf-11e9-a25d-8608290c688a" satisfied condition "success or failure"
+Jun 18 11:49:43.977: INFO: Trying to get logs from node 10.72.74.138 pod downwardapi-volume-277f5551-91bf-11e9-a25d-8608290c688a container client-container: 
+STEP: delete the pod
+Jun 18 11:49:44.059: INFO: Waiting for pod downwardapi-volume-277f5551-91bf-11e9-a25d-8608290c688a to disappear
+Jun 18 11:49:44.091: INFO: Pod downwardapi-volume-277f5551-91bf-11e9-a25d-8608290c688a no longer exists
+[AfterEach] [sig-storage] Projected downwardAPI
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 11:49:44.091: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "projected-9206" for this suite.
+Jun 18 11:49:50.132: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 11:49:50.515: INFO: namespace projected-9206 deletion completed in 6.409021084s
 
-• [SLOW TEST:27.517 seconds]
-[k8s.io] Container Runtime
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
-  blackbox test
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/runtime.go:37
-    when starting a container that exits
-    /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/runtime.go:38
-      should run with the expected status [NodeConformance] [Conformance]
-      /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+• [SLOW TEST:10.816 seconds]
+[sig-storage] Projected downwardAPI
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:33
+  should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-SSSSSSSSSSSSSSSS
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-api-machinery] Watchers 
-  should be able to start watching from a specific resource version [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [sig-api-machinery] Watchers
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+[sig-api-machinery] Garbage collector 
+  should delete RS created by deployment when not orphaning [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-api-machinery] Garbage collector
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 18:50:08.231: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-STEP: Building a namespace api object, basename watch
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in watch-1846
+Jun 18 11:49:50.515: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename gc
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in gc-8472
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should be able to start watching from a specific resource version [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: creating a new configmap
-STEP: modifying the configmap once
-STEP: modifying the configmap a second time
-STEP: deleting the configmap
-STEP: creating a watch on configmaps from the resource version returned by the first update
-STEP: Expecting to observe notifications for all changes to the configmap after the first update
-Jun  4 18:50:08.541: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-resource-version,GenerateName:,Namespace:watch-1846,SelfLink:/api/v1/namespaces/watch-1846/configmaps/e2e-watch-test-resource-version,UID:931a7a09-86f9-11e9-8318-1e5386706511,ResourceVersion:12855,Generation:0,CreationTimestamp:2019-06-04 18:50:08 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: from-resource-version,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},}
-Jun  4 18:50:08.542: INFO: Got : DELETED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-resource-version,GenerateName:,Namespace:watch-1846,SelfLink:/api/v1/namespaces/watch-1846/configmaps/e2e-watch-test-resource-version,UID:931a7a09-86f9-11e9-8318-1e5386706511,ResourceVersion:12856,Generation:0,CreationTimestamp:2019-06-04 18:50:08 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: from-resource-version,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},}
-[AfterEach] [sig-api-machinery] Watchers
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 18:50:08.542: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "watch-1846" for this suite.
-Jun  4 18:50:14.609: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 18:50:15.016: INFO: namespace watch-1846 deletion completed in 6.453898991s
+[It] should delete RS created by deployment when not orphaning [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: create the deployment
+STEP: Wait for the Deployment to create new ReplicaSet
+STEP: delete the deployment
+STEP: wait for all rs to be garbage collected
+STEP: expected 0 pods, got 2 pods
+STEP: expected 0 rs, got 1 rs
+STEP: Gathering metrics
+W0618 11:49:51.834872      20 metrics_grabber.go:79] Master node is not registered. Grabbing metrics from Scheduler, ControllerManager and ClusterAutoscaler is disabled.
+Jun 18 11:49:51.834: INFO: For apiserver_request_total:
+For apiserver_request_latencies_summary:
+For apiserver_init_events_total:
+For garbage_collector_attempt_to_delete_queue_latency:
+For garbage_collector_attempt_to_delete_work_duration:
+For garbage_collector_attempt_to_orphan_queue_latency:
+For garbage_collector_attempt_to_orphan_work_duration:
+For garbage_collector_dirty_processing_latency_microseconds:
+For garbage_collector_event_processing_latency_microseconds:
+For garbage_collector_graph_changes_queue_latency:
+For garbage_collector_graph_changes_work_duration:
+For garbage_collector_orphan_processing_latency_microseconds:
+For namespace_queue_latency:
+For namespace_queue_latency_sum:
+For namespace_queue_latency_count:
+For namespace_retries:
+For namespace_work_duration:
+For namespace_work_duration_sum:
+For namespace_work_duration_count:
+For function_duration_seconds:
+For errors_total:
+For evicted_pods_total:
 
-• [SLOW TEST:6.785 seconds]
-[sig-api-machinery] Watchers
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:22
-  should be able to start watching from a specific resource version [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[AfterEach] [sig-api-machinery] Garbage collector
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 11:49:51.835: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "gc-8472" for this suite.
+Jun 18 11:49:58.599: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 11:49:58.985: INFO: namespace gc-8472 deletion completed in 7.138943471s
+
+• [SLOW TEST:8.470 seconds]
+[sig-api-machinery] Garbage collector
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:22
+  should delete RS created by deployment when not orphaning [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-SSSS
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
 [sig-storage] Projected secret 
   should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 [BeforeEach] [sig-storage] Projected secret
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 18:50:15.016: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
+Jun 18 11:49:58.991: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
 STEP: Building a namespace api object, basename projected
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-6421
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-2138
 STEP: Waiting for a default service account to be provisioned in namespace
 [It] should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: Creating projection with secret that has name projected-secret-test-9728a29d-86f9-11e9-8d1b-467ee19922ac
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating projection with secret that has name projected-secret-test-32fe2ff1-91bf-11e9-a25d-8608290c688a
 STEP: Creating a pod to test consume secrets
-Jun  4 18:50:15.287: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-9729e5fc-86f9-11e9-8d1b-467ee19922ac" in namespace "projected-6421" to be "success or failure"
-Jun  4 18:50:15.298: INFO: Pod "pod-projected-secrets-9729e5fc-86f9-11e9-8d1b-467ee19922ac": Phase="Pending", Reason="", readiness=false. Elapsed: 11.167652ms
-Jun  4 18:50:17.312: INFO: Pod "pod-projected-secrets-9729e5fc-86f9-11e9-8d1b-467ee19922ac": Phase="Running", Reason="", readiness=true. Elapsed: 2.024654294s
-Jun  4 18:50:19.325: INFO: Pod "pod-projected-secrets-9729e5fc-86f9-11e9-8d1b-467ee19922ac": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.037557279s
+Jun 18 11:49:59.219: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-330005e0-91bf-11e9-a25d-8608290c688a" in namespace "projected-2138" to be "success or failure"
+Jun 18 11:49:59.232: INFO: Pod "pod-projected-secrets-330005e0-91bf-11e9-a25d-8608290c688a": Phase="Pending", Reason="", readiness=false. Elapsed: 13.370449ms
+Jun 18 11:50:01.248: INFO: Pod "pod-projected-secrets-330005e0-91bf-11e9-a25d-8608290c688a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.029062538s
 STEP: Saw pod success
-Jun  4 18:50:19.325: INFO: Pod "pod-projected-secrets-9729e5fc-86f9-11e9-8d1b-467ee19922ac" satisfied condition "success or failure"
-Jun  4 18:50:19.336: INFO: Trying to get logs from node 10.212.23.161 pod pod-projected-secrets-9729e5fc-86f9-11e9-8d1b-467ee19922ac container projected-secret-volume-test: 
+Jun 18 11:50:01.248: INFO: Pod "pod-projected-secrets-330005e0-91bf-11e9-a25d-8608290c688a" satisfied condition "success or failure"
+Jun 18 11:50:01.267: INFO: Trying to get logs from node 10.72.74.184 pod pod-projected-secrets-330005e0-91bf-11e9-a25d-8608290c688a container projected-secret-volume-test: 
 STEP: delete the pod
-Jun  4 18:50:19.406: INFO: Waiting for pod pod-projected-secrets-9729e5fc-86f9-11e9-8d1b-467ee19922ac to disappear
-Jun  4 18:50:19.420: INFO: Pod pod-projected-secrets-9729e5fc-86f9-11e9-8d1b-467ee19922ac no longer exists
+Jun 18 11:50:01.341: INFO: Waiting for pod pod-projected-secrets-330005e0-91bf-11e9-a25d-8608290c688a to disappear
+Jun 18 11:50:01.359: INFO: Pod pod-projected-secrets-330005e0-91bf-11e9-a25d-8608290c688a no longer exists
 [AfterEach] [sig-storage] Projected secret
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 18:50:19.421: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "projected-6421" for this suite.
-Jun  4 18:50:25.478: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 18:50:25.886: INFO: namespace projected-6421 deletion completed in 6.450929337s
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 11:50:01.359: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "projected-2138" for this suite.
+Jun 18 11:50:07.402: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 11:50:07.832: INFO: namespace projected-2138 deletion completed in 6.459098888s
 
-• [SLOW TEST:10.870 seconds]
+• [SLOW TEST:8.841 seconds]
 [sig-storage] Projected secret
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_secret.go:33
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_secret.go:33
   should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+SS
 ------------------------------
-[sig-storage] Projected configMap 
-  updates should be reflected in volume [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [sig-storage] Projected configMap
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+[sig-storage] Subpath Atomic writer volumes 
+  should support subpaths with secret pod [LinuxOnly] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-storage] Subpath
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 18:50:25.887: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-STEP: Building a namespace api object, basename projected
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-7932
+Jun 18 11:50:07.833: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename subpath
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in subpath-5322
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] updates should be reflected in volume [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: Creating projection with configMap that has name projected-configmap-test-upd-9da4c3cd-86f9-11e9-8d1b-467ee19922ac
-STEP: Creating the pod
-STEP: Updating configmap projected-configmap-test-upd-9da4c3cd-86f9-11e9-8d1b-467ee19922ac
-STEP: waiting to observe update in volume
-[AfterEach] [sig-storage] Projected configMap
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 18:50:30.326: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "projected-7932" for this suite.
-Jun  4 18:50:54.387: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 18:50:54.722: INFO: namespace projected-7932 deletion completed in 24.378846547s
+[BeforeEach] Atomic writer volumes
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:38
+STEP: Setting up data
+[It] should support subpaths with secret pod [LinuxOnly] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating pod pod-subpath-test-secret-5mq6
+STEP: Creating a pod to test atomic-volume-subpath
+Jun 18 11:50:08.085: INFO: Waiting up to 5m0s for pod "pod-subpath-test-secret-5mq6" in namespace "subpath-5322" to be "success or failure"
+Jun 18 11:50:08.098: INFO: Pod "pod-subpath-test-secret-5mq6": Phase="Pending", Reason="", readiness=false. Elapsed: 12.697522ms
+Jun 18 11:50:10.115: INFO: Pod "pod-subpath-test-secret-5mq6": Phase="Running", Reason="", readiness=true. Elapsed: 2.029576351s
+Jun 18 11:50:12.129: INFO: Pod "pod-subpath-test-secret-5mq6": Phase="Running", Reason="", readiness=true. Elapsed: 4.043452706s
+Jun 18 11:50:14.143: INFO: Pod "pod-subpath-test-secret-5mq6": Phase="Running", Reason="", readiness=true. Elapsed: 6.057357107s
+Jun 18 11:50:16.161: INFO: Pod "pod-subpath-test-secret-5mq6": Phase="Running", Reason="", readiness=true. Elapsed: 8.075248133s
+Jun 18 11:50:18.175: INFO: Pod "pod-subpath-test-secret-5mq6": Phase="Running", Reason="", readiness=true. Elapsed: 10.089346088s
+Jun 18 11:50:20.189: INFO: Pod "pod-subpath-test-secret-5mq6": Phase="Running", Reason="", readiness=true. Elapsed: 12.103324077s
+Jun 18 11:50:22.203: INFO: Pod "pod-subpath-test-secret-5mq6": Phase="Running", Reason="", readiness=true. Elapsed: 14.117637431s
+Jun 18 11:50:24.219: INFO: Pod "pod-subpath-test-secret-5mq6": Phase="Running", Reason="", readiness=true. Elapsed: 16.133782398s
+Jun 18 11:50:26.236: INFO: Pod "pod-subpath-test-secret-5mq6": Phase="Running", Reason="", readiness=true. Elapsed: 18.150876064s
+Jun 18 11:50:28.250: INFO: Pod "pod-subpath-test-secret-5mq6": Phase="Running", Reason="", readiness=true. Elapsed: 20.165200754s
+Jun 18 11:50:30.264: INFO: Pod "pod-subpath-test-secret-5mq6": Phase="Running", Reason="", readiness=true. Elapsed: 22.178647122s
+Jun 18 11:50:32.278: INFO: Pod "pod-subpath-test-secret-5mq6": Phase="Succeeded", Reason="", readiness=false. Elapsed: 24.1926271s
+STEP: Saw pod success
+Jun 18 11:50:32.278: INFO: Pod "pod-subpath-test-secret-5mq6" satisfied condition "success or failure"
+Jun 18 11:50:32.291: INFO: Trying to get logs from node 10.72.74.138 pod pod-subpath-test-secret-5mq6 container test-container-subpath-secret-5mq6: 
+STEP: delete the pod
+Jun 18 11:50:32.374: INFO: Waiting for pod pod-subpath-test-secret-5mq6 to disappear
+Jun 18 11:50:32.386: INFO: Pod pod-subpath-test-secret-5mq6 no longer exists
+STEP: Deleting pod pod-subpath-test-secret-5mq6
+Jun 18 11:50:32.387: INFO: Deleting pod "pod-subpath-test-secret-5mq6" in namespace "subpath-5322"
+[AfterEach] [sig-storage] Subpath
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 11:50:32.399: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "subpath-5322" for this suite.
+Jun 18 11:50:38.438: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 11:50:40.384: INFO: namespace subpath-5322 deletion completed in 7.971517056s
 
-• [SLOW TEST:28.836 seconds]
-[sig-storage] Projected configMap
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_configmap.go:33
-  updates should be reflected in volume [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+• [SLOW TEST:32.551 seconds]
+[sig-storage] Subpath
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:22
+  Atomic writer volumes
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:34
+    should support subpaths with secret pod [LinuxOnly] [Conformance]
+    /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-SSSSS
+SSS
 ------------------------------
-[sig-storage] Secrets 
-  should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [sig-storage] Secrets
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+[sig-storage] ConfigMap 
+  should be consumable from pods in volume as non-root [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-storage] ConfigMap
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 18:50:54.723: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-STEP: Building a namespace api object, basename secrets
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in secrets-1720
+Jun 18 11:50:40.384: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename configmap
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in configmap-8997
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: Creating secret with name secret-test-aed4fc34-86f9-11e9-8d1b-467ee19922ac
-STEP: Creating a pod to test consume secrets
-Jun  4 18:50:54.998: INFO: Waiting up to 5m0s for pod "pod-secrets-aed676d9-86f9-11e9-8d1b-467ee19922ac" in namespace "secrets-1720" to be "success or failure"
-Jun  4 18:50:55.008: INFO: Pod "pod-secrets-aed676d9-86f9-11e9-8d1b-467ee19922ac": Phase="Pending", Reason="", readiness=false. Elapsed: 10.215138ms
-Jun  4 18:50:57.022: INFO: Pod "pod-secrets-aed676d9-86f9-11e9-8d1b-467ee19922ac": Phase="Running", Reason="", readiness=true. Elapsed: 2.023911815s
-Jun  4 18:50:59.033: INFO: Pod "pod-secrets-aed676d9-86f9-11e9-8d1b-467ee19922ac": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.035715044s
+[It] should be consumable from pods in volume as non-root [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating configMap with name configmap-test-volume-4bab8dac-91bf-11e9-a25d-8608290c688a
+STEP: Creating a pod to test consume configMaps
+Jun 18 11:50:40.626: INFO: Waiting up to 5m0s for pod "pod-configmaps-4bae051f-91bf-11e9-a25d-8608290c688a" in namespace "configmap-8997" to be "success or failure"
+Jun 18 11:50:40.644: INFO: Pod "pod-configmaps-4bae051f-91bf-11e9-a25d-8608290c688a": Phase="Pending", Reason="", readiness=false. Elapsed: 18.657121ms
+Jun 18 11:50:42.658: INFO: Pod "pod-configmaps-4bae051f-91bf-11e9-a25d-8608290c688a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.032462022s
 STEP: Saw pod success
-Jun  4 18:50:59.033: INFO: Pod "pod-secrets-aed676d9-86f9-11e9-8d1b-467ee19922ac" satisfied condition "success or failure"
-Jun  4 18:50:59.044: INFO: Trying to get logs from node 10.212.23.189 pod pod-secrets-aed676d9-86f9-11e9-8d1b-467ee19922ac container secret-volume-test: 
+Jun 18 11:50:42.658: INFO: Pod "pod-configmaps-4bae051f-91bf-11e9-a25d-8608290c688a" satisfied condition "success or failure"
+Jun 18 11:50:42.677: INFO: Trying to get logs from node 10.72.74.184 pod pod-configmaps-4bae051f-91bf-11e9-a25d-8608290c688a container configmap-volume-test: 
 STEP: delete the pod
-Jun  4 18:50:59.192: INFO: Waiting for pod pod-secrets-aed676d9-86f9-11e9-8d1b-467ee19922ac to disappear
-Jun  4 18:50:59.207: INFO: Pod pod-secrets-aed676d9-86f9-11e9-8d1b-467ee19922ac no longer exists
-[AfterEach] [sig-storage] Secrets
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 18:50:59.207: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "secrets-1720" for this suite.
-Jun  4 18:51:05.423: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 18:51:05.947: INFO: namespace secrets-1720 deletion completed in 6.725028636s
+Jun 18 11:50:42.755: INFO: Waiting for pod pod-configmaps-4bae051f-91bf-11e9-a25d-8608290c688a to disappear
+Jun 18 11:50:42.778: INFO: Pod pod-configmaps-4bae051f-91bf-11e9-a25d-8608290c688a no longer exists
+[AfterEach] [sig-storage] ConfigMap
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 11:50:42.778: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "configmap-8997" for this suite.
+Jun 18 11:50:48.825: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 11:50:49.459: INFO: namespace configmap-8997 deletion completed in 6.664330121s
 
-• [SLOW TEST:11.224 seconds]
-[sig-storage] Secrets
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/secrets_volume.go:33
-  should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+• [SLOW TEST:9.075 seconds]
+[sig-storage] ConfigMap
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:32
+  should be consumable from pods in volume as non-root [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+SSSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-storage] EmptyDir wrapper volumes 
-  should not conflict [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [sig-storage] EmptyDir wrapper volumes
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+[sig-apps] Deployment 
+  RecreateDeployment should delete old pods and create new ones [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-apps] Deployment
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 18:51:05.948: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-STEP: Building a namespace api object, basename emptydir-wrapper
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in emptydir-wrapper-2141
+Jun 18 11:50:49.459: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename deployment
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in deployment-9807
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should not conflict [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: Cleaning up the secret
-STEP: Cleaning up the configmap
-STEP: Cleaning up the pod
-[AfterEach] [sig-storage] EmptyDir wrapper volumes
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 18:51:12.324: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "emptydir-wrapper-2141" for this suite.
-Jun  4 18:51:20.382: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 18:51:20.934: INFO: namespace emptydir-wrapper-2141 deletion completed in 8.592405479s
-
-• [SLOW TEST:14.987 seconds]
-[sig-storage] EmptyDir wrapper volumes
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:22
-  should not conflict [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-apps] Deployment
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:65
+[It] RecreateDeployment should delete old pods and create new ones [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+Jun 18 11:50:49.652: INFO: Creating deployment "test-recreate-deployment"
+Jun 18 11:50:49.661: INFO: Waiting deployment "test-recreate-deployment" to be updated to revision 1
+Jun 18 11:50:49.679: INFO: deployment "test-recreate-deployment" doesn't have the required revision set
+Jun 18 11:50:52.130: INFO: Waiting deployment "test-recreate-deployment" to complete
+Jun 18 11:50:52.138: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63696455449, loc:(*time.Location)(0x8a1a0e0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63696455449, loc:(*time.Location)(0x8a1a0e0)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63696455449, loc:(*time.Location)(0x8a1a0e0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63696455449, loc:(*time.Location)(0x8a1a0e0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-recreate-deployment-7d57d5ff7c\" is progressing."}}, CollisionCount:(*int32)(nil)}
+Jun 18 11:50:54.147: INFO: Triggering a new rollout for deployment "test-recreate-deployment"
+Jun 18 11:50:54.176: INFO: Updating deployment test-recreate-deployment
+Jun 18 11:50:54.176: INFO: Watching deployment "test-recreate-deployment" to verify that new pods will not run with olds pods
+[AfterEach] [sig-apps] Deployment
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:59
+Jun 18 11:50:54.325: INFO: Deployment "test-recreate-deployment":
+&Deployment{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-recreate-deployment,GenerateName:,Namespace:deployment-9807,SelfLink:/apis/apps/v1/namespaces/deployment-9807/deployments/test-recreate-deployment,UID:5113fec8-91bf-11e9-a08a-ee7a14707756,ResourceVersion:89279,Generation:2,CreationTimestamp:2019-06-18 11:50:49 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod-3,},Annotations:map[string]string{deployment.kubernetes.io/revision: 2,},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:DeploymentSpec{Replicas:*1,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod-3,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod-3,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] []  [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,},},Strategy:DeploymentStrategy{Type:Recreate,RollingUpdate:nil,},MinReadySeconds:0,RevisionHistoryLimit:*10,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:2,Replicas:1,UpdatedReplicas:1,AvailableReplicas:0,UnavailableReplicas:1,Conditions:[{Available False 2019-06-18 11:50:54 +0000 UTC 2019-06-18 11:50:54 +0000 UTC MinimumReplicasUnavailable Deployment does not have minimum availability.} {Progressing True 2019-06-18 11:50:54 +0000 UTC 2019-06-18 11:50:49 +0000 UTC ReplicaSetUpdated ReplicaSet "test-recreate-deployment-c9cbd8684" is progressing.}],ReadyReplicas:0,CollisionCount:nil,},}
+
+Jun 18 11:50:54.337: INFO: New ReplicaSet "test-recreate-deployment-c9cbd8684" of Deployment "test-recreate-deployment":
+&ReplicaSet{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-recreate-deployment-c9cbd8684,GenerateName:,Namespace:deployment-9807,SelfLink:/apis/apps/v1/namespaces/deployment-9807/replicasets/test-recreate-deployment-c9cbd8684,UID:53d3632e-91bf-11e9-a08a-ee7a14707756,ResourceVersion:89278,Generation:1,CreationTimestamp:2019-06-18 11:50:54 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod-3,pod-template-hash: c9cbd8684,},Annotations:map[string]string{deployment.kubernetes.io/desired-replicas: 1,deployment.kubernetes.io/max-replicas: 1,deployment.kubernetes.io/revision: 2,},OwnerReferences:[{apps/v1 Deployment test-recreate-deployment 5113fec8-91bf-11e9-a08a-ee7a14707756 0xc0018a8760 0xc0018a8761}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:ReplicaSetSpec{Replicas:*1,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod-3,pod-template-hash: c9cbd8684,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod-3,pod-template-hash: c9cbd8684,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] []  [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,},},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:1,FullyLabeledReplicas:1,ObservedGeneration:1,ReadyReplicas:0,AvailableReplicas:0,Conditions:[],},}
+Jun 18 11:50:54.337: INFO: All old ReplicaSets of Deployment "test-recreate-deployment":
+Jun 18 11:50:54.337: INFO: &ReplicaSet{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-recreate-deployment-7d57d5ff7c,GenerateName:,Namespace:deployment-9807,SelfLink:/apis/apps/v1/namespaces/deployment-9807/replicasets/test-recreate-deployment-7d57d5ff7c,UID:51151341-91bf-11e9-a08a-ee7a14707756,ResourceVersion:89269,Generation:2,CreationTimestamp:2019-06-18 11:50:49 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod-3,pod-template-hash: 7d57d5ff7c,},Annotations:map[string]string{deployment.kubernetes.io/desired-replicas: 1,deployment.kubernetes.io/max-replicas: 1,deployment.kubernetes.io/revision: 1,},OwnerReferences:[{apps/v1 Deployment test-recreate-deployment 5113fec8-91bf-11e9-a08a-ee7a14707756 0xc0018a8697 0xc0018a8698}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:ReplicaSetSpec{Replicas:*0,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod-3,pod-template-hash: 7d57d5ff7c,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod-3,pod-template-hash: 7d57d5ff7c,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[],Containers:[{redis gcr.io/kubernetes-e2e-test-images/redis:1.0 [] []  [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,},},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:0,FullyLabeledReplicas:0,ObservedGeneration:2,ReadyReplicas:0,AvailableReplicas:0,Conditions:[],},}
+Jun 18 11:50:54.349: INFO: Pod "test-recreate-deployment-c9cbd8684-c5xnj" is not available:
+&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-recreate-deployment-c9cbd8684-c5xnj,GenerateName:test-recreate-deployment-c9cbd8684-,Namespace:deployment-9807,SelfLink:/api/v1/namespaces/deployment-9807/pods/test-recreate-deployment-c9cbd8684-c5xnj,UID:53d54212-91bf-11e9-a08a-ee7a14707756,ResourceVersion:89281,Generation:0,CreationTimestamp:2019-06-18 11:50:54 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod-3,pod-template-hash: c9cbd8684,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet test-recreate-deployment-c9cbd8684 53d3632e-91bf-11e9-a08a-ee7a14707756 0xc0018a8fd0 0xc0018a8fd1}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-mrbd6 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-mrbd6,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] []  [] [] [] {map[] map[]} [{default-token-mrbd6 true /var/run/secrets/kubernetes.io/serviceaccount   }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.72.74.184,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists  NoExecute 0xc0018a9040} {node.kubernetes.io/unreachable Exists  NoExecute 0xc0018a9060}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:50:54 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:50:54 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:50:54 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:50:54 +0000 UTC  }],Message:,Reason:,HostIP:10.72.74.184,PodIP:,StartTime:2019-06-18 11:50:54 +0000 UTC,ContainerStatuses:[{nginx {ContainerStateWaiting{Reason:ContainerCreating,Message:,} nil nil} {nil nil nil} false 0 docker.io/library/nginx:1.14-alpine  }],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},}
+[AfterEach] [sig-apps] Deployment
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 11:50:54.349: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "deployment-9807" for this suite.
+Jun 18 11:51:02.395: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 11:51:03.217: INFO: namespace deployment-9807 deletion completed in 8.847741367s
+
+• [SLOW TEST:13.758 seconds]
+[sig-apps] Deployment
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22
+  RecreateDeployment should delete old pods and create new ones [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+SSSSS
 ------------------------------
-[sig-apps] Daemon set [Serial] 
-  should update pod when spec was updated and update strategy is RollingUpdate [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [sig-apps] Daemon set [Serial]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+[sig-storage] Subpath Atomic writer volumes 
+  should support subpaths with projected pod [LinuxOnly] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-storage] Subpath
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 18:51:20.935: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-STEP: Building a namespace api object, basename daemonsets
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in daemonsets-5749
+Jun 18 11:51:03.218: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename subpath
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in subpath-8784
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-apps] Daemon set [Serial]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:102
-[It] should update pod when spec was updated and update strategy is RollingUpdate [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-Jun  4 18:51:21.243: INFO: Creating simple daemon set daemon-set
-STEP: Check that daemon pods launch on every node of the cluster.
-Jun  4 18:51:21.277: INFO: Number of nodes with available pods: 0
-Jun  4 18:51:21.277: INFO: Node 10.212.23.161 is running more than one daemon pod
-Jun  4 18:51:22.307: INFO: Number of nodes with available pods: 0
-Jun  4 18:51:22.307: INFO: Node 10.212.23.161 is running more than one daemon pod
-Jun  4 18:51:23.367: INFO: Number of nodes with available pods: 3
-Jun  4 18:51:23.367: INFO: Number of running nodes: 3, number of available pods: 3
-STEP: Update daemon pods image.
-STEP: Check that daemon pods images are updated.
-Jun  4 18:51:23.449: INFO: Wrong image for pod: daemon-set-7rv74. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine.
-Jun  4 18:51:23.449: INFO: Wrong image for pod: daemon-set-9gz45. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine.
-Jun  4 18:51:23.449: INFO: Wrong image for pod: daemon-set-mc4bl. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine.
-Jun  4 18:51:24.473: INFO: Wrong image for pod: daemon-set-7rv74. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine.
-Jun  4 18:51:24.473: INFO: Wrong image for pod: daemon-set-9gz45. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine.
-Jun  4 18:51:24.473: INFO: Wrong image for pod: daemon-set-mc4bl. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine.
-Jun  4 18:51:25.479: INFO: Wrong image for pod: daemon-set-7rv74. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine.
-Jun  4 18:51:25.479: INFO: Wrong image for pod: daemon-set-9gz45. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine.
-Jun  4 18:51:25.479: INFO: Wrong image for pod: daemon-set-mc4bl. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine.
-Jun  4 18:51:26.473: INFO: Wrong image for pod: daemon-set-7rv74. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine.
-Jun  4 18:51:26.473: INFO: Pod daemon-set-7rv74 is not available
-Jun  4 18:51:26.473: INFO: Wrong image for pod: daemon-set-9gz45. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine.
-Jun  4 18:51:26.473: INFO: Wrong image for pod: daemon-set-mc4bl. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine.
-Jun  4 18:51:27.473: INFO: Wrong image for pod: daemon-set-7rv74. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine.
-Jun  4 18:51:27.473: INFO: Pod daemon-set-7rv74 is not available
-Jun  4 18:51:27.473: INFO: Wrong image for pod: daemon-set-9gz45. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine.
-Jun  4 18:51:27.473: INFO: Wrong image for pod: daemon-set-mc4bl. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine.
-Jun  4 18:51:28.474: INFO: Wrong image for pod: daemon-set-7rv74. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine.
-Jun  4 18:51:28.474: INFO: Pod daemon-set-7rv74 is not available
-Jun  4 18:51:28.474: INFO: Wrong image for pod: daemon-set-9gz45. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine.
-Jun  4 18:51:28.474: INFO: Wrong image for pod: daemon-set-mc4bl. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine.
-Jun  4 18:51:29.474: INFO: Wrong image for pod: daemon-set-7rv74. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine.
-Jun  4 18:51:29.474: INFO: Pod daemon-set-7rv74 is not available
-Jun  4 18:51:29.474: INFO: Wrong image for pod: daemon-set-9gz45. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine.
-Jun  4 18:51:29.474: INFO: Wrong image for pod: daemon-set-mc4bl. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine.
-Jun  4 18:51:30.473: INFO: Wrong image for pod: daemon-set-7rv74. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine.
-Jun  4 18:51:30.473: INFO: Pod daemon-set-7rv74 is not available
-Jun  4 18:51:30.473: INFO: Wrong image for pod: daemon-set-9gz45. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine.
-Jun  4 18:51:30.473: INFO: Wrong image for pod: daemon-set-mc4bl. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine.
-Jun  4 18:51:31.479: INFO: Wrong image for pod: daemon-set-7rv74. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine.
-Jun  4 18:51:31.479: INFO: Pod daemon-set-7rv74 is not available
-Jun  4 18:51:31.479: INFO: Wrong image for pod: daemon-set-9gz45. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine.
-Jun  4 18:51:31.479: INFO: Wrong image for pod: daemon-set-mc4bl. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine.
-Jun  4 18:51:32.502: INFO: Wrong image for pod: daemon-set-7rv74. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine.
-Jun  4 18:51:32.502: INFO: Pod daemon-set-7rv74 is not available
-Jun  4 18:51:32.502: INFO: Wrong image for pod: daemon-set-9gz45. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine.
-Jun  4 18:51:32.502: INFO: Wrong image for pod: daemon-set-mc4bl. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine.
-Jun  4 18:51:33.473: INFO: Wrong image for pod: daemon-set-7rv74. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine.
-Jun  4 18:51:33.473: INFO: Pod daemon-set-7rv74 is not available
-Jun  4 18:51:33.473: INFO: Wrong image for pod: daemon-set-9gz45. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine.
-Jun  4 18:51:33.473: INFO: Wrong image for pod: daemon-set-mc4bl. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine.
-Jun  4 18:51:34.475: INFO: Wrong image for pod: daemon-set-7rv74. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine.
-Jun  4 18:51:34.475: INFO: Pod daemon-set-7rv74 is not available
-Jun  4 18:51:34.475: INFO: Wrong image for pod: daemon-set-9gz45. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine.
-Jun  4 18:51:34.475: INFO: Wrong image for pod: daemon-set-mc4bl. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine.
-Jun  4 18:51:35.473: INFO: Wrong image for pod: daemon-set-7rv74. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine.
-Jun  4 18:51:35.473: INFO: Pod daemon-set-7rv74 is not available
-Jun  4 18:51:35.473: INFO: Wrong image for pod: daemon-set-9gz45. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine.
-Jun  4 18:51:35.473: INFO: Wrong image for pod: daemon-set-mc4bl. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine.
-Jun  4 18:51:36.556: INFO: Wrong image for pod: daemon-set-7rv74. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine.
-Jun  4 18:51:36.556: INFO: Pod daemon-set-7rv74 is not available
-Jun  4 18:51:36.556: INFO: Wrong image for pod: daemon-set-9gz45. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine.
-Jun  4 18:51:36.556: INFO: Wrong image for pod: daemon-set-mc4bl. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine.
-Jun  4 18:51:37.474: INFO: Pod daemon-set-29d6x is not available
-Jun  4 18:51:37.474: INFO: Wrong image for pod: daemon-set-9gz45. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine.
-Jun  4 18:51:37.474: INFO: Wrong image for pod: daemon-set-mc4bl. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine.
-Jun  4 18:51:38.473: INFO: Pod daemon-set-29d6x is not available
-Jun  4 18:51:38.474: INFO: Wrong image for pod: daemon-set-9gz45. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine.
-Jun  4 18:51:38.474: INFO: Wrong image for pod: daemon-set-mc4bl. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine.
-Jun  4 18:51:39.474: INFO: Pod daemon-set-29d6x is not available
-Jun  4 18:51:39.474: INFO: Wrong image for pod: daemon-set-9gz45. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine.
-Jun  4 18:51:39.474: INFO: Wrong image for pod: daemon-set-mc4bl. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine.
-Jun  4 18:51:40.474: INFO: Pod daemon-set-29d6x is not available
-Jun  4 18:51:40.474: INFO: Wrong image for pod: daemon-set-9gz45. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine.
-Jun  4 18:51:40.474: INFO: Wrong image for pod: daemon-set-mc4bl. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine.
-Jun  4 18:51:41.472: INFO: Pod daemon-set-29d6x is not available
-Jun  4 18:51:41.472: INFO: Wrong image for pod: daemon-set-9gz45. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine.
-Jun  4 18:51:41.472: INFO: Wrong image for pod: daemon-set-mc4bl. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine.
-Jun  4 18:51:42.474: INFO: Wrong image for pod: daemon-set-9gz45. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine.
-Jun  4 18:51:42.474: INFO: Pod daemon-set-9gz45 is not available
-Jun  4 18:51:42.474: INFO: Wrong image for pod: daemon-set-mc4bl. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine.
-Jun  4 18:51:43.473: INFO: Wrong image for pod: daemon-set-9gz45. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine.
-Jun  4 18:51:43.473: INFO: Pod daemon-set-9gz45 is not available
-Jun  4 18:51:43.473: INFO: Wrong image for pod: daemon-set-mc4bl. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine.
-Jun  4 18:51:44.474: INFO: Wrong image for pod: daemon-set-9gz45. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine.
-Jun  4 18:51:44.474: INFO: Pod daemon-set-9gz45 is not available
-Jun  4 18:51:44.474: INFO: Wrong image for pod: daemon-set-mc4bl. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine.
-Jun  4 18:51:45.473: INFO: Wrong image for pod: daemon-set-9gz45. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine.
-Jun  4 18:51:45.473: INFO: Pod daemon-set-9gz45 is not available
-Jun  4 18:51:45.473: INFO: Wrong image for pod: daemon-set-mc4bl. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine.
-Jun  4 18:51:46.474: INFO: Wrong image for pod: daemon-set-9gz45. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine.
-Jun  4 18:51:46.474: INFO: Pod daemon-set-9gz45 is not available
-Jun  4 18:51:46.474: INFO: Wrong image for pod: daemon-set-mc4bl. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine.
-Jun  4 18:51:47.474: INFO: Wrong image for pod: daemon-set-9gz45. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine.
-Jun  4 18:51:47.474: INFO: Pod daemon-set-9gz45 is not available
-Jun  4 18:51:47.474: INFO: Wrong image for pod: daemon-set-mc4bl. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine.
-Jun  4 18:51:48.475: INFO: Wrong image for pod: daemon-set-9gz45. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine.
-Jun  4 18:51:48.475: INFO: Pod daemon-set-9gz45 is not available
-Jun  4 18:51:48.475: INFO: Wrong image for pod: daemon-set-mc4bl. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine.
-Jun  4 18:51:49.473: INFO: Wrong image for pod: daemon-set-9gz45. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine.
-Jun  4 18:51:49.473: INFO: Pod daemon-set-9gz45 is not available
-Jun  4 18:51:49.473: INFO: Wrong image for pod: daemon-set-mc4bl. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine.
-Jun  4 18:51:50.473: INFO: Wrong image for pod: daemon-set-9gz45. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine.
-Jun  4 18:51:50.473: INFO: Pod daemon-set-9gz45 is not available
-Jun  4 18:51:50.473: INFO: Wrong image for pod: daemon-set-mc4bl. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine.
-Jun  4 18:51:51.474: INFO: Wrong image for pod: daemon-set-9gz45. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine.
-Jun  4 18:51:51.475: INFO: Pod daemon-set-9gz45 is not available
-Jun  4 18:51:51.475: INFO: Wrong image for pod: daemon-set-mc4bl. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine.
-Jun  4 18:51:52.475: INFO: Wrong image for pod: daemon-set-9gz45. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine.
-Jun  4 18:51:52.475: INFO: Pod daemon-set-9gz45 is not available
-Jun  4 18:51:52.475: INFO: Wrong image for pod: daemon-set-mc4bl. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine.
-Jun  4 18:51:53.474: INFO: Pod daemon-set-c9nmc is not available
-Jun  4 18:51:53.474: INFO: Wrong image for pod: daemon-set-mc4bl. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine.
-Jun  4 18:51:54.476: INFO: Wrong image for pod: daemon-set-mc4bl. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine.
-Jun  4 18:51:55.491: INFO: Wrong image for pod: daemon-set-mc4bl. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine.
-Jun  4 18:51:56.474: INFO: Wrong image for pod: daemon-set-mc4bl. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine.
-Jun  4 18:51:56.474: INFO: Pod daemon-set-mc4bl is not available
-Jun  4 18:51:57.473: INFO: Pod daemon-set-nrkkj is not available
-STEP: Check that daemon pods are still running on every node of the cluster.
-Jun  4 18:51:57.525: INFO: Number of nodes with available pods: 2
-Jun  4 18:51:57.525: INFO: Node 10.212.23.164 is running more than one daemon pod
-Jun  4 18:51:58.554: INFO: Number of nodes with available pods: 2
-Jun  4 18:51:58.554: INFO: Node 10.212.23.164 is running more than one daemon pod
-Jun  4 18:51:59.567: INFO: Number of nodes with available pods: 2
-Jun  4 18:51:59.567: INFO: Node 10.212.23.164 is running more than one daemon pod
-Jun  4 18:52:00.552: INFO: Number of nodes with available pods: 2
-Jun  4 18:52:00.552: INFO: Node 10.212.23.164 is running more than one daemon pod
-Jun  4 18:52:01.551: INFO: Number of nodes with available pods: 3
-Jun  4 18:52:01.551: INFO: Number of running nodes: 3, number of available pods: 3
-[AfterEach] [sig-apps] Daemon set [Serial]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:68
-STEP: Deleting DaemonSet "daemon-set"
-STEP: deleting DaemonSet.extensions daemon-set in namespace daemonsets-5749, will wait for the garbage collector to delete the pods
-Jun  4 18:52:01.694: INFO: Deleting DaemonSet.extensions daemon-set took: 27.893988ms
-Jun  4 18:52:01.794: INFO: Terminating DaemonSet.extensions daemon-set pods took: 100.290367ms
-Jun  4 18:52:12.653: INFO: Number of nodes with available pods: 0
-Jun  4 18:52:12.653: INFO: Number of running nodes: 0, number of available pods: 0
-Jun  4 18:52:12.664: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"selfLink":"/apis/apps/v1/namespaces/daemonsets-5749/daemonsets","resourceVersion":"13386"},"items":null}
-
-Jun  4 18:52:12.675: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"selfLink":"/api/v1/namespaces/daemonsets-5749/pods","resourceVersion":"13386"},"items":null}
-
-[AfterEach] [sig-apps] Daemon set [Serial]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 18:52:12.727: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "daemonsets-5749" for this suite.
-Jun  4 18:52:20.784: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 18:52:21.171: INFO: namespace daemonsets-5749 deletion completed in 8.431470046s
+[BeforeEach] Atomic writer volumes
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:38
+STEP: Setting up data
+[It] should support subpaths with projected pod [LinuxOnly] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating pod pod-subpath-test-projected-8xrd
+STEP: Creating a pod to test atomic-volume-subpath
+Jun 18 11:51:03.496: INFO: Waiting up to 5m0s for pod "pod-subpath-test-projected-8xrd" in namespace "subpath-8784" to be "success or failure"
+Jun 18 11:51:03.510: INFO: Pod "pod-subpath-test-projected-8xrd": Phase="Pending", Reason="", readiness=false. Elapsed: 13.323918ms
+Jun 18 11:51:05.523: INFO: Pod "pod-subpath-test-projected-8xrd": Phase="Pending", Reason="", readiness=false. Elapsed: 2.026801025s
+Jun 18 11:51:07.536: INFO: Pod "pod-subpath-test-projected-8xrd": Phase="Running", Reason="", readiness=true. Elapsed: 4.039667621s
+Jun 18 11:51:09.549: INFO: Pod "pod-subpath-test-projected-8xrd": Phase="Running", Reason="", readiness=true. Elapsed: 6.052819699s
+Jun 18 11:51:11.564: INFO: Pod "pod-subpath-test-projected-8xrd": Phase="Running", Reason="", readiness=true. Elapsed: 8.067378964s
+Jun 18 11:51:13.578: INFO: Pod "pod-subpath-test-projected-8xrd": Phase="Running", Reason="", readiness=true. Elapsed: 10.081669922s
+Jun 18 11:51:15.591: INFO: Pod "pod-subpath-test-projected-8xrd": Phase="Running", Reason="", readiness=true. Elapsed: 12.094880159s
+Jun 18 11:51:17.606: INFO: Pod "pod-subpath-test-projected-8xrd": Phase="Running", Reason="", readiness=true. Elapsed: 14.10958472s
+Jun 18 11:51:19.620: INFO: Pod "pod-subpath-test-projected-8xrd": Phase="Running", Reason="", readiness=true. Elapsed: 16.123659296s
+Jun 18 11:51:21.634: INFO: Pod "pod-subpath-test-projected-8xrd": Phase="Running", Reason="", readiness=true. Elapsed: 18.137366281s
+Jun 18 11:51:23.648: INFO: Pod "pod-subpath-test-projected-8xrd": Phase="Running", Reason="", readiness=true. Elapsed: 20.151728756s
+Jun 18 11:51:25.662: INFO: Pod "pod-subpath-test-projected-8xrd": Phase="Succeeded", Reason="", readiness=false. Elapsed: 22.165955037s
+STEP: Saw pod success
+Jun 18 11:51:25.662: INFO: Pod "pod-subpath-test-projected-8xrd" satisfied condition "success or failure"
+Jun 18 11:51:25.677: INFO: Trying to get logs from node 10.72.74.138 pod pod-subpath-test-projected-8xrd container test-container-subpath-projected-8xrd: 
+STEP: delete the pod
+Jun 18 11:51:25.773: INFO: Waiting for pod pod-subpath-test-projected-8xrd to disappear
+Jun 18 11:51:25.786: INFO: Pod pod-subpath-test-projected-8xrd no longer exists
+STEP: Deleting pod pod-subpath-test-projected-8xrd
+Jun 18 11:51:25.786: INFO: Deleting pod "pod-subpath-test-projected-8xrd" in namespace "subpath-8784"
+[AfterEach] [sig-storage] Subpath
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 11:51:25.799: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "subpath-8784" for this suite.
+Jun 18 11:51:31.844: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 11:51:32.240: INFO: namespace subpath-8784 deletion completed in 6.421793335s
 
-• [SLOW TEST:60.237 seconds]
-[sig-apps] Daemon set [Serial]
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22
-  should update pod when spec was updated and update strategy is RollingUpdate [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+• [SLOW TEST:29.022 seconds]
+[sig-storage] Subpath
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:22
+  Atomic writer volumes
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:34
+    should support subpaths with projected pod [LinuxOnly] [Conformance]
+    /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-SSSSSSSSSSSSSSSSS
+SSSS
 ------------------------------
-[sig-storage] Downward API volume 
-  should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [sig-storage] Downward API volume
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+[k8s.io] Probing container 
+  should be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [k8s.io] Probing container
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 18:52:21.172: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-STEP: Building a namespace api object, basename downward-api
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in downward-api-6373
+Jun 18 11:51:32.240: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename container-probe
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in container-probe-6349
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-storage] Downward API volume
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:39
-[It] should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: Creating a pod to test downward API volume plugin
-Jun  4 18:52:21.410: INFO: Waiting up to 5m0s for pod "downwardapi-volume-e257b36d-86f9-11e9-8d1b-467ee19922ac" in namespace "downward-api-6373" to be "success or failure"
-Jun  4 18:52:21.427: INFO: Pod "downwardapi-volume-e257b36d-86f9-11e9-8d1b-467ee19922ac": Phase="Pending", Reason="", readiness=false. Elapsed: 16.700599ms
-Jun  4 18:52:23.440: INFO: Pod "downwardapi-volume-e257b36d-86f9-11e9-8d1b-467ee19922ac": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.030217266s
-STEP: Saw pod success
-Jun  4 18:52:23.440: INFO: Pod "downwardapi-volume-e257b36d-86f9-11e9-8d1b-467ee19922ac" satisfied condition "success or failure"
-Jun  4 18:52:23.466: INFO: Trying to get logs from node 10.212.23.164 pod downwardapi-volume-e257b36d-86f9-11e9-8d1b-467ee19922ac container client-container: 
-STEP: delete the pod
-Jun  4 18:52:23.529: INFO: Waiting for pod downwardapi-volume-e257b36d-86f9-11e9-8d1b-467ee19922ac to disappear
-Jun  4 18:52:23.542: INFO: Pod downwardapi-volume-e257b36d-86f9-11e9-8d1b-467ee19922ac no longer exists
-[AfterEach] [sig-storage] Downward API volume
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 18:52:23.542: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "downward-api-6373" for this suite.
-Jun  4 18:52:29.601: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 18:52:30.180: INFO: namespace downward-api-6373 deletion completed in 6.621787879s
+[BeforeEach] [k8s.io] Probing container
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/container_probe.go:51
+[It] should be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating pod liveness-exec in namespace container-probe-6349
+Jun 18 11:51:34.488: INFO: Started pod liveness-exec in namespace container-probe-6349
+STEP: checking the pod's current state and verifying that restartCount is present
+Jun 18 11:51:34.501: INFO: Initial restart count of pod liveness-exec is 0
+Jun 18 11:52:23.112: INFO: Restart count of pod container-probe-6349/liveness-exec is now 1 (48.611453352s elapsed)
+STEP: deleting the pod
+[AfterEach] [k8s.io] Probing container
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 11:52:23.170: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "container-probe-6349" for this suite.
+Jun 18 11:52:29.212: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 11:52:29.618: INFO: namespace container-probe-6349 deletion completed in 6.432798436s
 
-• [SLOW TEST:9.008 seconds]
-[sig-storage] Downward API volume
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:34
-  should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+• [SLOW TEST:57.378 seconds]
+[k8s.io] Probing container
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
+  should be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
 [sig-storage] ConfigMap 
-  should be consumable from pods in volume with mappings as non-root [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+  should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 [BeforeEach] [sig-storage] ConfigMap
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 18:52:30.181: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
+Jun 18 11:52:29.619: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
 STEP: Building a namespace api object, basename configmap
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in configmap-8305
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in configmap-5568
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should be consumable from pods in volume with mappings as non-root [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: Creating configMap with name configmap-test-volume-map-e7b6d967-86f9-11e9-8d1b-467ee19922ac
+[It] should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating configMap with name configmap-test-volume-8cc97e4e-91bf-11e9-a25d-8608290c688a
 STEP: Creating a pod to test consume configMaps
-Jun  4 18:52:30.436: INFO: Waiting up to 5m0s for pod "pod-configmaps-e7b8ee01-86f9-11e9-8d1b-467ee19922ac" in namespace "configmap-8305" to be "success or failure"
-Jun  4 18:52:30.447: INFO: Pod "pod-configmaps-e7b8ee01-86f9-11e9-8d1b-467ee19922ac": Phase="Pending", Reason="", readiness=false. Elapsed: 11.01121ms
-Jun  4 18:52:32.459: INFO: Pod "pod-configmaps-e7b8ee01-86f9-11e9-8d1b-467ee19922ac": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.023298349s
+Jun 18 11:52:29.872: INFO: Waiting up to 5m0s for pod "pod-configmaps-8ccbe2f5-91bf-11e9-a25d-8608290c688a" in namespace "configmap-5568" to be "success or failure"
+Jun 18 11:52:29.907: INFO: Pod "pod-configmaps-8ccbe2f5-91bf-11e9-a25d-8608290c688a": Phase="Pending", Reason="", readiness=false. Elapsed: 34.588861ms
+Jun 18 11:52:31.921: INFO: Pod "pod-configmaps-8ccbe2f5-91bf-11e9-a25d-8608290c688a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.049077325s
 STEP: Saw pod success
-Jun  4 18:52:32.459: INFO: Pod "pod-configmaps-e7b8ee01-86f9-11e9-8d1b-467ee19922ac" satisfied condition "success or failure"
-Jun  4 18:52:32.469: INFO: Trying to get logs from node 10.212.23.189 pod pod-configmaps-e7b8ee01-86f9-11e9-8d1b-467ee19922ac container configmap-volume-test: 
+Jun 18 11:52:31.922: INFO: Pod "pod-configmaps-8ccbe2f5-91bf-11e9-a25d-8608290c688a" satisfied condition "success or failure"
+Jun 18 11:52:31.937: INFO: Trying to get logs from node 10.72.74.138 pod pod-configmaps-8ccbe2f5-91bf-11e9-a25d-8608290c688a container configmap-volume-test: 
 STEP: delete the pod
-Jun  4 18:52:32.594: INFO: Waiting for pod pod-configmaps-e7b8ee01-86f9-11e9-8d1b-467ee19922ac to disappear
-Jun  4 18:52:32.606: INFO: Pod pod-configmaps-e7b8ee01-86f9-11e9-8d1b-467ee19922ac no longer exists
+Jun 18 11:52:32.020: INFO: Waiting for pod pod-configmaps-8ccbe2f5-91bf-11e9-a25d-8608290c688a to disappear
+Jun 18 11:52:32.032: INFO: Pod pod-configmaps-8ccbe2f5-91bf-11e9-a25d-8608290c688a no longer exists
 [AfterEach] [sig-storage] ConfigMap
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 18:52:32.606: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "configmap-8305" for this suite.
-Jun  4 18:52:38.669: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 18:52:39.070: INFO: namespace configmap-8305 deletion completed in 6.446659154s
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 11:52:32.033: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "configmap-5568" for this suite.
+Jun 18 11:52:38.079: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 11:52:38.477: INFO: namespace configmap-5568 deletion completed in 6.430084727s
 
-• [SLOW TEST:8.889 seconds]
+• [SLOW TEST:8.858 seconds]
 [sig-storage] ConfigMap
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:32
-  should be consumable from pods in volume with mappings as non-root [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-------------------------------
-SSSSSSS
-------------------------------
-[sig-api-machinery] Watchers 
-  should observe add, update, and delete watch notifications on configmaps [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [sig-api-machinery] Watchers
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
-STEP: Creating a kubernetes client
-Jun  4 18:52:39.071: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-STEP: Building a namespace api object, basename watch
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in watch-5554
-STEP: Waiting for a default service account to be provisioned in namespace
-[It] should observe add, update, and delete watch notifications on configmaps [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: creating a watch on configmaps with label A
-STEP: creating a watch on configmaps with label B
-STEP: creating a watch on configmaps with label A or B
-STEP: creating a configmap with label A and ensuring the correct watchers observe the notification
-Jun  4 18:52:39.321: INFO: Got : ADDED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-configmap-a,GenerateName:,Namespace:watch-5554,SelfLink:/api/v1/namespaces/watch-5554/configmaps/e2e-watch-test-configmap-a,UID:ed067ee9-86f9-11e9-8318-1e5386706511,ResourceVersion:13563,Generation:0,CreationTimestamp:2019-06-04 18:52:39 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: multiple-watchers-A,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{},BinaryData:map[string][]byte{},}
-Jun  4 18:52:39.321: INFO: Got : ADDED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-configmap-a,GenerateName:,Namespace:watch-5554,SelfLink:/api/v1/namespaces/watch-5554/configmaps/e2e-watch-test-configmap-a,UID:ed067ee9-86f9-11e9-8318-1e5386706511,ResourceVersion:13563,Generation:0,CreationTimestamp:2019-06-04 18:52:39 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: multiple-watchers-A,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{},BinaryData:map[string][]byte{},}
-STEP: modifying configmap A and ensuring the correct watchers observe the notification
-Jun  4 18:52:49.346: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-configmap-a,GenerateName:,Namespace:watch-5554,SelfLink:/api/v1/namespaces/watch-5554/configmaps/e2e-watch-test-configmap-a,UID:ed067ee9-86f9-11e9-8318-1e5386706511,ResourceVersion:13580,Generation:0,CreationTimestamp:2019-06-04 18:52:39 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: multiple-watchers-A,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{mutation: 1,},BinaryData:map[string][]byte{},}
-Jun  4 18:52:49.346: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-configmap-a,GenerateName:,Namespace:watch-5554,SelfLink:/api/v1/namespaces/watch-5554/configmaps/e2e-watch-test-configmap-a,UID:ed067ee9-86f9-11e9-8318-1e5386706511,ResourceVersion:13580,Generation:0,CreationTimestamp:2019-06-04 18:52:39 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: multiple-watchers-A,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{mutation: 1,},BinaryData:map[string][]byte{},}
-STEP: modifying configmap A again and ensuring the correct watchers observe the notification
-Jun  4 18:52:59.455: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-configmap-a,GenerateName:,Namespace:watch-5554,SelfLink:/api/v1/namespaces/watch-5554/configmaps/e2e-watch-test-configmap-a,UID:ed067ee9-86f9-11e9-8318-1e5386706511,ResourceVersion:13597,Generation:0,CreationTimestamp:2019-06-04 18:52:39 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: multiple-watchers-A,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},}
-Jun  4 18:52:59.455: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-configmap-a,GenerateName:,Namespace:watch-5554,SelfLink:/api/v1/namespaces/watch-5554/configmaps/e2e-watch-test-configmap-a,UID:ed067ee9-86f9-11e9-8318-1e5386706511,ResourceVersion:13597,Generation:0,CreationTimestamp:2019-06-04 18:52:39 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: multiple-watchers-A,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},}
-STEP: deleting configmap A and ensuring the correct watchers observe the notification
-Jun  4 18:53:09.555: INFO: Got : DELETED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-configmap-a,GenerateName:,Namespace:watch-5554,SelfLink:/api/v1/namespaces/watch-5554/configmaps/e2e-watch-test-configmap-a,UID:ed067ee9-86f9-11e9-8318-1e5386706511,ResourceVersion:13615,Generation:0,CreationTimestamp:2019-06-04 18:52:39 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: multiple-watchers-A,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},}
-Jun  4 18:53:09.555: INFO: Got : DELETED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-configmap-a,GenerateName:,Namespace:watch-5554,SelfLink:/api/v1/namespaces/watch-5554/configmaps/e2e-watch-test-configmap-a,UID:ed067ee9-86f9-11e9-8318-1e5386706511,ResourceVersion:13615,Generation:0,CreationTimestamp:2019-06-04 18:52:39 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: multiple-watchers-A,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},}
-STEP: creating a configmap with label B and ensuring the correct watchers observe the notification
-Jun  4 18:53:19.578: INFO: Got : ADDED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-configmap-b,GenerateName:,Namespace:watch-5554,SelfLink:/api/v1/namespaces/watch-5554/configmaps/e2e-watch-test-configmap-b,UID:0503d374-86fa-11e9-8318-1e5386706511,ResourceVersion:13633,Generation:0,CreationTimestamp:2019-06-04 18:53:19 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: multiple-watchers-B,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{},BinaryData:map[string][]byte{},}
-Jun  4 18:53:19.578: INFO: Got : ADDED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-configmap-b,GenerateName:,Namespace:watch-5554,SelfLink:/api/v1/namespaces/watch-5554/configmaps/e2e-watch-test-configmap-b,UID:0503d374-86fa-11e9-8318-1e5386706511,ResourceVersion:13633,Generation:0,CreationTimestamp:2019-06-04 18:53:19 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: multiple-watchers-B,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{},BinaryData:map[string][]byte{},}
-STEP: deleting configmap B and ensuring the correct watchers observe the notification
-Jun  4 18:53:29.603: INFO: Got : DELETED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-configmap-b,GenerateName:,Namespace:watch-5554,SelfLink:/api/v1/namespaces/watch-5554/configmaps/e2e-watch-test-configmap-b,UID:0503d374-86fa-11e9-8318-1e5386706511,ResourceVersion:13651,Generation:0,CreationTimestamp:2019-06-04 18:53:19 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: multiple-watchers-B,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{},BinaryData:map[string][]byte{},}
-Jun  4 18:53:29.603: INFO: Got : DELETED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-configmap-b,GenerateName:,Namespace:watch-5554,SelfLink:/api/v1/namespaces/watch-5554/configmaps/e2e-watch-test-configmap-b,UID:0503d374-86fa-11e9-8318-1e5386706511,ResourceVersion:13651,Generation:0,CreationTimestamp:2019-06-04 18:53:19 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: multiple-watchers-B,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{},BinaryData:map[string][]byte{},}
-[AfterEach] [sig-api-machinery] Watchers
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 18:53:39.603: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "watch-5554" for this suite.
-Jun  4 18:53:45.662: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 18:53:46.064: INFO: namespace watch-5554 deletion completed in 6.444278715s
-
-• [SLOW TEST:66.994 seconds]
-[sig-api-machinery] Watchers
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:22
-  should observe add, update, and delete watch notifications on configmaps [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:32
+  should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-SSSSSSSSSSSSSSSSSS
+SSS
 ------------------------------
-[sig-scheduling] SchedulerPredicates [Serial] 
-  validates that NodeSelector is respected if not matching  [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [sig-scheduling] SchedulerPredicates [Serial]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+[sig-storage] EmptyDir volumes 
+  should support (root,0644,default) [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-storage] EmptyDir volumes
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 18:53:46.066: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-STEP: Building a namespace api object, basename sched-pred
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in sched-pred-8190
+Jun 18 11:52:38.478: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename emptydir
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in emptydir-5101
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-scheduling] SchedulerPredicates [Serial]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:79
-Jun  4 18:53:46.288: INFO: Waiting up to 1m0s for all (but 0) nodes to be ready
-Jun  4 18:53:46.314: INFO: Waiting for terminating namespaces to be deleted...
-Jun  4 18:53:46.325: INFO: 
-Logging pods the kubelet thinks is on node 10.212.23.161 before test
-Jun  4 18:53:46.360: INFO: coredns-autoscaler-5846b97547-t7mrd from kube-system started at 2019-06-04 17:44:57 +0000 UTC (1 container statuses recorded)
-Jun  4 18:53:46.360: INFO: 	Container autoscaler ready: true, restart count 0
-Jun  4 18:53:46.360: INFO: kubernetes-dashboard-b4f5d97fc-rrmfx from kube-system started at 2019-06-04 17:44:57 +0000 UTC (1 container statuses recorded)
-Jun  4 18:53:46.360: INFO: 	Container kubernetes-dashboard ready: true, restart count 0
-Jun  4 18:53:46.360: INFO: ibm-storage-watcher-f74547b86-jhnt2 from kube-system started at 2019-06-04 17:44:57 +0000 UTC (1 container statuses recorded)
-Jun  4 18:53:46.360: INFO: 	Container ibm-storage-watcher-container ready: true, restart count 0
-Jun  4 18:53:46.360: INFO: ibm-kube-fluentd-cj2l2 from kube-system started at 2019-06-04 17:45:04 +0000 UTC (1 container statuses recorded)
-Jun  4 18:53:46.360: INFO: 	Container fluentd ready: true, restart count 0
-Jun  4 18:53:46.360: INFO: calico-node-wtt96 from kube-system started at 2019-06-04 17:44:38 +0000 UTC (1 container statuses recorded)
-Jun  4 18:53:46.360: INFO: 	Container calico-node ready: true, restart count 0
-Jun  4 18:53:46.360: INFO: coredns-cc5d484bb-9nhk2 from kube-system started at 2019-06-04 17:44:57 +0000 UTC (1 container statuses recorded)
-Jun  4 18:53:46.360: INFO: 	Container coredns ready: true, restart count 0
-Jun  4 18:53:46.360: INFO: ibm-master-proxy-static-10.212.23.161 from kube-system started at  (0 container statuses recorded)
-Jun  4 18:53:46.360: INFO: ibm-file-plugin-69d4bc9c8f-jwhqt from kube-system started at 2019-06-04 17:44:57 +0000 UTC (1 container statuses recorded)
-Jun  4 18:53:46.360: INFO: 	Container ibm-file-plugin-container ready: true, restart count 0
-Jun  4 18:53:46.360: INFO: ibm-keepalived-watcher-vl7mx from kube-system started at 2019-06-04 17:44:38 +0000 UTC (1 container statuses recorded)
-Jun  4 18:53:46.360: INFO: 	Container keepalived-watcher ready: true, restart count 0
-Jun  4 18:53:46.360: INFO: calico-kube-controllers-64ccf5d5d4-k2qnd from kube-system started at 2019-06-04 17:44:57 +0000 UTC (1 container statuses recorded)
-Jun  4 18:53:46.360: INFO: 	Container calico-kube-controllers ready: true, restart count 0
-Jun  4 18:53:46.360: INFO: public-cra6d87a4394f440e3becc50f237563475-alb1-79855b4ffb-mqcml from kube-system started at 2019-06-04 17:49:42 +0000 UTC (4 container statuses recorded)
-Jun  4 18:53:46.360: INFO: 	Container ingress-auth-1 ready: true, restart count 0
-Jun  4 18:53:46.360: INFO: 	Container ingress-auth-2 ready: true, restart count 0
-Jun  4 18:53:46.360: INFO: 	Container ingress-auth-3 ready: true, restart count 0
-Jun  4 18:53:46.360: INFO: 	Container nginx-ingress ready: true, restart count 0
-Jun  4 18:53:46.360: INFO: sonobuoy-systemd-logs-daemon-set-d63a0136eef44275-l8n68 from heptio-sonobuoy started at 2019-06-04 18:27:33 +0000 UTC (2 container statuses recorded)
-Jun  4 18:53:46.360: INFO: 	Container sonobuoy-worker ready: true, restart count 0
-Jun  4 18:53:46.360: INFO: 	Container systemd-logs ready: true, restart count 0
-Jun  4 18:53:46.360: INFO: vpn-7bbfb5bc49-4mjxm from kube-system started at 2019-06-04 17:44:57 +0000 UTC (1 container statuses recorded)
-Jun  4 18:53:46.360: INFO: 	Container vpn ready: true, restart count 0
-Jun  4 18:53:46.360: INFO: 
-Logging pods the kubelet thinks is on node 10.212.23.164 before test
-Jun  4 18:53:46.398: INFO: ibm-master-proxy-static-10.212.23.164 from kube-system started at  (0 container statuses recorded)
-Jun  4 18:53:46.398: INFO: ibm-keepalived-watcher-lfldx from kube-system started at 2019-06-04 17:44:49 +0000 UTC (1 container statuses recorded)
-Jun  4 18:53:46.398: INFO: 	Container keepalived-watcher ready: true, restart count 0
-Jun  4 18:53:46.398: INFO: calico-node-sphpm from kube-system started at 2019-06-04 17:44:49 +0000 UTC (1 container statuses recorded)
-Jun  4 18:53:46.398: INFO: 	Container calico-node ready: true, restart count 0
-Jun  4 18:53:46.398: INFO: metrics-server-66cfcd4b8d-xg6wk from kube-system started at 2019-06-04 17:45:26 +0000 UTC (2 container statuses recorded)
-Jun  4 18:53:46.398: INFO: 	Container metrics-server ready: true, restart count 0
-Jun  4 18:53:46.398: INFO: 	Container metrics-server-nanny ready: true, restart count 0
-Jun  4 18:53:46.398: INFO: ibm-cloud-provider-ip-162-133-73-165-ffb96fbff-7g5nv from ibm-system started at 2019-06-04 17:47:12 +0000 UTC (1 container statuses recorded)
-Jun  4 18:53:46.398: INFO: 	Container ibm-cloud-provider-ip-162-133-73-165 ready: true, restart count 0
-Jun  4 18:53:46.398: INFO: test-k8s-e2e-pvg-master-verification from default started at 2019-06-04 18:27:15 +0000 UTC (1 container statuses recorded)
-Jun  4 18:53:46.398: INFO: 	Container test-k8s-e2e-pvg-master-verification ready: true, restart count 0
-Jun  4 18:53:46.398: INFO: sonobuoy-e2e-job-39546bc4ddc3410d from heptio-sonobuoy started at 2019-06-04 18:27:33 +0000 UTC (2 container statuses recorded)
-Jun  4 18:53:46.398: INFO: 	Container e2e ready: true, restart count 0
-Jun  4 18:53:46.398: INFO: 	Container sonobuoy-worker ready: true, restart count 0
-Jun  4 18:53:46.398: INFO: sonobuoy-systemd-logs-daemon-set-d63a0136eef44275-b6j5c from heptio-sonobuoy started at 2019-06-04 18:27:33 +0000 UTC (2 container statuses recorded)
-Jun  4 18:53:46.399: INFO: 	Container sonobuoy-worker ready: true, restart count 0
-Jun  4 18:53:46.399: INFO: 	Container systemd-logs ready: true, restart count 0
-Jun  4 18:53:46.399: INFO: ibm-kube-fluentd-nr9cm from kube-system started at 2019-06-04 17:45:04 +0000 UTC (1 container statuses recorded)
-Jun  4 18:53:46.399: INFO: 	Container fluentd ready: true, restart count 0
-Jun  4 18:53:46.399: INFO: sonobuoy from heptio-sonobuoy started at 2019-06-04 18:27:24 +0000 UTC (1 container statuses recorded)
-Jun  4 18:53:46.399: INFO: 	Container kube-sonobuoy ready: true, restart count 0
-Jun  4 18:53:46.399: INFO: 
-Logging pods the kubelet thinks is on node 10.212.23.189 before test
-Jun  4 18:53:46.431: INFO: ibm-keepalived-watcher-dqk2k from kube-system started at 2019-06-04 17:45:04 +0000 UTC (1 container statuses recorded)
-Jun  4 18:53:46.431: INFO: 	Container keepalived-watcher ready: true, restart count 0
-Jun  4 18:53:46.431: INFO: coredns-cc5d484bb-nlgcf from kube-system started at 2019-06-04 17:45:30 +0000 UTC (1 container statuses recorded)
-Jun  4 18:53:46.431: INFO: 	Container coredns ready: true, restart count 0
-Jun  4 18:53:46.431: INFO: calico-node-v6hns from kube-system started at 2019-06-04 17:45:04 +0000 UTC (1 container statuses recorded)
-Jun  4 18:53:46.431: INFO: 	Container calico-node ready: true, restart count 0
-Jun  4 18:53:46.431: INFO: sonobuoy-systemd-logs-daemon-set-d63a0136eef44275-hnnhz from heptio-sonobuoy started at 2019-06-04 18:27:33 +0000 UTC (2 container statuses recorded)
-Jun  4 18:53:46.431: INFO: 	Container sonobuoy-worker ready: true, restart count 0
-Jun  4 18:53:46.431: INFO: 	Container systemd-logs ready: true, restart count 0
-Jun  4 18:53:46.431: INFO: ibm-master-proxy-static-10.212.23.189 from kube-system started at  (0 container statuses recorded)
-Jun  4 18:53:46.431: INFO: ibm-kube-fluentd-k56t7 from kube-system started at 2019-06-04 17:45:04 +0000 UTC (1 container statuses recorded)
-Jun  4 18:53:46.431: INFO: 	Container fluentd ready: true, restart count 0
-Jun  4 18:53:46.431: INFO: ibm-cloud-provider-ip-162-133-73-165-ffb96fbff-jf4l4 from ibm-system started at 2019-06-04 17:47:12 +0000 UTC (1 container statuses recorded)
-Jun  4 18:53:46.431: INFO: 	Container ibm-cloud-provider-ip-162-133-73-165 ready: true, restart count 0
-Jun  4 18:53:46.431: INFO: public-cra6d87a4394f440e3becc50f237563475-alb1-79855b4ffb-89n6x from kube-system started at 2019-06-04 17:49:42 +0000 UTC (4 container statuses recorded)
-Jun  4 18:53:46.431: INFO: 	Container ingress-auth-1 ready: true, restart count 0
-Jun  4 18:53:46.431: INFO: 	Container ingress-auth-2 ready: true, restart count 0
-Jun  4 18:53:46.431: INFO: 	Container ingress-auth-3 ready: true, restart count 0
-Jun  4 18:53:46.432: INFO: 	Container nginx-ingress ready: true, restart count 0
-[It] validates that NodeSelector is respected if not matching  [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: Trying to schedule Pod with nonempty NodeSelector.
-STEP: Considering event: 
-Type = [Warning], Name = [restricted-pod.15a513a09bfc6969], Reason = [FailedScheduling], Message = [0/3 nodes are available: 3 node(s) didn't match node selector.]
-[AfterEach] [sig-scheduling] SchedulerPredicates [Serial]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 18:53:47.501: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "sched-pred-8190" for this suite.
-Jun  4 18:53:53.560: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 18:53:53.915: INFO: namespace sched-pred-8190 deletion completed in 6.397379505s
-[AfterEach] [sig-scheduling] SchedulerPredicates [Serial]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:70
+[It] should support (root,0644,default) [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating a pod to test emptydir 0644 on node default medium
+Jun 18 11:52:38.702: INFO: Waiting up to 5m0s for pod "pod-920efbc9-91bf-11e9-a25d-8608290c688a" in namespace "emptydir-5101" to be "success or failure"
+Jun 18 11:52:38.716: INFO: Pod "pod-920efbc9-91bf-11e9-a25d-8608290c688a": Phase="Pending", Reason="", readiness=false. Elapsed: 13.705306ms
+Jun 18 11:52:40.730: INFO: Pod "pod-920efbc9-91bf-11e9-a25d-8608290c688a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.027730974s
+STEP: Saw pod success
+Jun 18 11:52:40.730: INFO: Pod "pod-920efbc9-91bf-11e9-a25d-8608290c688a" satisfied condition "success or failure"
+Jun 18 11:52:40.743: INFO: Trying to get logs from node 10.72.74.184 pod pod-920efbc9-91bf-11e9-a25d-8608290c688a container test-container: 
+STEP: delete the pod
+Jun 18 11:52:40.831: INFO: Waiting for pod pod-920efbc9-91bf-11e9-a25d-8608290c688a to disappear
+Jun 18 11:52:40.847: INFO: Pod pod-920efbc9-91bf-11e9-a25d-8608290c688a no longer exists
+[AfterEach] [sig-storage] EmptyDir volumes
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 11:52:40.847: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "emptydir-5101" for this suite.
+Jun 18 11:52:46.888: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 11:52:47.497: INFO: namespace emptydir-5101 deletion completed in 6.635094028s
 
-• [SLOW TEST:7.849 seconds]
-[sig-scheduling] SchedulerPredicates [Serial]
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/framework.go:22
-  validates that NodeSelector is respected if not matching  [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+• [SLOW TEST:9.020 seconds]
+[sig-storage] EmptyDir volumes
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:41
+  should support (root,0644,default) [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-SSSSSSSSSSSSSSSSSSS
+SSSS
 ------------------------------
-[sig-storage] Projected downwardAPI 
-  should provide container's memory request [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [sig-storage] Projected downwardAPI
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+[sig-storage] Projected configMap 
+  should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-storage] Projected configMap
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 18:53:53.916: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
+Jun 18 11:52:47.499: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
 STEP: Building a namespace api object, basename projected
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-1684
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-753
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-storage] Projected downwardAPI
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:39
-[It] should provide container's memory request [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: Creating a pod to test downward API volume plugin
-Jun  4 18:53:54.161: INFO: Waiting up to 5m0s for pod "downwardapi-volume-19a08888-86fa-11e9-8d1b-467ee19922ac" in namespace "projected-1684" to be "success or failure"
-Jun  4 18:53:54.172: INFO: Pod "downwardapi-volume-19a08888-86fa-11e9-8d1b-467ee19922ac": Phase="Pending", Reason="", readiness=false. Elapsed: 10.890567ms
-Jun  4 18:53:56.186: INFO: Pod "downwardapi-volume-19a08888-86fa-11e9-8d1b-467ee19922ac": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.024069235s
+[It] should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating configMap with name projected-configmap-test-volume-977050f3-91bf-11e9-a25d-8608290c688a
+STEP: Creating a pod to test consume configMaps
+Jun 18 11:52:47.751: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-9772a365-91bf-11e9-a25d-8608290c688a" in namespace "projected-753" to be "success or failure"
+Jun 18 11:52:47.774: INFO: Pod "pod-projected-configmaps-9772a365-91bf-11e9-a25d-8608290c688a": Phase="Pending", Reason="", readiness=false. Elapsed: 23.521892ms
+Jun 18 11:52:49.788: INFO: Pod "pod-projected-configmaps-9772a365-91bf-11e9-a25d-8608290c688a": Phase="Pending", Reason="", readiness=false. Elapsed: 2.036791099s
+Jun 18 11:52:51.803: INFO: Pod "pod-projected-configmaps-9772a365-91bf-11e9-a25d-8608290c688a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.052371087s
 STEP: Saw pod success
-Jun  4 18:53:56.186: INFO: Pod "downwardapi-volume-19a08888-86fa-11e9-8d1b-467ee19922ac" satisfied condition "success or failure"
-Jun  4 18:53:56.197: INFO: Trying to get logs from node 10.212.23.164 pod downwardapi-volume-19a08888-86fa-11e9-8d1b-467ee19922ac container client-container: 
+Jun 18 11:52:51.803: INFO: Pod "pod-projected-configmaps-9772a365-91bf-11e9-a25d-8608290c688a" satisfied condition "success or failure"
+Jun 18 11:52:51.820: INFO: Trying to get logs from node 10.72.74.138 pod pod-projected-configmaps-9772a365-91bf-11e9-a25d-8608290c688a container projected-configmap-volume-test: 
 STEP: delete the pod
-Jun  4 18:53:56.258: INFO: Waiting for pod downwardapi-volume-19a08888-86fa-11e9-8d1b-467ee19922ac to disappear
-Jun  4 18:53:56.268: INFO: Pod downwardapi-volume-19a08888-86fa-11e9-8d1b-467ee19922ac no longer exists
-[AfterEach] [sig-storage] Projected downwardAPI
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 18:53:56.268: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "projected-1684" for this suite.
-Jun  4 18:54:02.352: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 18:54:02.741: INFO: namespace projected-1684 deletion completed in 6.457061556s
+Jun 18 11:52:51.895: INFO: Waiting for pod pod-projected-configmaps-9772a365-91bf-11e9-a25d-8608290c688a to disappear
+Jun 18 11:52:51.907: INFO: Pod pod-projected-configmaps-9772a365-91bf-11e9-a25d-8608290c688a no longer exists
+[AfterEach] [sig-storage] Projected configMap
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 11:52:51.907: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "projected-753" for this suite.
+Jun 18 11:52:57.947: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 11:52:58.387: INFO: namespace projected-753 deletion completed in 6.465593285s
 
-• [SLOW TEST:8.824 seconds]
-[sig-storage] Projected downwardAPI
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:33
-  should provide container's memory request [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+• [SLOW TEST:10.888 seconds]
+[sig-storage] Projected configMap
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_configmap.go:33
+  should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-SSSSSSSSSSSSSSSSSSSSS
+SSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
 [sig-storage] ConfigMap 
-  updates should be reflected in volume [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+  should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 [BeforeEach] [sig-storage] ConfigMap
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 18:54:02.741: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
+Jun 18 11:52:58.387: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
 STEP: Building a namespace api object, basename configmap
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in configmap-9849
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in configmap-9795
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] updates should be reflected in volume [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: Creating configMap with name configmap-test-upd-1ee30f83-86fa-11e9-8d1b-467ee19922ac
-STEP: Creating the pod
-STEP: Updating configmap configmap-test-upd-1ee30f83-86fa-11e9-8d1b-467ee19922ac
-STEP: waiting to observe update in volume
+[It] should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating configMap with name configmap-test-volume-9dfcda62-91bf-11e9-a25d-8608290c688a
+STEP: Creating a pod to test consume configMaps
+Jun 18 11:52:58.727: INFO: Waiting up to 5m0s for pod "pod-configmaps-9dff09a7-91bf-11e9-a25d-8608290c688a" in namespace "configmap-9795" to be "success or failure"
+Jun 18 11:52:58.741: INFO: Pod "pod-configmaps-9dff09a7-91bf-11e9-a25d-8608290c688a": Phase="Pending", Reason="", readiness=false. Elapsed: 13.193318ms
+Jun 18 11:53:00.754: INFO: Pod "pod-configmaps-9dff09a7-91bf-11e9-a25d-8608290c688a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.026890766s
+STEP: Saw pod success
+Jun 18 11:53:00.754: INFO: Pod "pod-configmaps-9dff09a7-91bf-11e9-a25d-8608290c688a" satisfied condition "success or failure"
+Jun 18 11:53:00.768: INFO: Trying to get logs from node 10.72.74.184 pod pod-configmaps-9dff09a7-91bf-11e9-a25d-8608290c688a container configmap-volume-test: 
+STEP: delete the pod
+Jun 18 11:53:00.840: INFO: Waiting for pod pod-configmaps-9dff09a7-91bf-11e9-a25d-8608290c688a to disappear
+Jun 18 11:53:00.852: INFO: Pod pod-configmaps-9dff09a7-91bf-11e9-a25d-8608290c688a no longer exists
 [AfterEach] [sig-storage] ConfigMap
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 18:54:07.138: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "configmap-9849" for this suite.
-Jun  4 18:54:31.199: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 18:54:31.587: INFO: namespace configmap-9849 deletion completed in 24.43141049s
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 11:53:00.853: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "configmap-9795" for this suite.
+Jun 18 11:53:06.892: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 11:53:07.313: INFO: namespace configmap-9795 deletion completed in 6.446925036s
 
-• [SLOW TEST:28.846 seconds]
+• [SLOW TEST:8.926 seconds]
 [sig-storage] ConfigMap
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:32
-  updates should be reflected in volume [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:32
+  should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+SSSSSSSS
 ------------------------------
-[sig-storage] Projected configMap 
-  should be consumable from pods in volume [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [sig-storage] Projected configMap
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+[k8s.io] [sig-node] Events 
+  should be sent by kubelets and the scheduler about pods scheduling and running  [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [k8s.io] [sig-node] Events
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 18:54:31.587: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-STEP: Building a namespace api object, basename projected
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-3195
+Jun 18 11:53:07.314: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename events
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in events-4073
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should be consumable from pods in volume [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: Creating configMap with name projected-configmap-test-volume-30130800-86fa-11e9-8d1b-467ee19922ac
+[It] should be sent by kubelets and the scheduler about pods scheduling and running  [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: creating the pod
+STEP: submitting the pod to kubernetes
+STEP: verifying the pod is in kubernetes
+STEP: retrieving the pod
+Jun 18 11:53:13.602: INFO: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:send-events-a340920a-91bf-11e9-a25d-8608290c688a,GenerateName:,Namespace:events-4073,SelfLink:/api/v1/namespaces/events-4073/pods/send-events-a340920a-91bf-11e9-a25d-8608290c688a,UID:a342ee1f-91bf-11e9-a08a-ee7a14707756,ResourceVersion:89809,Generation:0,CreationTimestamp:2019-06-18 11:53:07 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: foo,time: 520980362,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-l48s8 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-l48s8,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{p gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1 [] []  [{ 0 80 TCP }] [] [] {map[] map[]} [{default-token-l48s8 true /var/run/secrets/kubernetes.io/serviceaccount   }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*30,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.72.74.138,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists  NoExecute 0xc002f5d8c0} {node.kubernetes.io/unreachable Exists  NoExecute 0xc002f5d8e0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Running,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:53:07 +0000 UTC  } {Ready True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:53:11 +0000 UTC  } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:53:11 +0000 UTC  } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:53:07 +0000 UTC  }],Message:,Reason:,HostIP:10.72.74.138,PodIP:172.30.206.162,StartTime:2019-06-18 11:53:07 +0000 UTC,ContainerStatuses:[{p {nil ContainerStateRunning{StartedAt:2019-06-18 11:53:11 +0000 UTC,} nil} {nil nil nil} true 0 gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1 gcr.io/kubernetes-e2e-test-images/serve-hostname@sha256:bab70473a6d8ef65a22625dc9a1b0f0452e811530fdbe77e4408523460177ff1 containerd://3ee87e230ede86f480383360f7a5e666a4ae97df8b648f5294c0ec716990dc11}],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},}
+
+STEP: checking for scheduler event about the pod
+Jun 18 11:53:15.614: INFO: Saw scheduler event for our pod.
+STEP: checking for kubelet event about the pod
+Jun 18 11:53:17.627: INFO: Saw kubelet event for our pod.
+STEP: deleting the pod
+[AfterEach] [k8s.io] [sig-node] Events
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 11:53:17.652: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "events-4073" for this suite.
+Jun 18 11:53:57.709: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 11:53:58.130: INFO: namespace events-4073 deletion completed in 40.447290458s
+
+• [SLOW TEST:50.816 seconds]
+[k8s.io] [sig-node] Events
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
+  should be sent by kubelets and the scheduler about pods scheduling and running  [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+[sig-storage] ConfigMap 
+  should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-storage] ConfigMap
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Jun 18 11:53:58.130: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename configmap
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in configmap-4337
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating configMap with name configmap-test-volume-map-c188df8a-91bf-11e9-a25d-8608290c688a
 STEP: Creating a pod to test consume configMaps
-Jun  4 18:54:31.834: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-3015050b-86fa-11e9-8d1b-467ee19922ac" in namespace "projected-3195" to be "success or failure"
-Jun  4 18:54:31.846: INFO: Pod "pod-projected-configmaps-3015050b-86fa-11e9-8d1b-467ee19922ac": Phase="Pending", Reason="", readiness=false. Elapsed: 11.816384ms
-Jun  4 18:54:33.860: INFO: Pod "pod-projected-configmaps-3015050b-86fa-11e9-8d1b-467ee19922ac": Phase="Running", Reason="", readiness=true. Elapsed: 2.025373825s
-Jun  4 18:54:35.875: INFO: Pod "pod-projected-configmaps-3015050b-86fa-11e9-8d1b-467ee19922ac": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.040302747s
+Jun 18 11:53:58.367: INFO: Waiting up to 5m0s for pod "pod-configmaps-c18b10ca-91bf-11e9-a25d-8608290c688a" in namespace "configmap-4337" to be "success or failure"
+Jun 18 11:53:58.386: INFO: Pod "pod-configmaps-c18b10ca-91bf-11e9-a25d-8608290c688a": Phase="Pending", Reason="", readiness=false. Elapsed: 18.855616ms
+Jun 18 11:54:00.402: INFO: Pod "pod-configmaps-c18b10ca-91bf-11e9-a25d-8608290c688a": Phase="Running", Reason="", readiness=true. Elapsed: 2.034536698s
+Jun 18 11:54:02.416: INFO: Pod "pod-configmaps-c18b10ca-91bf-11e9-a25d-8608290c688a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.049018509s
 STEP: Saw pod success
-Jun  4 18:54:35.875: INFO: Pod "pod-projected-configmaps-3015050b-86fa-11e9-8d1b-467ee19922ac" satisfied condition "success or failure"
-Jun  4 18:54:35.886: INFO: Trying to get logs from node 10.212.23.161 pod pod-projected-configmaps-3015050b-86fa-11e9-8d1b-467ee19922ac container projected-configmap-volume-test: 
+Jun 18 11:54:02.416: INFO: Pod "pod-configmaps-c18b10ca-91bf-11e9-a25d-8608290c688a" satisfied condition "success or failure"
+Jun 18 11:54:02.429: INFO: Trying to get logs from node 10.72.74.184 pod pod-configmaps-c18b10ca-91bf-11e9-a25d-8608290c688a container configmap-volume-test: 
 STEP: delete the pod
-Jun  4 18:54:35.994: INFO: Waiting for pod pod-projected-configmaps-3015050b-86fa-11e9-8d1b-467ee19922ac to disappear
-Jun  4 18:54:36.005: INFO: Pod pod-projected-configmaps-3015050b-86fa-11e9-8d1b-467ee19922ac no longer exists
-[AfterEach] [sig-storage] Projected configMap
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 18:54:36.005: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "projected-3195" for this suite.
-Jun  4 18:54:42.061: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 18:54:42.498: INFO: namespace projected-3195 deletion completed in 6.478656191s
+Jun 18 11:54:02.507: INFO: Waiting for pod pod-configmaps-c18b10ca-91bf-11e9-a25d-8608290c688a to disappear
+Jun 18 11:54:02.519: INFO: Pod pod-configmaps-c18b10ca-91bf-11e9-a25d-8608290c688a no longer exists
+[AfterEach] [sig-storage] ConfigMap
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 11:54:02.519: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "configmap-4337" for this suite.
+Jun 18 11:54:10.560: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 11:54:11.255: INFO: namespace configmap-4337 deletion completed in 8.72135663s
 
-• [SLOW TEST:10.911 seconds]
-[sig-storage] Projected configMap
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_configmap.go:33
-  should be consumable from pods in volume [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-------------------------------
-SSSSSSSSSSSSS
+• [SLOW TEST:13.125 seconds]
+[sig-storage] ConfigMap
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:32
+  should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-[sig-network] Networking Granular Checks: Pods 
-  should function for intra-pod communication: udp [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [sig-network] Networking
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+[sig-apps] ReplicationController 
+  should serve a basic image on each replica with a public image  [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-apps] ReplicationController
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 18:54:42.498: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-STEP: Building a namespace api object, basename pod-network-test
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in pod-network-test-2894
+Jun 18 11:54:11.255: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename replication-controller
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in replication-controller-8271
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should function for intra-pod communication: udp [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: Performing setup for networking test in namespace pod-network-test-2894
-STEP: creating a selector
-STEP: Creating the service pods in kubernetes
-Jun  4 18:54:42.711: INFO: Waiting up to 10m0s for all (but 0) nodes to be schedulable
-STEP: Creating test pods
-Jun  4 18:55:07.005: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s 'http://172.30.193.35:8080/dial?request=hostName&protocol=udp&host=172.30.254.164&port=8081&tries=1'] Namespace:pod-network-test-2894 PodName:host-test-container-pod ContainerName:hostexec Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
-Jun  4 18:55:07.005: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-Jun  4 18:55:07.231: INFO: Waiting for endpoints: map[]
-Jun  4 18:55:07.245: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s 'http://172.30.193.35:8080/dial?request=hostName&protocol=udp&host=172.30.169.178&port=8081&tries=1'] Namespace:pod-network-test-2894 PodName:host-test-container-pod ContainerName:hostexec Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
-Jun  4 18:55:07.245: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-Jun  4 18:55:07.415: INFO: Waiting for endpoints: map[]
-Jun  4 18:55:07.426: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s 'http://172.30.193.35:8080/dial?request=hostName&protocol=udp&host=172.30.193.34&port=8081&tries=1'] Namespace:pod-network-test-2894 PodName:host-test-container-pod ContainerName:hostexec Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
-Jun  4 18:55:07.426: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-Jun  4 18:55:07.612: INFO: Waiting for endpoints: map[]
-[AfterEach] [sig-network] Networking
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 18:55:07.612: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "pod-network-test-2894" for this suite.
-Jun  4 18:55:31.673: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 18:55:32.068: INFO: namespace pod-network-test-2894 deletion completed in 24.438835122s
+[It] should serve a basic image on each replica with a public image  [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating replication controller my-hostname-basic-c95b87d7-91bf-11e9-a25d-8608290c688a
+Jun 18 11:54:11.485: INFO: Pod name my-hostname-basic-c95b87d7-91bf-11e9-a25d-8608290c688a: Found 0 pods out of 1
+Jun 18 11:54:16.499: INFO: Pod name my-hostname-basic-c95b87d7-91bf-11e9-a25d-8608290c688a: Found 1 pods out of 1
+Jun 18 11:54:16.499: INFO: Ensuring all pods for ReplicationController "my-hostname-basic-c95b87d7-91bf-11e9-a25d-8608290c688a" are running
+Jun 18 11:54:16.511: INFO: Pod "my-hostname-basic-c95b87d7-91bf-11e9-a25d-8608290c688a-djbwl" is running (conditions: [{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2019-06-18 11:54:11 +0000 UTC Reason: Message:} {Type:Ready Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2019-06-18 11:54:12 +0000 UTC Reason: Message:} {Type:ContainersReady Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2019-06-18 11:54:12 +0000 UTC Reason: Message:} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2019-06-18 11:54:11 +0000 UTC Reason: Message:}])
+Jun 18 11:54:16.511: INFO: Trying to dial the pod
+Jun 18 11:54:21.566: INFO: Controller my-hostname-basic-c95b87d7-91bf-11e9-a25d-8608290c688a: Got expected result from replica 1 [my-hostname-basic-c95b87d7-91bf-11e9-a25d-8608290c688a-djbwl]: "my-hostname-basic-c95b87d7-91bf-11e9-a25d-8608290c688a-djbwl", 1 of 1 required successes so far
+[AfterEach] [sig-apps] ReplicationController
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 11:54:21.566: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "replication-controller-8271" for this suite.
+Jun 18 11:54:27.879: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 11:54:28.278: INFO: namespace replication-controller-8271 deletion completed in 6.441638276s
 
-• [SLOW TEST:49.570 seconds]
-[sig-network] Networking
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:25
-  Granular Checks: Pods
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:28
-    should function for intra-pod communication: udp [LinuxOnly] [NodeConformance] [Conformance]
-    /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+• [SLOW TEST:17.023 seconds]
+[sig-apps] ReplicationController
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22
+  should serve a basic image on each replica with a public image  [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-SSSSSSSSS
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-storage] EmptyDir volumes 
-  should support (root,0644,default) [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [sig-storage] EmptyDir volumes
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+[k8s.io] Container Lifecycle Hook when create a pod with lifecycle hook 
+  should execute prestop exec hook properly [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [k8s.io] Container Lifecycle Hook
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 18:55:32.069: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-STEP: Building a namespace api object, basename emptydir
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in emptydir-6660
+Jun 18 11:54:28.285: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename container-lifecycle-hook
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in container-lifecycle-hook-9441
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should support (root,0644,default) [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: Creating a pod to test emptydir 0644 on node default medium
-Jun  4 18:55:32.382: INFO: Waiting up to 5m0s for pod "pod-542c3494-86fa-11e9-8d1b-467ee19922ac" in namespace "emptydir-6660" to be "success or failure"
-Jun  4 18:55:32.393: INFO: Pod "pod-542c3494-86fa-11e9-8d1b-467ee19922ac": Phase="Pending", Reason="", readiness=false. Elapsed: 11.274777ms
-Jun  4 18:55:34.405: INFO: Pod "pod-542c3494-86fa-11e9-8d1b-467ee19922ac": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.023344089s
-STEP: Saw pod success
-Jun  4 18:55:34.405: INFO: Pod "pod-542c3494-86fa-11e9-8d1b-467ee19922ac" satisfied condition "success or failure"
-Jun  4 18:55:34.456: INFO: Trying to get logs from node 10.212.23.161 pod pod-542c3494-86fa-11e9-8d1b-467ee19922ac container test-container: 
-STEP: delete the pod
-Jun  4 18:55:34.520: INFO: Waiting for pod pod-542c3494-86fa-11e9-8d1b-467ee19922ac to disappear
-Jun  4 18:55:34.531: INFO: Pod pod-542c3494-86fa-11e9-8d1b-467ee19922ac no longer exists
-[AfterEach] [sig-storage] EmptyDir volumes
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 18:55:34.531: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "emptydir-6660" for this suite.
-Jun  4 18:55:40.583: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 18:55:40.908: INFO: namespace emptydir-6660 deletion completed in 6.363997715s
+[BeforeEach] when create a pod with lifecycle hook
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:61
+STEP: create the container to handle the HTTPGet hook request.
+[It] should execute prestop exec hook properly [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: create the pod with lifecycle hook
+STEP: delete the pod with lifecycle hook
+Jun 18 11:54:38.632: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear
+Jun 18 11:54:38.659: INFO: Pod pod-with-prestop-exec-hook still exists
+Jun 18 11:54:40.659: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear
+Jun 18 11:54:40.674: INFO: Pod pod-with-prestop-exec-hook still exists
+Jun 18 11:54:42.660: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear
+Jun 18 11:54:42.673: INFO: Pod pod-with-prestop-exec-hook still exists
+Jun 18 11:54:44.659: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear
+Jun 18 11:54:44.673: INFO: Pod pod-with-prestop-exec-hook still exists
+Jun 18 11:54:46.659: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear
+Jun 18 11:54:46.673: INFO: Pod pod-with-prestop-exec-hook still exists
+Jun 18 11:54:48.660: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear
+Jun 18 11:54:48.672: INFO: Pod pod-with-prestop-exec-hook still exists
+Jun 18 11:54:50.659: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear
+Jun 18 11:54:50.673: INFO: Pod pod-with-prestop-exec-hook still exists
+Jun 18 11:54:52.659: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear
+Jun 18 11:54:52.673: INFO: Pod pod-with-prestop-exec-hook still exists
+Jun 18 11:54:54.660: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear
+Jun 18 11:54:54.674: INFO: Pod pod-with-prestop-exec-hook still exists
+Jun 18 11:54:56.659: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear
+Jun 18 11:54:56.673: INFO: Pod pod-with-prestop-exec-hook still exists
+Jun 18 11:54:58.659: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear
+Jun 18 11:54:58.673: INFO: Pod pod-with-prestop-exec-hook still exists
+Jun 18 11:55:00.659: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear
+Jun 18 11:55:00.673: INFO: Pod pod-with-prestop-exec-hook still exists
+Jun 18 11:55:02.659: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear
+Jun 18 11:55:02.673: INFO: Pod pod-with-prestop-exec-hook still exists
+Jun 18 11:55:04.659: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear
+Jun 18 11:55:04.673: INFO: Pod pod-with-prestop-exec-hook no longer exists
+STEP: check prestop hook
+[AfterEach] [k8s.io] Container Lifecycle Hook
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 11:55:04.702: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "container-lifecycle-hook-9441" for this suite.
+Jun 18 11:55:28.762: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 11:55:29.265: INFO: namespace container-lifecycle-hook-9441 deletion completed in 24.547837244s
 
-• [SLOW TEST:8.840 seconds]
-[sig-storage] EmptyDir volumes
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:41
-  should support (root,0644,default) [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+• [SLOW TEST:60.981 seconds]
+[k8s.io] Container Lifecycle Hook
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
+  when create a pod with lifecycle hook
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:40
+    should execute prestop exec hook properly [NodeConformance] [Conformance]
+    /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-SSSSSSSSSSSSSSSSSSSS
+SSS
 ------------------------------
 [sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] 
-  Scaling should happen in predictable order and halt if any stateful pod is unhealthy [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+  Should recreate evicted statefulset [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 [BeforeEach] [sig-apps] StatefulSet
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 18:55:40.908: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
+Jun 18 11:55:29.265: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
 STEP: Building a namespace api object, basename statefulset
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in statefulset-3974
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in statefulset-4962
 STEP: Waiting for a default service account to be provisioned in namespace
 [BeforeEach] [sig-apps] StatefulSet
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:59
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:59
 [BeforeEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:74
-STEP: Creating service test in namespace statefulset-3974
-[It] Scaling should happen in predictable order and halt if any stateful pod is unhealthy [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: Initializing watcher for selector baz=blah,foo=bar
-STEP: Creating stateful set ss in namespace statefulset-3974
-STEP: Waiting until all stateful set ss replicas will be running in namespace statefulset-3974
-Jun  4 18:55:41.159: INFO: Found 0 stateful pods, waiting for 1
-Jun  4 18:55:51.173: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true
-STEP: Confirming that stateful set scale up will halt with unhealthy stateful pod
-Jun  4 18:55:51.185: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 exec --namespace=statefulset-3974 ss-0 -- /bin/sh -x -c mv -v /usr/share/nginx/html/index.html /tmp/ || true'
-Jun  4 18:55:51.505: INFO: stderr: "+ mv -v /usr/share/nginx/html/index.html /tmp/\n"
-Jun  4 18:55:51.505: INFO: stdout: "'/usr/share/nginx/html/index.html' -> '/tmp/index.html'\n"
-Jun  4 18:55:51.505: INFO: stdout of mv -v /usr/share/nginx/html/index.html /tmp/ || true on ss-0: '/usr/share/nginx/html/index.html' -> '/tmp/index.html'
-
-Jun  4 18:55:51.518: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=true
-Jun  4 18:56:01.542: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=false
-Jun  4 18:56:01.542: INFO: Waiting for statefulset status.replicas updated to 0
-Jun  4 18:56:01.655: INFO: Verifying statefulset ss doesn't scale past 1 for another 9.999998586s
-Jun  4 18:56:02.670: INFO: Verifying statefulset ss doesn't scale past 1 for another 8.93307819s
-Jun  4 18:56:03.683: INFO: Verifying statefulset ss doesn't scale past 1 for another 7.918628948s
-Jun  4 18:56:04.696: INFO: Verifying statefulset ss doesn't scale past 1 for another 6.905201212s
-Jun  4 18:56:05.719: INFO: Verifying statefulset ss doesn't scale past 1 for another 5.892043002s
-Jun  4 18:56:06.732: INFO: Verifying statefulset ss doesn't scale past 1 for another 4.869041508s
-Jun  4 18:56:07.745: INFO: Verifying statefulset ss doesn't scale past 1 for another 3.855980336s
-Jun  4 18:56:08.758: INFO: Verifying statefulset ss doesn't scale past 1 for another 2.843231937s
-Jun  4 18:56:09.770: INFO: Verifying statefulset ss doesn't scale past 1 for another 1.829858512s
-Jun  4 18:56:10.783: INFO: Verifying statefulset ss doesn't scale past 1 for another 818.225453ms
-STEP: Scaling up stateful set ss to 3 replicas and waiting until all of them will be running in namespace statefulset-3974
-Jun  4 18:56:11.810: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 exec --namespace=statefulset-3974 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
-Jun  4 18:56:12.222: INFO: stderr: "+ mv -v /tmp/index.html /usr/share/nginx/html/\n"
-Jun  4 18:56:12.222: INFO: stdout: "'/tmp/index.html' -> '/usr/share/nginx/html/index.html'\n"
-Jun  4 18:56:12.222: INFO: stdout of mv -v /tmp/index.html /usr/share/nginx/html/ || true on ss-0: '/tmp/index.html' -> '/usr/share/nginx/html/index.html'
-
-Jun  4 18:56:12.239: INFO: Found 1 stateful pods, waiting for 3
-Jun  4 18:56:22.290: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true
-Jun  4 18:56:22.290: INFO: Waiting for pod ss-1 to enter Running - Ready=true, currently Running - Ready=true
-Jun  4 18:56:22.290: INFO: Waiting for pod ss-2 to enter Running - Ready=true, currently Running - Ready=true
-STEP: Verifying that stateful set ss was scaled up in order
-STEP: Scale down will halt with unhealthy stateful pod
-Jun  4 18:56:22.372: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 exec --namespace=statefulset-3974 ss-0 -- /bin/sh -x -c mv -v /usr/share/nginx/html/index.html /tmp/ || true'
-Jun  4 18:56:22.624: INFO: stderr: "+ mv -v /usr/share/nginx/html/index.html /tmp/\n"
-Jun  4 18:56:22.624: INFO: stdout: "'/usr/share/nginx/html/index.html' -> '/tmp/index.html'\n"
-Jun  4 18:56:22.624: INFO: stdout of mv -v /usr/share/nginx/html/index.html /tmp/ || true on ss-0: '/usr/share/nginx/html/index.html' -> '/tmp/index.html'
-
-Jun  4 18:56:22.624: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 exec --namespace=statefulset-3974 ss-1 -- /bin/sh -x -c mv -v /usr/share/nginx/html/index.html /tmp/ || true'
-Jun  4 18:56:22.906: INFO: stderr: "+ mv -v /usr/share/nginx/html/index.html /tmp/\n"
-Jun  4 18:56:22.906: INFO: stdout: "'/usr/share/nginx/html/index.html' -> '/tmp/index.html'\n"
-Jun  4 18:56:22.906: INFO: stdout of mv -v /usr/share/nginx/html/index.html /tmp/ || true on ss-1: '/usr/share/nginx/html/index.html' -> '/tmp/index.html'
-
-Jun  4 18:56:22.906: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 exec --namespace=statefulset-3974 ss-2 -- /bin/sh -x -c mv -v /usr/share/nginx/html/index.html /tmp/ || true'
-Jun  4 18:56:23.193: INFO: stderr: "+ mv -v /usr/share/nginx/html/index.html /tmp/\n"
-Jun  4 18:56:23.193: INFO: stdout: "'/usr/share/nginx/html/index.html' -> '/tmp/index.html'\n"
-Jun  4 18:56:23.193: INFO: stdout of mv -v /usr/share/nginx/html/index.html /tmp/ || true on ss-2: '/usr/share/nginx/html/index.html' -> '/tmp/index.html'
-
-Jun  4 18:56:23.193: INFO: Waiting for statefulset status.replicas updated to 0
-Jun  4 18:56:23.201: INFO: Waiting for stateful set status.readyReplicas to become 0, currently 3
-Jun  4 18:56:33.224: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=false
-Jun  4 18:56:33.224: INFO: Waiting for pod ss-1 to enter Running - Ready=false, currently Running - Ready=false
-Jun  4 18:56:33.224: INFO: Waiting for pod ss-2 to enter Running - Ready=false, currently Running - Ready=false
-Jun  4 18:56:33.285: INFO: Verifying statefulset ss doesn't scale past 3 for another 9.999998878s
-Jun  4 18:56:34.299: INFO: Verifying statefulset ss doesn't scale past 3 for another 8.985839144s
-Jun  4 18:56:35.313: INFO: Verifying statefulset ss doesn't scale past 3 for another 7.972515536s
-Jun  4 18:56:36.327: INFO: Verifying statefulset ss doesn't scale past 3 for another 6.958030694s
-Jun  4 18:56:37.340: INFO: Verifying statefulset ss doesn't scale past 3 for another 5.944685952s
-Jun  4 18:56:38.355: INFO: Verifying statefulset ss doesn't scale past 3 for another 4.931294543s
-Jun  4 18:56:39.367: INFO: Verifying statefulset ss doesn't scale past 3 for another 3.916001005s
-Jun  4 18:56:40.382: INFO: Verifying statefulset ss doesn't scale past 3 for another 2.903873854s
-Jun  4 18:56:41.395: INFO: Verifying statefulset ss doesn't scale past 3 for another 1.888995353s
-Jun  4 18:56:42.408: INFO: Verifying statefulset ss doesn't scale past 3 for another 876.329341ms
-STEP: Scaling down stateful set ss to 0 replicas and waiting until none of pods will run in namespacestatefulset-3974
-Jun  4 18:56:43.421: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 exec --namespace=statefulset-3974 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
-Jun  4 18:56:43.724: INFO: stderr: "+ mv -v /tmp/index.html /usr/share/nginx/html/\n"
-Jun  4 18:56:43.724: INFO: stdout: "'/tmp/index.html' -> '/usr/share/nginx/html/index.html'\n"
-Jun  4 18:56:43.724: INFO: stdout of mv -v /tmp/index.html /usr/share/nginx/html/ || true on ss-0: '/tmp/index.html' -> '/usr/share/nginx/html/index.html'
-
-Jun  4 18:56:43.724: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 exec --namespace=statefulset-3974 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
-Jun  4 18:56:44.005: INFO: stderr: "+ mv -v /tmp/index.html /usr/share/nginx/html/\n"
-Jun  4 18:56:44.005: INFO: stdout: "'/tmp/index.html' -> '/usr/share/nginx/html/index.html'\n"
-Jun  4 18:56:44.005: INFO: stdout of mv -v /tmp/index.html /usr/share/nginx/html/ || true on ss-1: '/tmp/index.html' -> '/usr/share/nginx/html/index.html'
-
-Jun  4 18:56:44.005: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 exec --namespace=statefulset-3974 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
-Jun  4 18:56:44.294: INFO: stderr: "+ mv -v /tmp/index.html /usr/share/nginx/html/\n"
-Jun  4 18:56:44.294: INFO: stdout: "'/tmp/index.html' -> '/usr/share/nginx/html/index.html'\n"
-Jun  4 18:56:44.294: INFO: stdout of mv -v /tmp/index.html /usr/share/nginx/html/ || true on ss-2: '/tmp/index.html' -> '/usr/share/nginx/html/index.html'
-
-Jun  4 18:56:44.294: INFO: Scaling statefulset ss to 0
-STEP: Verifying that stateful set ss was scaled down in reverse order
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:74
+STEP: Creating service test in namespace statefulset-4962
+[It] Should recreate evicted statefulset [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Looking for a node to schedule stateful set and pod
+STEP: Creating pod with conflicting port in namespace statefulset-4962
+STEP: Creating statefulset with conflicting port in namespace statefulset-4962
+STEP: Waiting until pod test-pod will start running in namespace statefulset-4962
+STEP: Waiting until stateful pod ss-0 will be recreated and deleted at least once in namespace statefulset-4962
+Jun 18 11:55:33.563: INFO: Observed stateful pod in namespace: statefulset-4962, name: ss-0, uid: fa138602-91bf-11e9-a08a-ee7a14707756, status phase: Pending. Waiting for statefulset controller to delete.
+Jun 18 11:55:33.733: INFO: Observed stateful pod in namespace: statefulset-4962, name: ss-0, uid: fa138602-91bf-11e9-a08a-ee7a14707756, status phase: Failed. Waiting for statefulset controller to delete.
+Jun 18 11:55:33.755: INFO: Observed stateful pod in namespace: statefulset-4962, name: ss-0, uid: fa138602-91bf-11e9-a08a-ee7a14707756, status phase: Failed. Waiting for statefulset controller to delete.
+Jun 18 11:55:33.774: INFO: Observed delete event for stateful pod ss-0 in namespace statefulset-4962
+STEP: Removing pod with conflicting port in namespace statefulset-4962
+STEP: Waiting when stateful pod ss-0 will be recreated in namespace statefulset-4962 and will be in running state
 [AfterEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:85
-Jun  4 18:57:14.362: INFO: Deleting all statefulset in ns statefulset-3974
-Jun  4 18:57:14.377: INFO: Scaling statefulset ss to 0
-Jun  4 18:57:14.419: INFO: Waiting for statefulset status.replicas updated to 0
-Jun  4 18:57:14.428: INFO: Deleting statefulset ss
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:85
+Jun 18 11:55:38.011: INFO: Deleting all statefulset in ns statefulset-4962
+Jun 18 11:55:38.022: INFO: Scaling statefulset ss to 0
+Jun 18 11:55:48.084: INFO: Waiting for statefulset status.replicas updated to 0
+Jun 18 11:55:48.098: INFO: Deleting statefulset ss
 [AfterEach] [sig-apps] StatefulSet
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 18:57:14.481: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "statefulset-3974" for this suite.
-Jun  4 18:57:22.557: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 18:57:23.065: INFO: namespace statefulset-3974 deletion completed in 8.550563947s
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 11:55:48.149: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "statefulset-4962" for this suite.
+Jun 18 11:55:56.194: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 11:55:57.374: INFO: namespace statefulset-4962 deletion completed in 9.211276556s
 
-• [SLOW TEST:102.157 seconds]
+• [SLOW TEST:28.109 seconds]
 [sig-apps] StatefulSet
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22
   [k8s.io] Basic StatefulSet functionality [StatefulSetBasic]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
-    Scaling should happen in predictable order and halt if any stateful pod is unhealthy [Conformance]
-    /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
+    Should recreate evicted statefulset [Conformance]
+    /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-SSSSSSSSSSSSSSSSSSSSSSS
-------------------------------
-[k8s.io] Kubelet when scheduling a busybox command that always fails in a pod 
-  should have an terminated reason [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [k8s.io] Kubelet
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
-STEP: Creating a kubernetes client
-Jun  4 18:57:23.067: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-STEP: Building a namespace api object, basename kubelet-test
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubelet-test-1081
-STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [k8s.io] Kubelet
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:37
-[BeforeEach] when scheduling a busybox command that always fails in a pod
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:81
-[It] should have an terminated reason [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[AfterEach] [k8s.io] Kubelet
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 18:57:27.405: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "kubelet-test-1081" for this suite.
-Jun  4 18:57:33.461: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 18:57:33.889: INFO: namespace kubelet-test-1081 deletion completed in 6.469693204s
-
-• [SLOW TEST:10.822 seconds]
-[k8s.io] Kubelet
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
-  when scheduling a busybox command that always fails in a pod
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:78
-    should have an terminated reason [NodeConformance] [Conformance]
-    /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-------------------------------
-SSSSSSSSSSSS
-------------------------------
-[sig-scheduling] SchedulerPredicates [Serial] 
-  validates that NodeSelector is respected if matching  [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [sig-scheduling] SchedulerPredicates [Serial]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
-STEP: Creating a kubernetes client
-Jun  4 18:57:33.889: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-STEP: Building a namespace api object, basename sched-pred
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in sched-pred-6011
-STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-scheduling] SchedulerPredicates [Serial]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:79
-Jun  4 18:57:34.106: INFO: Waiting up to 1m0s for all (but 0) nodes to be ready
-Jun  4 18:57:34.130: INFO: Waiting for terminating namespaces to be deleted...
-Jun  4 18:57:34.142: INFO: 
-Logging pods the kubelet thinks is on node 10.212.23.161 before test
-Jun  4 18:57:34.183: INFO: calico-node-wtt96 from kube-system started at 2019-06-04 17:44:38 +0000 UTC (1 container statuses recorded)
-Jun  4 18:57:34.183: INFO: 	Container calico-node ready: true, restart count 0
-Jun  4 18:57:34.183: INFO: coredns-autoscaler-5846b97547-t7mrd from kube-system started at 2019-06-04 17:44:57 +0000 UTC (1 container statuses recorded)
-Jun  4 18:57:34.183: INFO: 	Container autoscaler ready: true, restart count 0
-Jun  4 18:57:34.183: INFO: kubernetes-dashboard-b4f5d97fc-rrmfx from kube-system started at 2019-06-04 17:44:57 +0000 UTC (1 container statuses recorded)
-Jun  4 18:57:34.183: INFO: 	Container kubernetes-dashboard ready: true, restart count 0
-Jun  4 18:57:34.183: INFO: ibm-storage-watcher-f74547b86-jhnt2 from kube-system started at 2019-06-04 17:44:57 +0000 UTC (1 container statuses recorded)
-Jun  4 18:57:34.183: INFO: 	Container ibm-storage-watcher-container ready: true, restart count 0
-Jun  4 18:57:34.183: INFO: ibm-kube-fluentd-cj2l2 from kube-system started at 2019-06-04 17:45:04 +0000 UTC (1 container statuses recorded)
-Jun  4 18:57:34.183: INFO: 	Container fluentd ready: true, restart count 0
-Jun  4 18:57:34.183: INFO: ibm-master-proxy-static-10.212.23.161 from kube-system started at  (0 container statuses recorded)
-Jun  4 18:57:34.183: INFO: coredns-cc5d484bb-9nhk2 from kube-system started at 2019-06-04 17:44:57 +0000 UTC (1 container statuses recorded)
-Jun  4 18:57:34.183: INFO: 	Container coredns ready: true, restart count 0
-Jun  4 18:57:34.183: INFO: ibm-keepalived-watcher-vl7mx from kube-system started at 2019-06-04 17:44:38 +0000 UTC (1 container statuses recorded)
-Jun  4 18:57:34.183: INFO: 	Container keepalived-watcher ready: true, restart count 0
-Jun  4 18:57:34.183: INFO: ibm-file-plugin-69d4bc9c8f-jwhqt from kube-system started at 2019-06-04 17:44:57 +0000 UTC (1 container statuses recorded)
-Jun  4 18:57:34.183: INFO: 	Container ibm-file-plugin-container ready: true, restart count 0
-Jun  4 18:57:34.183: INFO: vpn-7bbfb5bc49-4mjxm from kube-system started at 2019-06-04 17:44:57 +0000 UTC (1 container statuses recorded)
-Jun  4 18:57:34.183: INFO: 	Container vpn ready: true, restart count 0
-Jun  4 18:57:34.183: INFO: calico-kube-controllers-64ccf5d5d4-k2qnd from kube-system started at 2019-06-04 17:44:57 +0000 UTC (1 container statuses recorded)
-Jun  4 18:57:34.183: INFO: 	Container calico-kube-controllers ready: true, restart count 0
-Jun  4 18:57:34.183: INFO: public-cra6d87a4394f440e3becc50f237563475-alb1-79855b4ffb-mqcml from kube-system started at 2019-06-04 17:49:42 +0000 UTC (4 container statuses recorded)
-Jun  4 18:57:34.183: INFO: 	Container ingress-auth-1 ready: true, restart count 0
-Jun  4 18:57:34.183: INFO: 	Container ingress-auth-2 ready: true, restart count 0
-Jun  4 18:57:34.183: INFO: 	Container ingress-auth-3 ready: true, restart count 0
-Jun  4 18:57:34.183: INFO: 	Container nginx-ingress ready: true, restart count 0
-Jun  4 18:57:34.183: INFO: sonobuoy-systemd-logs-daemon-set-d63a0136eef44275-l8n68 from heptio-sonobuoy started at 2019-06-04 18:27:33 +0000 UTC (2 container statuses recorded)
-Jun  4 18:57:34.183: INFO: 	Container sonobuoy-worker ready: true, restart count 0
-Jun  4 18:57:34.183: INFO: 	Container systemd-logs ready: true, restart count 0
-Jun  4 18:57:34.183: INFO: 
-Logging pods the kubelet thinks is on node 10.212.23.164 before test
-Jun  4 18:57:34.220: INFO: ibm-kube-fluentd-nr9cm from kube-system started at 2019-06-04 17:45:04 +0000 UTC (1 container statuses recorded)
-Jun  4 18:57:34.220: INFO: 	Container fluentd ready: true, restart count 0
-Jun  4 18:57:34.220: INFO: sonobuoy from heptio-sonobuoy started at 2019-06-04 18:27:24 +0000 UTC (1 container statuses recorded)
-Jun  4 18:57:34.220: INFO: 	Container kube-sonobuoy ready: true, restart count 0
-Jun  4 18:57:34.220: INFO: calico-node-sphpm from kube-system started at 2019-06-04 17:44:49 +0000 UTC (1 container statuses recorded)
-Jun  4 18:57:34.220: INFO: 	Container calico-node ready: true, restart count 0
-Jun  4 18:57:34.220: INFO: metrics-server-66cfcd4b8d-xg6wk from kube-system started at 2019-06-04 17:45:26 +0000 UTC (2 container statuses recorded)
-Jun  4 18:57:34.220: INFO: 	Container metrics-server ready: true, restart count 0
-Jun  4 18:57:34.220: INFO: 	Container metrics-server-nanny ready: true, restart count 0
-Jun  4 18:57:34.220: INFO: ibm-master-proxy-static-10.212.23.164 from kube-system started at  (0 container statuses recorded)
-Jun  4 18:57:34.220: INFO: ibm-keepalived-watcher-lfldx from kube-system started at 2019-06-04 17:44:49 +0000 UTC (1 container statuses recorded)
-Jun  4 18:57:34.220: INFO: 	Container keepalived-watcher ready: true, restart count 0
-Jun  4 18:57:34.220: INFO: sonobuoy-e2e-job-39546bc4ddc3410d from heptio-sonobuoy started at 2019-06-04 18:27:33 +0000 UTC (2 container statuses recorded)
-Jun  4 18:57:34.220: INFO: 	Container e2e ready: true, restart count 0
-Jun  4 18:57:34.220: INFO: 	Container sonobuoy-worker ready: true, restart count 0
-Jun  4 18:57:34.220: INFO: sonobuoy-systemd-logs-daemon-set-d63a0136eef44275-b6j5c from heptio-sonobuoy started at 2019-06-04 18:27:33 +0000 UTC (2 container statuses recorded)
-Jun  4 18:57:34.220: INFO: 	Container sonobuoy-worker ready: true, restart count 0
-Jun  4 18:57:34.220: INFO: 	Container systemd-logs ready: true, restart count 0
-Jun  4 18:57:34.220: INFO: ibm-cloud-provider-ip-162-133-73-165-ffb96fbff-7g5nv from ibm-system started at 2019-06-04 17:47:12 +0000 UTC (1 container statuses recorded)
-Jun  4 18:57:34.220: INFO: 	Container ibm-cloud-provider-ip-162-133-73-165 ready: true, restart count 0
-Jun  4 18:57:34.220: INFO: test-k8s-e2e-pvg-master-verification from default started at 2019-06-04 18:27:15 +0000 UTC (1 container statuses recorded)
-Jun  4 18:57:34.220: INFO: 	Container test-k8s-e2e-pvg-master-verification ready: true, restart count 0
-Jun  4 18:57:34.220: INFO: 
-Logging pods the kubelet thinks is on node 10.212.23.189 before test
-Jun  4 18:57:34.255: INFO: ibm-kube-fluentd-k56t7 from kube-system started at 2019-06-04 17:45:04 +0000 UTC (1 container statuses recorded)
-Jun  4 18:57:34.255: INFO: 	Container fluentd ready: true, restart count 0
-Jun  4 18:57:34.255: INFO: ibm-cloud-provider-ip-162-133-73-165-ffb96fbff-jf4l4 from ibm-system started at 2019-06-04 17:47:12 +0000 UTC (1 container statuses recorded)
-Jun  4 18:57:34.255: INFO: 	Container ibm-cloud-provider-ip-162-133-73-165 ready: true, restart count 0
-Jun  4 18:57:34.255: INFO: public-cra6d87a4394f440e3becc50f237563475-alb1-79855b4ffb-89n6x from kube-system started at 2019-06-04 17:49:42 +0000 UTC (4 container statuses recorded)
-Jun  4 18:57:34.255: INFO: 	Container ingress-auth-1 ready: true, restart count 0
-Jun  4 18:57:34.255: INFO: 	Container ingress-auth-2 ready: true, restart count 0
-Jun  4 18:57:34.255: INFO: 	Container ingress-auth-3 ready: true, restart count 0
-Jun  4 18:57:34.255: INFO: 	Container nginx-ingress ready: true, restart count 0
-Jun  4 18:57:34.255: INFO: sonobuoy-systemd-logs-daemon-set-d63a0136eef44275-hnnhz from heptio-sonobuoy started at 2019-06-04 18:27:33 +0000 UTC (2 container statuses recorded)
-Jun  4 18:57:34.255: INFO: 	Container sonobuoy-worker ready: true, restart count 0
-Jun  4 18:57:34.255: INFO: 	Container systemd-logs ready: true, restart count 0
-Jun  4 18:57:34.255: INFO: ibm-master-proxy-static-10.212.23.189 from kube-system started at  (0 container statuses recorded)
-Jun  4 18:57:34.255: INFO: coredns-cc5d484bb-nlgcf from kube-system started at 2019-06-04 17:45:30 +0000 UTC (1 container statuses recorded)
-Jun  4 18:57:34.255: INFO: 	Container coredns ready: true, restart count 0
-Jun  4 18:57:34.255: INFO: calico-node-v6hns from kube-system started at 2019-06-04 17:45:04 +0000 UTC (1 container statuses recorded)
-Jun  4 18:57:34.255: INFO: 	Container calico-node ready: true, restart count 0
-Jun  4 18:57:34.255: INFO: ibm-keepalived-watcher-dqk2k from kube-system started at 2019-06-04 17:45:04 +0000 UTC (1 container statuses recorded)
-Jun  4 18:57:34.255: INFO: 	Container keepalived-watcher ready: true, restart count 0
-[It] validates that NodeSelector is respected if matching  [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: Trying to launch a pod without a label to get a node which can launch it.
-STEP: Explicitly delete pod here to free the resource it takes.
-STEP: Trying to apply a random label on the found node.
-STEP: verifying the node has the label kubernetes.io/e2e-9e12d895-86fa-11e9-8d1b-467ee19922ac 42
-STEP: Trying to relaunch the pod, now with labels.
-STEP: removing the label kubernetes.io/e2e-9e12d895-86fa-11e9-8d1b-467ee19922ac off the node 10.212.23.164
-STEP: verifying the node doesn't have the label kubernetes.io/e2e-9e12d895-86fa-11e9-8d1b-467ee19922ac
-[AfterEach] [sig-scheduling] SchedulerPredicates [Serial]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 18:57:38.549: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "sched-pred-6011" for this suite.
-Jun  4 18:57:48.606: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 18:57:49.007: INFO: namespace sched-pred-6011 deletion completed in 10.443831742s
-[AfterEach] [sig-scheduling] SchedulerPredicates [Serial]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:70
-
-• [SLOW TEST:15.118 seconds]
-[sig-scheduling] SchedulerPredicates [Serial]
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/framework.go:22
-  validates that NodeSelector is respected if matching  [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-------------------------------
-SSSS
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-network] Proxy version v1 
-  should proxy logs on node using proxy subresource  [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] version v1
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+[sig-cli] Kubectl client [k8s.io] Proxy server 
+  should support --unix-socket=/path  [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-cli] Kubectl client
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 18:57:49.007: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-STEP: Building a namespace api object, basename proxy
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in proxy-4251
+Jun 18 11:55:57.376: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename kubectl
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-8957
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should proxy logs on node using proxy subresource  [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-Jun  4 18:57:49.260: INFO: (0) /api/v1/nodes/10.212.23.161/proxy/logs/: 
-alb/
-alternatives.log
-apt/... (200; 23.983293ms)
-Jun  4 18:57:49.276: INFO: (1) /api/v1/nodes/10.212.23.161/proxy/logs/: 
-alb/
-alternatives.log
-apt/... (200; 15.484852ms)
-Jun  4 18:57:49.291: INFO: (2) /api/v1/nodes/10.212.23.161/proxy/logs/: 
-alb/
-alternatives.log
-apt/... (200; 15.591516ms)
-Jun  4 18:57:49.308: INFO: (3) /api/v1/nodes/10.212.23.161/proxy/logs/: 
-alb/
-alternatives.log
-apt/... (200; 16.186099ms)
-Jun  4 18:57:49.323: INFO: (4) /api/v1/nodes/10.212.23.161/proxy/logs/: 
-alb/
-alternatives.log
-apt/... (200; 15.370616ms)
-Jun  4 18:57:49.343: INFO: (5) /api/v1/nodes/10.212.23.161/proxy/logs/: 
-alb/
-alternatives.log
-apt/... (200; 19.895753ms)
-Jun  4 18:57:49.359: INFO: (6) /api/v1/nodes/10.212.23.161/proxy/logs/: 
-alb/
-alternatives.log
-apt/... (200; 16.50187ms)
-Jun  4 18:57:49.376: INFO: (7) /api/v1/nodes/10.212.23.161/proxy/logs/: 
-alb/
-alternatives.log
-apt/... (200; 16.989935ms)
-Jun  4 18:57:49.399: INFO: (8) /api/v1/nodes/10.212.23.161/proxy/logs/: 
-alb/
-alternatives.log
-apt/... (200; 22.47681ms)
-Jun  4 18:57:49.414: INFO: (9) /api/v1/nodes/10.212.23.161/proxy/logs/: 
-alb/
-alternatives.log
-apt/... (200; 14.537501ms)
-Jun  4 18:57:49.429: INFO: (10) /api/v1/nodes/10.212.23.161/proxy/logs/: 
-alb/
-alternatives.log
-apt/... (200; 15.007944ms)
-Jun  4 18:57:49.444: INFO: (11) /api/v1/nodes/10.212.23.161/proxy/logs/: 
-alb/
-alternatives.log
-apt/... (200; 15.515404ms)
-Jun  4 18:57:49.459: INFO: (12) /api/v1/nodes/10.212.23.161/proxy/logs/: 
-alb/
-alternatives.log
-apt/... (200; 15.369216ms)
-Jun  4 18:57:49.571: INFO: (13) /api/v1/nodes/10.212.23.161/proxy/logs/: 
-alb/
-alternatives.log
-apt/... (200; 111.655091ms)
-Jun  4 18:57:49.587: INFO: (14) /api/v1/nodes/10.212.23.161/proxy/logs/: 
-alb/
-alternatives.log
-apt/... (200; 15.353434ms)
-Jun  4 18:57:49.602: INFO: (15) /api/v1/nodes/10.212.23.161/proxy/logs/: 
-alb/
-alternatives.log
-apt/... (200; 15.027637ms)
-Jun  4 18:57:49.617: INFO: (16) /api/v1/nodes/10.212.23.161/proxy/logs/: 
-alb/
-alternatives.log
-apt/... (200; 15.826155ms)
-Jun  4 18:57:49.634: INFO: (17) /api/v1/nodes/10.212.23.161/proxy/logs/: 
-alb/
-alternatives.log
-apt/... (200; 16.405442ms)
-Jun  4 18:57:49.651: INFO: (18) /api/v1/nodes/10.212.23.161/proxy/logs/: 
-alb/
-alternatives.log
-apt/... (200; 17.435409ms)
-Jun  4 18:57:49.670: INFO: (19) /api/v1/nodes/10.212.23.161/proxy/logs/: 
-alb/
-alternatives.log
-apt/... (200; 18.471234ms)
-[AfterEach] version v1
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 18:57:49.670: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "proxy-4251" for this suite.
-Jun  4 18:57:55.727: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 18:57:56.173: INFO: namespace proxy-4251 deletion completed in 6.489941389s
+[BeforeEach] [sig-cli] Kubectl client
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:213
+[It] should support --unix-socket=/path  [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Starting the proxy
+Jun 18 11:55:57.574: INFO: Asynchronously running '/usr/local/bin/kubectl kubectl --kubeconfig=/tmp/kubeconfig-426929150 proxy --unix-socket=/tmp/kubectl-proxy-unix227447621/test'
+STEP: retrieving proxy /api/ output
+[AfterEach] [sig-cli] Kubectl client
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 11:55:57.641: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "kubectl-8957" for this suite.
+Jun 18 11:56:03.683: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 11:56:04.076: INFO: namespace kubectl-8957 deletion completed in 6.419474681s
 
-• [SLOW TEST:7.166 seconds]
-[sig-network] Proxy
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:22
-  version v1
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/proxy.go:56
-    should proxy logs on node using proxy subresource  [Conformance]
-    /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+• [SLOW TEST:6.701 seconds]
+[sig-cli] Kubectl client
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23
+  [k8s.io] Proxy server
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
+    should support --unix-socket=/path  [Conformance]
+    /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-SSSSSS
+SSSSSSSSSSS
 ------------------------------
 [sig-api-machinery] Garbage collector 
-  should keep the rc around until all its pods are deleted if the deleteOptions says so [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+  should delete pods created by rc when not orphaning [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 [BeforeEach] [sig-api-machinery] Garbage collector
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 18:57:56.174: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
+Jun 18 11:56:04.077: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
 STEP: Building a namespace api object, basename gc
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in gc-953
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in gc-1457
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should keep the rc around until all its pods are deleted if the deleteOptions says so [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[It] should delete pods created by rc when not orphaning [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 STEP: create the rc
 STEP: delete the rc
-STEP: wait for the rc to be deleted
+STEP: wait for all pods to be garbage collected
 STEP: Gathering metrics
-W0604 18:58:02.551000      18 metrics_grabber.go:79] Master node is not registered. Grabbing metrics from Scheduler, ControllerManager and ClusterAutoscaler is disabled.
-Jun  4 18:58:02.551: INFO: For apiserver_request_total:
+W0618 11:56:14.385336      20 metrics_grabber.go:79] Master node is not registered. Grabbing metrics from Scheduler, ControllerManager and ClusterAutoscaler is disabled.
+Jun 18 11:56:14.385: INFO: For apiserver_request_total:
 For apiserver_request_latencies_summary:
 For apiserver_init_events_total:
 For garbage_collector_attempt_to_delete_queue_latency:
@@ -3504,663 +3058,1096 @@ For errors_total:
 For evicted_pods_total:
 
 [AfterEach] [sig-api-machinery] Garbage collector
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 18:58:02.551: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "gc-953" for this suite.
-Jun  4 18:58:10.609: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 18:58:10.984: INFO: namespace gc-953 deletion completed in 8.420469403s
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 11:56:14.385: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "gc-1457" for this suite.
+Jun 18 11:56:22.441: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 11:56:22.822: INFO: namespace gc-1457 deletion completed in 8.42571235s
 
-• [SLOW TEST:14.810 seconds]
+• [SLOW TEST:18.745 seconds]
 [sig-api-machinery] Garbage collector
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:22
-  should keep the rc around until all its pods are deleted if the deleteOptions says so [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-------------------------------
-SSSSSSSSSSSSSSSSSSSSS
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:22
+  should delete pods created by rc when not orphaning [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-[sig-network] Networking Granular Checks: Pods 
-  should function for node-pod communication: http [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [sig-network] Networking
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+[sig-storage] Secrets 
+  should be consumable from pods in volume with mappings and Item Mode set [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-storage] Secrets
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 18:58:10.985: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-STEP: Building a namespace api object, basename pod-network-test
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in pod-network-test-728
+Jun 18 11:56:22.822: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename secrets
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in secrets-2278
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should function for node-pod communication: http [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: Performing setup for networking test in namespace pod-network-test-728
-STEP: creating a selector
-STEP: Creating the service pods in kubernetes
-Jun  4 18:58:11.202: INFO: Waiting up to 10m0s for all (but 0) nodes to be schedulable
-STEP: Creating test pods
-Jun  4 18:58:33.473: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s --max-time 15 --connect-timeout 1 http://172.30.169.184:8080/hostName | grep -v '^\s*$'] Namespace:pod-network-test-728 PodName:host-test-container-pod ContainerName:hostexec Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
-Jun  4 18:58:33.473: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-Jun  4 18:58:33.655: INFO: Found all expected endpoints: [netserver-0]
-Jun  4 18:58:33.667: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s --max-time 15 --connect-timeout 1 http://172.30.254.172:8080/hostName | grep -v '^\s*$'] Namespace:pod-network-test-728 PodName:host-test-container-pod ContainerName:hostexec Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
-Jun  4 18:58:33.667: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-Jun  4 18:58:33.858: INFO: Found all expected endpoints: [netserver-1]
-Jun  4 18:58:33.870: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s --max-time 15 --connect-timeout 1 http://172.30.193.41:8080/hostName | grep -v '^\s*$'] Namespace:pod-network-test-728 PodName:host-test-container-pod ContainerName:hostexec Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
-Jun  4 18:58:33.870: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-Jun  4 18:58:34.046: INFO: Found all expected endpoints: [netserver-2]
-[AfterEach] [sig-network] Networking
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 18:58:34.046: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "pod-network-test-728" for this suite.
-Jun  4 18:58:58.103: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 18:58:58.436: INFO: namespace pod-network-test-728 deletion completed in 24.374498428s
+[It] should be consumable from pods in volume with mappings and Item Mode set [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating secret with name secret-test-map-17cc20d2-91c0-11e9-a25d-8608290c688a
+STEP: Creating a pod to test consume secrets
+Jun 18 11:56:23.096: INFO: Waiting up to 5m0s for pod "pod-secrets-17cf1edb-91c0-11e9-a25d-8608290c688a" in namespace "secrets-2278" to be "success or failure"
+Jun 18 11:56:23.110: INFO: Pod "pod-secrets-17cf1edb-91c0-11e9-a25d-8608290c688a": Phase="Pending", Reason="", readiness=false. Elapsed: 13.592706ms
+Jun 18 11:56:25.127: INFO: Pod "pod-secrets-17cf1edb-91c0-11e9-a25d-8608290c688a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.031367053s
+STEP: Saw pod success
+Jun 18 11:56:25.128: INFO: Pod "pod-secrets-17cf1edb-91c0-11e9-a25d-8608290c688a" satisfied condition "success or failure"
+Jun 18 11:56:25.141: INFO: Trying to get logs from node 10.72.74.184 pod pod-secrets-17cf1edb-91c0-11e9-a25d-8608290c688a container secret-volume-test: 
+STEP: delete the pod
+Jun 18 11:56:25.222: INFO: Waiting for pod pod-secrets-17cf1edb-91c0-11e9-a25d-8608290c688a to disappear
+Jun 18 11:56:25.241: INFO: Pod pod-secrets-17cf1edb-91c0-11e9-a25d-8608290c688a no longer exists
+[AfterEach] [sig-storage] Secrets
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 11:56:25.241: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "secrets-2278" for this suite.
+Jun 18 11:56:33.283: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 11:56:33.661: INFO: namespace secrets-2278 deletion completed in 8.402133781s
 
-• [SLOW TEST:47.450 seconds]
-[sig-network] Networking
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:25
-  Granular Checks: Pods
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:28
-    should function for node-pod communication: http [LinuxOnly] [NodeConformance] [Conformance]
-    /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-------------------------------
-SSSSSSSSSSSSSSSSSSS
+• [SLOW TEST:10.839 seconds]
+[sig-storage] Secrets
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/secrets_volume.go:33
+  should be consumable from pods in volume with mappings and Item Mode set [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-[sig-node] ConfigMap 
-  should be consumable via the environment [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [sig-node] ConfigMap
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+[sig-storage] Projected configMap 
+  should be consumable from pods in volume [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-storage] Projected configMap
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 18:58:58.437: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-STEP: Building a namespace api object, basename configmap
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in configmap-6039
+Jun 18 11:56:33.661: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename projected
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-8671
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should be consumable via the environment [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: Creating configMap configmap-6039/configmap-test-cf21f691-86fa-11e9-8d1b-467ee19922ac
+[It] should be consumable from pods in volume [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating configMap with name projected-configmap-test-volume-1e3d9753-91c0-11e9-a25d-8608290c688a
 STEP: Creating a pod to test consume configMaps
-Jun  4 18:58:58.689: INFO: Waiting up to 5m0s for pod "pod-configmaps-cf240c5d-86fa-11e9-8d1b-467ee19922ac" in namespace "configmap-6039" to be "success or failure"
-Jun  4 18:58:58.712: INFO: Pod "pod-configmaps-cf240c5d-86fa-11e9-8d1b-467ee19922ac": Phase="Pending", Reason="", readiness=false. Elapsed: 23.457244ms
-Jun  4 18:59:00.725: INFO: Pod "pod-configmaps-cf240c5d-86fa-11e9-8d1b-467ee19922ac": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.036323511s
+Jun 18 11:56:33.913: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-1e414810-91c0-11e9-a25d-8608290c688a" in namespace "projected-8671" to be "success or failure"
+Jun 18 11:56:33.932: INFO: Pod "pod-projected-configmaps-1e414810-91c0-11e9-a25d-8608290c688a": Phase="Pending", Reason="", readiness=false. Elapsed: 19.118019ms
+Jun 18 11:56:35.945: INFO: Pod "pod-projected-configmaps-1e414810-91c0-11e9-a25d-8608290c688a": Phase="Running", Reason="", readiness=true. Elapsed: 2.032750128s
+Jun 18 11:56:37.960: INFO: Pod "pod-projected-configmaps-1e414810-91c0-11e9-a25d-8608290c688a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.046976534s
 STEP: Saw pod success
-Jun  4 18:59:00.725: INFO: Pod "pod-configmaps-cf240c5d-86fa-11e9-8d1b-467ee19922ac" satisfied condition "success or failure"
-Jun  4 18:59:00.738: INFO: Trying to get logs from node 10.212.23.161 pod pod-configmaps-cf240c5d-86fa-11e9-8d1b-467ee19922ac container env-test: 
+Jun 18 11:56:37.960: INFO: Pod "pod-projected-configmaps-1e414810-91c0-11e9-a25d-8608290c688a" satisfied condition "success or failure"
+Jun 18 11:56:37.972: INFO: Trying to get logs from node 10.72.74.138 pod pod-projected-configmaps-1e414810-91c0-11e9-a25d-8608290c688a container projected-configmap-volume-test: 
 STEP: delete the pod
-Jun  4 18:59:00.893: INFO: Waiting for pod pod-configmaps-cf240c5d-86fa-11e9-8d1b-467ee19922ac to disappear
-Jun  4 18:59:00.903: INFO: Pod pod-configmaps-cf240c5d-86fa-11e9-8d1b-467ee19922ac no longer exists
-[AfterEach] [sig-node] ConfigMap
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 18:59:00.903: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "configmap-6039" for this suite.
-Jun  4 18:59:06.988: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 18:59:07.376: INFO: namespace configmap-6039 deletion completed in 6.430896516s
+Jun 18 11:56:38.051: INFO: Waiting for pod pod-projected-configmaps-1e414810-91c0-11e9-a25d-8608290c688a to disappear
+Jun 18 11:56:38.064: INFO: Pod pod-projected-configmaps-1e414810-91c0-11e9-a25d-8608290c688a no longer exists
+[AfterEach] [sig-storage] Projected configMap
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 11:56:38.064: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "projected-8671" for this suite.
+Jun 18 11:56:44.110: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 11:56:44.922: INFO: namespace projected-8671 deletion completed in 6.844259113s
 
-• [SLOW TEST:8.939 seconds]
-[sig-node] ConfigMap
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/configmap.go:32
-  should be consumable via the environment [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+• [SLOW TEST:11.262 seconds]
+[sig-storage] Projected configMap
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_configmap.go:33
+  should be consumable from pods in volume [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-apps] ReplicationController 
-  should release no longer matching pods [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [sig-apps] ReplicationController
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+[sig-apps] ReplicaSet 
+  should adopt matching pods on creation and release no longer matching pods [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-apps] ReplicaSet
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 18:59:07.376: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-STEP: Building a namespace api object, basename replication-controller
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in replication-controller-1636
+Jun 18 11:56:44.927: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename replicaset
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in replicaset-2709
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should release no longer matching pods [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: Given a ReplicationController is created
+[It] should adopt matching pods on creation and release no longer matching pods [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Given a Pod with a 'name' label pod-adoption-release is created
+STEP: When a replicaset with a matching selector is created
+STEP: Then the orphan pod is adopted
 STEP: When the matched label of one of its pods change
-Jun  4 18:59:07.687: INFO: Pod name pod-release: Found 0 pods out of 1
-Jun  4 18:59:12.699: INFO: Pod name pod-release: Found 1 pods out of 1
+Jun 18 11:56:48.240: INFO: Pod name pod-adoption-release: Found 1 pods out of 1
 STEP: Then the pod is released
-[AfterEach] [sig-apps] ReplicationController
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 18:59:12.755: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "replication-controller-1636" for this suite.
-Jun  4 18:59:18.815: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 18:59:19.158: INFO: namespace replication-controller-1636 deletion completed in 6.386162426s
+[AfterEach] [sig-apps] ReplicaSet
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 11:56:49.301: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "replicaset-2709" for this suite.
+Jun 18 11:57:13.348: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 11:57:13.748: INFO: namespace replicaset-2709 deletion completed in 24.428914589s
 
-• [SLOW TEST:11.782 seconds]
-[sig-apps] ReplicationController
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22
-  should release no longer matching pods [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+• [SLOW TEST:28.821 seconds]
+[sig-apps] ReplicaSet
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22
+  should adopt matching pods on creation and release no longer matching pods [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+SSSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[k8s.io] InitContainer [NodeConformance] 
-  should invoke init containers on a RestartNever pod [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [k8s.io] InitContainer [NodeConformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+[sig-storage] Subpath Atomic writer volumes 
+  should support subpaths with downward pod [LinuxOnly] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-storage] Subpath
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 18:59:19.158: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-STEP: Building a namespace api object, basename init-container
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in init-container-162
+Jun 18 11:57:13.749: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename subpath
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in subpath-8933
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [k8s.io] InitContainer [NodeConformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/init_container.go:43
-[It] should invoke init containers on a RestartNever pod [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: creating the pod
-Jun  4 18:59:19.377: INFO: PodSpec: initContainers in spec.initContainers
-[AfterEach] [k8s.io] InitContainer [NodeConformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 18:59:23.592: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "init-container-162" for this suite.
-Jun  4 18:59:29.717: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 18:59:30.062: INFO: namespace init-container-162 deletion completed in 6.391147813s
+[BeforeEach] Atomic writer volumes
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:38
+STEP: Setting up data
+[It] should support subpaths with downward pod [LinuxOnly] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating pod pod-subpath-test-downwardapi-4f6m
+STEP: Creating a pod to test atomic-volume-subpath
+Jun 18 11:57:14.005: INFO: Waiting up to 5m0s for pod "pod-subpath-test-downwardapi-4f6m" in namespace "subpath-8933" to be "success or failure"
+Jun 18 11:57:14.019: INFO: Pod "pod-subpath-test-downwardapi-4f6m": Phase="Pending", Reason="", readiness=false. Elapsed: 14.132057ms
+Jun 18 11:57:16.034: INFO: Pod "pod-subpath-test-downwardapi-4f6m": Phase="Pending", Reason="", readiness=false. Elapsed: 2.028721333s
+Jun 18 11:57:18.048: INFO: Pod "pod-subpath-test-downwardapi-4f6m": Phase="Running", Reason="", readiness=true. Elapsed: 4.042759829s
+Jun 18 11:57:20.062: INFO: Pod "pod-subpath-test-downwardapi-4f6m": Phase="Running", Reason="", readiness=true. Elapsed: 6.056949936s
+Jun 18 11:57:22.075: INFO: Pod "pod-subpath-test-downwardapi-4f6m": Phase="Running", Reason="", readiness=true. Elapsed: 8.070430818s
+Jun 18 11:57:24.089: INFO: Pod "pod-subpath-test-downwardapi-4f6m": Phase="Running", Reason="", readiness=true. Elapsed: 10.083969776s
+Jun 18 11:57:26.223: INFO: Pod "pod-subpath-test-downwardapi-4f6m": Phase="Running", Reason="", readiness=true. Elapsed: 12.21841117s
+Jun 18 11:57:28.237: INFO: Pod "pod-subpath-test-downwardapi-4f6m": Phase="Running", Reason="", readiness=true. Elapsed: 14.231913917s
+Jun 18 11:57:30.250: INFO: Pod "pod-subpath-test-downwardapi-4f6m": Phase="Running", Reason="", readiness=true. Elapsed: 16.245619249s
+Jun 18 11:57:32.264: INFO: Pod "pod-subpath-test-downwardapi-4f6m": Phase="Running", Reason="", readiness=true. Elapsed: 18.259087212s
+Jun 18 11:57:34.279: INFO: Pod "pod-subpath-test-downwardapi-4f6m": Phase="Running", Reason="", readiness=true. Elapsed: 20.273797716s
+Jun 18 11:57:36.293: INFO: Pod "pod-subpath-test-downwardapi-4f6m": Phase="Succeeded", Reason="", readiness=false. Elapsed: 22.287738032s
+STEP: Saw pod success
+Jun 18 11:57:36.293: INFO: Pod "pod-subpath-test-downwardapi-4f6m" satisfied condition "success or failure"
+Jun 18 11:57:36.306: INFO: Trying to get logs from node 10.72.74.184 pod pod-subpath-test-downwardapi-4f6m container test-container-subpath-downwardapi-4f6m: 
+STEP: delete the pod
+Jun 18 11:57:36.391: INFO: Waiting for pod pod-subpath-test-downwardapi-4f6m to disappear
+Jun 18 11:57:36.403: INFO: Pod pod-subpath-test-downwardapi-4f6m no longer exists
+STEP: Deleting pod pod-subpath-test-downwardapi-4f6m
+Jun 18 11:57:36.403: INFO: Deleting pod "pod-subpath-test-downwardapi-4f6m" in namespace "subpath-8933"
+[AfterEach] [sig-storage] Subpath
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 11:57:36.417: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "subpath-8933" for this suite.
+Jun 18 11:57:42.458: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 11:57:43.086: INFO: namespace subpath-8933 deletion completed in 6.655892746s
 
-• [SLOW TEST:10.904 seconds]
-[k8s.io] InitContainer [NodeConformance]
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
-  should invoke init containers on a RestartNever pod [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+• [SLOW TEST:29.338 seconds]
+[sig-storage] Subpath
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:22
+  Atomic writer volumes
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:34
+    should support subpaths with downward pod [LinuxOnly] [Conformance]
+    /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-SSSSSS
+S
 ------------------------------
-[sig-cli] Kubectl client [k8s.io] Kubectl run pod 
-  should create a pod from an image when restart is Never  [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[sig-cli] Kubectl client [k8s.io] Proxy server 
+  should support proxy with --port 0  [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 [BeforeEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 18:59:30.063: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
+Jun 18 11:57:43.088: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
 STEP: Building a namespace api object, basename kubectl
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-4274
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-209
 STEP: Waiting for a default service account to be provisioned in namespace
 [BeforeEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:213
-[BeforeEach] [k8s.io] Kubectl run pod
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1583
-[It] should create a pod from an image when restart is Never  [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: running the image docker.io/library/nginx:1.14-alpine
-Jun  4 18:59:30.373: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 run e2e-test-nginx-pod --restart=Never --generator=run-pod/v1 --image=docker.io/library/nginx:1.14-alpine --namespace=kubectl-4274'
-Jun  4 18:59:30.544: INFO: stderr: ""
-Jun  4 18:59:30.544: INFO: stdout: "pod/e2e-test-nginx-pod created\n"
-STEP: verifying the pod e2e-test-nginx-pod was created
-[AfterEach] [k8s.io] Kubectl run pod
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1588
-Jun  4 18:59:30.555: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 delete pods e2e-test-nginx-pod --namespace=kubectl-4274'
-Jun  4 18:59:46.893: INFO: stderr: ""
-Jun  4 18:59:46.893: INFO: stdout: "pod \"e2e-test-nginx-pod\" deleted\n"
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:213
+[It] should support proxy with --port 0  [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: starting the proxy server
+Jun 18 11:57:43.294: INFO: Asynchronously running '/usr/local/bin/kubectl kubectl --kubeconfig=/tmp/kubeconfig-426929150 proxy -p 0 --disable-filter'
+STEP: curling proxy /api/ output
 [AfterEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 18:59:46.893: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "kubectl-4274" for this suite.
-Jun  4 18:59:52.984: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 18:59:53.405: INFO: namespace kubectl-4274 deletion completed in 6.495672735s
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 11:57:43.395: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "kubectl-209" for this suite.
+Jun 18 11:57:49.438: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 11:57:50.046: INFO: namespace kubectl-209 deletion completed in 6.636672436s
 
-• [SLOW TEST:23.342 seconds]
+• [SLOW TEST:6.959 seconds]
 [sig-cli] Kubectl client
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23
-  [k8s.io] Kubectl run pod
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
-    should create a pod from an image when restart is Never  [Conformance]
-    /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23
+  [k8s.io] Proxy server
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
+    should support proxy with --port 0  [Conformance]
+    /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-SSSSSSSSS
+SSSSS
 ------------------------------
-[sig-storage] Projected downwardAPI 
-  should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [sig-storage] Projected downwardAPI
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+[sig-storage] Projected secret 
+  should be consumable from pods in volume with mappings and Item Mode set [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-storage] Projected secret
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 18:59:53.405: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
+Jun 18 11:57:50.046: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
 STEP: Building a namespace api object, basename projected
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-3705
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-7463
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-storage] Projected downwardAPI
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:39
-[It] should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: Creating a pod to test downward API volume plugin
-Jun  4 18:59:53.792: INFO: Waiting up to 5m0s for pod "downwardapi-volume-effbbc5e-86fa-11e9-8d1b-467ee19922ac" in namespace "projected-3705" to be "success or failure"
-Jun  4 18:59:53.803: INFO: Pod "downwardapi-volume-effbbc5e-86fa-11e9-8d1b-467ee19922ac": Phase="Pending", Reason="", readiness=false. Elapsed: 11.354245ms
-Jun  4 18:59:55.816: INFO: Pod "downwardapi-volume-effbbc5e-86fa-11e9-8d1b-467ee19922ac": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.024790498s
+[It] should be consumable from pods in volume with mappings and Item Mode set [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating projection with secret that has name projected-secret-test-map-4bc50d95-91c0-11e9-a25d-8608290c688a
+STEP: Creating a pod to test consume secrets
+Jun 18 11:57:50.285: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-4bc6f28a-91c0-11e9-a25d-8608290c688a" in namespace "projected-7463" to be "success or failure"
+Jun 18 11:57:50.303: INFO: Pod "pod-projected-secrets-4bc6f28a-91c0-11e9-a25d-8608290c688a": Phase="Pending", Reason="", readiness=false. Elapsed: 17.78267ms
+Jun 18 11:57:52.317: INFO: Pod "pod-projected-secrets-4bc6f28a-91c0-11e9-a25d-8608290c688a": Phase="Pending", Reason="", readiness=false. Elapsed: 2.031570864s
+Jun 18 11:57:54.330: INFO: Pod "pod-projected-secrets-4bc6f28a-91c0-11e9-a25d-8608290c688a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.045206566s
 STEP: Saw pod success
-Jun  4 18:59:55.816: INFO: Pod "downwardapi-volume-effbbc5e-86fa-11e9-8d1b-467ee19922ac" satisfied condition "success or failure"
-Jun  4 18:59:55.829: INFO: Trying to get logs from node 10.212.23.164 pod downwardapi-volume-effbbc5e-86fa-11e9-8d1b-467ee19922ac container client-container: 
+Jun 18 11:57:54.330: INFO: Pod "pod-projected-secrets-4bc6f28a-91c0-11e9-a25d-8608290c688a" satisfied condition "success or failure"
+Jun 18 11:57:54.344: INFO: Trying to get logs from node 10.72.74.138 pod pod-projected-secrets-4bc6f28a-91c0-11e9-a25d-8608290c688a container projected-secret-volume-test: 
 STEP: delete the pod
-Jun  4 18:59:55.891: INFO: Waiting for pod downwardapi-volume-effbbc5e-86fa-11e9-8d1b-467ee19922ac to disappear
-Jun  4 18:59:55.901: INFO: Pod downwardapi-volume-effbbc5e-86fa-11e9-8d1b-467ee19922ac no longer exists
-[AfterEach] [sig-storage] Projected downwardAPI
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 18:59:55.902: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "projected-3705" for this suite.
-Jun  4 19:00:01.958: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 19:00:02.386: INFO: namespace projected-3705 deletion completed in 6.470483765s
+Jun 18 11:57:54.438: INFO: Waiting for pod pod-projected-secrets-4bc6f28a-91c0-11e9-a25d-8608290c688a to disappear
+Jun 18 11:57:54.453: INFO: Pod pod-projected-secrets-4bc6f28a-91c0-11e9-a25d-8608290c688a no longer exists
+[AfterEach] [sig-storage] Projected secret
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 11:57:54.453: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "projected-7463" for this suite.
+Jun 18 11:58:00.497: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 11:58:01.337: INFO: namespace projected-7463 deletion completed in 6.871124237s
 
-• [SLOW TEST:8.981 seconds]
-[sig-storage] Projected downwardAPI
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:33
-  should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+• [SLOW TEST:11.291 seconds]
+[sig-storage] Projected secret
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_secret.go:33
+  should be consumable from pods in volume with mappings and Item Mode set [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-SSSSSSS
+SSSSSSSSSSSS
 ------------------------------
-[sig-cli] Kubectl client [k8s.io] Kubectl api-versions 
-  should check if v1 is in available api versions  [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+[sig-scheduling] SchedulerPredicates [Serial] 
+  validates that NodeSelector is respected if not matching  [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-scheduling] SchedulerPredicates [Serial]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 19:00:02.387: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-STEP: Building a namespace api object, basename kubectl
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-3324
+Jun 18 11:58:01.337: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename sched-pred
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in sched-pred-8916
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:213
-[It] should check if v1 is in available api versions  [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: validating api versions
-Jun  4 19:00:02.605: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 api-versions'
-Jun  4 19:00:02.690: INFO: stderr: ""
-Jun  4 19:00:02.690: INFO: stdout: "admissionregistration.k8s.io/v1beta1\napiextensions.k8s.io/v1beta1\napiregistration.k8s.io/v1\napiregistration.k8s.io/v1beta1\napps/v1\napps/v1beta1\napps/v1beta2\nauthentication.k8s.io/v1\nauthentication.k8s.io/v1beta1\nauthorization.k8s.io/v1\nauthorization.k8s.io/v1beta1\nautoscaling/v1\nautoscaling/v2beta1\nautoscaling/v2beta2\nbatch/v1\nbatch/v1beta1\nbatch/v2alpha1\ncertificates.k8s.io/v1beta1\ncoordination.k8s.io/v1\ncoordination.k8s.io/v1beta1\nevents.k8s.io/v1beta1\nextensions/v1beta1\nmetrics.k8s.io/v1beta1\nnetworking.k8s.io/v1\nnetworking.k8s.io/v1beta1\npolicy/v1beta1\nrbac.authorization.k8s.io/v1\nrbac.authorization.k8s.io/v1beta1\nscheduling.k8s.io/v1\nscheduling.k8s.io/v1beta1\nstorage.k8s.io/v1\nstorage.k8s.io/v1beta1\nv1\n"
-[AfterEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 19:00:02.690: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "kubectl-3324" for this suite.
-Jun  4 19:00:08.764: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 19:00:09.208: INFO: namespace kubectl-3324 deletion completed in 6.501265493s
-
-• [SLOW TEST:6.821 seconds]
-[sig-cli] Kubectl client
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23
-  [k8s.io] Kubectl api-versions
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
-    should check if v1 is in available api versions  [Conformance]
-    /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-------------------------------
-[k8s.io] [sig-node] Events 
-  should be sent by kubelets and the scheduler about pods scheduling and running  [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [k8s.io] [sig-node] Events
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+[BeforeEach] [sig-scheduling] SchedulerPredicates [Serial]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:79
+Jun 18 11:58:01.534: INFO: Waiting up to 1m0s for all (but 0) nodes to be ready
+Jun 18 11:58:01.560: INFO: Waiting for terminating namespaces to be deleted...
+Jun 18 11:58:01.571: INFO: 
+Logging pods the kubelet thinks is on node 10.72.74.138 before test
+Jun 18 11:58:01.607: INFO: ibm-cloud-provider-ip-158-176-120-140-d8f5f45f5-4tksx from ibm-system started at 2019-06-17 21:41:02 +0000 UTC (1 container statuses recorded)
+Jun 18 11:58:01.607: INFO: 	Container ibm-cloud-provider-ip-158-176-120-140 ready: true, restart count 0
+Jun 18 11:58:01.607: INFO: public-crd9ae0c6aaa554167a4f1d2c76b9ad9f3-alb1-fd9fb76d-q646z from kube-system started at 2019-06-17 21:41:07 +0000 UTC (4 container statuses recorded)
+Jun 18 11:58:01.607: INFO: 	Container ingress-auth-1 ready: true, restart count 0
+Jun 18 11:58:01.607: INFO: 	Container ingress-auth-2 ready: true, restart count 0
+Jun 18 11:58:01.607: INFO: 	Container ingress-auth-3 ready: true, restart count 0
+Jun 18 11:58:01.607: INFO: 	Container nginx-ingress ready: true, restart count 0
+Jun 18 11:58:01.607: INFO: ibm-keepalived-watcher-4l788 from kube-system started at 2019-06-17 21:37:25 +0000 UTC (1 container statuses recorded)
+Jun 18 11:58:01.607: INFO: 	Container keepalived-watcher ready: true, restart count 0
+Jun 18 11:58:01.607: INFO: ibm-kube-fluentd-gmp54 from kube-system started at 2019-06-17 21:37:34 +0000 UTC (1 container statuses recorded)
+Jun 18 11:58:01.607: INFO: 	Container fluentd ready: true, restart count 0
+Jun 18 11:58:01.607: INFO: sonobuoy-systemd-logs-daemon-set-376e58a3dd534c11-ccq8b from heptio-sonobuoy started at 2019-06-18 11:33:56 +0000 UTC (2 container statuses recorded)
+Jun 18 11:58:01.607: INFO: 	Container sonobuoy-worker ready: true, restart count 0
+Jun 18 11:58:01.608: INFO: 	Container systemd-logs ready: true, restart count 0
+Jun 18 11:58:01.608: INFO: ibm-master-proxy-static-10.72.74.138 from kube-system started at  (0 container statuses recorded)
+Jun 18 11:58:01.608: INFO: calico-node-prxhz from kube-system started at 2019-06-17 21:37:25 +0000 UTC (1 container statuses recorded)
+Jun 18 11:58:01.608: INFO: 	Container calico-node ready: true, restart count 0
+Jun 18 11:58:01.608: INFO: 
+Logging pods the kubelet thinks is on node 10.72.74.184 before test
+Jun 18 11:58:01.650: INFO: ibm-master-proxy-static-10.72.74.184 from kube-system started at  (0 container statuses recorded)
+Jun 18 11:58:01.650: INFO: test-k8s-e2e-pvg-master-verification from default started at 2019-06-18 11:33:41 +0000 UTC (1 container statuses recorded)
+Jun 18 11:58:01.650: INFO: 	Container test-k8s-e2e-pvg-master-verification ready: true, restart count 0
+Jun 18 11:58:01.650: INFO: sonobuoy-systemd-logs-daemon-set-376e58a3dd534c11-x9mmr from heptio-sonobuoy started at 2019-06-18 11:33:56 +0000 UTC (2 container statuses recorded)
+Jun 18 11:58:01.650: INFO: 	Container sonobuoy-worker ready: true, restart count 0
+Jun 18 11:58:01.650: INFO: 	Container systemd-logs ready: true, restart count 0
+Jun 18 11:58:01.650: INFO: calico-node-2jll6 from kube-system started at 2019-06-17 21:37:03 +0000 UTC (1 container statuses recorded)
+Jun 18 11:58:01.650: INFO: 	Container calico-node ready: true, restart count 0
+Jun 18 11:58:01.650: INFO: ibm-kube-fluentd-66f2t from kube-system started at 2019-06-17 21:37:34 +0000 UTC (1 container statuses recorded)
+Jun 18 11:58:01.650: INFO: 	Container fluentd ready: true, restart count 0
+Jun 18 11:58:01.650: INFO: ibm-cloud-provider-ip-158-176-120-140-d8f5f45f5-kszmp from ibm-system started at 2019-06-17 21:41:02 +0000 UTC (1 container statuses recorded)
+Jun 18 11:58:01.650: INFO: 	Container ibm-cloud-provider-ip-158-176-120-140 ready: true, restart count 0
+Jun 18 11:58:01.650: INFO: sonobuoy from heptio-sonobuoy started at 2019-06-18 11:33:48 +0000 UTC (1 container statuses recorded)
+Jun 18 11:58:01.650: INFO: 	Container kube-sonobuoy ready: true, restart count 0
+Jun 18 11:58:01.650: INFO: metrics-server-5cdb4c5b-67974 from kube-system started at 2019-06-17 21:37:27 +0000 UTC (2 container statuses recorded)
+Jun 18 11:58:01.650: INFO: 	Container metrics-server ready: true, restart count 0
+Jun 18 11:58:01.650: INFO: 	Container metrics-server-nanny ready: true, restart count 0
+Jun 18 11:58:01.650: INFO: ibm-keepalived-watcher-xkcvg from kube-system started at 2019-06-17 21:37:03 +0000 UTC (1 container statuses recorded)
+Jun 18 11:58:01.650: INFO: 	Container keepalived-watcher ready: true, restart count 0
+Jun 18 11:58:01.650: INFO: coredns-78cff85d65-jvwmh from kube-system started at 2019-06-17 21:37:29 +0000 UTC (1 container statuses recorded)
+Jun 18 11:58:01.650: INFO: 	Container coredns ready: true, restart count 0
+Jun 18 11:58:01.650: INFO: 
+Logging pods the kubelet thinks is on node 10.72.74.189 before test
+Jun 18 11:58:02.750: INFO: ibm-master-proxy-static-10.72.74.189 from kube-system started at  (0 container statuses recorded)
+Jun 18 11:58:02.750: INFO: ibm-keepalived-watcher-klm4c from kube-system started at 2019-06-17 21:36:59 +0000 UTC (1 container statuses recorded)
+Jun 18 11:58:02.750: INFO: 	Container keepalived-watcher ready: true, restart count 0
+Jun 18 11:58:02.750: INFO: vpn-cf6ff59b-gqxpw from kube-system started at 2019-06-17 21:37:09 +0000 UTC (1 container statuses recorded)
+Jun 18 11:58:02.750: INFO: 	Container vpn ready: true, restart count 0
+Jun 18 11:58:02.750: INFO: kubernetes-dashboard-6f5f8c6896-rbvmt from kube-system started at 2019-06-17 21:37:09 +0000 UTC (1 container statuses recorded)
+Jun 18 11:58:02.750: INFO: 	Container kubernetes-dashboard ready: true, restart count 0
+Jun 18 11:58:02.750: INFO: sonobuoy-e2e-job-68f893029cbd431f from heptio-sonobuoy started at 2019-06-18 11:33:56 +0000 UTC (2 container statuses recorded)
+Jun 18 11:58:02.750: INFO: 	Container e2e ready: true, restart count 0
+Jun 18 11:58:02.750: INFO: 	Container sonobuoy-worker ready: true, restart count 0
+Jun 18 11:58:02.750: INFO: public-crd9ae0c6aaa554167a4f1d2c76b9ad9f3-alb1-fd9fb76d-sf25n from kube-system started at 2019-06-17 21:41:07 +0000 UTC (4 container statuses recorded)
+Jun 18 11:58:02.750: INFO: 	Container ingress-auth-1 ready: true, restart count 0
+Jun 18 11:58:02.750: INFO: 	Container ingress-auth-2 ready: true, restart count 0
+Jun 18 11:58:02.750: INFO: 	Container ingress-auth-3 ready: true, restart count 0
+Jun 18 11:58:02.750: INFO: 	Container nginx-ingress ready: true, restart count 0
+Jun 18 11:58:02.750: INFO: ibm-storage-watcher-964758dd-rsgxw from kube-system started at 2019-06-17 21:37:09 +0000 UTC (1 container statuses recorded)
+Jun 18 11:58:02.750: INFO: 	Container ibm-storage-watcher-container ready: true, restart count 0
+Jun 18 11:58:02.750: INFO: coredns-78cff85d65-2mm72 from kube-system started at 2019-06-17 21:37:09 +0000 UTC (1 container statuses recorded)
+Jun 18 11:58:02.750: INFO: 	Container coredns ready: true, restart count 0
+Jun 18 11:58:02.750: INFO: calico-kube-controllers-85d6c48f7-ggvwd from kube-system started at 2019-06-17 21:37:09 +0000 UTC (1 container statuses recorded)
+Jun 18 11:58:02.750: INFO: 	Container calico-kube-controllers ready: true, restart count 0
+Jun 18 11:58:02.750: INFO: coredns-autoscaler-6854575d6-p69qn from kube-system started at 2019-06-17 21:37:09 +0000 UTC (1 container statuses recorded)
+Jun 18 11:58:02.750: INFO: 	Container autoscaler ready: true, restart count 0
+Jun 18 11:58:02.750: INFO: calico-node-htlqc from kube-system started at 2019-06-17 21:36:59 +0000 UTC (1 container statuses recorded)
+Jun 18 11:58:02.750: INFO: 	Container calico-node ready: true, restart count 0
+Jun 18 11:58:02.750: INFO: ibm-file-plugin-98f5986d7-bbzfx from kube-system started at 2019-06-17 21:37:09 +0000 UTC (1 container statuses recorded)
+Jun 18 11:58:02.750: INFO: 	Container ibm-file-plugin-container ready: true, restart count 0
+Jun 18 11:58:02.750: INFO: ibm-kube-fluentd-qg9mz from kube-system started at 2019-06-17 21:37:34 +0000 UTC (1 container statuses recorded)
+Jun 18 11:58:02.750: INFO: 	Container fluentd ready: true, restart count 0
+Jun 18 11:58:02.750: INFO: sonobuoy-systemd-logs-daemon-set-376e58a3dd534c11-rn8dw from heptio-sonobuoy started at 2019-06-18 11:33:56 +0000 UTC (2 container statuses recorded)
+Jun 18 11:58:02.750: INFO: 	Container sonobuoy-worker ready: true, restart count 0
+Jun 18 11:58:02.750: INFO: 	Container systemd-logs ready: true, restart count 0
+[It] validates that NodeSelector is respected if not matching  [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Trying to schedule Pod with nonempty NodeSelector.
+STEP: Considering event: 
+Type = [Warning], Name = [restricted-pod.15a94910e897ed06], Reason = [FailedScheduling], Message = [0/3 nodes are available: 3 node(s) didn't match node selector.]
+[AfterEach] [sig-scheduling] SchedulerPredicates [Serial]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 11:58:03.845: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "sched-pred-8916" for this suite.
+Jun 18 11:58:09.891: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 11:58:10.299: INFO: namespace sched-pred-8916 deletion completed in 6.436836167s
+[AfterEach] [sig-scheduling] SchedulerPredicates [Serial]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:70
+
+• [SLOW TEST:8.962 seconds]
+[sig-scheduling] SchedulerPredicates [Serial]
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/framework.go:22
+  validates that NodeSelector is respected if not matching  [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSSSSSS
+------------------------------
+[k8s.io] Variable Expansion 
+  should allow substituting values in a container's command [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [k8s.io] Variable Expansion
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 19:00:09.208: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-STEP: Building a namespace api object, basename events
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in events-4094
+Jun 18 11:58:10.301: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename var-expansion
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in var-expansion-6228
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should be sent by kubelets and the scheduler about pods scheduling and running  [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: creating the pod
-STEP: submitting the pod to kubernetes
-STEP: verifying the pod is in kubernetes
-STEP: retrieving the pod
-Jun  4 19:00:11.491: INFO: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:send-events-f950df68-86fa-11e9-8d1b-467ee19922ac,GenerateName:,Namespace:events-4094,SelfLink:/api/v1/namespaces/events-4094/pods/send-events-f950df68-86fa-11e9-8d1b-467ee19922ac,UID:f952267c-86fa-11e9-8318-1e5386706511,ResourceVersion:15603,Generation:0,CreationTimestamp:2019-06-04 19:00:09 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: foo,time: 428366750,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-974t7 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-974t7,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{p gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1 [] []  [{ 0 80 TCP }] [] [] {map[] map[]} [{default-token-974t7 true /var/run/secrets/kubernetes.io/serviceaccount   }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*30,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.212.23.189,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists  NoExecute 0xc003198f70} {node.kubernetes.io/unreachable Exists  NoExecute 0xc003198f90}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Running,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 19:00:09 +0000 UTC  } {Ready True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 19:00:11 +0000 UTC  } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 19:00:11 +0000 UTC  } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 19:00:09 +0000 UTC  }],Message:,Reason:,HostIP:10.212.23.189,PodIP:172.30.193.43,StartTime:2019-06-04 19:00:09 +0000 UTC,ContainerStatuses:[{p {nil ContainerStateRunning{StartedAt:2019-06-04 19:00:10 +0000 UTC,} nil} {nil nil nil} true 0 gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1 gcr.io/kubernetes-e2e-test-images/serve-hostname@sha256:bab70473a6d8ef65a22625dc9a1b0f0452e811530fdbe77e4408523460177ff1 containerd://1376d35fb9bdaa41af0756d5790b057ff65046eedf34f6540ebab05c6b7ad15b}],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},}
-
-STEP: checking for scheduler event about the pod
-Jun  4 19:00:13.502: INFO: Saw scheduler event for our pod.
-STEP: checking for kubelet event about the pod
-Jun  4 19:00:15.511: INFO: Saw kubelet event for our pod.
-STEP: deleting the pod
-[AfterEach] [k8s.io] [sig-node] Events
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 19:00:15.532: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "events-4094" for this suite.
-Jun  4 19:00:55.604: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 19:00:55.990: INFO: namespace events-4094 deletion completed in 40.428896671s
+[It] should allow substituting values in a container's command [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating a pod to test substitution in container's command
+Jun 18 11:58:10.527: INFO: Waiting up to 5m0s for pod "var-expansion-57d74487-91c0-11e9-a25d-8608290c688a" in namespace "var-expansion-6228" to be "success or failure"
+Jun 18 11:58:10.550: INFO: Pod "var-expansion-57d74487-91c0-11e9-a25d-8608290c688a": Phase="Pending", Reason="", readiness=false. Elapsed: 22.588718ms
+Jun 18 11:58:12.564: INFO: Pod "var-expansion-57d74487-91c0-11e9-a25d-8608290c688a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.036379811s
+STEP: Saw pod success
+Jun 18 11:58:12.564: INFO: Pod "var-expansion-57d74487-91c0-11e9-a25d-8608290c688a" satisfied condition "success or failure"
+Jun 18 11:58:12.577: INFO: Trying to get logs from node 10.72.74.184 pod var-expansion-57d74487-91c0-11e9-a25d-8608290c688a container dapi-container: 
+STEP: delete the pod
+Jun 18 11:58:12.662: INFO: Waiting for pod var-expansion-57d74487-91c0-11e9-a25d-8608290c688a to disappear
+Jun 18 11:58:12.681: INFO: Pod var-expansion-57d74487-91c0-11e9-a25d-8608290c688a no longer exists
+[AfterEach] [k8s.io] Variable Expansion
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 11:58:12.681: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "var-expansion-6228" for this suite.
+Jun 18 11:58:18.733: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 11:58:19.127: INFO: namespace var-expansion-6228 deletion completed in 6.432729952s
 
-• [SLOW TEST:46.782 seconds]
-[k8s.io] [sig-node] Events
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
-  should be sent by kubelets and the scheduler about pods scheduling and running  [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+• [SLOW TEST:8.827 seconds]
+[k8s.io] Variable Expansion
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
+  should allow substituting values in a container's command [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-SSSSSSS
+SSSSSSSSS
 ------------------------------
-[sig-storage] ConfigMap 
-  optional updates should be reflected in volume [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [sig-storage] ConfigMap
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+[sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] 
+  Burst scaling should run to completion even with unhealthy pods [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-apps] StatefulSet
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 19:00:55.991: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-STEP: Building a namespace api object, basename configmap
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in configmap-7537
+Jun 18 11:58:19.128: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename statefulset
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in statefulset-6365
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] optional updates should be reflected in volume [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: Creating configMap with name cm-test-opt-del-153542b9-86fb-11e9-8d1b-467ee19922ac
-STEP: Creating configMap with name cm-test-opt-upd-153542f0-86fb-11e9-8d1b-467ee19922ac
-STEP: Creating the pod
-STEP: Deleting configmap cm-test-opt-del-153542b9-86fb-11e9-8d1b-467ee19922ac
-STEP: Updating configmap cm-test-opt-upd-153542f0-86fb-11e9-8d1b-467ee19922ac
-STEP: Creating configMap with name cm-test-opt-create-1535430f-86fb-11e9-8d1b-467ee19922ac
-STEP: waiting to observe update in volume
-[AfterEach] [sig-storage] ConfigMap
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 19:02:28.079: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "configmap-7537" for this suite.
-Jun  4 19:02:52.132: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 19:02:52.483: INFO: namespace configmap-7537 deletion completed in 24.390285645s
+[BeforeEach] [sig-apps] StatefulSet
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:59
+[BeforeEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:74
+STEP: Creating service test in namespace statefulset-6365
+[It] Burst scaling should run to completion even with unhealthy pods [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating stateful set ss in namespace statefulset-6365
+STEP: Waiting until all stateful set ss replicas will be running in namespace statefulset-6365
+Jun 18 11:58:19.367: INFO: Found 0 stateful pods, waiting for 1
+Jun 18 11:58:29.382: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true
+STEP: Confirming that stateful set scale up will not halt with unhealthy stateful pod
+Jun 18 11:58:29.396: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-6365 ss-0 -- /bin/sh -x -c mv -v /usr/share/nginx/html/index.html /tmp/ || true'
+Jun 18 11:58:30.571: INFO: stderr: "+ mv -v /usr/share/nginx/html/index.html /tmp/\n"
+Jun 18 11:58:30.571: INFO: stdout: "'/usr/share/nginx/html/index.html' -> '/tmp/index.html'\n"
+Jun 18 11:58:30.571: INFO: stdout of mv -v /usr/share/nginx/html/index.html /tmp/ || true on ss-0: '/usr/share/nginx/html/index.html' -> '/tmp/index.html'
+
+Jun 18 11:58:30.584: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=true
+Jun 18 11:58:40.599: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=false
+Jun 18 11:58:40.600: INFO: Waiting for statefulset status.replicas updated to 0
+Jun 18 11:58:40.650: INFO: POD   NODE          PHASE    GRACE  CONDITIONS
+Jun 18 11:58:40.650: INFO: ss-0  10.72.74.138  Running         [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:58:19 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:58:31 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:58:31 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:58:19 +0000 UTC  }]
+Jun 18 11:58:40.650: INFO: 
+Jun 18 11:58:40.650: INFO: StatefulSet ss has not reached scale 3, at 1
+Jun 18 11:58:41.664: INFO: Verifying statefulset ss doesn't scale past 3 for another 8.986945539s
+Jun 18 11:58:42.678: INFO: Verifying statefulset ss doesn't scale past 3 for another 7.972699239s
+Jun 18 11:58:43.694: INFO: Verifying statefulset ss doesn't scale past 3 for another 6.958526257s
+Jun 18 11:58:44.709: INFO: Verifying statefulset ss doesn't scale past 3 for another 5.942646894s
+Jun 18 11:58:45.724: INFO: Verifying statefulset ss doesn't scale past 3 for another 4.928018775s
+Jun 18 11:58:46.738: INFO: Verifying statefulset ss doesn't scale past 3 for another 3.912241975s
+Jun 18 11:58:47.753: INFO: Verifying statefulset ss doesn't scale past 3 for another 2.898559531s
+Jun 18 11:58:48.767: INFO: Verifying statefulset ss doesn't scale past 3 for another 1.88403401s
+Jun 18 11:58:49.780: INFO: Verifying statefulset ss doesn't scale past 3 for another 869.668633ms
+STEP: Scaling up stateful set ss to 3 replicas and waiting until all of them will be running in namespace statefulset-6365
+Jun 18 11:58:50.794: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-6365 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
+Jun 18 11:58:51.129: INFO: stderr: "+ mv -v /tmp/index.html /usr/share/nginx/html/\n"
+Jun 18 11:58:51.129: INFO: stdout: "'/tmp/index.html' -> '/usr/share/nginx/html/index.html'\n"
+Jun 18 11:58:51.129: INFO: stdout of mv -v /tmp/index.html /usr/share/nginx/html/ || true on ss-0: '/tmp/index.html' -> '/usr/share/nginx/html/index.html'
+
+Jun 18 11:58:51.129: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-6365 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
+Jun 18 11:58:51.537: INFO: stderr: "+ mv -v /tmp/index.html /usr/share/nginx/html/\nmv: can't rename '/tmp/index.html': No such file or directory\n+ true\n"
+Jun 18 11:58:51.537: INFO: stdout: "'/tmp/index.html' -> '/usr/share/nginx/html/index.html'\n"
+Jun 18 11:58:51.537: INFO: stdout of mv -v /tmp/index.html /usr/share/nginx/html/ || true on ss-1: '/tmp/index.html' -> '/usr/share/nginx/html/index.html'
+
+Jun 18 11:58:51.537: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-6365 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
+Jun 18 11:58:52.432: INFO: stderr: "+ mv -v /tmp/index.html /usr/share/nginx/html/\nmv: can't rename '/tmp/index.html': No such file or directory\n+ true\n"
+Jun 18 11:58:52.432: INFO: stdout: "'/tmp/index.html' -> '/usr/share/nginx/html/index.html'\n"
+Jun 18 11:58:52.432: INFO: stdout of mv -v /tmp/index.html /usr/share/nginx/html/ || true on ss-2: '/tmp/index.html' -> '/usr/share/nginx/html/index.html'
+
+Jun 18 11:58:52.445: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true
+Jun 18 11:58:52.445: INFO: Waiting for pod ss-1 to enter Running - Ready=true, currently Running - Ready=true
+Jun 18 11:58:52.445: INFO: Waiting for pod ss-2 to enter Running - Ready=true, currently Running - Ready=true
+STEP: Scale down will not halt with unhealthy stateful pod
+Jun 18 11:58:52.458: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-6365 ss-0 -- /bin/sh -x -c mv -v /usr/share/nginx/html/index.html /tmp/ || true'
+Jun 18 11:58:52.777: INFO: stderr: "+ mv -v /usr/share/nginx/html/index.html /tmp/\n"
+Jun 18 11:58:52.777: INFO: stdout: "'/usr/share/nginx/html/index.html' -> '/tmp/index.html'\n"
+Jun 18 11:58:52.777: INFO: stdout of mv -v /usr/share/nginx/html/index.html /tmp/ || true on ss-0: '/usr/share/nginx/html/index.html' -> '/tmp/index.html'
+
+Jun 18 11:58:52.777: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-6365 ss-1 -- /bin/sh -x -c mv -v /usr/share/nginx/html/index.html /tmp/ || true'
+Jun 18 11:58:53.119: INFO: stderr: "+ mv -v /usr/share/nginx/html/index.html /tmp/\n"
+Jun 18 11:58:53.119: INFO: stdout: "'/usr/share/nginx/html/index.html' -> '/tmp/index.html'\n"
+Jun 18 11:58:53.119: INFO: stdout of mv -v /usr/share/nginx/html/index.html /tmp/ || true on ss-1: '/usr/share/nginx/html/index.html' -> '/tmp/index.html'
+
+Jun 18 11:58:53.119: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-6365 ss-2 -- /bin/sh -x -c mv -v /usr/share/nginx/html/index.html /tmp/ || true'
+Jun 18 11:58:53.479: INFO: stderr: "+ mv -v /usr/share/nginx/html/index.html /tmp/\n"
+Jun 18 11:58:53.479: INFO: stdout: "'/usr/share/nginx/html/index.html' -> '/tmp/index.html'\n"
+Jun 18 11:58:53.479: INFO: stdout of mv -v /usr/share/nginx/html/index.html /tmp/ || true on ss-2: '/usr/share/nginx/html/index.html' -> '/tmp/index.html'
+
+Jun 18 11:58:53.479: INFO: Waiting for statefulset status.replicas updated to 0
+Jun 18 11:58:53.491: INFO: Waiting for stateful set status.readyReplicas to become 0, currently 1
+Jun 18 11:59:03.521: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=false
+Jun 18 11:59:03.521: INFO: Waiting for pod ss-1 to enter Running - Ready=false, currently Running - Ready=false
+Jun 18 11:59:03.521: INFO: Waiting for pod ss-2 to enter Running - Ready=false, currently Running - Ready=false
+Jun 18 11:59:03.581: INFO: POD   NODE          PHASE    GRACE  CONDITIONS
+Jun 18 11:59:03.581: INFO: ss-0  10.72.74.138  Running         [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:58:19 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:58:53 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:58:53 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:58:19 +0000 UTC  }]
+Jun 18 11:59:03.581: INFO: ss-1  10.72.74.184  Running         [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:58:40 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:58:53 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:58:53 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:58:40 +0000 UTC  }]
+Jun 18 11:59:03.581: INFO: ss-2  10.72.74.189  Running         [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:58:40 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:58:53 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:58:53 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:58:40 +0000 UTC  }]
+Jun 18 11:59:03.582: INFO: 
+Jun 18 11:59:03.582: INFO: StatefulSet ss has not reached scale 0, at 3
+Jun 18 11:59:04.595: INFO: POD   NODE          PHASE    GRACE  CONDITIONS
+Jun 18 11:59:04.595: INFO: ss-0  10.72.74.138  Running  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:58:19 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:58:53 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:58:53 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:58:19 +0000 UTC  }]
+Jun 18 11:59:04.595: INFO: ss-1  10.72.74.184  Running  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:58:40 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:58:53 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:58:53 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:58:40 +0000 UTC  }]
+Jun 18 11:59:04.595: INFO: ss-2  10.72.74.189  Running  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:58:40 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:58:53 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:58:53 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:58:40 +0000 UTC  }]
+Jun 18 11:59:04.595: INFO: 
+Jun 18 11:59:04.595: INFO: StatefulSet ss has not reached scale 0, at 3
+Jun 18 11:59:05.610: INFO: POD   NODE          PHASE    GRACE  CONDITIONS
+Jun 18 11:59:05.610: INFO: ss-0  10.72.74.138  Running  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:58:19 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:58:53 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:58:53 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:58:19 +0000 UTC  }]
+Jun 18 11:59:05.610: INFO: ss-1  10.72.74.184  Running  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:58:40 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:58:53 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:58:53 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:58:40 +0000 UTC  }]
+Jun 18 11:59:05.610: INFO: ss-2  10.72.74.189  Running  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:58:40 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:58:53 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:58:53 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:58:40 +0000 UTC  }]
+Jun 18 11:59:05.610: INFO: 
+Jun 18 11:59:05.610: INFO: StatefulSet ss has not reached scale 0, at 3
+Jun 18 11:59:06.624: INFO: POD   NODE          PHASE    GRACE  CONDITIONS
+Jun 18 11:59:06.624: INFO: ss-0  10.72.74.138  Running  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:58:19 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:58:53 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:58:53 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:58:19 +0000 UTC  }]
+Jun 18 11:59:06.624: INFO: ss-1  10.72.74.184  Running  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:58:40 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:58:53 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:58:53 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:58:40 +0000 UTC  }]
+Jun 18 11:59:06.624: INFO: 
+Jun 18 11:59:06.624: INFO: StatefulSet ss has not reached scale 0, at 2
+Jun 18 11:59:07.642: INFO: POD   NODE          PHASE    GRACE  CONDITIONS
+Jun 18 11:59:07.642: INFO: ss-0  10.72.74.138  Running  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:58:19 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:58:53 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:58:53 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:58:19 +0000 UTC  }]
+Jun 18 11:59:07.642: INFO: ss-1  10.72.74.184  Running  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:58:40 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:58:53 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:58:53 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:58:40 +0000 UTC  }]
+Jun 18 11:59:07.642: INFO: 
+Jun 18 11:59:07.642: INFO: StatefulSet ss has not reached scale 0, at 2
+Jun 18 11:59:08.656: INFO: POD   NODE          PHASE    GRACE  CONDITIONS
+Jun 18 11:59:08.656: INFO: ss-0  10.72.74.138  Running  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:58:19 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:58:53 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:58:53 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:58:19 +0000 UTC  }]
+Jun 18 11:59:08.656: INFO: ss-1  10.72.74.184  Running  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:58:40 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:58:53 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:58:53 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:58:40 +0000 UTC  }]
+Jun 18 11:59:08.656: INFO: 
+Jun 18 11:59:08.656: INFO: StatefulSet ss has not reached scale 0, at 2
+Jun 18 11:59:09.670: INFO: POD   NODE          PHASE    GRACE  CONDITIONS
+Jun 18 11:59:09.670: INFO: ss-0  10.72.74.138  Running  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:58:19 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:58:53 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:58:53 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:58:19 +0000 UTC  }]
+Jun 18 11:59:09.670: INFO: ss-1  10.72.74.184  Running  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:58:40 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:58:53 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:58:53 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:58:40 +0000 UTC  }]
+Jun 18 11:59:09.670: INFO: 
+Jun 18 11:59:09.670: INFO: StatefulSet ss has not reached scale 0, at 2
+Jun 18 11:59:10.686: INFO: POD   NODE          PHASE    GRACE  CONDITIONS
+Jun 18 11:59:10.686: INFO: ss-0  10.72.74.138  Running  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:58:19 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:58:53 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:58:53 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:58:19 +0000 UTC  }]
+Jun 18 11:59:10.686: INFO: ss-1  10.72.74.184  Running  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:58:40 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:58:53 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:58:53 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:58:40 +0000 UTC  }]
+Jun 18 11:59:10.686: INFO: 
+Jun 18 11:59:10.686: INFO: StatefulSet ss has not reached scale 0, at 2
+Jun 18 11:59:11.704: INFO: POD   NODE          PHASE    GRACE  CONDITIONS
+Jun 18 11:59:11.704: INFO: ss-0  10.72.74.138  Running  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:58:19 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:58:53 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:58:53 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:58:19 +0000 UTC  }]
+Jun 18 11:59:11.704: INFO: 
+Jun 18 11:59:11.704: INFO: StatefulSet ss has not reached scale 0, at 1
+Jun 18 11:59:12.718: INFO: POD   NODE          PHASE    GRACE  CONDITIONS
+Jun 18 11:59:12.718: INFO: ss-0  10.72.74.138  Running  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:58:19 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:58:53 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:58:53 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 11:58:19 +0000 UTC  }]
+Jun 18 11:59:12.718: INFO: 
+Jun 18 11:59:12.718: INFO: StatefulSet ss has not reached scale 0, at 1
+STEP: Scaling down stateful set ss to 0 replicas and waiting until none of pods will run in namespacestatefulset-6365
+Jun 18 11:59:13.732: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-6365 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
+Jun 18 11:59:13.948: INFO: rc: 1
+Jun 18 11:59:13.948: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-6365 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] []    error: unable to upgrade connection: pod does not exist
+ []  0xc002cfb950 exit status 1   true [0xc001954180 0xc001954198 0xc0019541b0] [0xc001954180 0xc001954198 0xc0019541b0] [0xc001954190 0xc0019541a8] [0x9c00a0 0x9c00a0] 0xc00294f320 }:
+Command stdout:
+
+stderr:
+error: unable to upgrade connection: pod does not exist
+
+error:
+exit status 1
+
+Jun 18 11:59:23.948: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-6365 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
+Jun 18 11:59:24.090: INFO: rc: 1
+Jun 18 11:59:24.090: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-6365 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] []    Error from server (NotFound): pods "ss-0" not found
+ []  0xc002cfbc80 exit status 1   true [0xc0019541b8 0xc0019541d8 0xc0019541f0] [0xc0019541b8 0xc0019541d8 0xc0019541f0] [0xc0019541d0 0xc0019541e8] [0x9c00a0 0x9c00a0] 0xc00294f680 }:
+Command stdout:
+
+stderr:
+Error from server (NotFound): pods "ss-0" not found
+
+error:
+exit status 1
+
+Jun 18 11:59:34.090: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-6365 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
+Jun 18 11:59:34.206: INFO: rc: 1
+Jun 18 11:59:34.206: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-6365 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] []    Error from server (NotFound): pods "ss-0" not found
+ []  0xc001cf6960 exit status 1   true [0xc0028d4338 0xc0028d4350 0xc0028d4368] [0xc0028d4338 0xc0028d4350 0xc0028d4368] [0xc0028d4348 0xc0028d4360] [0x9c00a0 0x9c00a0] 0xc00313af00 }:
+Command stdout:
+
+stderr:
+Error from server (NotFound): pods "ss-0" not found
+
+error:
+exit status 1
+
+Jun 18 11:59:44.206: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-6365 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
+Jun 18 11:59:44.349: INFO: rc: 1
+Jun 18 11:59:44.350: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-6365 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] []    Error from server (NotFound): pods "ss-0" not found
+ []  0xc001cf6cc0 exit status 1   true [0xc0028d4370 0xc0028d4388 0xc0028d43a0] [0xc0028d4370 0xc0028d4388 0xc0028d43a0] [0xc0028d4380 0xc0028d4398] [0x9c00a0 0x9c00a0] 0xc00313b5c0 }:
+Command stdout:
+
+stderr:
+Error from server (NotFound): pods "ss-0" not found
+
+error:
+exit status 1
+
+Jun 18 11:59:54.350: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-6365 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
+Jun 18 11:59:54.483: INFO: rc: 1
+Jun 18 11:59:54.483: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-6365 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] []    Error from server (NotFound): pods "ss-0" not found
+ []  0xc001e5e1b0 exit status 1   true [0xc002ac83a8 0xc002ac83c0 0xc002ac83d8] [0xc002ac83a8 0xc002ac83c0 0xc002ac83d8] [0xc002ac83b8 0xc002ac83d0] [0x9c00a0 0x9c00a0] 0xc0021ed9e0 }:
+Command stdout:
+
+stderr:
+Error from server (NotFound): pods "ss-0" not found
+
+error:
+exit status 1
+
+Jun 18 12:00:04.483: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-6365 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
+Jun 18 12:00:04.623: INFO: rc: 1
+Jun 18 12:00:04.623: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-6365 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] []    Error from server (NotFound): pods "ss-0" not found
+ []  0xc0029216e0 exit status 1   true [0xc0030e2150 0xc0030e2168 0xc0030e2180] [0xc0030e2150 0xc0030e2168 0xc0030e2180] [0xc0030e2160 0xc0030e2178] [0x9c00a0 0x9c00a0] 0xc002bd3260 }:
+Command stdout:
+
+stderr:
+Error from server (NotFound): pods "ss-0" not found
+
+error:
+exit status 1
+
+Jun 18 12:00:14.623: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-6365 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
+Jun 18 12:00:14.740: INFO: rc: 1
+Jun 18 12:00:14.740: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-6365 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] []    Error from server (NotFound): pods "ss-0" not found
+ []  0xc001cf6ff0 exit status 1   true [0xc0028d43a8 0xc0028d43c0 0xc0028d43d8] [0xc0028d43a8 0xc0028d43c0 0xc0028d43d8] [0xc0028d43b8 0xc0028d43d0] [0x9c00a0 0x9c00a0] 0xc00313baa0 }:
+Command stdout:
+
+stderr:
+Error from server (NotFound): pods "ss-0" not found
+
+error:
+exit status 1
+
+Jun 18 12:00:24.740: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-6365 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
+Jun 18 12:00:24.862: INFO: rc: 1
+Jun 18 12:00:24.862: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-6365 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] []    Error from server (NotFound): pods "ss-0" not found
+ []  0xc001cf75f0 exit status 1   true [0xc0028d43e0 0xc0028d43f8 0xc0028d4410] [0xc0028d43e0 0xc0028d43f8 0xc0028d4410] [0xc0028d43f0 0xc0028d4408] [0x9c00a0 0x9c00a0] 0xc00251c0c0 }:
+Command stdout:
+
+stderr:
+Error from server (NotFound): pods "ss-0" not found
+
+error:
+exit status 1
+
+Jun 18 12:00:34.863: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-6365 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
+Jun 18 12:00:35.005: INFO: rc: 1
+Jun 18 12:00:35.005: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-6365 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] []    Error from server (NotFound): pods "ss-0" not found
+ []  0xc002938300 exit status 1   true [0xc0028d4020 0xc0028d4038 0xc0028d4050] [0xc0028d4020 0xc0028d4038 0xc0028d4050] [0xc0028d4030 0xc0028d4048] [0x9c00a0 0x9c00a0] 0xc00313a720 }:
+Command stdout:
 
-• [SLOW TEST:116.492 seconds]
-[sig-storage] ConfigMap
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:32
-  optional updates should be reflected in volume [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+stderr:
+Error from server (NotFound): pods "ss-0" not found
+
+error:
+exit status 1
+
+Jun 18 12:00:45.005: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-6365 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
+Jun 18 12:00:45.130: INFO: rc: 1
+Jun 18 12:00:45.130: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-6365 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] []    Error from server (NotFound): pods "ss-0" not found
+ []  0xc002938660 exit status 1   true [0xc0028d4068 0xc0028d40a8 0xc0028d40c0] [0xc0028d4068 0xc0028d40a8 0xc0028d40c0] [0xc0028d4090 0xc0028d40b8] [0x9c00a0 0x9c00a0] 0xc00313af60 }:
+Command stdout:
+
+stderr:
+Error from server (NotFound): pods "ss-0" not found
+
+error:
+exit status 1
+
+Jun 18 12:00:55.130: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-6365 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
+Jun 18 12:00:55.257: INFO: rc: 1
+Jun 18 12:00:55.257: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-6365 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] []    Error from server (NotFound): pods "ss-0" not found
+ []  0xc0024c0570 exit status 1   true [0xc001954000 0xc001954018 0xc001954030] [0xc001954000 0xc001954018 0xc001954030] [0xc001954010 0xc001954028] [0x9c00a0 0x9c00a0] 0xc002f242a0 }:
+Command stdout:
+
+stderr:
+Error from server (NotFound): pods "ss-0" not found
+
+error:
+exit status 1
+
+Jun 18 12:01:05.257: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-6365 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
+Jun 18 12:01:05.382: INFO: rc: 1
+Jun 18 12:01:05.382: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-6365 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] []    Error from server (NotFound): pods "ss-0" not found
+ []  0xc002782330 exit status 1   true [0xc002ac8008 0xc002ac8068 0xc002ac8080] [0xc002ac8008 0xc002ac8068 0xc002ac8080] [0xc002ac8040 0xc002ac8078] [0x9c00a0 0x9c00a0] 0xc00251c540 }:
+Command stdout:
+
+stderr:
+Error from server (NotFound): pods "ss-0" not found
+
+error:
+exit status 1
+
+Jun 18 12:01:15.382: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-6365 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
+Jun 18 12:01:15.522: INFO: rc: 1
+Jun 18 12:01:15.522: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-6365 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] []    Error from server (NotFound): pods "ss-0" not found
+ []  0xc0024c0900 exit status 1   true [0xc001954038 0xc001954050 0xc001954068] [0xc001954038 0xc001954050 0xc001954068] [0xc001954048 0xc001954060] [0x9c00a0 0x9c00a0] 0xc002f24600 }:
+Command stdout:
+
+stderr:
+Error from server (NotFound): pods "ss-0" not found
+
+error:
+exit status 1
+
+Jun 18 12:01:25.522: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-6365 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
+Jun 18 12:01:25.658: INFO: rc: 1
+Jun 18 12:01:25.658: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-6365 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] []    Error from server (NotFound): pods "ss-0" not found
+ []  0xc002ac6390 exit status 1   true [0xc0030e2000 0xc0030e2018 0xc0030e2030] [0xc0030e2000 0xc0030e2018 0xc0030e2030] [0xc0030e2010 0xc0030e2028] [0x9c00a0 0x9c00a0] 0xc00294e480 }:
+Command stdout:
+
+stderr:
+Error from server (NotFound): pods "ss-0" not found
+
+error:
+exit status 1
+
+Jun 18 12:01:35.658: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-6365 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
+Jun 18 12:01:35.785: INFO: rc: 1
+Jun 18 12:01:35.785: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-6365 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] []    Error from server (NotFound): pods "ss-0" not found
+ []  0xc0024c0c30 exit status 1   true [0xc001954070 0xc001954088 0xc0019540a0] [0xc001954070 0xc001954088 0xc0019540a0] [0xc001954080 0xc001954098] [0x9c00a0 0x9c00a0] 0xc002f24960 }:
+Command stdout:
+
+stderr:
+Error from server (NotFound): pods "ss-0" not found
+
+error:
+exit status 1
+
+Jun 18 12:01:45.786: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-6365 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
+Jun 18 12:01:45.906: INFO: rc: 1
+Jun 18 12:01:45.906: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-6365 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] []    Error from server (NotFound): pods "ss-0" not found
+ []  0xc002ac6720 exit status 1   true [0xc0030e2038 0xc0030e2050 0xc0030e2068] [0xc0030e2038 0xc0030e2050 0xc0030e2068] [0xc0030e2048 0xc0030e2060] [0x9c00a0 0x9c00a0] 0xc00294e840 }:
+Command stdout:
+
+stderr:
+Error from server (NotFound): pods "ss-0" not found
+
+error:
+exit status 1
+
+Jun 18 12:01:55.906: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-6365 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
+Jun 18 12:01:56.027: INFO: rc: 1
+Jun 18 12:01:56.027: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-6365 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] []    Error from server (NotFound): pods "ss-0" not found
+ []  0xc0024c0f60 exit status 1   true [0xc0019540a8 0xc0019540c0 0xc0019540d8] [0xc0019540a8 0xc0019540c0 0xc0019540d8] [0xc0019540b8 0xc0019540d0] [0x9c00a0 0x9c00a0] 0xc002f24e40 }:
+Command stdout:
+
+stderr:
+Error from server (NotFound): pods "ss-0" not found
+
+error:
+exit status 1
+
+Jun 18 12:02:06.027: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-6365 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
+Jun 18 12:02:07.192: INFO: rc: 1
+Jun 18 12:02:07.192: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-6365 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] []    Error from server (NotFound): pods "ss-0" not found
+ []  0xc002ac6ae0 exit status 1   true [0xc0030e2070 0xc0030e2088 0xc0030e20a0] [0xc0030e2070 0xc0030e2088 0xc0030e20a0] [0xc0030e2080 0xc0030e2098] [0x9c00a0 0x9c00a0] 0xc00294eba0 }:
+Command stdout:
+
+stderr:
+Error from server (NotFound): pods "ss-0" not found
+
+error:
+exit status 1
+
+Jun 18 12:02:17.192: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-6365 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
+Jun 18 12:02:17.793: INFO: rc: 1
+Jun 18 12:02:17.793: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-6365 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] []    Error from server (NotFound): pods "ss-0" not found
+ []  0xc0029389f0 exit status 1   true [0xc0028d40c8 0xc0028d40e0 0xc0028d40f8] [0xc0028d40c8 0xc0028d40e0 0xc0028d40f8] [0xc0028d40d8 0xc0028d40f0] [0x9c00a0 0x9c00a0] 0xc00313b620 }:
+Command stdout:
+
+stderr:
+Error from server (NotFound): pods "ss-0" not found
+
+error:
+exit status 1
+
+Jun 18 12:02:27.793: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-6365 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
+Jun 18 12:02:28.382: INFO: rc: 1
+Jun 18 12:02:28.382: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-6365 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] []    Error from server (NotFound): pods "ss-0" not found
+ []  0xc002ac6ea0 exit status 1   true [0xc0030e20a8 0xc0030e20c0 0xc0030e20d8] [0xc0030e20a8 0xc0030e20c0 0xc0030e20d8] [0xc0030e20b8 0xc0030e20d0] [0x9c00a0 0x9c00a0] 0xc00294ef00 }:
+Command stdout:
+
+stderr:
+Error from server (NotFound): pods "ss-0" not found
+
+error:
+exit status 1
+
+Jun 18 12:02:38.382: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-6365 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
+Jun 18 12:02:38.510: INFO: rc: 1
+Jun 18 12:02:38.510: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-6365 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] []    Error from server (NotFound): pods "ss-0" not found
+ []  0xc002ac6360 exit status 1   true [0xc0030e2008 0xc0030e2020 0xc0030e2038] [0xc0030e2008 0xc0030e2020 0xc0030e2038] [0xc0030e2018 0xc0030e2030] [0x9c00a0 0x9c00a0] 0xc00294e480 }:
+Command stdout:
+
+stderr:
+Error from server (NotFound): pods "ss-0" not found
+
+error:
+exit status 1
+
+Jun 18 12:02:48.510: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-6365 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
+Jun 18 12:02:48.661: INFO: rc: 1
+Jun 18 12:02:48.662: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-6365 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] []    Error from server (NotFound): pods "ss-0" not found
+ []  0xc002ac66f0 exit status 1   true [0xc0030e2040 0xc0030e2058 0xc0030e2070] [0xc0030e2040 0xc0030e2058 0xc0030e2070] [0xc0030e2050 0xc0030e2068] [0x9c00a0 0x9c00a0] 0xc00294e840 }:
+Command stdout:
+
+stderr:
+Error from server (NotFound): pods "ss-0" not found
+
+error:
+exit status 1
+
+Jun 18 12:02:58.662: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-6365 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
+Jun 18 12:02:58.789: INFO: rc: 1
+Jun 18 12:02:58.789: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-6365 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] []    Error from server (NotFound): pods "ss-0" not found
+ []  0xc002782300 exit status 1   true [0xc002ac8008 0xc002ac8068 0xc002ac8080] [0xc002ac8008 0xc002ac8068 0xc002ac8080] [0xc002ac8040 0xc002ac8078] [0x9c00a0 0x9c00a0] 0xc00251c540 }:
+Command stdout:
+
+stderr:
+Error from server (NotFound): pods "ss-0" not found
+
+error:
+exit status 1
+
+Jun 18 12:03:08.790: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-6365 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
+Jun 18 12:03:08.911: INFO: rc: 1
+Jun 18 12:03:08.911: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-6365 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] []    Error from server (NotFound): pods "ss-0" not found
+ []  0xc002ac6a80 exit status 1   true [0xc0030e2078 0xc0030e2090 0xc0030e20a8] [0xc0030e2078 0xc0030e2090 0xc0030e20a8] [0xc0030e2088 0xc0030e20a0] [0x9c00a0 0x9c00a0] 0xc00294eba0 }:
+Command stdout:
+
+stderr:
+Error from server (NotFound): pods "ss-0" not found
+
+error:
+exit status 1
+
+Jun 18 12:03:18.911: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-6365 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
+Jun 18 12:03:19.038: INFO: rc: 1
+Jun 18 12:03:19.038: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-6365 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] []    Error from server (NotFound): pods "ss-0" not found
+ []  0xc002ac6e40 exit status 1   true [0xc0030e20b0 0xc0030e20c8 0xc0030e20e0] [0xc0030e20b0 0xc0030e20c8 0xc0030e20e0] [0xc0030e20c0 0xc0030e20d8] [0x9c00a0 0x9c00a0] 0xc00294ef00 }:
+Command stdout:
+
+stderr:
+Error from server (NotFound): pods "ss-0" not found
+
+error:
+exit status 1
+
+Jun 18 12:03:29.038: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-6365 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
+Jun 18 12:03:29.167: INFO: rc: 1
+Jun 18 12:03:29.167: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-6365 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] []    Error from server (NotFound): pods "ss-0" not found
+ []  0xc0024c05d0 exit status 1   true [0xc001954000 0xc001954018 0xc001954030] [0xc001954000 0xc001954018 0xc001954030] [0xc001954010 0xc001954028] [0x9c00a0 0x9c00a0] 0xc002f242a0 }:
+Command stdout:
+
+stderr:
+Error from server (NotFound): pods "ss-0" not found
+
+error:
+exit status 1
+
+Jun 18 12:03:39.168: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-6365 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
+Jun 18 12:03:39.302: INFO: rc: 1
+Jun 18 12:03:39.303: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-6365 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] []    Error from server (NotFound): pods "ss-0" not found
+ []  0xc002ac7230 exit status 1   true [0xc0030e20e8 0xc0030e2100 0xc0030e2118] [0xc0030e20e8 0xc0030e2100 0xc0030e2118] [0xc0030e20f8 0xc0030e2110] [0x9c00a0 0x9c00a0] 0xc00294f2c0 }:
+Command stdout:
+
+stderr:
+Error from server (NotFound): pods "ss-0" not found
+
+error:
+exit status 1
+
+Jun 18 12:03:49.303: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-6365 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
+Jun 18 12:03:49.418: INFO: rc: 1
+Jun 18 12:03:49.418: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-6365 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] []    Error from server (NotFound): pods "ss-0" not found
+ []  0xc0027826f0 exit status 1   true [0xc002ac8088 0xc002ac80a8 0xc002ac80f8] [0xc002ac8088 0xc002ac80a8 0xc002ac80f8] [0xc002ac8098 0xc002ac80e0] [0x9c00a0 0x9c00a0] 0xc00251cc00 }:
+Command stdout:
+
+stderr:
+Error from server (NotFound): pods "ss-0" not found
+
+error:
+exit status 1
+
+Jun 18 12:03:59.418: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-6365 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
+Jun 18 12:03:59.544: INFO: rc: 1
+Jun 18 12:03:59.544: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-6365 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] []    Error from server (NotFound): pods "ss-0" not found
+ []  0xc002782ba0 exit status 1   true [0xc002ac8100 0xc002ac8118 0xc002ac8150] [0xc002ac8100 0xc002ac8118 0xc002ac8150] [0xc002ac8110 0xc002ac8148] [0x9c00a0 0x9c00a0] 0xc00251d260 }:
+Command stdout:
+
+stderr:
+Error from server (NotFound): pods "ss-0" not found
+
+error:
+exit status 1
+
+Jun 18 12:04:09.545: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-6365 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
+Jun 18 12:04:09.685: INFO: rc: 1
+Jun 18 12:04:09.685: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-6365 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] []    Error from server (NotFound): pods "ss-0" not found
+ []  0xc002ac7590 exit status 1   true [0xc0030e2120 0xc0030e2138 0xc0030e2150] [0xc0030e2120 0xc0030e2138 0xc0030e2150] [0xc0030e2130 0xc0030e2148] [0x9c00a0 0x9c00a0] 0xc00294f620 }:
+Command stdout:
+
+stderr:
+Error from server (NotFound): pods "ss-0" not found
+
+error:
+exit status 1
+
+Jun 18 12:04:19.685: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-6365 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
+Jun 18 12:04:19.824: INFO: rc: 1
+Jun 18 12:04:19.824: INFO: stdout of mv -v /tmp/index.html /usr/share/nginx/html/ || true on ss-0: 
+Jun 18 12:04:19.824: INFO: Scaling statefulset ss to 0
+Jun 18 12:04:19.861: INFO: Waiting for statefulset status.replicas updated to 0
+[AfterEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:85
+Jun 18 12:04:19.873: INFO: Deleting all statefulset in ns statefulset-6365
+Jun 18 12:04:19.883: INFO: Scaling statefulset ss to 0
+Jun 18 12:04:19.918: INFO: Waiting for statefulset status.replicas updated to 0
+Jun 18 12:04:19.929: INFO: Deleting statefulset ss
+[AfterEach] [sig-apps] StatefulSet
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 12:04:19.980: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "statefulset-6365" for this suite.
+Jun 18 12:04:28.024: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 12:04:28.418: INFO: namespace statefulset-6365 deletion completed in 8.420445369s
+
+• [SLOW TEST:369.290 seconds]
+[sig-apps] StatefulSet
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22
+  [k8s.io] Basic StatefulSet functionality [StatefulSetBasic]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
+    Burst scaling should run to completion even with unhealthy pods [Conformance]
+    /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-SSSSSSSSSSSSSSSSSSSSSSSSSSS
+SSSSSSSS
 ------------------------------
-[sig-storage] ConfigMap 
-  binary data should be reflected in volume [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [sig-storage] ConfigMap
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+[k8s.io] Pods 
+  should be updated [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [k8s.io] Pods
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 19:02:52.483: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-STEP: Building a namespace api object, basename configmap
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in configmap-5439
+Jun 18 12:04:28.419: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename pods
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in pods-5555
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] binary data should be reflected in volume [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: Creating configMap with name configmap-test-upd-5aaddaf3-86fb-11e9-8d1b-467ee19922ac
-STEP: Creating the pod
-STEP: Waiting for pod with text data
-STEP: Waiting for pod with binary data
-[AfterEach] [sig-storage] ConfigMap
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 19:02:56.998: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "configmap-5439" for this suite.
-Jun  4 19:03:21.057: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 19:03:21.573: INFO: namespace configmap-5439 deletion completed in 24.559180292s
+[BeforeEach] [k8s.io] Pods
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:135
+[It] should be updated [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: creating the pod
+STEP: submitting the pod to kubernetes
+STEP: verifying the pod is in kubernetes
+STEP: updating the pod
+Jun 18 12:04:31.230: INFO: Successfully updated pod "pod-update-3936e9c6-91c1-11e9-a25d-8608290c688a"
+STEP: verifying the updated pod is in kubernetes
+Jun 18 12:04:31.257: INFO: Pod update OK
+[AfterEach] [k8s.io] Pods
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 12:04:31.257: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "pods-5555" for this suite.
+Jun 18 12:04:55.297: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 12:04:55.677: INFO: namespace pods-5555 deletion completed in 24.405856395s
 
-• [SLOW TEST:29.090 seconds]
-[sig-storage] ConfigMap
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:32
-  binary data should be reflected in volume [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+• [SLOW TEST:27.258 seconds]
+[k8s.io] Pods
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
+  should be updated [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-SSSSSSSSSSSSSSSSSSSSSS
+SSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-storage] Secrets 
-  should be consumable in multiple volumes in a pod [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [sig-storage] Secrets
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+[sig-storage] Downward API volume 
+  should provide container's cpu limit [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-storage] Downward API volume
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 19:03:21.573: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-STEP: Building a namespace api object, basename secrets
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in secrets-1251
+Jun 18 12:04:55.678: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename downward-api
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in downward-api-7150
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should be consumable in multiple volumes in a pod [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: Creating secret with name secret-test-6bf87737-86fb-11e9-8d1b-467ee19922ac
-STEP: Creating a pod to test consume secrets
-Jun  4 19:03:21.814: INFO: Waiting up to 5m0s for pod "pod-secrets-6bf9d478-86fb-11e9-8d1b-467ee19922ac" in namespace "secrets-1251" to be "success or failure"
-Jun  4 19:03:21.830: INFO: Pod "pod-secrets-6bf9d478-86fb-11e9-8d1b-467ee19922ac": Phase="Pending", Reason="", readiness=false. Elapsed: 15.539586ms
-Jun  4 19:03:23.842: INFO: Pod "pod-secrets-6bf9d478-86fb-11e9-8d1b-467ee19922ac": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.027401999s
+[BeforeEach] [sig-storage] Downward API volume
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:39
+[It] should provide container's cpu limit [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating a pod to test downward API volume plugin
+Jun 18 12:04:55.909: INFO: Waiting up to 5m0s for pod "downwardapi-volume-49781dda-91c1-11e9-a25d-8608290c688a" in namespace "downward-api-7150" to be "success or failure"
+Jun 18 12:04:55.933: INFO: Pod "downwardapi-volume-49781dda-91c1-11e9-a25d-8608290c688a": Phase="Pending", Reason="", readiness=false. Elapsed: 23.464074ms
+Jun 18 12:04:57.953: INFO: Pod "downwardapi-volume-49781dda-91c1-11e9-a25d-8608290c688a": Phase="Running", Reason="", readiness=true. Elapsed: 2.043197206s
+Jun 18 12:04:59.967: INFO: Pod "downwardapi-volume-49781dda-91c1-11e9-a25d-8608290c688a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.057698851s
 STEP: Saw pod success
-Jun  4 19:03:23.842: INFO: Pod "pod-secrets-6bf9d478-86fb-11e9-8d1b-467ee19922ac" satisfied condition "success or failure"
-Jun  4 19:03:23.856: INFO: Trying to get logs from node 10.212.23.161 pod pod-secrets-6bf9d478-86fb-11e9-8d1b-467ee19922ac container secret-volume-test: 
+Jun 18 12:04:59.967: INFO: Pod "downwardapi-volume-49781dda-91c1-11e9-a25d-8608290c688a" satisfied condition "success or failure"
+Jun 18 12:04:59.980: INFO: Trying to get logs from node 10.72.74.138 pod downwardapi-volume-49781dda-91c1-11e9-a25d-8608290c688a container client-container: 
 STEP: delete the pod
-Jun  4 19:03:23.955: INFO: Waiting for pod pod-secrets-6bf9d478-86fb-11e9-8d1b-467ee19922ac to disappear
-Jun  4 19:03:23.966: INFO: Pod pod-secrets-6bf9d478-86fb-11e9-8d1b-467ee19922ac no longer exists
-[AfterEach] [sig-storage] Secrets
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 19:03:23.966: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "secrets-1251" for this suite.
-Jun  4 19:03:30.024: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 19:03:30.502: INFO: namespace secrets-1251 deletion completed in 6.521094323s
+Jun 18 12:05:00.065: INFO: Waiting for pod downwardapi-volume-49781dda-91c1-11e9-a25d-8608290c688a to disappear
+Jun 18 12:05:00.079: INFO: Pod downwardapi-volume-49781dda-91c1-11e9-a25d-8608290c688a no longer exists
+[AfterEach] [sig-storage] Downward API volume
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 12:05:00.079: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "downward-api-7150" for this suite.
+Jun 18 12:05:06.842: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 12:05:07.261: INFO: namespace downward-api-7150 deletion completed in 6.455007982s
 
-• [SLOW TEST:8.929 seconds]
-[sig-storage] Secrets
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/secrets_volume.go:33
-  should be consumable in multiple volumes in a pod [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-------------------------------
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+• [SLOW TEST:11.583 seconds]
+[sig-storage] Downward API volume
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:34
+  should provide container's cpu limit [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
 [sig-storage] EmptyDir volumes 
-  should support (root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+  volume on default medium should have the correct mode [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 [BeforeEach] [sig-storage] EmptyDir volumes
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 19:03:30.504: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
+Jun 18 12:05:07.261: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
 STEP: Building a namespace api object, basename emptydir
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in emptydir-303
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in emptydir-9037
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should support (root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: Creating a pod to test emptydir 0777 on tmpfs
-Jun  4 19:03:30.744: INFO: Waiting up to 5m0s for pod "pod-714c546f-86fb-11e9-8d1b-467ee19922ac" in namespace "emptydir-303" to be "success or failure"
-Jun  4 19:03:30.755: INFO: Pod "pod-714c546f-86fb-11e9-8d1b-467ee19922ac": Phase="Pending", Reason="", readiness=false. Elapsed: 10.881215ms
-Jun  4 19:03:32.770: INFO: Pod "pod-714c546f-86fb-11e9-8d1b-467ee19922ac": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.025287216s
+[It] volume on default medium should have the correct mode [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating a pod to test emptydir volume type on node default medium
+Jun 18 12:05:07.483: INFO: Waiting up to 5m0s for pod "pod-505df898-91c1-11e9-a25d-8608290c688a" in namespace "emptydir-9037" to be "success or failure"
+Jun 18 12:05:07.500: INFO: Pod "pod-505df898-91c1-11e9-a25d-8608290c688a": Phase="Pending", Reason="", readiness=false. Elapsed: 16.750543ms
+Jun 18 12:05:09.515: INFO: Pod "pod-505df898-91c1-11e9-a25d-8608290c688a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.03218869s
 STEP: Saw pod success
-Jun  4 19:03:32.770: INFO: Pod "pod-714c546f-86fb-11e9-8d1b-467ee19922ac" satisfied condition "success or failure"
-Jun  4 19:03:32.782: INFO: Trying to get logs from node 10.212.23.161 pod pod-714c546f-86fb-11e9-8d1b-467ee19922ac container test-container: 
+Jun 18 12:05:09.515: INFO: Pod "pod-505df898-91c1-11e9-a25d-8608290c688a" satisfied condition "success or failure"
+Jun 18 12:05:09.528: INFO: Trying to get logs from node 10.72.74.184 pod pod-505df898-91c1-11e9-a25d-8608290c688a container test-container: 
 STEP: delete the pod
-Jun  4 19:03:32.844: INFO: Waiting for pod pod-714c546f-86fb-11e9-8d1b-467ee19922ac to disappear
-Jun  4 19:03:32.855: INFO: Pod pod-714c546f-86fb-11e9-8d1b-467ee19922ac no longer exists
+Jun 18 12:05:09.613: INFO: Waiting for pod pod-505df898-91c1-11e9-a25d-8608290c688a to disappear
+Jun 18 12:05:09.625: INFO: Pod pod-505df898-91c1-11e9-a25d-8608290c688a no longer exists
 [AfterEach] [sig-storage] EmptyDir volumes
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 19:03:32.855: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "emptydir-303" for this suite.
-Jun  4 19:03:38.911: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 19:03:39.283: INFO: namespace emptydir-303 deletion completed in 6.413396691s
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 12:05:09.625: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "emptydir-9037" for this suite.
+Jun 18 12:05:15.664: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 12:05:16.052: INFO: namespace emptydir-9037 deletion completed in 6.413106318s
 
-• [SLOW TEST:8.779 seconds]
+• [SLOW TEST:8.791 seconds]
 [sig-storage] EmptyDir volumes
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:41
-  should support (root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-------------------------------
-SSSSSSSSSSSSSS
-------------------------------
-[sig-cli] Kubectl client [k8s.io] Kubectl rolling-update 
-  should support rolling-update to same image  [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
-STEP: Creating a kubernetes client
-Jun  4 19:03:39.283: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-STEP: Building a namespace api object, basename kubectl
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-9572
-STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:213
-[BeforeEach] [k8s.io] Kubectl rolling-update
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1414
-[It] should support rolling-update to same image  [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: running the image docker.io/library/nginx:1.14-alpine
-Jun  4 19:03:39.559: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 run e2e-test-nginx-rc --image=docker.io/library/nginx:1.14-alpine --generator=run/v1 --namespace=kubectl-9572'
-Jun  4 19:03:39.681: INFO: stderr: "kubectl run --generator=run/v1 is DEPRECATED and will be removed in a future version. Use kubectl run --generator=run-pod/v1 or kubectl create instead.\n"
-Jun  4 19:03:39.681: INFO: stdout: "replicationcontroller/e2e-test-nginx-rc created\n"
-STEP: verifying the rc e2e-test-nginx-rc was created
-Jun  4 19:03:39.695: INFO: Waiting for rc e2e-test-nginx-rc to stabilize, generation 1 observed generation 0 spec.replicas 1 status.replicas 0
-Jun  4 19:03:39.698: INFO: Waiting for rc e2e-test-nginx-rc to stabilize, generation 1 observed generation 1 spec.replicas 1 status.replicas 0
-STEP: rolling-update to same image controller
-Jun  4 19:03:39.708: INFO: scanned /root for discovery docs: 
-Jun  4 19:03:39.708: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 rolling-update e2e-test-nginx-rc --update-period=1s --image=docker.io/library/nginx:1.14-alpine --image-pull-policy=IfNotPresent --namespace=kubectl-9572'
-Jun  4 19:03:55.672: INFO: stderr: "Command \"rolling-update\" is deprecated, use \"rollout\" instead\n"
-Jun  4 19:03:55.672: INFO: stdout: "Created e2e-test-nginx-rc-0876314a617b23d26b833eabf9a80919\nScaling up e2e-test-nginx-rc-0876314a617b23d26b833eabf9a80919 from 0 to 1, scaling down e2e-test-nginx-rc from 1 to 0 (keep 1 pods available, don't exceed 2 pods)\nScaling e2e-test-nginx-rc-0876314a617b23d26b833eabf9a80919 up to 1\nScaling e2e-test-nginx-rc down to 0\nUpdate succeeded. Deleting old controller: e2e-test-nginx-rc\nRenaming e2e-test-nginx-rc-0876314a617b23d26b833eabf9a80919 to e2e-test-nginx-rc\nreplicationcontroller/e2e-test-nginx-rc rolling updated\n"
-Jun  4 19:03:55.672: INFO: stdout: "Created e2e-test-nginx-rc-0876314a617b23d26b833eabf9a80919\nScaling up e2e-test-nginx-rc-0876314a617b23d26b833eabf9a80919 from 0 to 1, scaling down e2e-test-nginx-rc from 1 to 0 (keep 1 pods available, don't exceed 2 pods)\nScaling e2e-test-nginx-rc-0876314a617b23d26b833eabf9a80919 up to 1\nScaling e2e-test-nginx-rc down to 0\nUpdate succeeded. Deleting old controller: e2e-test-nginx-rc\nRenaming e2e-test-nginx-rc-0876314a617b23d26b833eabf9a80919 to e2e-test-nginx-rc\nreplicationcontroller/e2e-test-nginx-rc rolling updated\n"
-STEP: waiting for all containers in run=e2e-test-nginx-rc pods to come up.
-Jun  4 19:03:55.672: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l run=e2e-test-nginx-rc --namespace=kubectl-9572'
-Jun  4 19:03:55.758: INFO: stderr: ""
-Jun  4 19:03:55.758: INFO: stdout: "e2e-test-nginx-rc-0876314a617b23d26b833eabf9a80919-qftp2 e2e-test-nginx-rc-wzk69 "
-STEP: Replicas for run=e2e-test-nginx-rc: expected=1 actual=2
-Jun  4 19:04:00.759: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l run=e2e-test-nginx-rc --namespace=kubectl-9572'
-Jun  4 19:04:00.837: INFO: stderr: ""
-Jun  4 19:04:00.837: INFO: stdout: "e2e-test-nginx-rc-0876314a617b23d26b833eabf9a80919-qftp2 "
-Jun  4 19:04:00.838: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 get pods e2e-test-nginx-rc-0876314a617b23d26b833eabf9a80919-qftp2 -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "e2e-test-nginx-rc") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-9572'
-Jun  4 19:04:00.995: INFO: stderr: ""
-Jun  4 19:04:00.995: INFO: stdout: "true"
-Jun  4 19:04:00.995: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 get pods e2e-test-nginx-rc-0876314a617b23d26b833eabf9a80919-qftp2 -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "e2e-test-nginx-rc"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-9572'
-Jun  4 19:04:01.290: INFO: stderr: ""
-Jun  4 19:04:01.290: INFO: stdout: "docker.io/library/nginx:1.14-alpine"
-Jun  4 19:04:01.290: INFO: e2e-test-nginx-rc-0876314a617b23d26b833eabf9a80919-qftp2 is verified up and running
-[AfterEach] [k8s.io] Kubectl rolling-update
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1420
-Jun  4 19:04:01.290: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 delete rc e2e-test-nginx-rc --namespace=kubectl-9572'
-Jun  4 19:04:01.416: INFO: stderr: ""
-Jun  4 19:04:01.416: INFO: stdout: "replicationcontroller \"e2e-test-nginx-rc\" deleted\n"
-[AfterEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 19:04:01.416: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "kubectl-9572" for this suite.
-Jun  4 19:04:25.553: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 19:04:25.969: INFO: namespace kubectl-9572 deletion completed in 24.530115639s
-
-• [SLOW TEST:46.686 seconds]
-[sig-cli] Kubectl client
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23
-  [k8s.io] Kubectl rolling-update
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
-    should support rolling-update to same image  [Conformance]
-    /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:41
+  volume on default medium should have the correct mode [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-SSSSS
+SSSSSSS
 ------------------------------
-[sig-storage] Projected downwardAPI 
-  should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [sig-storage] Projected downwardAPI
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+[sig-storage] Projected secret 
+  should be consumable in multiple volumes in a pod [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-storage] Projected secret
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 19:04:25.970: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
+Jun 18 12:05:16.052: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
 STEP: Building a namespace api object, basename projected
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-51
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-7938
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-storage] Projected downwardAPI
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:39
-[It] should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: Creating a pod to test downward API volume plugin
-Jun  4 19:04:26.208: INFO: Waiting up to 5m0s for pod "downwardapi-volume-925b51fc-86fb-11e9-8d1b-467ee19922ac" in namespace "projected-51" to be "success or failure"
-Jun  4 19:04:26.220: INFO: Pod "downwardapi-volume-925b51fc-86fb-11e9-8d1b-467ee19922ac": Phase="Pending", Reason="", readiness=false. Elapsed: 11.896694ms
-Jun  4 19:04:28.234: INFO: Pod "downwardapi-volume-925b51fc-86fb-11e9-8d1b-467ee19922ac": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.025240778s
+[It] should be consumable in multiple volumes in a pod [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating secret with name projected-secret-test-55a7285f-91c1-11e9-a25d-8608290c688a
+STEP: Creating a pod to test consume secrets
+Jun 18 12:05:16.362: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-55a9203d-91c1-11e9-a25d-8608290c688a" in namespace "projected-7938" to be "success or failure"
+Jun 18 12:05:16.376: INFO: Pod "pod-projected-secrets-55a9203d-91c1-11e9-a25d-8608290c688a": Phase="Pending", Reason="", readiness=false. Elapsed: 13.385481ms
+Jun 18 12:05:18.395: INFO: Pod "pod-projected-secrets-55a9203d-91c1-11e9-a25d-8608290c688a": Phase="Pending", Reason="", readiness=false. Elapsed: 2.0329297s
+Jun 18 12:05:20.409: INFO: Pod "pod-projected-secrets-55a9203d-91c1-11e9-a25d-8608290c688a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.046452139s
 STEP: Saw pod success
-Jun  4 19:04:28.234: INFO: Pod "downwardapi-volume-925b51fc-86fb-11e9-8d1b-467ee19922ac" satisfied condition "success or failure"
-Jun  4 19:04:28.245: INFO: Trying to get logs from node 10.212.23.161 pod downwardapi-volume-925b51fc-86fb-11e9-8d1b-467ee19922ac container client-container: 
+Jun 18 12:05:20.409: INFO: Pod "pod-projected-secrets-55a9203d-91c1-11e9-a25d-8608290c688a" satisfied condition "success or failure"
+Jun 18 12:05:20.424: INFO: Trying to get logs from node 10.72.74.138 pod pod-projected-secrets-55a9203d-91c1-11e9-a25d-8608290c688a container secret-volume-test: 
 STEP: delete the pod
-Jun  4 19:04:28.305: INFO: Waiting for pod downwardapi-volume-925b51fc-86fb-11e9-8d1b-467ee19922ac to disappear
-Jun  4 19:04:28.322: INFO: Pod downwardapi-volume-925b51fc-86fb-11e9-8d1b-467ee19922ac no longer exists
-[AfterEach] [sig-storage] Projected downwardAPI
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 19:04:28.322: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "projected-51" for this suite.
-Jun  4 19:04:34.385: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 19:04:34.742: INFO: namespace projected-51 deletion completed in 6.400959405s
-
-• [SLOW TEST:8.772 seconds]
-[sig-storage] Projected downwardAPI
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:33
-  should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-------------------------------
-[sig-network] Networking Granular Checks: Pods 
-  should function for intra-pod communication: http [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [sig-network] Networking
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
-STEP: Creating a kubernetes client
-Jun  4 19:04:34.742: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-STEP: Building a namespace api object, basename pod-network-test
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in pod-network-test-7626
-STEP: Waiting for a default service account to be provisioned in namespace
-[It] should function for intra-pod communication: http [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: Performing setup for networking test in namespace pod-network-test-7626
-STEP: creating a selector
-STEP: Creating the service pods in kubernetes
-Jun  4 19:04:34.957: INFO: Waiting up to 10m0s for all (but 0) nodes to be schedulable
-STEP: Creating test pods
-Jun  4 19:04:59.459: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s 'http://172.30.254.179:8080/dial?request=hostName&protocol=http&host=172.30.254.178&port=8080&tries=1'] Namespace:pod-network-test-7626 PodName:host-test-container-pod ContainerName:hostexec Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
-Jun  4 19:04:59.459: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-Jun  4 19:04:59.647: INFO: Waiting for endpoints: map[]
-Jun  4 19:04:59.660: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s 'http://172.30.254.179:8080/dial?request=hostName&protocol=http&host=172.30.169.130&port=8080&tries=1'] Namespace:pod-network-test-7626 PodName:host-test-container-pod ContainerName:hostexec Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
-Jun  4 19:04:59.660: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-Jun  4 19:04:59.838: INFO: Waiting for endpoints: map[]
-Jun  4 19:04:59.850: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s 'http://172.30.254.179:8080/dial?request=hostName&protocol=http&host=172.30.193.45&port=8080&tries=1'] Namespace:pod-network-test-7626 PodName:host-test-container-pod ContainerName:hostexec Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
-Jun  4 19:04:59.850: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-Jun  4 19:05:00.130: INFO: Waiting for endpoints: map[]
-[AfterEach] [sig-network] Networking
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 19:05:00.130: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "pod-network-test-7626" for this suite.
-Jun  4 19:05:24.199: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 19:05:24.540: INFO: namespace pod-network-test-7626 deletion completed in 24.393404292s
+Jun 18 12:05:20.518: INFO: Waiting for pod pod-projected-secrets-55a9203d-91c1-11e9-a25d-8608290c688a to disappear
+Jun 18 12:05:20.531: INFO: Pod pod-projected-secrets-55a9203d-91c1-11e9-a25d-8608290c688a no longer exists
+[AfterEach] [sig-storage] Projected secret
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 12:05:20.531: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "projected-7938" for this suite.
+Jun 18 12:05:26.570: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 12:05:28.502: INFO: namespace projected-7938 deletion completed in 7.958589831s
 
-• [SLOW TEST:49.798 seconds]
-[sig-network] Networking
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:25
-  Granular Checks: Pods
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:28
-    should function for intra-pod communication: http [LinuxOnly] [NodeConformance] [Conformance]
-    /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+• [SLOW TEST:12.450 seconds]
+[sig-storage] Projected secret
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_secret.go:33
+  should be consumable in multiple volumes in a pod [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+SSSSSSSSSSSS
 ------------------------------
-[sig-apps] Deployment 
-  deployment should delete old replica sets [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [sig-apps] Deployment
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+[sig-api-machinery] Garbage collector 
+  should not delete dependents that have both valid owner and owner that's waiting for dependents to be deleted [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-api-machinery] Garbage collector
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 19:05:24.541: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-STEP: Building a namespace api object, basename deployment
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in deployment-1886
+Jun 18 12:05:28.504: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename gc
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in gc-8196
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-apps] Deployment
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:65
-[It] deployment should delete old replica sets [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-Jun  4 19:05:24.784: INFO: Pod name cleanup-pod: Found 0 pods out of 1
-Jun  4 19:05:29.797: INFO: Pod name cleanup-pod: Found 1 pods out of 1
-STEP: ensuring each pod is running
-Jun  4 19:05:29.797: INFO: Creating deployment test-cleanup-deployment
-STEP: Waiting for deployment test-cleanup-deployment history to be cleaned up
-[AfterEach] [sig-apps] Deployment
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:59
-Jun  4 19:05:31.871: INFO: Deployment "test-cleanup-deployment":
-&Deployment{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-cleanup-deployment,GenerateName:,Namespace:deployment-1886,SelfLink:/apis/apps/v1/namespaces/deployment-1886/deployments/test-cleanup-deployment,UID:b8481cb1-86fb-11e9-8318-1e5386706511,ResourceVersion:16633,Generation:1,CreationTimestamp:2019-06-04 19:05:29 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: cleanup-pod,},Annotations:map[string]string{deployment.kubernetes.io/revision: 1,},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:DeploymentSpec{Replicas:*1,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: cleanup-pod,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: cleanup-pod,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[],Containers:[{redis gcr.io/kubernetes-e2e-test-images/redis:1.0 [] []  [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,},},Strategy:DeploymentStrategy{Type:RollingUpdate,RollingUpdate:&RollingUpdateDeployment{MaxUnavailable:25%!,(MISSING)MaxSurge:25%!,(MISSING)},},MinReadySeconds:0,RevisionHistoryLimit:*0,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:1,Replicas:1,UpdatedReplicas:1,AvailableReplicas:1,UnavailableReplicas:0,Conditions:[{Available True 2019-06-04 19:05:29 +0000 UTC 2019-06-04 19:05:29 +0000 UTC MinimumReplicasAvailable Deployment has minimum availability.} {Progressing True 2019-06-04 19:05:31 +0000 UTC 2019-06-04 19:05:29 +0000 UTC NewReplicaSetAvailable ReplicaSet "test-cleanup-deployment-55cbfbc8f5" has successfully progressed.}],ReadyReplicas:1,CollisionCount:nil,},}
-
-Jun  4 19:05:31.881: INFO: New ReplicaSet "test-cleanup-deployment-55cbfbc8f5" of Deployment "test-cleanup-deployment":
-&ReplicaSet{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-cleanup-deployment-55cbfbc8f5,GenerateName:,Namespace:deployment-1886,SelfLink:/apis/apps/v1/namespaces/deployment-1886/replicasets/test-cleanup-deployment-55cbfbc8f5,UID:b84bfb7a-86fb-11e9-9509-923735b172b1,ResourceVersion:16622,Generation:1,CreationTimestamp:2019-06-04 19:05:29 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: cleanup-pod,pod-template-hash: 55cbfbc8f5,},Annotations:map[string]string{deployment.kubernetes.io/desired-replicas: 1,deployment.kubernetes.io/max-replicas: 2,deployment.kubernetes.io/revision: 1,},OwnerReferences:[{apps/v1 Deployment test-cleanup-deployment b8481cb1-86fb-11e9-8318-1e5386706511 0xc0029bb3c7 0xc0029bb3c8}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:ReplicaSetSpec{Replicas:*1,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: cleanup-pod,pod-template-hash: 55cbfbc8f5,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: cleanup-pod,pod-template-hash: 55cbfbc8f5,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[],Containers:[{redis gcr.io/kubernetes-e2e-test-images/redis:1.0 [] []  [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,},},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:1,FullyLabeledReplicas:1,ObservedGeneration:1,ReadyReplicas:1,AvailableReplicas:1,Conditions:[],},}
-Jun  4 19:05:31.891: INFO: Pod "test-cleanup-deployment-55cbfbc8f5-2qhcf" is available:
-&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-cleanup-deployment-55cbfbc8f5-2qhcf,GenerateName:test-cleanup-deployment-55cbfbc8f5-,Namespace:deployment-1886,SelfLink:/api/v1/namespaces/deployment-1886/pods/test-cleanup-deployment-55cbfbc8f5-2qhcf,UID:b84d5c0c-86fb-11e9-9509-923735b172b1,ResourceVersion:16621,Generation:0,CreationTimestamp:2019-06-04 19:05:29 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: cleanup-pod,pod-template-hash: 55cbfbc8f5,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet test-cleanup-deployment-55cbfbc8f5 b84bfb7a-86fb-11e9-9509-923735b172b1 0xc002978087 0xc002978088}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-b8bth {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-b8bth,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{redis gcr.io/kubernetes-e2e-test-images/redis:1.0 [] []  [] [] [] {map[] map[]} [{default-token-b8bth true /var/run/secrets/kubernetes.io/serviceaccount   }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.212.23.164,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists  NoExecute 0xc002978100} {node.kubernetes.io/unreachable Exists  NoExecute 0xc002978120}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Running,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 19:05:29 +0000 UTC  } {Ready True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 19:05:31 +0000 UTC  } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 19:05:31 +0000 UTC  } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 19:05:29 +0000 UTC  }],Message:,Reason:,HostIP:10.212.23.164,PodIP:172.30.254.180,StartTime:2019-06-04 19:05:29 +0000 UTC,ContainerStatuses:[{redis {nil ContainerStateRunning{StartedAt:2019-06-04 19:05:31 +0000 UTC,} nil} {nil nil nil} true 0 gcr.io/kubernetes-e2e-test-images/redis:1.0 gcr.io/kubernetes-e2e-test-images/redis@sha256:af4748d1655c08dc54d4be5182135395db9ce87aba2d4699b26b14ae197c5830 containerd://98f1c5743ef955803723a436c6fb87126a1bcacff024e87ec45503f3078b317a}],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},}
-[AfterEach] [sig-apps] Deployment
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 19:05:31.891: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "deployment-1886" for this suite.
-Jun  4 19:05:37.967: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 19:05:38.371: INFO: namespace deployment-1886 deletion completed in 6.463335995s
-
-• [SLOW TEST:13.830 seconds]
-[sig-apps] Deployment
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22
-  deployment should delete old replica sets [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-------------------------------
-[sig-api-machinery] Garbage collector 
-  should orphan pods created by rc if delete options say so [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [sig-api-machinery] Garbage collector
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
-STEP: Creating a kubernetes client
-Jun  4 19:05:38.371: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-STEP: Building a namespace api object, basename gc
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in gc-4770
-STEP: Waiting for a default service account to be provisioned in namespace
-[It] should orphan pods created by rc if delete options say so [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: create the rc
-STEP: delete the rc
+[It] should not delete dependents that have both valid owner and owner that's waiting for dependents to be deleted [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: create the rc1
+STEP: create the rc2
+STEP: set half of pods created by rc simpletest-rc-to-be-deleted to have rc simpletest-rc-to-stay as owner as well
+STEP: delete the rc simpletest-rc-to-be-deleted
 STEP: wait for the rc to be deleted
-STEP: wait for 30 seconds to see if the garbage collector mistakenly deletes the pods
 STEP: Gathering metrics
-W0604 19:06:18.767357      18 metrics_grabber.go:79] Master node is not registered. Grabbing metrics from Scheduler, ControllerManager and ClusterAutoscaler is disabled.
-Jun  4 19:06:18.767: INFO: For apiserver_request_total:
+W0618 12:05:38.937785      20 metrics_grabber.go:79] Master node is not registered. Grabbing metrics from Scheduler, ControllerManager and ClusterAutoscaler is disabled.
+Jun 18 12:05:38.937: INFO: For apiserver_request_total:
 For apiserver_request_latencies_summary:
 For apiserver_init_events_total:
 For garbage_collector_attempt_to_delete_queue_latency:
@@ -4184,6214 +4171,6924 @@ For errors_total:
 For evicted_pods_total:
 
 [AfterEach] [sig-api-machinery] Garbage collector
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 19:06:18.767: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "gc-4770" for this suite.
-Jun  4 19:06:26.898: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 19:06:27.239: INFO: namespace gc-4770 deletion completed in 8.451759773s
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 12:05:38.937: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "gc-8196" for this suite.
+Jun 18 12:05:46.982: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 12:05:48.540: INFO: namespace gc-8196 deletion completed in 9.589182306s
 
-• [SLOW TEST:48.868 seconds]
+• [SLOW TEST:20.036 seconds]
 [sig-api-machinery] Garbage collector
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:22
-  should orphan pods created by rc if delete options say so [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:22
+  should not delete dependents that have both valid owner and owner that's waiting for dependents to be deleted [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-SSSSSSSSSSSSSSS
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-storage] Projected configMap 
-  should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [sig-storage] Projected configMap
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+[k8s.io] Docker Containers 
+  should use the image defaults if command and args are blank [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [k8s.io] Docker Containers
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 19:06:27.240: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-STEP: Building a namespace api object, basename projected
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-6667
+Jun 18 12:05:48.540: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename containers
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in containers-2277
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: Creating configMap with name projected-configmap-test-volume-daa42f12-86fb-11e9-8d1b-467ee19922ac
-STEP: Creating a pod to test consume configMaps
-Jun  4 19:06:27.493: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-daa60a77-86fb-11e9-8d1b-467ee19922ac" in namespace "projected-6667" to be "success or failure"
-Jun  4 19:06:27.503: INFO: Pod "pod-projected-configmaps-daa60a77-86fb-11e9-8d1b-467ee19922ac": Phase="Pending", Reason="", readiness=false. Elapsed: 9.885755ms
-Jun  4 19:06:29.515: INFO: Pod "pod-projected-configmaps-daa60a77-86fb-11e9-8d1b-467ee19922ac": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.02235696s
+[It] should use the image defaults if command and args are blank [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating a pod to test use defaults
+Jun 18 12:05:48.762: INFO: Waiting up to 5m0s for pod "client-containers-68f8d798-91c1-11e9-a25d-8608290c688a" in namespace "containers-2277" to be "success or failure"
+Jun 18 12:05:48.775: INFO: Pod "client-containers-68f8d798-91c1-11e9-a25d-8608290c688a": Phase="Pending", Reason="", readiness=false. Elapsed: 12.899139ms
+Jun 18 12:05:50.788: INFO: Pod "client-containers-68f8d798-91c1-11e9-a25d-8608290c688a": Phase="Pending", Reason="", readiness=false. Elapsed: 2.026199526s
+Jun 18 12:05:52.802: INFO: Pod "client-containers-68f8d798-91c1-11e9-a25d-8608290c688a": Phase="Pending", Reason="", readiness=false. Elapsed: 4.039691135s
+Jun 18 12:05:54.816: INFO: Pod "client-containers-68f8d798-91c1-11e9-a25d-8608290c688a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.053909646s
 STEP: Saw pod success
-Jun  4 19:06:29.515: INFO: Pod "pod-projected-configmaps-daa60a77-86fb-11e9-8d1b-467ee19922ac" satisfied condition "success or failure"
-Jun  4 19:06:29.526: INFO: Trying to get logs from node 10.212.23.189 pod pod-projected-configmaps-daa60a77-86fb-11e9-8d1b-467ee19922ac container projected-configmap-volume-test: 
+Jun 18 12:05:54.816: INFO: Pod "client-containers-68f8d798-91c1-11e9-a25d-8608290c688a" satisfied condition "success or failure"
+Jun 18 12:05:54.828: INFO: Trying to get logs from node 10.72.74.184 pod client-containers-68f8d798-91c1-11e9-a25d-8608290c688a container test-container: 
 STEP: delete the pod
-Jun  4 19:06:29.613: INFO: Waiting for pod pod-projected-configmaps-daa60a77-86fb-11e9-8d1b-467ee19922ac to disappear
-Jun  4 19:06:29.625: INFO: Pod pod-projected-configmaps-daa60a77-86fb-11e9-8d1b-467ee19922ac no longer exists
-[AfterEach] [sig-storage] Projected configMap
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 19:06:29.625: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "projected-6667" for this suite.
-Jun  4 19:06:35.695: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 19:06:36.073: INFO: namespace projected-6667 deletion completed in 6.434599537s
+Jun 18 12:05:54.908: INFO: Waiting for pod client-containers-68f8d798-91c1-11e9-a25d-8608290c688a to disappear
+Jun 18 12:05:54.927: INFO: Pod client-containers-68f8d798-91c1-11e9-a25d-8608290c688a no longer exists
+[AfterEach] [k8s.io] Docker Containers
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 12:05:54.927: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "containers-2277" for this suite.
+Jun 18 12:06:00.965: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 12:06:01.594: INFO: namespace containers-2277 deletion completed in 6.65404728s
 
-• [SLOW TEST:8.834 seconds]
-[sig-storage] Projected configMap
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_configmap.go:33
-  should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+• [SLOW TEST:13.054 seconds]
+[k8s.io] Docker Containers
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
+  should use the image defaults if command and args are blank [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-SSS
+SSSSSS
 ------------------------------
-[sig-storage] Projected secret 
-  should be consumable from pods in volume with mappings [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [sig-storage] Projected secret
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+[sig-apps] Daemon set [Serial] 
+  should run and stop simple daemon [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-apps] Daemon set [Serial]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Jun 18 12:06:01.594: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename daemonsets
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in daemonsets-4231
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [sig-apps] Daemon set [Serial]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:102
+[It] should run and stop simple daemon [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating simple DaemonSet "daemon-set"
+STEP: Check that daemon pods launch on every node of the cluster.
+Jun 18 12:06:01.898: INFO: Number of nodes with available pods: 0
+Jun 18 12:06:01.898: INFO: Node 10.72.74.138 is running more than one daemon pod
+Jun 18 12:06:02.932: INFO: Number of nodes with available pods: 0
+Jun 18 12:06:02.932: INFO: Node 10.72.74.138 is running more than one daemon pod
+Jun 18 12:06:03.927: INFO: Number of nodes with available pods: 3
+Jun 18 12:06:03.927: INFO: Number of running nodes: 3, number of available pods: 3
+STEP: Stop a daemon pod, check that the daemon pod is revived.
+Jun 18 12:06:04.003: INFO: Number of nodes with available pods: 2
+Jun 18 12:06:04.003: INFO: Node 10.72.74.184 is running more than one daemon pod
+Jun 18 12:06:05.033: INFO: Number of nodes with available pods: 2
+Jun 18 12:06:05.033: INFO: Node 10.72.74.184 is running more than one daemon pod
+Jun 18 12:06:06.035: INFO: Number of nodes with available pods: 2
+Jun 18 12:06:06.035: INFO: Node 10.72.74.184 is running more than one daemon pod
+Jun 18 12:06:07.041: INFO: Number of nodes with available pods: 2
+Jun 18 12:06:07.041: INFO: Node 10.72.74.184 is running more than one daemon pod
+Jun 18 12:06:08.034: INFO: Number of nodes with available pods: 2
+Jun 18 12:06:08.034: INFO: Node 10.72.74.184 is running more than one daemon pod
+Jun 18 12:06:09.032: INFO: Number of nodes with available pods: 2
+Jun 18 12:06:09.033: INFO: Node 10.72.74.184 is running more than one daemon pod
+Jun 18 12:06:10.032: INFO: Number of nodes with available pods: 3
+Jun 18 12:06:10.033: INFO: Number of running nodes: 3, number of available pods: 3
+[AfterEach] [sig-apps] Daemon set [Serial]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:68
+STEP: Deleting DaemonSet "daemon-set"
+STEP: deleting DaemonSet.extensions daemon-set in namespace daemonsets-4231, will wait for the garbage collector to delete the pods
+Jun 18 12:06:10.146: INFO: Deleting DaemonSet.extensions daemon-set took: 36.825556ms
+Jun 18 12:06:10.446: INFO: Terminating DaemonSet.extensions daemon-set pods took: 300.451181ms
+Jun 18 12:06:23.960: INFO: Number of nodes with available pods: 0
+Jun 18 12:06:23.960: INFO: Number of running nodes: 0, number of available pods: 0
+Jun 18 12:06:23.974: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"selfLink":"/apis/apps/v1/namespaces/daemonsets-4231/daemonsets","resourceVersion":"92570"},"items":null}
+
+Jun 18 12:06:23.990: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"selfLink":"/api/v1/namespaces/daemonsets-4231/pods","resourceVersion":"92570"},"items":null}
+
+[AfterEach] [sig-apps] Daemon set [Serial]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 12:06:24.061: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "daemonsets-4231" for this suite.
+Jun 18 12:06:32.099: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 12:06:32.491: INFO: namespace daemonsets-4231 deletion completed in 8.417563148s
+
+• [SLOW TEST:30.896 seconds]
+[sig-apps] Daemon set [Serial]
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22
+  should run and stop simple daemon [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+------------------------------
+[k8s.io] Probing container 
+  should *not* be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [k8s.io] Probing container
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Jun 18 12:06:32.494: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename container-probe
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in container-probe-7078
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [k8s.io] Probing container
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/container_probe.go:51
+[It] should *not* be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating pod liveness-exec in namespace container-probe-7078
+Jun 18 12:06:36.745: INFO: Started pod liveness-exec in namespace container-probe-7078
+STEP: checking the pod's current state and verifying that restartCount is present
+Jun 18 12:06:36.758: INFO: Initial restart count of pod liveness-exec is 0
+STEP: deleting the pod
+[AfterEach] [k8s.io] Probing container
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 12:10:37.649: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "container-probe-7078" for this suite.
+Jun 18 12:10:43.694: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 12:10:44.154: INFO: namespace container-probe-7078 deletion completed in 6.490891677s
+
+• [SLOW TEST:251.660 seconds]
+[k8s.io] Probing container
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
+  should *not* be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSSSSSSSSSSSSSS
+------------------------------
+[sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] 
+  should perform canary updates and phased rolling updates of template modifications [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-apps] StatefulSet
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Jun 18 12:10:44.154: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename statefulset
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in statefulset-5138
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [sig-apps] StatefulSet
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:59
+[BeforeEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:74
+STEP: Creating service test in namespace statefulset-5138
+[It] should perform canary updates and phased rolling updates of template modifications [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating a new StatefulSet
+Jun 18 12:10:44.395: INFO: Found 0 stateful pods, waiting for 3
+Jun 18 12:10:54.409: INFO: Waiting for pod ss2-0 to enter Running - Ready=true, currently Running - Ready=true
+Jun 18 12:10:54.410: INFO: Waiting for pod ss2-1 to enter Running - Ready=true, currently Running - Ready=true
+Jun 18 12:10:54.410: INFO: Waiting for pod ss2-2 to enter Running - Ready=true, currently Running - Ready=true
+STEP: Updating stateful set template: update image from docker.io/library/nginx:1.14-alpine to docker.io/library/nginx:1.15-alpine
+Jun 18 12:10:54.496: INFO: Updating stateful set ss2
+STEP: Creating a new revision
+STEP: Not applying an update when the partition is greater than the number of replicas
+STEP: Performing a canary update
+Jun 18 12:11:04.591: INFO: Updating stateful set ss2
+Jun 18 12:11:04.628: INFO: Waiting for Pod statefulset-5138/ss2-2 to have revision ss2-c79899b9 update revision ss2-787997d666
+Jun 18 12:11:14.654: INFO: Waiting for Pod statefulset-5138/ss2-2 to have revision ss2-c79899b9 update revision ss2-787997d666
+STEP: Restoring Pods to the correct revision when they are deleted
+Jun 18 12:11:24.786: INFO: Found 2 stateful pods, waiting for 3
+Jun 18 12:11:34.800: INFO: Waiting for pod ss2-0 to enter Running - Ready=true, currently Running - Ready=true
+Jun 18 12:11:34.800: INFO: Waiting for pod ss2-1 to enter Running - Ready=true, currently Running - Ready=true
+Jun 18 12:11:34.800: INFO: Waiting for pod ss2-2 to enter Running - Ready=true, currently Running - Ready=true
+STEP: Performing a phased rolling update
+Jun 18 12:11:34.858: INFO: Updating stateful set ss2
+Jun 18 12:11:34.881: INFO: Waiting for Pod statefulset-5138/ss2-1 to have revision ss2-c79899b9 update revision ss2-787997d666
+Jun 18 12:11:44.940: INFO: Updating stateful set ss2
+Jun 18 12:11:44.964: INFO: Waiting for StatefulSet statefulset-5138/ss2 to complete update
+Jun 18 12:11:44.964: INFO: Waiting for Pod statefulset-5138/ss2-0 to have revision ss2-c79899b9 update revision ss2-787997d666
+Jun 18 12:11:54.989: INFO: Waiting for StatefulSet statefulset-5138/ss2 to complete update
+Jun 18 12:11:54.989: INFO: Waiting for Pod statefulset-5138/ss2-0 to have revision ss2-c79899b9 update revision ss2-787997d666
+Jun 18 12:12:04.991: INFO: Waiting for StatefulSet statefulset-5138/ss2 to complete update
+[AfterEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:85
+Jun 18 12:12:14.993: INFO: Deleting all statefulset in ns statefulset-5138
+Jun 18 12:12:15.005: INFO: Scaling statefulset ss2 to 0
+Jun 18 12:12:35.076: INFO: Waiting for statefulset status.replicas updated to 0
+Jun 18 12:12:35.088: INFO: Deleting statefulset ss2
+[AfterEach] [sig-apps] StatefulSet
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 12:12:35.139: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "statefulset-5138" for this suite.
+Jun 18 12:12:43.179: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 12:12:43.563: INFO: namespace statefulset-5138 deletion completed in 8.41070133s
+
+• [SLOW TEST:119.409 seconds]
+[sig-apps] StatefulSet
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22
+  [k8s.io] Basic StatefulSet functionality [StatefulSetBasic]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
+    should perform canary updates and phased rolling updates of template modifications [Conformance]
+    /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSSSSSSSSSSSSS
+------------------------------
+[sig-storage] Projected downwardAPI 
+  should provide podname only [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-storage] Projected downwardAPI
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 19:06:36.074: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
+Jun 18 12:12:43.568: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
 STEP: Building a namespace api object, basename projected
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-2674
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-5470
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should be consumable from pods in volume with mappings [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: Creating projection with secret that has name projected-secret-test-map-dfe7955d-86fb-11e9-8d1b-467ee19922ac
-STEP: Creating a pod to test consume secrets
-Jun  4 19:06:36.323: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-dfe93048-86fb-11e9-8d1b-467ee19922ac" in namespace "projected-2674" to be "success or failure"
-Jun  4 19:06:36.335: INFO: Pod "pod-projected-secrets-dfe93048-86fb-11e9-8d1b-467ee19922ac": Phase="Pending", Reason="", readiness=false. Elapsed: 11.671018ms
-Jun  4 19:06:38.347: INFO: Pod "pod-projected-secrets-dfe93048-86fb-11e9-8d1b-467ee19922ac": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.024246721s
+[BeforeEach] [sig-storage] Projected downwardAPI
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:39
+[It] should provide podname only [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating a pod to test downward API volume plugin
+Jun 18 12:12:43.802: INFO: Waiting up to 5m0s for pod "downwardapi-volume-605977c6-91c2-11e9-a25d-8608290c688a" in namespace "projected-5470" to be "success or failure"
+Jun 18 12:12:43.823: INFO: Pod "downwardapi-volume-605977c6-91c2-11e9-a25d-8608290c688a": Phase="Pending", Reason="", readiness=false. Elapsed: 21.365035ms
+Jun 18 12:12:45.838: INFO: Pod "downwardapi-volume-605977c6-91c2-11e9-a25d-8608290c688a": Phase="Running", Reason="", readiness=true. Elapsed: 2.036051314s
+Jun 18 12:12:48.101: INFO: Pod "downwardapi-volume-605977c6-91c2-11e9-a25d-8608290c688a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.29921862s
 STEP: Saw pod success
-Jun  4 19:06:38.348: INFO: Pod "pod-projected-secrets-dfe93048-86fb-11e9-8d1b-467ee19922ac" satisfied condition "success or failure"
-Jun  4 19:06:38.360: INFO: Trying to get logs from node 10.212.23.189 pod pod-projected-secrets-dfe93048-86fb-11e9-8d1b-467ee19922ac container projected-secret-volume-test: 
+Jun 18 12:12:48.101: INFO: Pod "downwardapi-volume-605977c6-91c2-11e9-a25d-8608290c688a" satisfied condition "success or failure"
+Jun 18 12:12:48.125: INFO: Trying to get logs from node 10.72.74.184 pod downwardapi-volume-605977c6-91c2-11e9-a25d-8608290c688a container client-container: 
 STEP: delete the pod
-Jun  4 19:06:38.417: INFO: Waiting for pod pod-projected-secrets-dfe93048-86fb-11e9-8d1b-467ee19922ac to disappear
-Jun  4 19:06:38.427: INFO: Pod pod-projected-secrets-dfe93048-86fb-11e9-8d1b-467ee19922ac no longer exists
-[AfterEach] [sig-storage] Projected secret
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 19:06:38.427: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "projected-2674" for this suite.
-Jun  4 19:06:44.510: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 19:06:45.254: INFO: namespace projected-2674 deletion completed in 6.783917812s
+Jun 18 12:12:48.214: INFO: Waiting for pod downwardapi-volume-605977c6-91c2-11e9-a25d-8608290c688a to disappear
+Jun 18 12:12:48.227: INFO: Pod downwardapi-volume-605977c6-91c2-11e9-a25d-8608290c688a no longer exists
+[AfterEach] [sig-storage] Projected downwardAPI
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 12:12:48.227: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "projected-5470" for this suite.
+Jun 18 12:12:54.658: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 12:12:55.056: INFO: namespace projected-5470 deletion completed in 6.811834785s
 
-• [SLOW TEST:9.180 seconds]
-[sig-storage] Projected secret
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_secret.go:33
-  should be consumable from pods in volume with mappings [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-------------------------------
-SSSS
+• [SLOW TEST:11.488 seconds]
+[sig-storage] Projected downwardAPI
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:33
+  should provide podname only [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
 [sig-storage] EmptyDir volumes 
-  should support (root,0777,default) [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+  should support (root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 [BeforeEach] [sig-storage] EmptyDir volumes
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 19:06:45.254: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
+Jun 18 12:12:55.057: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
 STEP: Building a namespace api object, basename emptydir
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in emptydir-8006
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in emptydir-1574
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should support (root,0777,default) [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: Creating a pod to test emptydir 0777 on node default medium
-Jun  4 19:06:45.495: INFO: Waiting up to 5m0s for pod "pod-e5603a30-86fb-11e9-8d1b-467ee19922ac" in namespace "emptydir-8006" to be "success or failure"
-Jun  4 19:06:45.506: INFO: Pod "pod-e5603a30-86fb-11e9-8d1b-467ee19922ac": Phase="Pending", Reason="", readiness=false. Elapsed: 11.128355ms
-Jun  4 19:06:47.522: INFO: Pod "pod-e5603a30-86fb-11e9-8d1b-467ee19922ac": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.027170433s
+[It] should support (root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating a pod to test emptydir 0777 on tmpfs
+Jun 18 12:12:55.730: INFO: Waiting up to 5m0s for pod "pod-6776fd60-91c2-11e9-a25d-8608290c688a" in namespace "emptydir-1574" to be "success or failure"
+Jun 18 12:12:55.743: INFO: Pod "pod-6776fd60-91c2-11e9-a25d-8608290c688a": Phase="Pending", Reason="", readiness=false. Elapsed: 13.436787ms
+Jun 18 12:12:57.756: INFO: Pod "pod-6776fd60-91c2-11e9-a25d-8608290c688a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.02643586s
 STEP: Saw pod success
-Jun  4 19:06:47.523: INFO: Pod "pod-e5603a30-86fb-11e9-8d1b-467ee19922ac" satisfied condition "success or failure"
-Jun  4 19:06:47.555: INFO: Trying to get logs from node 10.212.23.161 pod pod-e5603a30-86fb-11e9-8d1b-467ee19922ac container test-container: 
+Jun 18 12:12:57.756: INFO: Pod "pod-6776fd60-91c2-11e9-a25d-8608290c688a" satisfied condition "success or failure"
+Jun 18 12:12:57.770: INFO: Trying to get logs from node 10.72.74.138 pod pod-6776fd60-91c2-11e9-a25d-8608290c688a container test-container: 
 STEP: delete the pod
-Jun  4 19:06:47.614: INFO: Waiting for pod pod-e5603a30-86fb-11e9-8d1b-467ee19922ac to disappear
-Jun  4 19:06:47.626: INFO: Pod pod-e5603a30-86fb-11e9-8d1b-467ee19922ac no longer exists
+Jun 18 12:12:57.848: INFO: Waiting for pod pod-6776fd60-91c2-11e9-a25d-8608290c688a to disappear
+Jun 18 12:12:57.861: INFO: Pod pod-6776fd60-91c2-11e9-a25d-8608290c688a no longer exists
 [AfterEach] [sig-storage] EmptyDir volumes
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 19:06:47.627: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "emptydir-8006" for this suite.
-Jun  4 19:06:53.685: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 19:06:54.089: INFO: namespace emptydir-8006 deletion completed in 6.44764857s
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 12:12:57.861: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "emptydir-1574" for this suite.
+Jun 18 12:13:03.904: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 12:13:04.322: INFO: namespace emptydir-1574 deletion completed in 6.447538672s
 
-• [SLOW TEST:8.835 seconds]
+• [SLOW TEST:9.266 seconds]
 [sig-storage] EmptyDir volumes
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:41
-  should support (root,0777,default) [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:41
+  should support (root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-SSSSSSSSSSSSSSSS
+SSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[k8s.io] Container Lifecycle Hook when create a pod with lifecycle hook 
-  should execute poststart exec hook properly [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [k8s.io] Container Lifecycle Hook
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+[sig-storage] Projected configMap 
+  should be consumable from pods in volume as non-root [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-storage] Projected configMap
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 19:06:54.090: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-STEP: Building a namespace api object, basename container-lifecycle-hook
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in container-lifecycle-hook-5975
+Jun 18 12:13:04.323: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename projected
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-5413
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] when create a pod with lifecycle hook
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:61
-STEP: create the container to handle the HTTPGet hook request.
-[It] should execute poststart exec hook properly [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: create the pod with lifecycle hook
-STEP: check poststart hook
-STEP: delete the pod with lifecycle hook
-Jun  4 19:06:58.556: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear
-Jun  4 19:06:58.567: INFO: Pod pod-with-poststart-exec-hook still exists
-Jun  4 19:07:00.567: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear
-Jun  4 19:07:00.580: INFO: Pod pod-with-poststart-exec-hook still exists
-Jun  4 19:07:02.567: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear
-Jun  4 19:07:02.581: INFO: Pod pod-with-poststart-exec-hook still exists
-Jun  4 19:07:04.568: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear
-Jun  4 19:07:04.580: INFO: Pod pod-with-poststart-exec-hook still exists
-Jun  4 19:07:06.568: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear
-Jun  4 19:07:06.581: INFO: Pod pod-with-poststart-exec-hook still exists
-Jun  4 19:07:08.567: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear
-Jun  4 19:07:08.581: INFO: Pod pod-with-poststart-exec-hook still exists
-Jun  4 19:07:10.567: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear
-Jun  4 19:07:10.580: INFO: Pod pod-with-poststart-exec-hook still exists
-Jun  4 19:07:12.567: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear
-Jun  4 19:07:12.580: INFO: Pod pod-with-poststart-exec-hook still exists
-Jun  4 19:07:14.568: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear
-Jun  4 19:07:14.597: INFO: Pod pod-with-poststart-exec-hook still exists
-Jun  4 19:07:16.568: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear
-Jun  4 19:07:16.579: INFO: Pod pod-with-poststart-exec-hook still exists
-Jun  4 19:07:18.568: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear
-Jun  4 19:07:18.581: INFO: Pod pod-with-poststart-exec-hook no longer exists
-[AfterEach] [k8s.io] Container Lifecycle Hook
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 19:07:18.581: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "container-lifecycle-hook-5975" for this suite.
-Jun  4 19:07:42.639: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 19:07:42.994: INFO: namespace container-lifecycle-hook-5975 deletion completed in 24.397356921s
+[It] should be consumable from pods in volume as non-root [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating configMap with name projected-configmap-test-volume-6cbb3848-91c2-11e9-a25d-8608290c688a
+STEP: Creating a pod to test consume configMaps
+Jun 18 12:13:04.583: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-6cbd7006-91c2-11e9-a25d-8608290c688a" in namespace "projected-5413" to be "success or failure"
+Jun 18 12:13:04.596: INFO: Pod "pod-projected-configmaps-6cbd7006-91c2-11e9-a25d-8608290c688a": Phase="Pending", Reason="", readiness=false. Elapsed: 13.104747ms
+Jun 18 12:13:06.610: INFO: Pod "pod-projected-configmaps-6cbd7006-91c2-11e9-a25d-8608290c688a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.026609184s
+STEP: Saw pod success
+Jun 18 12:13:06.610: INFO: Pod "pod-projected-configmaps-6cbd7006-91c2-11e9-a25d-8608290c688a" satisfied condition "success or failure"
+Jun 18 12:13:06.623: INFO: Trying to get logs from node 10.72.74.184 pod pod-projected-configmaps-6cbd7006-91c2-11e9-a25d-8608290c688a container projected-configmap-volume-test: 
+STEP: delete the pod
+Jun 18 12:13:06.701: INFO: Waiting for pod pod-projected-configmaps-6cbd7006-91c2-11e9-a25d-8608290c688a to disappear
+Jun 18 12:13:06.721: INFO: Pod pod-projected-configmaps-6cbd7006-91c2-11e9-a25d-8608290c688a no longer exists
+[AfterEach] [sig-storage] Projected configMap
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 12:13:06.721: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "projected-5413" for this suite.
+Jun 18 12:13:12.771: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 12:13:13.200: INFO: namespace projected-5413 deletion completed in 6.461593244s
 
-• [SLOW TEST:48.904 seconds]
-[k8s.io] Container Lifecycle Hook
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
-  when create a pod with lifecycle hook
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:40
-    should execute poststart exec hook properly [NodeConformance] [Conformance]
-    /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+• [SLOW TEST:8.877 seconds]
+[sig-storage] Projected configMap
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_configmap.go:33
+  should be consumable from pods in volume as non-root [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-SSSSSSSSSSSSSSSSS
+SSSS
 ------------------------------
 [sig-storage] EmptyDir volumes 
-  should support (non-root,0644,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+  should support (root,0777,default) [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 [BeforeEach] [sig-storage] EmptyDir volumes
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 19:07:42.994: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
+Jun 18 12:13:13.200: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
 STEP: Building a namespace api object, basename emptydir
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in emptydir-1086
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in emptydir-7827
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should support (non-root,0644,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: Creating a pod to test emptydir 0644 on tmpfs
-Jun  4 19:07:43.232: INFO: Waiting up to 5m0s for pod "pod-07caf125-86fc-11e9-8d1b-467ee19922ac" in namespace "emptydir-1086" to be "success or failure"
-Jun  4 19:07:43.243: INFO: Pod "pod-07caf125-86fc-11e9-8d1b-467ee19922ac": Phase="Pending", Reason="", readiness=false. Elapsed: 10.764059ms
-Jun  4 19:07:45.259: INFO: Pod "pod-07caf125-86fc-11e9-8d1b-467ee19922ac": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.026825324s
+[It] should support (root,0777,default) [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating a pod to test emptydir 0777 on node default medium
+Jun 18 12:13:13.422: INFO: Waiting up to 5m0s for pod "pod-7202808f-91c2-11e9-a25d-8608290c688a" in namespace "emptydir-7827" to be "success or failure"
+Jun 18 12:13:13.442: INFO: Pod "pod-7202808f-91c2-11e9-a25d-8608290c688a": Phase="Pending", Reason="", readiness=false. Elapsed: 19.813455ms
+Jun 18 12:13:15.456: INFO: Pod "pod-7202808f-91c2-11e9-a25d-8608290c688a": Phase="Pending", Reason="", readiness=false. Elapsed: 2.033425263s
+Jun 18 12:13:17.471: INFO: Pod "pod-7202808f-91c2-11e9-a25d-8608290c688a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.048470239s
 STEP: Saw pod success
-Jun  4 19:07:45.259: INFO: Pod "pod-07caf125-86fc-11e9-8d1b-467ee19922ac" satisfied condition "success or failure"
-Jun  4 19:07:45.270: INFO: Trying to get logs from node 10.212.23.161 pod pod-07caf125-86fc-11e9-8d1b-467ee19922ac container test-container: 
+Jun 18 12:13:17.471: INFO: Pod "pod-7202808f-91c2-11e9-a25d-8608290c688a" satisfied condition "success or failure"
+Jun 18 12:13:17.485: INFO: Trying to get logs from node 10.72.74.138 pod pod-7202808f-91c2-11e9-a25d-8608290c688a container test-container: 
 STEP: delete the pod
-Jun  4 19:07:45.355: INFO: Waiting for pod pod-07caf125-86fc-11e9-8d1b-467ee19922ac to disappear
-Jun  4 19:07:45.370: INFO: Pod pod-07caf125-86fc-11e9-8d1b-467ee19922ac no longer exists
+Jun 18 12:13:17.707: INFO: Waiting for pod pod-7202808f-91c2-11e9-a25d-8608290c688a to disappear
+Jun 18 12:13:17.722: INFO: Pod pod-7202808f-91c2-11e9-a25d-8608290c688a no longer exists
 [AfterEach] [sig-storage] EmptyDir volumes
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 19:07:45.370: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "emptydir-1086" for this suite.
-Jun  4 19:07:51.423: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 19:07:51.757: INFO: namespace emptydir-1086 deletion completed in 6.373460121s
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 12:13:17.722: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "emptydir-7827" for this suite.
+Jun 18 12:13:25.769: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 12:13:26.223: INFO: namespace emptydir-7827 deletion completed in 8.487630106s
 
-• [SLOW TEST:8.762 seconds]
+• [SLOW TEST:13.024 seconds]
 [sig-storage] EmptyDir volumes
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:41
-  should support (non-root,0644,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-------------------------------
-SSSSSSSSSSSSSSSSSSSSSSSSS
-------------------------------
-[sig-api-machinery] Garbage collector 
-  should orphan RS created by deployment when deleteOptions.PropagationPolicy is Orphan [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [sig-api-machinery] Garbage collector
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
-STEP: Creating a kubernetes client
-Jun  4 19:07:51.757: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-STEP: Building a namespace api object, basename gc
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in gc-2748
-STEP: Waiting for a default service account to be provisioned in namespace
-[It] should orphan RS created by deployment when deleteOptions.PropagationPolicy is Orphan [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: create the deployment
-STEP: Wait for the Deployment to create new ReplicaSet
-STEP: delete the deployment
-STEP: wait for 30 seconds to see if the garbage collector mistakenly deletes the rs
-STEP: Gathering metrics
-W0604 19:08:22.689955      18 metrics_grabber.go:79] Master node is not registered. Grabbing metrics from Scheduler, ControllerManager and ClusterAutoscaler is disabled.
-Jun  4 19:08:22.689: INFO: For apiserver_request_total:
-For apiserver_request_latencies_summary:
-For apiserver_init_events_total:
-For garbage_collector_attempt_to_delete_queue_latency:
-For garbage_collector_attempt_to_delete_work_duration:
-For garbage_collector_attempt_to_orphan_queue_latency:
-For garbage_collector_attempt_to_orphan_work_duration:
-For garbage_collector_dirty_processing_latency_microseconds:
-For garbage_collector_event_processing_latency_microseconds:
-For garbage_collector_graph_changes_queue_latency:
-For garbage_collector_graph_changes_work_duration:
-For garbage_collector_orphan_processing_latency_microseconds:
-For namespace_queue_latency:
-For namespace_queue_latency_sum:
-For namespace_queue_latency_count:
-For namespace_retries:
-For namespace_work_duration:
-For namespace_work_duration_sum:
-For namespace_work_duration_count:
-For function_duration_seconds:
-For errors_total:
-For evicted_pods_total:
-
-[AfterEach] [sig-api-machinery] Garbage collector
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 19:08:22.690: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "gc-2748" for this suite.
-Jun  4 19:08:28.749: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 19:08:29.186: INFO: namespace gc-2748 deletion completed in 6.480830957s
-
-• [SLOW TEST:37.429 seconds]
-[sig-api-machinery] Garbage collector
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:22
-  should orphan RS created by deployment when deleteOptions.PropagationPolicy is Orphan [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:41
+  should support (root,0777,default) [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-SSSSSSSSSSSSS
+SS
 ------------------------------
 [sig-auth] ServiceAccounts 
   should mount an API token into pods  [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 [BeforeEach] [sig-auth] ServiceAccounts
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 19:08:29.186: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
+Jun 18 12:13:26.224: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
 STEP: Building a namespace api object, basename svcaccounts
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in svcaccounts-4055
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in svcaccounts-6255
 STEP: Waiting for a default service account to be provisioned in namespace
 [It] should mount an API token into pods  [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 STEP: getting the auto-created API token
 STEP: reading a file in the container
-Jun  4 19:08:32.045: INFO: Running '/usr/local/bin/kubectl exec --namespace=svcaccounts-4055 pod-service-account-23ad6455-86fc-11e9-8d1b-467ee19922ac -c=test -- cat /var/run/secrets/kubernetes.io/serviceaccount/token'
+Jun 18 12:13:29.016: INFO: Running '/usr/local/bin/kubectl exec --namespace=svcaccounts-6255 pod-service-account-7a181ad3-91c2-11e9-a25d-8608290c688a -c=test -- cat /var/run/secrets/kubernetes.io/serviceaccount/token'
 STEP: reading a file in the container
-Jun  4 19:08:32.427: INFO: Running '/usr/local/bin/kubectl exec --namespace=svcaccounts-4055 pod-service-account-23ad6455-86fc-11e9-8d1b-467ee19922ac -c=test -- cat /var/run/secrets/kubernetes.io/serviceaccount/ca.crt'
+Jun 18 12:13:29.460: INFO: Running '/usr/local/bin/kubectl exec --namespace=svcaccounts-6255 pod-service-account-7a181ad3-91c2-11e9-a25d-8608290c688a -c=test -- cat /var/run/secrets/kubernetes.io/serviceaccount/ca.crt'
 STEP: reading a file in the container
-Jun  4 19:08:32.680: INFO: Running '/usr/local/bin/kubectl exec --namespace=svcaccounts-4055 pod-service-account-23ad6455-86fc-11e9-8d1b-467ee19922ac -c=test -- cat /var/run/secrets/kubernetes.io/serviceaccount/namespace'
+Jun 18 12:13:29.865: INFO: Running '/usr/local/bin/kubectl exec --namespace=svcaccounts-6255 pod-service-account-7a181ad3-91c2-11e9-a25d-8608290c688a -c=test -- cat /var/run/secrets/kubernetes.io/serviceaccount/namespace'
 [AfterEach] [sig-auth] ServiceAccounts
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 19:08:32.959: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "svcaccounts-4055" for this suite.
-Jun  4 19:08:39.017: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 19:08:39.478: INFO: namespace svcaccounts-4055 deletion completed in 6.502633007s
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 12:13:30.233: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "svcaccounts-6255" for this suite.
+Jun 18 12:13:36.273: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 12:13:36.683: INFO: namespace svcaccounts-6255 deletion completed in 6.436816343s
 
-• [SLOW TEST:10.292 seconds]
+• [SLOW TEST:10.460 seconds]
 [sig-auth] ServiceAccounts
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/auth/framework.go:22
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/auth/framework.go:22
   should mount an API token into pods  [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-[sig-storage] Projected configMap 
-  should be consumable from pods in volume with mappings [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [sig-storage] Projected configMap
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+------------------------------
+[sig-storage] ConfigMap 
+  should be consumable from pods in volume [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-storage] ConfigMap
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 19:08:39.478: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-STEP: Building a namespace api object, basename projected
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-7029
+Jun 18 12:13:36.685: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename configmap
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in configmap-134
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should be consumable from pods in volume with mappings [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: Creating configMap with name projected-configmap-test-volume-map-297582f5-86fc-11e9-8d1b-467ee19922ac
+[It] should be consumable from pods in volume [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating configMap with name configmap-test-volume-8001f9e7-91c2-11e9-a25d-8608290c688a
 STEP: Creating a pod to test consume configMaps
-Jun  4 19:08:39.729: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-297766c4-86fc-11e9-8d1b-467ee19922ac" in namespace "projected-7029" to be "success or failure"
-Jun  4 19:08:39.741: INFO: Pod "pod-projected-configmaps-297766c4-86fc-11e9-8d1b-467ee19922ac": Phase="Pending", Reason="", readiness=false. Elapsed: 11.458448ms
-Jun  4 19:08:41.753: INFO: Pod "pod-projected-configmaps-297766c4-86fc-11e9-8d1b-467ee19922ac": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.023303653s
+Jun 18 12:13:36.923: INFO: Waiting up to 5m0s for pod "pod-configmaps-80041deb-91c2-11e9-a25d-8608290c688a" in namespace "configmap-134" to be "success or failure"
+Jun 18 12:13:36.946: INFO: Pod "pod-configmaps-80041deb-91c2-11e9-a25d-8608290c688a": Phase="Pending", Reason="", readiness=false. Elapsed: 23.113445ms
+Jun 18 12:13:38.960: INFO: Pod "pod-configmaps-80041deb-91c2-11e9-a25d-8608290c688a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.036661576s
 STEP: Saw pod success
-Jun  4 19:08:41.753: INFO: Pod "pod-projected-configmaps-297766c4-86fc-11e9-8d1b-467ee19922ac" satisfied condition "success or failure"
-Jun  4 19:08:41.765: INFO: Trying to get logs from node 10.212.23.161 pod pod-projected-configmaps-297766c4-86fc-11e9-8d1b-467ee19922ac container projected-configmap-volume-test: 
+Jun 18 12:13:38.960: INFO: Pod "pod-configmaps-80041deb-91c2-11e9-a25d-8608290c688a" satisfied condition "success or failure"
+Jun 18 12:13:38.972: INFO: Trying to get logs from node 10.72.74.138 pod pod-configmaps-80041deb-91c2-11e9-a25d-8608290c688a container configmap-volume-test: 
 STEP: delete the pod
-Jun  4 19:08:41.917: INFO: Waiting for pod pod-projected-configmaps-297766c4-86fc-11e9-8d1b-467ee19922ac to disappear
-Jun  4 19:08:41.926: INFO: Pod pod-projected-configmaps-297766c4-86fc-11e9-8d1b-467ee19922ac no longer exists
-[AfterEach] [sig-storage] Projected configMap
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 19:08:41.926: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "projected-7029" for this suite.
-Jun  4 19:08:47.983: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 19:08:48.371: INFO: namespace projected-7029 deletion completed in 6.431112321s
+Jun 18 12:13:39.050: INFO: Waiting for pod pod-configmaps-80041deb-91c2-11e9-a25d-8608290c688a to disappear
+Jun 18 12:13:39.063: INFO: Pod pod-configmaps-80041deb-91c2-11e9-a25d-8608290c688a no longer exists
+[AfterEach] [sig-storage] ConfigMap
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 12:13:39.063: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "configmap-134" for this suite.
+Jun 18 12:13:45.108: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 12:13:45.494: INFO: namespace configmap-134 deletion completed in 6.411394734s
 
-• [SLOW TEST:8.893 seconds]
-[sig-storage] Projected configMap
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_configmap.go:33
-  should be consumable from pods in volume with mappings [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+• [SLOW TEST:8.809 seconds]
+[sig-storage] ConfigMap
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:32
+  should be consumable from pods in volume [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-SSSSSSSSSSSSSSS
+SS
 ------------------------------
-[sig-storage] Projected downwardAPI 
-  should set mode on item file [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [sig-storage] Projected downwardAPI
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+[sig-cli] Kubectl client [k8s.io] Kubectl rolling-update 
+  should support rolling-update to same image  [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-cli] Kubectl client
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 19:08:48.372: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-STEP: Building a namespace api object, basename projected
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-1587
+Jun 18 12:13:45.494: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename kubectl
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-3308
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-storage] Projected downwardAPI
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:39
-[It] should set mode on item file [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: Creating a pod to test downward API volume plugin
-Jun  4 19:08:48.610: INFO: Waiting up to 5m0s for pod "downwardapi-volume-2ec2ae08-86fc-11e9-8d1b-467ee19922ac" in namespace "projected-1587" to be "success or failure"
-Jun  4 19:08:48.621: INFO: Pod "downwardapi-volume-2ec2ae08-86fc-11e9-8d1b-467ee19922ac": Phase="Pending", Reason="", readiness=false. Elapsed: 11.472931ms
-Jun  4 19:08:50.635: INFO: Pod "downwardapi-volume-2ec2ae08-86fc-11e9-8d1b-467ee19922ac": Phase="Pending", Reason="", readiness=false. Elapsed: 2.025076339s
-Jun  4 19:08:52.648: INFO: Pod "downwardapi-volume-2ec2ae08-86fc-11e9-8d1b-467ee19922ac": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.037934895s
-STEP: Saw pod success
-Jun  4 19:08:52.648: INFO: Pod "downwardapi-volume-2ec2ae08-86fc-11e9-8d1b-467ee19922ac" satisfied condition "success or failure"
-Jun  4 19:08:52.667: INFO: Trying to get logs from node 10.212.23.164 pod downwardapi-volume-2ec2ae08-86fc-11e9-8d1b-467ee19922ac container client-container: 
-STEP: delete the pod
-Jun  4 19:08:52.795: INFO: Waiting for pod downwardapi-volume-2ec2ae08-86fc-11e9-8d1b-467ee19922ac to disappear
-Jun  4 19:08:52.805: INFO: Pod downwardapi-volume-2ec2ae08-86fc-11e9-8d1b-467ee19922ac no longer exists
-[AfterEach] [sig-storage] Projected downwardAPI
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 19:08:52.805: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "projected-1587" for this suite.
-Jun  4 19:08:58.869: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 19:08:59.217: INFO: namespace projected-1587 deletion completed in 6.391257388s
+[BeforeEach] [sig-cli] Kubectl client
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:213
+[BeforeEach] [k8s.io] Kubectl rolling-update
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1414
+[It] should support rolling-update to same image  [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: running the image docker.io/library/nginx:1.14-alpine
+Jun 18 12:13:45.690: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 run e2e-test-nginx-rc --image=docker.io/library/nginx:1.14-alpine --generator=run/v1 --namespace=kubectl-3308'
+Jun 18 12:13:45.932: INFO: stderr: "kubectl run --generator=run/v1 is DEPRECATED and will be removed in a future version. Use kubectl run --generator=run-pod/v1 or kubectl create instead.\n"
+Jun 18 12:13:45.932: INFO: stdout: "replicationcontroller/e2e-test-nginx-rc created\n"
+STEP: verifying the rc e2e-test-nginx-rc was created
+Jun 18 12:13:45.951: INFO: Waiting for rc e2e-test-nginx-rc to stabilize, generation 1 observed generation 0 spec.replicas 1 status.replicas 0
+Jun 18 12:13:45.985: INFO: Waiting for rc e2e-test-nginx-rc to stabilize, generation 1 observed generation 1 spec.replicas 1 status.replicas 0
+STEP: rolling-update to same image controller
+Jun 18 12:13:46.464: INFO: scanned /root for discovery docs: 
+Jun 18 12:13:46.464: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 rolling-update e2e-test-nginx-rc --update-period=1s --image=docker.io/library/nginx:1.14-alpine --image-pull-policy=IfNotPresent --namespace=kubectl-3308'
+Jun 18 12:14:02.511: INFO: stderr: "Command \"rolling-update\" is deprecated, use \"rollout\" instead\n"
+Jun 18 12:14:02.511: INFO: stdout: "Created e2e-test-nginx-rc-f6f24dcccc248a7a6145e45acaf203e8\nScaling up e2e-test-nginx-rc-f6f24dcccc248a7a6145e45acaf203e8 from 0 to 1, scaling down e2e-test-nginx-rc from 1 to 0 (keep 1 pods available, don't exceed 2 pods)\nScaling e2e-test-nginx-rc-f6f24dcccc248a7a6145e45acaf203e8 up to 1\nScaling e2e-test-nginx-rc down to 0\nUpdate succeeded. Deleting old controller: e2e-test-nginx-rc\nRenaming e2e-test-nginx-rc-f6f24dcccc248a7a6145e45acaf203e8 to e2e-test-nginx-rc\nreplicationcontroller/e2e-test-nginx-rc rolling updated\n"
+Jun 18 12:14:02.511: INFO: stdout: "Created e2e-test-nginx-rc-f6f24dcccc248a7a6145e45acaf203e8\nScaling up e2e-test-nginx-rc-f6f24dcccc248a7a6145e45acaf203e8 from 0 to 1, scaling down e2e-test-nginx-rc from 1 to 0 (keep 1 pods available, don't exceed 2 pods)\nScaling e2e-test-nginx-rc-f6f24dcccc248a7a6145e45acaf203e8 up to 1\nScaling e2e-test-nginx-rc down to 0\nUpdate succeeded. Deleting old controller: e2e-test-nginx-rc\nRenaming e2e-test-nginx-rc-f6f24dcccc248a7a6145e45acaf203e8 to e2e-test-nginx-rc\nreplicationcontroller/e2e-test-nginx-rc rolling updated\n"
+STEP: waiting for all containers in run=e2e-test-nginx-rc pods to come up.
+Jun 18 12:14:02.511: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l run=e2e-test-nginx-rc --namespace=kubectl-3308'
+Jun 18 12:14:02.627: INFO: stderr: ""
+Jun 18 12:14:02.627: INFO: stdout: "e2e-test-nginx-rc-f6f24dcccc248a7a6145e45acaf203e8-j2lnt "
+Jun 18 12:14:02.627: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 get pods e2e-test-nginx-rc-f6f24dcccc248a7a6145e45acaf203e8-j2lnt -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "e2e-test-nginx-rc") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-3308'
+Jun 18 12:14:02.745: INFO: stderr: ""
+Jun 18 12:14:02.745: INFO: stdout: "true"
+Jun 18 12:14:02.745: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 get pods e2e-test-nginx-rc-f6f24dcccc248a7a6145e45acaf203e8-j2lnt -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "e2e-test-nginx-rc"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-3308'
+Jun 18 12:14:02.867: INFO: stderr: ""
+Jun 18 12:14:02.867: INFO: stdout: "docker.io/library/nginx:1.14-alpine"
+Jun 18 12:14:02.867: INFO: e2e-test-nginx-rc-f6f24dcccc248a7a6145e45acaf203e8-j2lnt is verified up and running
+[AfterEach] [k8s.io] Kubectl rolling-update
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1420
+Jun 18 12:14:02.867: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 delete rc e2e-test-nginx-rc --namespace=kubectl-3308'
+Jun 18 12:14:03.033: INFO: stderr: ""
+Jun 18 12:14:03.033: INFO: stdout: "replicationcontroller \"e2e-test-nginx-rc\" deleted\n"
+[AfterEach] [sig-cli] Kubectl client
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 12:14:03.033: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "kubectl-3308" for this suite.
+Jun 18 12:14:27.074: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 12:14:27.465: INFO: namespace kubectl-3308 deletion completed in 24.417672293s
 
-• [SLOW TEST:10.845 seconds]
-[sig-storage] Projected downwardAPI
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:33
-  should set mode on item file [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-------------------------------
-SSSSSS
+• [SLOW TEST:41.971 seconds]
+[sig-cli] Kubectl client
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23
+  [k8s.io] Kubectl rolling-update
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
+    should support rolling-update to same image  [Conformance]
+    /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-[sig-apps] Deployment 
-  deployment should support proportional scaling [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [sig-apps] Deployment
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+SSSSSSSSSSSSSSS
+------------------------------
+[sig-apps] ReplicaSet 
+  should serve a basic image on each replica with a public image  [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-apps] ReplicaSet
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 19:08:59.217: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-STEP: Building a namespace api object, basename deployment
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in deployment-4196
+Jun 18 12:14:27.466: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename replicaset
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in replicaset-3213
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-apps] Deployment
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:65
-[It] deployment should support proportional scaling [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-Jun  4 19:08:59.439: INFO: Creating deployment "nginx-deployment"
-Jun  4 19:08:59.448: INFO: Waiting for observed generation 1
-Jun  4 19:09:01.469: INFO: Waiting for all required pods to come up
-Jun  4 19:09:01.482: INFO: Pod name nginx: Found 10 pods out of 10
-STEP: ensuring each pod is running
-Jun  4 19:09:03.510: INFO: Waiting for deployment "nginx-deployment" to complete
-Jun  4 19:09:03.522: INFO: Updating deployment "nginx-deployment" with a non-existent image
-Jun  4 19:09:03.539: INFO: Updating deployment nginx-deployment
-Jun  4 19:09:03.539: INFO: Waiting for observed generation 2
-Jun  4 19:09:05.590: INFO: Waiting for the first rollout's replicaset to have .status.availableReplicas = 8
-Jun  4 19:09:05.603: INFO: Waiting for the first rollout's replicaset to have .spec.replicas = 8
-Jun  4 19:09:05.666: INFO: Waiting for the first rollout's replicaset of deployment "nginx-deployment" to have desired number of replicas
-Jun  4 19:09:05.691: INFO: Verifying that the second rollout's replicaset has .status.availableReplicas = 0
-Jun  4 19:09:05.691: INFO: Waiting for the second rollout's replicaset to have .spec.replicas = 5
-Jun  4 19:09:05.702: INFO: Waiting for the second rollout's replicaset of deployment "nginx-deployment" to have desired number of replicas
-Jun  4 19:09:05.717: INFO: Verifying that deployment "nginx-deployment" has minimum required number of available replicas
-Jun  4 19:09:05.717: INFO: Scaling up the deployment "nginx-deployment" from 10 to 30
-Jun  4 19:09:05.736: INFO: Updating deployment nginx-deployment
-Jun  4 19:09:05.736: INFO: Waiting for the replicasets of deployment "nginx-deployment" to have desired number of replicas
-Jun  4 19:09:05.753: INFO: Verifying that first rollout's replicaset has .spec.replicas = 20
-Jun  4 19:09:05.760: INFO: Verifying that second rollout's replicaset has .spec.replicas = 13
-[AfterEach] [sig-apps] Deployment
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:59
-Jun  4 19:09:05.780: INFO: Deployment "nginx-deployment":
-&Deployment{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment,GenerateName:,Namespace:deployment-4196,SelfLink:/apis/apps/v1/namespaces/deployment-4196/deployments/nginx-deployment,UID:353a6583-86fc-11e9-8318-1e5386706511,ResourceVersion:17931,Generation:3,CreationTimestamp:2019-06-04 19:08:59 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,},Annotations:map[string]string{deployment.kubernetes.io/revision: 2,},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:DeploymentSpec{Replicas:*30,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: nginx,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[],Containers:[{nginx nginx:404 [] []  [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,},},Strategy:DeploymentStrategy{Type:RollingUpdate,RollingUpdate:&RollingUpdateDeployment{MaxUnavailable:2,MaxSurge:3,},},MinReadySeconds:0,RevisionHistoryLimit:*10,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:3,Replicas:13,UpdatedReplicas:5,AvailableReplicas:8,UnavailableReplicas:5,Conditions:[{Progressing True 2019-06-04 19:09:03 +0000 UTC 2019-06-04 19:08:59 +0000 UTC ReplicaSetUpdated ReplicaSet "nginx-deployment-5f9595f595" is progressing.} {Available False 2019-06-04 19:09:05 +0000 UTC 2019-06-04 19:09:05 +0000 UTC MinimumReplicasUnavailable Deployment does not have minimum availability.}],ReadyReplicas:8,CollisionCount:nil,},}
-
-Jun  4 19:09:05.788: INFO: New ReplicaSet "nginx-deployment-5f9595f595" of Deployment "nginx-deployment":
-&ReplicaSet{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-5f9595f595,GenerateName:,Namespace:deployment-4196,SelfLink:/apis/apps/v1/namespaces/deployment-4196/replicasets/nginx-deployment-5f9595f595,UID:37abadc3-86fc-11e9-9509-923735b172b1,ResourceVersion:17927,Generation:3,CreationTimestamp:2019-06-04 19:09:03 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 5f9595f595,},Annotations:map[string]string{deployment.kubernetes.io/desired-replicas: 30,deployment.kubernetes.io/max-replicas: 33,deployment.kubernetes.io/revision: 2,},OwnerReferences:[{apps/v1 Deployment nginx-deployment 353a6583-86fc-11e9-8318-1e5386706511 0xc0029787b7 0xc0029787b8}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:ReplicaSetSpec{Replicas:*13,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: nginx,pod-template-hash: 5f9595f595,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 5f9595f595,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[],Containers:[{nginx nginx:404 [] []  [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,},},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:5,FullyLabeledReplicas:5,ObservedGeneration:2,ReadyReplicas:0,AvailableReplicas:0,Conditions:[],},}
-Jun  4 19:09:05.788: INFO: All old ReplicaSets of Deployment "nginx-deployment":
-Jun  4 19:09:05.788: INFO: &ReplicaSet{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-6f478d8d8,GenerateName:,Namespace:deployment-4196,SelfLink:/apis/apps/v1/namespaces/deployment-4196/replicasets/nginx-deployment-6f478d8d8,UID:353b97c5-86fc-11e9-9509-923735b172b1,ResourceVersion:17924,Generation:3,CreationTimestamp:2019-06-04 19:08:59 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 6f478d8d8,},Annotations:map[string]string{deployment.kubernetes.io/desired-replicas: 30,deployment.kubernetes.io/max-replicas: 33,deployment.kubernetes.io/revision: 1,},OwnerReferences:[{apps/v1 Deployment nginx-deployment 353a6583-86fc-11e9-8318-1e5386706511 0xc0029788a7 0xc0029788a8}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:ReplicaSetSpec{Replicas:*20,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: nginx,pod-template-hash: 6f478d8d8,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 6f478d8d8,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] []  [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,},},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:8,FullyLabeledReplicas:8,ObservedGeneration:2,ReadyReplicas:8,AvailableReplicas:8,Conditions:[],},}
-Jun  4 19:09:05.803: INFO: Pod "nginx-deployment-5f9595f595-4lzr2" is not available:
-&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-5f9595f595-4lzr2,GenerateName:nginx-deployment-5f9595f595-,Namespace:deployment-4196,SelfLink:/api/v1/namespaces/deployment-4196/pods/nginx-deployment-5f9595f595-4lzr2,UID:38fe8192-86fc-11e9-9509-923735b172b1,ResourceVersion:17935,Generation:0,CreationTimestamp:2019-06-04 19:09:05 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 5f9595f595,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-5f9595f595 37abadc3-86fc-11e9-9509-923735b172b1 0xc0029791d7 0xc0029791d8}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-tbnzl {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-tbnzl,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] []  [] [] [] {map[] map[]} [{default-token-tbnzl true /var/run/secrets/kubernetes.io/serviceaccount   }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.212.23.161,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists  NoExecute 0xc002979250} {node.kubernetes.io/unreachable Exists  NoExecute 0xc002979280}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 19:09:05 +0000 UTC  }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},}
-Jun  4 19:09:05.804: INFO: Pod "nginx-deployment-5f9595f595-7f7xv" is not available:
-&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-5f9595f595-7f7xv,GenerateName:nginx-deployment-5f9595f595-,Namespace:deployment-4196,SelfLink:/api/v1/namespaces/deployment-4196/pods/nginx-deployment-5f9595f595-7f7xv,UID:37b5e305-86fc-11e9-9509-923735b172b1,ResourceVersion:17889,Generation:0,CreationTimestamp:2019-06-04 19:09:03 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 5f9595f595,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-5f9595f595 37abadc3-86fc-11e9-9509-923735b172b1 0xc002979300 0xc002979301}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-tbnzl {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-tbnzl,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] []  [] [] [] {map[] map[]} [{default-token-tbnzl true /var/run/secrets/kubernetes.io/serviceaccount   }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.212.23.164,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists  NoExecute 0xc002979380} {node.kubernetes.io/unreachable Exists  NoExecute 0xc0029793a0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 19:09:03 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-04 19:09:03 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-04 19:09:03 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 19:09:03 +0000 UTC  }],Message:,Reason:,HostIP:10.212.23.164,PodIP:,StartTime:2019-06-04 19:09:03 +0000 UTC,ContainerStatuses:[{nginx {ContainerStateWaiting{Reason:ContainerCreating,Message:,} nil nil} {nil nil nil} false 0 nginx:404  }],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},}
-Jun  4 19:09:05.804: INFO: Pod "nginx-deployment-5f9595f595-8lw2q" is not available:
-&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-5f9595f595-8lw2q,GenerateName:nginx-deployment-5f9595f595-,Namespace:deployment-4196,SelfLink:/api/v1/namespaces/deployment-4196/pods/nginx-deployment-5f9595f595-8lw2q,UID:39000e39-86fc-11e9-9509-923735b172b1,ResourceVersion:17946,Generation:0,CreationTimestamp:2019-06-04 19:09:05 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 5f9595f595,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-5f9595f595 37abadc3-86fc-11e9-9509-923735b172b1 0xc002979470 0xc002979471}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-tbnzl {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-tbnzl,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] []  [] [] [] {map[] map[]} [{default-token-tbnzl true /var/run/secrets/kubernetes.io/serviceaccount   }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.212.23.164,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists  NoExecute 0xc0029794f0} {node.kubernetes.io/unreachable Exists  NoExecute 0xc002979510}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 19:09:05 +0000 UTC  }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},}
-Jun  4 19:09:05.804: INFO: Pod "nginx-deployment-5f9595f595-9gfb5" is not available:
-&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-5f9595f595-9gfb5,GenerateName:nginx-deployment-5f9595f595-,Namespace:deployment-4196,SelfLink:/api/v1/namespaces/deployment-4196/pods/nginx-deployment-5f9595f595-9gfb5,UID:390248ad-86fc-11e9-9509-923735b172b1,ResourceVersion:17957,Generation:0,CreationTimestamp:2019-06-04 19:09:05 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 5f9595f595,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-5f9595f595 37abadc3-86fc-11e9-9509-923735b172b1 0xc002979590 0xc002979591}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-tbnzl {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-tbnzl,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] []  [] [] [] {map[] map[]} [{default-token-tbnzl true /var/run/secrets/kubernetes.io/serviceaccount   }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists  NoExecute 0xc002979600} {node.kubernetes.io/unreachable Exists  NoExecute 0xc002979620}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},}
-Jun  4 19:09:05.804: INFO: Pod "nginx-deployment-5f9595f595-crbj5" is not available:
-&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-5f9595f595-crbj5,GenerateName:nginx-deployment-5f9595f595-,Namespace:deployment-4196,SelfLink:/api/v1/namespaces/deployment-4196/pods/nginx-deployment-5f9595f595-crbj5,UID:390242a0-86fc-11e9-9509-923735b172b1,ResourceVersion:17954,Generation:0,CreationTimestamp:2019-06-04 19:09:05 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 5f9595f595,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-5f9595f595 37abadc3-86fc-11e9-9509-923735b172b1 0xc002979687 0xc002979688}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-tbnzl {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-tbnzl,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] []  [] [] [] {map[] map[]} [{default-token-tbnzl true /var/run/secrets/kubernetes.io/serviceaccount   }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists  NoExecute 0xc0029796f0} {node.kubernetes.io/unreachable Exists  NoExecute 0xc002979710}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},}
-Jun  4 19:09:05.804: INFO: Pod "nginx-deployment-5f9595f595-gbjg5" is not available:
-&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-5f9595f595-gbjg5,GenerateName:nginx-deployment-5f9595f595-,Namespace:deployment-4196,SelfLink:/api/v1/namespaces/deployment-4196/pods/nginx-deployment-5f9595f595-gbjg5,UID:3900176f-86fc-11e9-9509-923735b172b1,ResourceVersion:17948,Generation:0,CreationTimestamp:2019-06-04 19:09:05 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 5f9595f595,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-5f9595f595 37abadc3-86fc-11e9-9509-923735b172b1 0xc002979777 0xc002979778}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-tbnzl {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-tbnzl,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] []  [] [] [] {map[] map[]} [{default-token-tbnzl true /var/run/secrets/kubernetes.io/serviceaccount   }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.212.23.189,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists  NoExecute 0xc0029797f0} {node.kubernetes.io/unreachable Exists  NoExecute 0xc002979810}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 19:09:05 +0000 UTC  }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},}
-Jun  4 19:09:05.804: INFO: Pod "nginx-deployment-5f9595f595-mq59d" is not available:
-&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-5f9595f595-mq59d,GenerateName:nginx-deployment-5f9595f595-,Namespace:deployment-4196,SelfLink:/api/v1/namespaces/deployment-4196/pods/nginx-deployment-5f9595f595-mq59d,UID:37ae79f5-86fc-11e9-9509-923735b172b1,ResourceVersion:17863,Generation:0,CreationTimestamp:2019-06-04 19:09:03 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 5f9595f595,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-5f9595f595 37abadc3-86fc-11e9-9509-923735b172b1 0xc002979890 0xc002979891}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-tbnzl {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-tbnzl,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] []  [] [] [] {map[] map[]} [{default-token-tbnzl true /var/run/secrets/kubernetes.io/serviceaccount   }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.212.23.161,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists  NoExecute 0xc002979910} {node.kubernetes.io/unreachable Exists  NoExecute 0xc002979930}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 19:09:03 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-04 19:09:03 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-04 19:09:03 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 19:09:03 +0000 UTC  }],Message:,Reason:,HostIP:10.212.23.161,PodIP:,StartTime:2019-06-04 19:09:03 +0000 UTC,ContainerStatuses:[{nginx {ContainerStateWaiting{Reason:ContainerCreating,Message:,} nil nil} {nil nil nil} false 0 nginx:404  }],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},}
-Jun  4 19:09:05.805: INFO: Pod "nginx-deployment-5f9595f595-nmd26" is not available:
-&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-5f9595f595-nmd26,GenerateName:nginx-deployment-5f9595f595-,Namespace:deployment-4196,SelfLink:/api/v1/namespaces/deployment-4196/pods/nginx-deployment-5f9595f595-nmd26,UID:37ae7586-86fc-11e9-9509-923735b172b1,ResourceVersion:17865,Generation:0,CreationTimestamp:2019-06-04 19:09:03 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 5f9595f595,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-5f9595f595 37abadc3-86fc-11e9-9509-923735b172b1 0xc002979a00 0xc002979a01}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-tbnzl {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-tbnzl,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] []  [] [] [] {map[] map[]} [{default-token-tbnzl true /var/run/secrets/kubernetes.io/serviceaccount   }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.212.23.189,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists  NoExecute 0xc002979a80} {node.kubernetes.io/unreachable Exists  NoExecute 0xc002979aa0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 19:09:03 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-04 19:09:03 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-04 19:09:03 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 19:09:03 +0000 UTC  }],Message:,Reason:,HostIP:10.212.23.189,PodIP:,StartTime:2019-06-04 19:09:03 +0000 UTC,ContainerStatuses:[{nginx {ContainerStateWaiting{Reason:ContainerCreating,Message:,} nil nil} {nil nil nil} false 0 nginx:404  }],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},}
-Jun  4 19:09:05.805: INFO: Pod "nginx-deployment-5f9595f595-q7mw9" is not available:
-&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-5f9595f595-q7mw9,GenerateName:nginx-deployment-5f9595f595-,Namespace:deployment-4196,SelfLink:/api/v1/namespaces/deployment-4196/pods/nginx-deployment-5f9595f595-q7mw9,UID:37ad0e86-86fc-11e9-9509-923735b172b1,ResourceVersion:17869,Generation:0,CreationTimestamp:2019-06-04 19:09:03 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 5f9595f595,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-5f9595f595 37abadc3-86fc-11e9-9509-923735b172b1 0xc002979b70 0xc002979b71}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-tbnzl {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-tbnzl,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] []  [] [] [] {map[] map[]} [{default-token-tbnzl true /var/run/secrets/kubernetes.io/serviceaccount   }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.212.23.164,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists  NoExecute 0xc002979bf0} {node.kubernetes.io/unreachable Exists  NoExecute 0xc002979c10}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 19:09:03 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-04 19:09:03 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-04 19:09:03 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 19:09:03 +0000 UTC  }],Message:,Reason:,HostIP:10.212.23.164,PodIP:,StartTime:2019-06-04 19:09:03 +0000 UTC,ContainerStatuses:[{nginx {ContainerStateWaiting{Reason:ContainerCreating,Message:,} nil nil} {nil nil nil} false 0 nginx:404  }],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},}
-Jun  4 19:09:05.805: INFO: Pod "nginx-deployment-5f9595f595-tdp9l" is not available:
-&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-5f9595f595-tdp9l,GenerateName:nginx-deployment-5f9595f595-,Namespace:deployment-4196,SelfLink:/api/v1/namespaces/deployment-4196/pods/nginx-deployment-5f9595f595-tdp9l,UID:3902404b-86fc-11e9-9509-923735b172b1,ResourceVersion:17955,Generation:0,CreationTimestamp:2019-06-04 19:09:05 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 5f9595f595,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-5f9595f595 37abadc3-86fc-11e9-9509-923735b172b1 0xc002979ce0 0xc002979ce1}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-tbnzl {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-tbnzl,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] []  [] [] [] {map[] map[]} [{default-token-tbnzl true /var/run/secrets/kubernetes.io/serviceaccount   }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists  NoExecute 0xc002979d50} {node.kubernetes.io/unreachable Exists  NoExecute 0xc002979d70}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},}
-Jun  4 19:09:05.805: INFO: Pod "nginx-deployment-5f9595f595-tmb6v" is not available:
-&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-5f9595f595-tmb6v,GenerateName:nginx-deployment-5f9595f595-,Namespace:deployment-4196,SelfLink:/api/v1/namespaces/deployment-4196/pods/nginx-deployment-5f9595f595-tmb6v,UID:37b77069-86fc-11e9-9509-923735b172b1,ResourceVersion:17888,Generation:0,CreationTimestamp:2019-06-04 19:09:03 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 5f9595f595,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-5f9595f595 37abadc3-86fc-11e9-9509-923735b172b1 0xc002979dd7 0xc002979dd8}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-tbnzl {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-tbnzl,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] []  [] [] [] {map[] map[]} [{default-token-tbnzl true /var/run/secrets/kubernetes.io/serviceaccount   }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.212.23.189,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists  NoExecute 0xc002979e50} {node.kubernetes.io/unreachable Exists  NoExecute 0xc002979e70}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 19:09:03 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-04 19:09:03 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-04 19:09:03 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 19:09:03 +0000 UTC  }],Message:,Reason:,HostIP:10.212.23.189,PodIP:,StartTime:2019-06-04 19:09:03 +0000 UTC,ContainerStatuses:[{nginx {ContainerStateWaiting{Reason:ContainerCreating,Message:,} nil nil} {nil nil nil} false 0 nginx:404  }],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},}
-Jun  4 19:09:05.805: INFO: Pod "nginx-deployment-5f9595f595-zqnxq" is not available:
-&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-5f9595f595-zqnxq,GenerateName:nginx-deployment-5f9595f595-,Namespace:deployment-4196,SelfLink:/api/v1/namespaces/deployment-4196/pods/nginx-deployment-5f9595f595-zqnxq,UID:39023231-86fc-11e9-9509-923735b172b1,ResourceVersion:17950,Generation:0,CreationTimestamp:2019-06-04 19:09:05 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 5f9595f595,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-5f9595f595 37abadc3-86fc-11e9-9509-923735b172b1 0xc002979f40 0xc002979f41}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-tbnzl {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-tbnzl,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] []  [] [] [] {map[] map[]} [{default-token-tbnzl true /var/run/secrets/kubernetes.io/serviceaccount   }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists  NoExecute 0xc002979fb0} {node.kubernetes.io/unreachable Exists  NoExecute 0xc002979fd0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},}
-Jun  4 19:09:05.805: INFO: Pod "nginx-deployment-6f478d8d8-4pcrp" is not available:
-&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-6f478d8d8-4pcrp,GenerateName:nginx-deployment-6f478d8d8-,Namespace:deployment-4196,SelfLink:/api/v1/namespaces/deployment-4196/pods/nginx-deployment-6f478d8d8-4pcrp,UID:39023f24-86fc-11e9-9509-923735b172b1,ResourceVersion:17952,Generation:0,CreationTimestamp:2019-06-04 19:09:05 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 6f478d8d8,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-6f478d8d8 353b97c5-86fc-11e9-9509-923735b172b1 0xc0026680a7 0xc0026680a8}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-tbnzl {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-tbnzl,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] []  [] [] [] {map[] map[]} [{default-token-tbnzl true /var/run/secrets/kubernetes.io/serviceaccount   }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists  NoExecute 0xc0026681b0} {node.kubernetes.io/unreachable Exists  NoExecute 0xc0026681d0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},}
-Jun  4 19:09:05.805: INFO: Pod "nginx-deployment-6f478d8d8-5sxlt" is available:
-&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-6f478d8d8-5sxlt,GenerateName:nginx-deployment-6f478d8d8-,Namespace:deployment-4196,SelfLink:/api/v1/namespaces/deployment-4196/pods/nginx-deployment-6f478d8d8-5sxlt,UID:353facff-86fc-11e9-9509-923735b172b1,ResourceVersion:17793,Generation:0,CreationTimestamp:2019-06-04 19:08:59 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 6f478d8d8,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-6f478d8d8 353b97c5-86fc-11e9-9509-923735b172b1 0xc002668237 0xc002668238}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-tbnzl {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-tbnzl,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] []  [] [] [] {map[] map[]} [{default-token-tbnzl true /var/run/secrets/kubernetes.io/serviceaccount   }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.212.23.164,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists  NoExecute 0xc0026682b0} {node.kubernetes.io/unreachable Exists  NoExecute 0xc0026682d0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Running,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 19:08:59 +0000 UTC  } {Ready True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 19:09:00 +0000 UTC  } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 19:09:00 +0000 UTC  } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 19:08:59 +0000 UTC  }],Message:,Reason:,HostIP:10.212.23.164,PodIP:172.30.254.188,StartTime:2019-06-04 19:08:59 +0000 UTC,ContainerStatuses:[{nginx {nil ContainerStateRunning{StartedAt:2019-06-04 19:09:00 +0000 UTC,} nil} {nil nil nil} true 0 docker.io/library/nginx:1.14-alpine docker.io/library/nginx@sha256:485b610fefec7ff6c463ced9623314a04ed67e3945b9c08d7e53a47f6d108dc7 containerd://7a6c4d7785e36f570f01963f57fadc431314fc4d96031401da88afe293253f2f}],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},}
-Jun  4 19:09:05.806: INFO: Pod "nginx-deployment-6f478d8d8-6b54v" is not available:
-&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-6f478d8d8-6b54v,GenerateName:nginx-deployment-6f478d8d8-,Namespace:deployment-4196,SelfLink:/api/v1/namespaces/deployment-4196/pods/nginx-deployment-6f478d8d8-6b54v,UID:38fe2e7d-86fc-11e9-9509-923735b172b1,ResourceVersion:17934,Generation:0,CreationTimestamp:2019-06-04 19:09:05 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 6f478d8d8,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-6f478d8d8 353b97c5-86fc-11e9-9509-923735b172b1 0xc002668557 0xc002668558}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-tbnzl {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-tbnzl,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] []  [] [] [] {map[] map[]} [{default-token-tbnzl true /var/run/secrets/kubernetes.io/serviceaccount   }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.212.23.161,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists  NoExecute 0xc0026686a0} {node.kubernetes.io/unreachable Exists  NoExecute 0xc002668720}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 19:09:05 +0000 UTC  }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},}
-Jun  4 19:09:05.806: INFO: Pod "nginx-deployment-6f478d8d8-fkb4t" is available:
-&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-6f478d8d8-fkb4t,GenerateName:nginx-deployment-6f478d8d8-,Namespace:deployment-4196,SelfLink:/api/v1/namespaces/deployment-4196/pods/nginx-deployment-6f478d8d8-fkb4t,UID:3542f73c-86fc-11e9-9509-923735b172b1,ResourceVersion:17818,Generation:0,CreationTimestamp:2019-06-04 19:08:59 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 6f478d8d8,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-6f478d8d8 353b97c5-86fc-11e9-9509-923735b172b1 0xc0026688f0 0xc0026688f1}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-tbnzl {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-tbnzl,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] []  [] [] [] {map[] map[]} [{default-token-tbnzl true /var/run/secrets/kubernetes.io/serviceaccount   }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.212.23.189,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists  NoExecute 0xc002668a40} {node.kubernetes.io/unreachable Exists  NoExecute 0xc002668ab0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Running,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 19:08:59 +0000 UTC  } {Ready True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 19:09:01 +0000 UTC  } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 19:09:01 +0000 UTC  } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 19:08:59 +0000 UTC  }],Message:,Reason:,HostIP:10.212.23.189,PodIP:172.30.193.57,StartTime:2019-06-04 19:08:59 +0000 UTC,ContainerStatuses:[{nginx {nil ContainerStateRunning{StartedAt:2019-06-04 19:09:00 +0000 UTC,} nil} {nil nil nil} true 0 docker.io/library/nginx:1.14-alpine docker.io/library/nginx@sha256:485b610fefec7ff6c463ced9623314a04ed67e3945b9c08d7e53a47f6d108dc7 containerd://c6b9cfb18fd507c9a67f70c768850842b567f1809074e670a7ef85c4232af986}],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},}
-Jun  4 19:09:05.806: INFO: Pod "nginx-deployment-6f478d8d8-fmpm4" is not available:
-&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-6f478d8d8-fmpm4,GenerateName:nginx-deployment-6f478d8d8-,Namespace:deployment-4196,SelfLink:/api/v1/namespaces/deployment-4196/pods/nginx-deployment-6f478d8d8-fmpm4,UID:38fca38e-86fc-11e9-9509-923735b172b1,ResourceVersion:17962,Generation:0,CreationTimestamp:2019-06-04 19:09:05 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 6f478d8d8,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-6f478d8d8 353b97c5-86fc-11e9-9509-923735b172b1 0xc002668c77 0xc002668c78}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-tbnzl {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-tbnzl,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] []  [] [] [] {map[] map[]} [{default-token-tbnzl true /var/run/secrets/kubernetes.io/serviceaccount   }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.212.23.161,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists  NoExecute 0xc002668e00} {node.kubernetes.io/unreachable Exists  NoExecute 0xc002668e70}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 19:09:05 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-04 19:09:05 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-04 19:09:05 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 19:09:05 +0000 UTC  }],Message:,Reason:,HostIP:10.212.23.161,PodIP:,StartTime:2019-06-04 19:09:05 +0000 UTC,ContainerStatuses:[{nginx {ContainerStateWaiting{Reason:ContainerCreating,Message:,} nil nil} {nil nil nil} false 0 docker.io/library/nginx:1.14-alpine  }],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},}
-Jun  4 19:09:05.806: INFO: Pod "nginx-deployment-6f478d8d8-g2m5f" is available:
-&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-6f478d8d8-g2m5f,GenerateName:nginx-deployment-6f478d8d8-,Namespace:deployment-4196,SelfLink:/api/v1/namespaces/deployment-4196/pods/nginx-deployment-6f478d8d8-g2m5f,UID:35415ddd-86fc-11e9-9509-923735b172b1,ResourceVersion:17830,Generation:0,CreationTimestamp:2019-06-04 19:08:59 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 6f478d8d8,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-6f478d8d8 353b97c5-86fc-11e9-9509-923735b172b1 0xc002668ff7 0xc002668ff8}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-tbnzl {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-tbnzl,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] []  [] [] [] {map[] map[]} [{default-token-tbnzl true /var/run/secrets/kubernetes.io/serviceaccount   }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.212.23.164,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists  NoExecute 0xc002669190} {node.kubernetes.io/unreachable Exists  NoExecute 0xc002669290}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Running,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 19:08:59 +0000 UTC  } {Ready True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 19:09:01 +0000 UTC  } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 19:09:01 +0000 UTC  } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 19:08:59 +0000 UTC  }],Message:,Reason:,HostIP:10.212.23.164,PodIP:172.30.254.189,StartTime:2019-06-04 19:08:59 +0000 UTC,ContainerStatuses:[{nginx {nil ContainerStateRunning{StartedAt:2019-06-04 19:09:00 +0000 UTC,} nil} {nil nil nil} true 0 docker.io/library/nginx:1.14-alpine docker.io/library/nginx@sha256:485b610fefec7ff6c463ced9623314a04ed67e3945b9c08d7e53a47f6d108dc7 containerd://baa79b21813e4b8f510ce92039dbdef05eb694b1bceb7b97f261053cdb5aeba2}],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},}
-Jun  4 19:09:05.806: INFO: Pod "nginx-deployment-6f478d8d8-gcbcw" is not available:
-&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-6f478d8d8-gcbcw,GenerateName:nginx-deployment-6f478d8d8-,Namespace:deployment-4196,SelfLink:/api/v1/namespaces/deployment-4196/pods/nginx-deployment-6f478d8d8-gcbcw,UID:39001df2-86fc-11e9-9509-923735b172b1,ResourceVersion:17961,Generation:0,CreationTimestamp:2019-06-04 19:09:05 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 6f478d8d8,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-6f478d8d8 353b97c5-86fc-11e9-9509-923735b172b1 0xc002669367 0xc002669368}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-tbnzl {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-tbnzl,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] []  [] [] [] {map[] map[]} [{default-token-tbnzl true /var/run/secrets/kubernetes.io/serviceaccount   }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.212.23.164,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists  NoExecute 0xc0026693e0} {node.kubernetes.io/unreachable Exists  NoExecute 0xc0026694d0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 19:09:05 +0000 UTC  }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},}
-Jun  4 19:09:05.806: INFO: Pod "nginx-deployment-6f478d8d8-grqwj" is not available:
-&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-6f478d8d8-grqwj,GenerateName:nginx-deployment-6f478d8d8-,Namespace:deployment-4196,SelfLink:/api/v1/namespaces/deployment-4196/pods/nginx-deployment-6f478d8d8-grqwj,UID:390244b2-86fc-11e9-9509-923735b172b1,ResourceVersion:17956,Generation:0,CreationTimestamp:2019-06-04 19:09:05 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 6f478d8d8,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-6f478d8d8 353b97c5-86fc-11e9-9509-923735b172b1 0xc002669650 0xc002669651}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-tbnzl {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-tbnzl,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] []  [] [] [] {map[] map[]} [{default-token-tbnzl true /var/run/secrets/kubernetes.io/serviceaccount   }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists  NoExecute 0xc002669890} {node.kubernetes.io/unreachable Exists  NoExecute 0xc0026699d0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},}
-Jun  4 19:09:05.807: INFO: Pod "nginx-deployment-6f478d8d8-jm2h6" is available:
-&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-6f478d8d8-jm2h6,GenerateName:nginx-deployment-6f478d8d8-,Namespace:deployment-4196,SelfLink:/api/v1/namespaces/deployment-4196/pods/nginx-deployment-6f478d8d8-jm2h6,UID:3542ebc7-86fc-11e9-9509-923735b172b1,ResourceVersion:17821,Generation:0,CreationTimestamp:2019-06-04 19:08:59 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 6f478d8d8,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-6f478d8d8 353b97c5-86fc-11e9-9509-923735b172b1 0xc002669a37 0xc002669a38}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-tbnzl {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-tbnzl,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] []  [] [] [] {map[] map[]} [{default-token-tbnzl true /var/run/secrets/kubernetes.io/serviceaccount   }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.212.23.189,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists  NoExecute 0xc002669d60} {node.kubernetes.io/unreachable Exists  NoExecute 0xc002669dc0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Running,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 19:08:59 +0000 UTC  } {Ready True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 19:09:01 +0000 UTC  } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 19:09:01 +0000 UTC  } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 19:08:59 +0000 UTC  }],Message:,Reason:,HostIP:10.212.23.189,PodIP:172.30.193.58,StartTime:2019-06-04 19:08:59 +0000 UTC,ContainerStatuses:[{nginx {nil ContainerStateRunning{StartedAt:2019-06-04 19:09:01 +0000 UTC,} nil} {nil nil nil} true 0 docker.io/library/nginx:1.14-alpine docker.io/library/nginx@sha256:485b610fefec7ff6c463ced9623314a04ed67e3945b9c08d7e53a47f6d108dc7 containerd://954dd7ffab1716fe93de8be86187da195471f949e1f271e7e8e5fe4807f220af}],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},}
-Jun  4 19:09:05.807: INFO: Pod "nginx-deployment-6f478d8d8-jzgnf" is available:
-&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-6f478d8d8-jzgnf,GenerateName:nginx-deployment-6f478d8d8-,Namespace:deployment-4196,SelfLink:/api/v1/namespaces/deployment-4196/pods/nginx-deployment-6f478d8d8-jzgnf,UID:353f9759-86fc-11e9-9509-923735b172b1,ResourceVersion:17787,Generation:0,CreationTimestamp:2019-06-04 19:08:59 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 6f478d8d8,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-6f478d8d8 353b97c5-86fc-11e9-9509-923735b172b1 0xc000ca4167 0xc000ca4168}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-tbnzl {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-tbnzl,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] []  [] [] [] {map[] map[]} [{default-token-tbnzl true /var/run/secrets/kubernetes.io/serviceaccount   }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.212.23.161,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists  NoExecute 0xc000ca4440} {node.kubernetes.io/unreachable Exists  NoExecute 0xc000ca46d0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Running,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 19:08:59 +0000 UTC  } {Ready True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 19:09:00 +0000 UTC  } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 19:09:00 +0000 UTC  } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 19:08:59 +0000 UTC  }],Message:,Reason:,HostIP:10.212.23.161,PodIP:172.30.169.143,StartTime:2019-06-04 19:08:59 +0000 UTC,ContainerStatuses:[{nginx {nil ContainerStateRunning{StartedAt:2019-06-04 19:09:00 +0000 UTC,} nil} {nil nil nil} true 0 docker.io/library/nginx:1.14-alpine docker.io/library/nginx@sha256:485b610fefec7ff6c463ced9623314a04ed67e3945b9c08d7e53a47f6d108dc7 containerd://168c7b606353201ef9951c9ae40c5ae9cdad71d9a81e18e5fe1ddb87c9874e67}],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},}
-Jun  4 19:09:05.807: INFO: Pod "nginx-deployment-6f478d8d8-lvmql" is not available:
-&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-6f478d8d8-lvmql,GenerateName:nginx-deployment-6f478d8d8-,Namespace:deployment-4196,SelfLink:/api/v1/namespaces/deployment-4196/pods/nginx-deployment-6f478d8d8-lvmql,UID:38fe2ae3-86fc-11e9-9509-923735b172b1,ResourceVersion:17939,Generation:0,CreationTimestamp:2019-06-04 19:09:05 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 6f478d8d8,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-6f478d8d8 353b97c5-86fc-11e9-9509-923735b172b1 0xc000ca4c37 0xc000ca4c38}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-tbnzl {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-tbnzl,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] []  [] [] [] {map[] map[]} [{default-token-tbnzl true /var/run/secrets/kubernetes.io/serviceaccount   }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.212.23.189,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists  NoExecute 0xc000ca4e20} {node.kubernetes.io/unreachable Exists  NoExecute 0xc000ca4e40}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 19:09:05 +0000 UTC  }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},}
-Jun  4 19:09:05.807: INFO: Pod "nginx-deployment-6f478d8d8-m995k" is not available:
-&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-6f478d8d8-m995k,GenerateName:nginx-deployment-6f478d8d8-,Namespace:deployment-4196,SelfLink:/api/v1/namespaces/deployment-4196/pods/nginx-deployment-6f478d8d8-m995k,UID:390241f5-86fc-11e9-9509-923735b172b1,ResourceVersion:17953,Generation:0,CreationTimestamp:2019-06-04 19:09:05 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 6f478d8d8,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-6f478d8d8 353b97c5-86fc-11e9-9509-923735b172b1 0xc000ca5630 0xc000ca5631}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-tbnzl {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-tbnzl,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] []  [] [] [] {map[] map[]} [{default-token-tbnzl true /var/run/secrets/kubernetes.io/serviceaccount   }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists  NoExecute 0xc000ca5960} {node.kubernetes.io/unreachable Exists  NoExecute 0xc000ca5980}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},}
-Jun  4 19:09:05.807: INFO: Pod "nginx-deployment-6f478d8d8-qt7qk" is available:
-&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-6f478d8d8-qt7qk,GenerateName:nginx-deployment-6f478d8d8-,Namespace:deployment-4196,SelfLink:/api/v1/namespaces/deployment-4196/pods/nginx-deployment-6f478d8d8-qt7qk,UID:35416fd2-86fc-11e9-9509-923735b172b1,ResourceVersion:17833,Generation:0,CreationTimestamp:2019-06-04 19:08:59 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 6f478d8d8,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-6f478d8d8 353b97c5-86fc-11e9-9509-923735b172b1 0xc000ca5b07 0xc000ca5b08}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-tbnzl {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-tbnzl,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] []  [] [] [] {map[] map[]} [{default-token-tbnzl true /var/run/secrets/kubernetes.io/serviceaccount   }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.212.23.164,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists  NoExecute 0xc000ca5dc0} {node.kubernetes.io/unreachable Exists  NoExecute 0xc000ca5de0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Running,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 19:08:59 +0000 UTC  } {Ready True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 19:09:01 +0000 UTC  } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 19:09:01 +0000 UTC  } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 19:08:59 +0000 UTC  }],Message:,Reason:,HostIP:10.212.23.164,PodIP:172.30.254.190,StartTime:2019-06-04 19:08:59 +0000 UTC,ContainerStatuses:[{nginx {nil ContainerStateRunning{StartedAt:2019-06-04 19:09:00 +0000 UTC,} nil} {nil nil nil} true 0 docker.io/library/nginx:1.14-alpine docker.io/library/nginx@sha256:485b610fefec7ff6c463ced9623314a04ed67e3945b9c08d7e53a47f6d108dc7 containerd://7b714fec1a8e9d7983a08cac0b99c581e6b3d5bc3aa5c0176db64b2f017b027e}],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},}
-Jun  4 19:09:05.807: INFO: Pod "nginx-deployment-6f478d8d8-qz6dg" is not available:
-&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-6f478d8d8-qz6dg,GenerateName:nginx-deployment-6f478d8d8-,Namespace:deployment-4196,SelfLink:/api/v1/namespaces/deployment-4196/pods/nginx-deployment-6f478d8d8-qz6dg,UID:390065ff-86fc-11e9-9509-923735b172b1,ResourceVersion:17960,Generation:0,CreationTimestamp:2019-06-04 19:09:05 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 6f478d8d8,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-6f478d8d8 353b97c5-86fc-11e9-9509-923735b172b1 0xc000705347 0xc000705348}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-tbnzl {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-tbnzl,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] []  [] [] [] {map[] map[]} [{default-token-tbnzl true /var/run/secrets/kubernetes.io/serviceaccount   }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.212.23.189,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists  NoExecute 0xc0007056d0} {node.kubernetes.io/unreachable Exists  NoExecute 0xc0007056f0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 19:09:05 +0000 UTC  }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},}
-Jun  4 19:09:05.807: INFO: Pod "nginx-deployment-6f478d8d8-rmshb" is available:
-&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-6f478d8d8-rmshb,GenerateName:nginx-deployment-6f478d8d8-,Namespace:deployment-4196,SelfLink:/api/v1/namespaces/deployment-4196/pods/nginx-deployment-6f478d8d8-rmshb,UID:3542fb4b-86fc-11e9-9509-923735b172b1,ResourceVersion:17827,Generation:0,CreationTimestamp:2019-06-04 19:08:59 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 6f478d8d8,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-6f478d8d8 353b97c5-86fc-11e9-9509-923735b172b1 0xc000705da0 0xc000705da1}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-tbnzl {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-tbnzl,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] []  [] [] [] {map[] map[]} [{default-token-tbnzl true /var/run/secrets/kubernetes.io/serviceaccount   }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.212.23.164,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists  NoExecute 0xc000705fb0} {node.kubernetes.io/unreachable Exists  NoExecute 0xc001f94180}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Running,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 19:08:59 +0000 UTC  } {Ready True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 19:09:01 +0000 UTC  } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 19:09:01 +0000 UTC  } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 19:08:59 +0000 UTC  }],Message:,Reason:,HostIP:10.212.23.164,PodIP:172.30.254.191,StartTime:2019-06-04 19:08:59 +0000 UTC,ContainerStatuses:[{nginx {nil ContainerStateRunning{StartedAt:2019-06-04 19:09:01 +0000 UTC,} nil} {nil nil nil} true 0 docker.io/library/nginx:1.14-alpine docker.io/library/nginx@sha256:485b610fefec7ff6c463ced9623314a04ed67e3945b9c08d7e53a47f6d108dc7 containerd://f94afc1f774d5543c00724a0f53539ff52995bc40b0ae896af1bd7164bda5e4d}],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},}
-Jun  4 19:09:05.807: INFO: Pod "nginx-deployment-6f478d8d8-tknrc" is available:
-&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-6f478d8d8-tknrc,GenerateName:nginx-deployment-6f478d8d8-,Namespace:deployment-4196,SelfLink:/api/v1/namespaces/deployment-4196/pods/nginx-deployment-6f478d8d8-tknrc,UID:353e566d-86fc-11e9-9509-923735b172b1,ResourceVersion:17815,Generation:0,CreationTimestamp:2019-06-04 19:08:59 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 6f478d8d8,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-6f478d8d8 353b97c5-86fc-11e9-9509-923735b172b1 0xc001f945a7 0xc001f945a8}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-tbnzl {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-tbnzl,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] []  [] [] [] {map[] map[]} [{default-token-tbnzl true /var/run/secrets/kubernetes.io/serviceaccount   }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.212.23.189,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists  NoExecute 0xc001f94aa0} {node.kubernetes.io/unreachable Exists  NoExecute 0xc001f94b00}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Running,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 19:08:59 +0000 UTC  } {Ready True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 19:09:01 +0000 UTC  } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 19:09:01 +0000 UTC  } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 19:08:59 +0000 UTC  }],Message:,Reason:,HostIP:10.212.23.189,PodIP:172.30.193.55,StartTime:2019-06-04 19:08:59 +0000 UTC,ContainerStatuses:[{nginx {nil ContainerStateRunning{StartedAt:2019-06-04 19:09:00 +0000 UTC,} nil} {nil nil nil} true 0 docker.io/library/nginx:1.14-alpine docker.io/library/nginx@sha256:485b610fefec7ff6c463ced9623314a04ed67e3945b9c08d7e53a47f6d108dc7 containerd://13adb9683e88ac4113605c880223869b65eb0405d10dbffbb5fde3f30ce6b100}],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},}
-Jun  4 19:09:05.807: INFO: Pod "nginx-deployment-6f478d8d8-v9x77" is not available:
-&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-6f478d8d8-v9x77,GenerateName:nginx-deployment-6f478d8d8-,Namespace:deployment-4196,SelfLink:/api/v1/namespaces/deployment-4196/pods/nginx-deployment-6f478d8d8-v9x77,UID:390268db-86fc-11e9-9509-923735b172b1,ResourceVersion:17959,Generation:0,CreationTimestamp:2019-06-04 19:09:05 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 6f478d8d8,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-6f478d8d8 353b97c5-86fc-11e9-9509-923735b172b1 0xc001f95817 0xc001f95818}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-tbnzl {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-tbnzl,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] []  [] [] [] {map[] map[]} [{default-token-tbnzl true /var/run/secrets/kubernetes.io/serviceaccount   }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists  NoExecute 0xc001f95990} {node.kubernetes.io/unreachable Exists  NoExecute 0xc001f95b30}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},}
-Jun  4 19:09:05.808: INFO: Pod "nginx-deployment-6f478d8d8-w762t" is not available:
-&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-6f478d8d8-w762t,GenerateName:nginx-deployment-6f478d8d8-,Namespace:deployment-4196,SelfLink:/api/v1/namespaces/deployment-4196/pods/nginx-deployment-6f478d8d8-w762t,UID:390012fb-86fc-11e9-9509-923735b172b1,ResourceVersion:17947,Generation:0,CreationTimestamp:2019-06-04 19:09:05 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 6f478d8d8,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-6f478d8d8 353b97c5-86fc-11e9-9509-923735b172b1 0xc001f95cc7 0xc001f95cc8}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-tbnzl {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-tbnzl,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] []  [] [] [] {map[] map[]} [{default-token-tbnzl true /var/run/secrets/kubernetes.io/serviceaccount   }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.212.23.161,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists  NoExecute 0xc00237a0d0} {node.kubernetes.io/unreachable Exists  NoExecute 0xc00237a3c0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 19:09:05 +0000 UTC  }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},}
-Jun  4 19:09:05.808: INFO: Pod "nginx-deployment-6f478d8d8-zmmfl" is not available:
-&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-6f478d8d8-zmmfl,GenerateName:nginx-deployment-6f478d8d8-,Namespace:deployment-4196,SelfLink:/api/v1/namespaces/deployment-4196/pods/nginx-deployment-6f478d8d8-zmmfl,UID:39005a21-86fc-11e9-9509-923735b172b1,ResourceVersion:17958,Generation:0,CreationTimestamp:2019-06-04 19:09:05 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 6f478d8d8,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-6f478d8d8 353b97c5-86fc-11e9-9509-923735b172b1 0xc00237a4f0 0xc00237a4f1}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-tbnzl {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-tbnzl,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] []  [] [] [] {map[] map[]} [{default-token-tbnzl true /var/run/secrets/kubernetes.io/serviceaccount   }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.212.23.189,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists  NoExecute 0xc00237a5d0} {node.kubernetes.io/unreachable Exists  NoExecute 0xc00237a6f0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 19:09:05 +0000 UTC  }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},}
-Jun  4 19:09:05.808: INFO: Pod "nginx-deployment-6f478d8d8-zwvvt" is not available:
-&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-6f478d8d8-zwvvt,GenerateName:nginx-deployment-6f478d8d8-,Namespace:deployment-4196,SelfLink:/api/v1/namespaces/deployment-4196/pods/nginx-deployment-6f478d8d8-zwvvt,UID:39023e24-86fc-11e9-9509-923735b172b1,ResourceVersion:17951,Generation:0,CreationTimestamp:2019-06-04 19:09:05 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 6f478d8d8,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-6f478d8d8 353b97c5-86fc-11e9-9509-923735b172b1 0xc00237a850 0xc00237a851}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-tbnzl {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-tbnzl,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] []  [] [] [] {map[] map[]} [{default-token-tbnzl true /var/run/secrets/kubernetes.io/serviceaccount   }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists  NoExecute 0xc00237aa50} {node.kubernetes.io/unreachable Exists  NoExecute 0xc00237aaf0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},}
-[AfterEach] [sig-apps] Deployment
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 19:09:05.808: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "deployment-4196" for this suite.
-Jun  4 19:09:13.879: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 19:09:14.316: INFO: namespace deployment-4196 deletion completed in 8.495219787s
+[It] should serve a basic image on each replica with a public image  [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+Jun 18 12:14:27.665: INFO: Creating ReplicaSet my-hostname-basic-9e4710f8-91c2-11e9-a25d-8608290c688a
+Jun 18 12:14:27.700: INFO: Pod name my-hostname-basic-9e4710f8-91c2-11e9-a25d-8608290c688a: Found 0 pods out of 1
+Jun 18 12:14:32.714: INFO: Pod name my-hostname-basic-9e4710f8-91c2-11e9-a25d-8608290c688a: Found 1 pods out of 1
+Jun 18 12:14:32.714: INFO: Ensuring a pod for ReplicaSet "my-hostname-basic-9e4710f8-91c2-11e9-a25d-8608290c688a" is running
+Jun 18 12:14:32.727: INFO: Pod "my-hostname-basic-9e4710f8-91c2-11e9-a25d-8608290c688a-8zmks" is running (conditions: [{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2019-06-18 12:14:27 +0000 UTC Reason: Message:} {Type:Ready Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2019-06-18 12:14:32 +0000 UTC Reason: Message:} {Type:ContainersReady Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2019-06-18 12:14:32 +0000 UTC Reason: Message:} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2019-06-18 12:14:27 +0000 UTC Reason: Message:}])
+Jun 18 12:14:32.727: INFO: Trying to dial the pod
+Jun 18 12:14:37.785: INFO: Controller my-hostname-basic-9e4710f8-91c2-11e9-a25d-8608290c688a: Got expected result from replica 1 [my-hostname-basic-9e4710f8-91c2-11e9-a25d-8608290c688a-8zmks]: "my-hostname-basic-9e4710f8-91c2-11e9-a25d-8608290c688a-8zmks", 1 of 1 required successes so far
+[AfterEach] [sig-apps] ReplicaSet
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 12:14:37.785: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "replicaset-3213" for this suite.
+Jun 18 12:14:45.825: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 12:14:46.233: INFO: namespace replicaset-3213 deletion completed in 8.434694829s
 
-• [SLOW TEST:15.099 seconds]
-[sig-apps] Deployment
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22
-  deployment should support proportional scaling [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+• [SLOW TEST:18.768 seconds]
+[sig-apps] ReplicaSet
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22
+  should serve a basic image on each replica with a public image  [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-SSSSSS
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[k8s.io] Probing container 
-  with readiness probe that fails should never be ready and never restart [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [k8s.io] Probing container
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+[sig-storage] EmptyDir volumes 
+  should support (root,0666,default) [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-storage] EmptyDir volumes
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 19:09:14.316: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-STEP: Building a namespace api object, basename container-probe
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in container-probe-6675
+Jun 18 12:14:46.234: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename emptydir
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in emptydir-2366
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [k8s.io] Probing container
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/container_probe.go:51
-[It] with readiness probe that fails should never be ready and never restart [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[AfterEach] [k8s.io] Probing container
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 19:10:14.570: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "container-probe-6675" for this suite.
-Jun  4 19:10:38.637: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 19:10:38.989: INFO: namespace container-probe-6675 deletion completed in 24.398976288s
+[It] should support (root,0666,default) [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating a pod to test emptydir 0666 on node default medium
+Jun 18 12:14:46.459: INFO: Waiting up to 5m0s for pod "pod-a9767ecb-91c2-11e9-a25d-8608290c688a" in namespace "emptydir-2366" to be "success or failure"
+Jun 18 12:14:46.471: INFO: Pod "pod-a9767ecb-91c2-11e9-a25d-8608290c688a": Phase="Pending", Reason="", readiness=false. Elapsed: 12.015012ms
+Jun 18 12:14:48.487: INFO: Pod "pod-a9767ecb-91c2-11e9-a25d-8608290c688a": Phase="Pending", Reason="", readiness=false. Elapsed: 2.02779703s
+Jun 18 12:14:50.502: INFO: Pod "pod-a9767ecb-91c2-11e9-a25d-8608290c688a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.043029769s
+STEP: Saw pod success
+Jun 18 12:14:50.502: INFO: Pod "pod-a9767ecb-91c2-11e9-a25d-8608290c688a" satisfied condition "success or failure"
+Jun 18 12:14:50.515: INFO: Trying to get logs from node 10.72.74.138 pod pod-a9767ecb-91c2-11e9-a25d-8608290c688a container test-container: 
+STEP: delete the pod
+Jun 18 12:14:50.590: INFO: Waiting for pod pod-a9767ecb-91c2-11e9-a25d-8608290c688a to disappear
+Jun 18 12:14:50.604: INFO: Pod pod-a9767ecb-91c2-11e9-a25d-8608290c688a no longer exists
+[AfterEach] [sig-storage] EmptyDir volumes
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 12:14:50.604: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "emptydir-2366" for this suite.
+Jun 18 12:14:56.646: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 12:14:57.132: INFO: namespace emptydir-2366 deletion completed in 6.513982339s
 
-• [SLOW TEST:84.673 seconds]
-[k8s.io] Probing container
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
-  with readiness probe that fails should never be ready and never restart [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+• [SLOW TEST:10.898 seconds]
+[sig-storage] EmptyDir volumes
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:41
+  should support (root,0666,default) [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-SSSSSSSSSSSSSSSS
+SSSSSS
 ------------------------------
-[sig-api-machinery] Namespaces [Serial] 
-  should ensure that all services are removed when a namespace is deleted [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [sig-api-machinery] Namespaces [Serial]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+[sig-api-machinery] Aggregator 
+  Should be able to support the 1.10 Sample API Server using the current Aggregator [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-api-machinery] Aggregator
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 19:10:38.991: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-STEP: Building a namespace api object, basename namespaces
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in namespaces-5012
+Jun 18 12:14:57.133: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename aggregator
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in aggregator-7667
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should ensure that all services are removed when a namespace is deleted [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: Creating a test namespace
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in nsdeletetest-2886
+[BeforeEach] [sig-api-machinery] Aggregator
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/aggregator.go:69
+[It] Should be able to support the 1.10 Sample API Server using the current Aggregator [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Registering the sample API server.
+Jun 18 12:14:57.739: INFO: deployment "sample-apiserver-deployment" doesn't have the required revision set
+Jun 18 12:14:59.852: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63696456897, loc:(*time.Location)(0x8a1a0e0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63696456897, loc:(*time.Location)(0x8a1a0e0)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63696456897, loc:(*time.Location)(0x8a1a0e0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63696456897, loc:(*time.Location)(0x8a1a0e0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-65db6755fc\" is progressing."}}, CollisionCount:(*int32)(nil)}
+Jun 18 12:15:01.860: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63696456897, loc:(*time.Location)(0x8a1a0e0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63696456897, loc:(*time.Location)(0x8a1a0e0)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63696456897, loc:(*time.Location)(0x8a1a0e0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63696456897, loc:(*time.Location)(0x8a1a0e0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-65db6755fc\" is progressing."}}, CollisionCount:(*int32)(nil)}
+Jun 18 12:15:03.860: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63696456897, loc:(*time.Location)(0x8a1a0e0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63696456897, loc:(*time.Location)(0x8a1a0e0)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63696456897, loc:(*time.Location)(0x8a1a0e0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63696456897, loc:(*time.Location)(0x8a1a0e0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-65db6755fc\" is progressing."}}, CollisionCount:(*int32)(nil)}
+Jun 18 12:15:06.758: INFO: Waited 881.066401ms for the sample-apiserver to be ready to handle requests.
+[AfterEach] [sig-api-machinery] Aggregator
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/aggregator.go:60
+[AfterEach] [sig-api-machinery] Aggregator
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 12:15:07.278: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "aggregator-7667" for this suite.
+Jun 18 12:15:15.315: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 12:15:15.697: INFO: namespace aggregator-7667 deletion completed in 8.40660739s
+
+• [SLOW TEST:18.564 seconds]
+[sig-api-machinery] Aggregator
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:22
+  Should be able to support the 1.10 Sample API Server using the current Aggregator [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSSSSSS
+------------------------------
+[sig-apps] ReplicationController 
+  should release no longer matching pods [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-apps] ReplicationController
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Jun 18 12:15:15.699: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename replication-controller
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in replication-controller-2152
 STEP: Waiting for a default service account to be provisioned in namespace
-STEP: Creating a service in the namespace
-STEP: Deleting the namespace
-STEP: Waiting for the namespace to be removed.
-STEP: Recreating the namespace
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in nsdeletetest-9444
-STEP: Verifying there is no service in the namespace
-[AfterEach] [sig-api-machinery] Namespaces [Serial]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 19:10:45.780: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "namespaces-5012" for this suite.
-Jun  4 19:10:51.838: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 19:10:52.184: INFO: namespace namespaces-5012 deletion completed in 6.387904621s
-STEP: Destroying namespace "nsdeletetest-2886" for this suite.
-Jun  4 19:10:52.195: INFO: Namespace nsdeletetest-2886 was already deleted
-STEP: Destroying namespace "nsdeletetest-9444" for this suite.
-Jun  4 19:10:58.239: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 19:10:58.571: INFO: namespace nsdeletetest-9444 deletion completed in 6.376299427s
-
-• [SLOW TEST:19.581 seconds]
-[sig-api-machinery] Namespaces [Serial]
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:22
-  should ensure that all services are removed when a namespace is deleted [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[It] should release no longer matching pods [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Given a ReplicationController is created
+STEP: When the matched label of one of its pods change
+Jun 18 12:15:15.941: INFO: Pod name pod-release: Found 0 pods out of 1
+Jun 18 12:15:20.954: INFO: Pod name pod-release: Found 1 pods out of 1
+STEP: Then the pod is released
+[AfterEach] [sig-apps] ReplicationController
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 12:15:21.005: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "replication-controller-2152" for this suite.
+Jun 18 12:15:29.046: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 12:15:29.437: INFO: namespace replication-controller-2152 deletion completed in 8.417084804s
+
+• [SLOW TEST:13.738 seconds]
+[sig-apps] ReplicationController
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22
+  should release no longer matching pods [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-[sig-cli] Kubectl client [k8s.io] Kubectl run --rm job 
-  should create a job from an image, then delete the job  [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+[sig-api-machinery] Secrets 
+  should be consumable from pods in env vars [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-api-machinery] Secrets
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 19:10:58.572: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-STEP: Building a namespace api object, basename kubectl
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-7184
+Jun 18 12:15:29.438: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename secrets
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in secrets-7089
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:213
-[It] should create a job from an image, then delete the job  [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: executing a command with run --rm and attach with stdin
-Jun  4 19:10:58.792: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 --namespace=kubectl-7184 run e2e-test-rm-busybox-job --image=docker.io/library/busybox:1.29 --rm=true --generator=job/v1 --restart=OnFailure --attach=true --stdin -- sh -c cat && echo 'stdin closed''
-Jun  4 19:11:00.342: INFO: stderr: "kubectl run --generator=job/v1 is DEPRECATED and will be removed in a future version. Use kubectl run --generator=run-pod/v1 or kubectl create instead.\nIf you don't see a command prompt, try pressing enter.\n"
-Jun  4 19:11:00.342: INFO: stdout: "abcd1234stdin closed\njob.batch \"e2e-test-rm-busybox-job\" deleted\n"
-STEP: verifying the job e2e-test-rm-busybox-job was deleted
-[AfterEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 19:11:02.367: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "kubectl-7184" for this suite.
-Jun  4 19:11:14.508: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 19:11:14.876: INFO: namespace kubectl-7184 deletion completed in 12.420416483s
+[It] should be consumable from pods in env vars [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating secret with name secret-test-c3364feb-91c2-11e9-a25d-8608290c688a
+STEP: Creating a pod to test consume secrets
+Jun 18 12:15:29.668: INFO: Waiting up to 5m0s for pod "pod-secrets-c3381cfc-91c2-11e9-a25d-8608290c688a" in namespace "secrets-7089" to be "success or failure"
+Jun 18 12:15:29.688: INFO: Pod "pod-secrets-c3381cfc-91c2-11e9-a25d-8608290c688a": Phase="Pending", Reason="", readiness=false. Elapsed: 19.823944ms
+Jun 18 12:15:31.701: INFO: Pod "pod-secrets-c3381cfc-91c2-11e9-a25d-8608290c688a": Phase="Pending", Reason="", readiness=false. Elapsed: 2.033290177s
+Jun 18 12:15:33.715: INFO: Pod "pod-secrets-c3381cfc-91c2-11e9-a25d-8608290c688a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.046985401s
+STEP: Saw pod success
+Jun 18 12:15:33.715: INFO: Pod "pod-secrets-c3381cfc-91c2-11e9-a25d-8608290c688a" satisfied condition "success or failure"
+Jun 18 12:15:33.727: INFO: Trying to get logs from node 10.72.74.138 pod pod-secrets-c3381cfc-91c2-11e9-a25d-8608290c688a container secret-env-test: 
+STEP: delete the pod
+Jun 18 12:15:33.980: INFO: Waiting for pod pod-secrets-c3381cfc-91c2-11e9-a25d-8608290c688a to disappear
+Jun 18 12:15:33.994: INFO: Pod pod-secrets-c3381cfc-91c2-11e9-a25d-8608290c688a no longer exists
+[AfterEach] [sig-api-machinery] Secrets
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 12:15:33.994: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "secrets-7089" for this suite.
+Jun 18 12:15:40.041: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 12:15:40.439: INFO: namespace secrets-7089 deletion completed in 6.430496177s
 
-• [SLOW TEST:16.304 seconds]
-[sig-cli] Kubectl client
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23
-  [k8s.io] Kubectl run --rm job
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
-    should create a job from an image, then delete the job  [Conformance]
-    /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+• [SLOW TEST:11.001 seconds]
+[sig-api-machinery] Secrets
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/secrets.go:32
+  should be consumable from pods in env vars [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-SSSSSSSSSSSSSSSSSSS
+SSSSSSSSSSS
 ------------------------------
 [k8s.io] Pods 
-  should be submitted and removed [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+  should allow activeDeadlineSeconds to be updated [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 [BeforeEach] [k8s.io] Pods
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 19:11:14.876: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
+Jun 18 12:15:40.441: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
 STEP: Building a namespace api object, basename pods
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in pods-9871
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in pods-3579
 STEP: Waiting for a default service account to be provisioned in namespace
 [BeforeEach] [k8s.io] Pods
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:135
-[It] should be submitted and removed [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:135
+[It] should allow activeDeadlineSeconds to be updated [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 STEP: creating the pod
-STEP: setting up watch
 STEP: submitting the pod to kubernetes
-Jun  4 19:11:15.121: INFO: observed the pod list
 STEP: verifying the pod is in kubernetes
-STEP: verifying pod creation was observed
-STEP: deleting the pod gracefully
-STEP: verifying the kubelet observed the termination notice
-STEP: verifying pod deletion was observed
+STEP: updating the pod
+Jun 18 12:15:45.250: INFO: Successfully updated pod "pod-update-activedeadlineseconds-c9c5f8b4-91c2-11e9-a25d-8608290c688a"
+Jun 18 12:15:45.250: INFO: Waiting up to 5m0s for pod "pod-update-activedeadlineseconds-c9c5f8b4-91c2-11e9-a25d-8608290c688a" in namespace "pods-3579" to be "terminated due to deadline exceeded"
+Jun 18 12:15:45.262: INFO: Pod "pod-update-activedeadlineseconds-c9c5f8b4-91c2-11e9-a25d-8608290c688a": Phase="Running", Reason="", readiness=true. Elapsed: 12.151654ms
+Jun 18 12:15:47.276: INFO: Pod "pod-update-activedeadlineseconds-c9c5f8b4-91c2-11e9-a25d-8608290c688a": Phase="Failed", Reason="DeadlineExceeded", readiness=false. Elapsed: 2.025899381s
+Jun 18 12:15:47.276: INFO: Pod "pod-update-activedeadlineseconds-c9c5f8b4-91c2-11e9-a25d-8608290c688a" satisfied condition "terminated due to deadline exceeded"
 [AfterEach] [k8s.io] Pods
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 19:11:26.900: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "pods-9871" for this suite.
-Jun  4 19:11:33.001: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 19:11:33.333: INFO: namespace pods-9871 deletion completed in 6.377212289s
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 12:15:47.276: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "pods-3579" for this suite.
+Jun 18 12:15:53.338: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 12:15:53.735: INFO: namespace pods-3579 deletion completed in 6.443632238s
 
-• [SLOW TEST:18.457 seconds]
+• [SLOW TEST:13.294 seconds]
 [k8s.io] Pods
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
-  should be submitted and removed [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
+  should allow activeDeadlineSeconds to be updated [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-SS
+SSSSSSSS
 ------------------------------
-[sig-storage] ConfigMap 
-  should be consumable from pods in volume as non-root [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [sig-storage] ConfigMap
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+[sig-network] Proxy version v1 
+  should proxy logs on node using proxy subresource  [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] version v1
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 19:11:33.334: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-STEP: Building a namespace api object, basename configmap
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in configmap-2908
+Jun 18 12:15:53.735: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename proxy
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in proxy-1339
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should be consumable from pods in volume as non-root [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: Creating configMap with name configmap-test-volume-91158463-86fc-11e9-8d1b-467ee19922ac
-STEP: Creating a pod to test consume configMaps
-Jun  4 19:11:33.581: INFO: Waiting up to 5m0s for pod "pod-configmaps-9117a1e2-86fc-11e9-8d1b-467ee19922ac" in namespace "configmap-2908" to be "success or failure"
-Jun  4 19:11:33.592: INFO: Pod "pod-configmaps-9117a1e2-86fc-11e9-8d1b-467ee19922ac": Phase="Pending", Reason="", readiness=false. Elapsed: 10.994834ms
-Jun  4 19:11:35.605: INFO: Pod "pod-configmaps-9117a1e2-86fc-11e9-8d1b-467ee19922ac": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.023406711s
-STEP: Saw pod success
-Jun  4 19:11:35.605: INFO: Pod "pod-configmaps-9117a1e2-86fc-11e9-8d1b-467ee19922ac" satisfied condition "success or failure"
-Jun  4 19:11:35.619: INFO: Trying to get logs from node 10.212.23.164 pod pod-configmaps-9117a1e2-86fc-11e9-8d1b-467ee19922ac container configmap-volume-test: 
-STEP: delete the pod
-Jun  4 19:11:35.679: INFO: Waiting for pod pod-configmaps-9117a1e2-86fc-11e9-8d1b-467ee19922ac to disappear
-Jun  4 19:11:35.689: INFO: Pod pod-configmaps-9117a1e2-86fc-11e9-8d1b-467ee19922ac no longer exists
-[AfterEach] [sig-storage] ConfigMap
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 19:11:35.689: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "configmap-2908" for this suite.
-Jun  4 19:11:41.801: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 19:11:42.219: INFO: namespace configmap-2908 deletion completed in 6.51522762s
+[It] should proxy logs on node using proxy subresource  [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+Jun 18 12:15:53.989: INFO: (0) /api/v1/nodes/10.72.74.138/proxy/logs/: 
+alb/
+alternatives.log
+apt/... (200; 34.702289ms)
+Jun 18 12:15:54.010: INFO: (1) /api/v1/nodes/10.72.74.138/proxy/logs/: 
+alb/
+alternatives.log
+apt/... (200; 20.703466ms)
+Jun 18 12:15:54.028: INFO: (2) /api/v1/nodes/10.72.74.138/proxy/logs/: 
+alb/
+alternatives.log
+apt/... (200; 18.234063ms)
+Jun 18 12:15:54.046: INFO: (3) /api/v1/nodes/10.72.74.138/proxy/logs/: 
+alb/
+alternatives.log
+apt/... (200; 18.37955ms)
+Jun 18 12:15:54.064: INFO: (4) /api/v1/nodes/10.72.74.138/proxy/logs/: 
+alb/
+alternatives.log
+apt/... (200; 17.578169ms)
+Jun 18 12:15:54.084: INFO: (5) /api/v1/nodes/10.72.74.138/proxy/logs/: 
+alb/
+alternatives.log
+apt/... (200; 19.848338ms)
+Jun 18 12:15:54.111: INFO: (6) /api/v1/nodes/10.72.74.138/proxy/logs/: 
+alb/
+alternatives.log
+apt/... (200; 27.246391ms)
+Jun 18 12:15:54.131: INFO: (7) /api/v1/nodes/10.72.74.138/proxy/logs/: 
+alb/
+alternatives.log
+apt/... (200; 19.158284ms)
+Jun 18 12:15:54.148: INFO: (8) /api/v1/nodes/10.72.74.138/proxy/logs/: 
+alb/
+alternatives.log
+apt/... (200; 17.680198ms)
+Jun 18 12:15:54.170: INFO: (9) /api/v1/nodes/10.72.74.138/proxy/logs/: 
+alb/
+alternatives.log
+apt/... (200; 21.001172ms)
+Jun 18 12:15:54.188: INFO: (10) /api/v1/nodes/10.72.74.138/proxy/logs/: 
+alb/
+alternatives.log
+apt/... (200; 18.682978ms)
+Jun 18 12:15:54.206: INFO: (11) /api/v1/nodes/10.72.74.138/proxy/logs/: 
+alb/
+alternatives.log
+apt/... (200; 17.422017ms)
+Jun 18 12:15:54.223: INFO: (12) /api/v1/nodes/10.72.74.138/proxy/logs/: 
+alb/
+alternatives.log
+apt/... (200; 16.713498ms)
+Jun 18 12:15:54.239: INFO: (13) /api/v1/nodes/10.72.74.138/proxy/logs/: 
+alb/
+alternatives.log
+apt/... (200; 16.327536ms)
+Jun 18 12:15:54.259: INFO: (14) /api/v1/nodes/10.72.74.138/proxy/logs/: 
+alb/
+alternatives.log
+apt/... (200; 19.743563ms)
+Jun 18 12:15:54.277: INFO: (15) /api/v1/nodes/10.72.74.138/proxy/logs/: 
+alb/
+alternatives.log
+apt/... (200; 17.933985ms)
+Jun 18 12:15:54.296: INFO: (16) /api/v1/nodes/10.72.74.138/proxy/logs/: 
+alb/
+alternatives.log
+apt/... (200; 19.049843ms)
+Jun 18 12:15:54.316: INFO: (17) /api/v1/nodes/10.72.74.138/proxy/logs/: 
+alb/
+alternatives.log
+apt/... (200; 19.989215ms)
+Jun 18 12:15:54.334: INFO: (18) /api/v1/nodes/10.72.74.138/proxy/logs/: 
+alb/
+alternatives.log
+apt/... (200; 17.969537ms)
+Jun 18 12:15:54.351: INFO: (19) /api/v1/nodes/10.72.74.138/proxy/logs/: 
+alb/
+alternatives.log
+apt/... (200; 17.30402ms)
+[AfterEach] version v1
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 12:15:54.352: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "proxy-1339" for this suite.
+Jun 18 12:16:00.399: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 12:16:01.683: INFO: namespace proxy-1339 deletion completed in 7.318969644s
 
-• [SLOW TEST:8.886 seconds]
-[sig-storage] ConfigMap
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:32
-  should be consumable from pods in volume as non-root [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+• [SLOW TEST:7.948 seconds]
+[sig-network] Proxy
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:22
+  version v1
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/proxy.go:56
+    should proxy logs on node using proxy subresource  [Conformance]
+    /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-SSSSSSSSSSSSSSSSSSSSSSS
+SSSSSSSSSSSSS
 ------------------------------
-[sig-storage] ConfigMap 
-  should be consumable from pods in volume with mappings [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [sig-storage] ConfigMap
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+[sig-cli] Kubectl client [k8s.io] Kubectl logs 
+  should be able to retrieve and filter logs  [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-cli] Kubectl client
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 19:11:42.219: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-STEP: Building a namespace api object, basename configmap
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in configmap-4630
+Jun 18 12:16:01.684: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename kubectl
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-4243
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should be consumable from pods in volume with mappings [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: Creating configMap with name configmap-test-volume-map-9665e319-86fc-11e9-8d1b-467ee19922ac
-STEP: Creating a pod to test consume configMaps
-Jun  4 19:11:42.499: INFO: Waiting up to 5m0s for pod "pod-configmaps-9667fba4-86fc-11e9-8d1b-467ee19922ac" in namespace "configmap-4630" to be "success or failure"
-Jun  4 19:11:42.510: INFO: Pod "pod-configmaps-9667fba4-86fc-11e9-8d1b-467ee19922ac": Phase="Pending", Reason="", readiness=false. Elapsed: 10.851813ms
-Jun  4 19:11:44.523: INFO: Pod "pod-configmaps-9667fba4-86fc-11e9-8d1b-467ee19922ac": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.023914322s
-STEP: Saw pod success
-Jun  4 19:11:44.523: INFO: Pod "pod-configmaps-9667fba4-86fc-11e9-8d1b-467ee19922ac" satisfied condition "success or failure"
-Jun  4 19:11:44.536: INFO: Trying to get logs from node 10.212.23.189 pod pod-configmaps-9667fba4-86fc-11e9-8d1b-467ee19922ac container configmap-volume-test: 
-STEP: delete the pod
-Jun  4 19:11:44.759: INFO: Waiting for pod pod-configmaps-9667fba4-86fc-11e9-8d1b-467ee19922ac to disappear
-Jun  4 19:11:44.770: INFO: Pod pod-configmaps-9667fba4-86fc-11e9-8d1b-467ee19922ac no longer exists
-[AfterEach] [sig-storage] ConfigMap
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 19:11:44.770: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "configmap-4630" for this suite.
-Jun  4 19:11:50.833: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 19:11:51.263: INFO: namespace configmap-4630 deletion completed in 6.473955435s
+[BeforeEach] [sig-cli] Kubectl client
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:213
+[BeforeEach] [k8s.io] Kubectl logs
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1190
+STEP: creating an rc
+Jun 18 12:16:01.885: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 create -f - --namespace=kubectl-4243'
+Jun 18 12:16:02.226: INFO: stderr: ""
+Jun 18 12:16:02.226: INFO: stdout: "replicationcontroller/redis-master created\n"
+[It] should be able to retrieve and filter logs  [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Waiting for Redis master to start.
+Jun 18 12:16:03.240: INFO: Selector matched 1 pods for map[app:redis]
+Jun 18 12:16:03.240: INFO: Found 0 / 1
+Jun 18 12:16:04.240: INFO: Selector matched 1 pods for map[app:redis]
+Jun 18 12:16:04.240: INFO: Found 1 / 1
+Jun 18 12:16:04.240: INFO: WaitFor completed with timeout 5m0s.  Pods found = 1 out of 1
+Jun 18 12:16:04.254: INFO: Selector matched 1 pods for map[app:redis]
+Jun 18 12:16:04.254: INFO: ForEach: Found 1 pods from the filter.  Now looping through them.
+STEP: checking for a matching strings
+Jun 18 12:16:04.254: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 logs redis-master-xj6nb redis-master --namespace=kubectl-4243'
+Jun 18 12:16:04.784: INFO: stderr: ""
+Jun 18 12:16:04.784: INFO: stdout: "                _._                                                  \n           _.-``__ ''-._                                             \n      _.-``    `.  `_.  ''-._           Redis 3.2.12 (35a5711f/0) 64 bit\n  .-`` .-```.  ```\\/    _.,_ ''-._                                   \n (    '      ,       .-`  | `,    )     Running in standalone mode\n |`-._`-...-` __...-.``-._|'` _.-'|     Port: 6379\n |    `-._   `._    /     _.-'    |     PID: 1\n  `-._    `-._  `-./  _.-'    _.-'                                   \n |`-._`-._    `-.__.-'    _.-'_.-'|                                  \n |    `-._`-._        _.-'_.-'    |           http://redis.io        \n  `-._    `-._`-.__.-'_.-'    _.-'                                   \n |`-._`-._    `-.__.-'    _.-'_.-'|                                  \n |    `-._`-._        _.-'_.-'    |                                  \n  `-._    `-._`-.__.-'_.-'    _.-'                                   \n      `-._    `-.__.-'    _.-'                                       \n          `-._        _.-'                                           \n              `-.__.-'                                               \n\n1:M 18 Jun 12:16:04.071 # WARNING: The TCP backlog setting of 511 cannot be enforced because /proc/sys/net/core/somaxconn is set to the lower value of 128.\n1:M 18 Jun 12:16:04.071 # Server started, Redis version 3.2.12\n1:M 18 Jun 12:16:04.071 # WARNING you have Transparent Huge Pages (THP) support enabled in your kernel. This will create latency and memory usage issues with Redis. To fix this issue run the command 'echo never > /sys/kernel/mm/transparent_hugepage/enabled' as root, and add it to your /etc/rc.local in order to retain the setting after a reboot. Redis must be restarted after THP is disabled.\n1:M 18 Jun 12:16:04.071 * The server is now ready to accept connections on port 6379\n"
+STEP: limiting log lines
+Jun 18 12:16:04.784: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 log redis-master-xj6nb redis-master --namespace=kubectl-4243 --tail=1'
+Jun 18 12:16:04.936: INFO: stderr: ""
+Jun 18 12:16:04.936: INFO: stdout: "1:M 18 Jun 12:16:04.071 * The server is now ready to accept connections on port 6379\n"
+STEP: limiting log bytes
+Jun 18 12:16:04.936: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 log redis-master-xj6nb redis-master --namespace=kubectl-4243 --limit-bytes=1'
+Jun 18 12:16:05.154: INFO: stderr: ""
+Jun 18 12:16:05.155: INFO: stdout: " "
+STEP: exposing timestamps
+Jun 18 12:16:05.155: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 log redis-master-xj6nb redis-master --namespace=kubectl-4243 --tail=1 --timestamps'
+Jun 18 12:16:05.397: INFO: stderr: ""
+Jun 18 12:16:05.397: INFO: stdout: "2019-06-18T12:16:04.072511557Z 1:M 18 Jun 12:16:04.071 * The server is now ready to accept connections on port 6379\n"
+STEP: restricting to a time range
+Jun 18 12:16:07.897: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 log redis-master-xj6nb redis-master --namespace=kubectl-4243 --since=1s'
+Jun 18 12:16:08.084: INFO: stderr: ""
+Jun 18 12:16:08.084: INFO: stdout: ""
+Jun 18 12:16:08.084: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 log redis-master-xj6nb redis-master --namespace=kubectl-4243 --since=24h'
+Jun 18 12:16:08.243: INFO: stderr: ""
+Jun 18 12:16:08.243: INFO: stdout: "                _._                                                  \n           _.-``__ ''-._                                             \n      _.-``    `.  `_.  ''-._           Redis 3.2.12 (35a5711f/0) 64 bit\n  .-`` .-```.  ```\\/    _.,_ ''-._                                   \n (    '      ,       .-`  | `,    )     Running in standalone mode\n |`-._`-...-` __...-.``-._|'` _.-'|     Port: 6379\n |    `-._   `._    /     _.-'    |     PID: 1\n  `-._    `-._  `-./  _.-'    _.-'                                   \n |`-._`-._    `-.__.-'    _.-'_.-'|                                  \n |    `-._`-._        _.-'_.-'    |           http://redis.io        \n  `-._    `-._`-.__.-'_.-'    _.-'                                   \n |`-._`-._    `-.__.-'    _.-'_.-'|                                  \n |    `-._`-._        _.-'_.-'    |                                  \n  `-._    `-._`-.__.-'_.-'    _.-'                                   \n      `-._    `-.__.-'    _.-'                                       \n          `-._        _.-'                                           \n              `-.__.-'                                               \n\n1:M 18 Jun 12:16:04.071 # WARNING: The TCP backlog setting of 511 cannot be enforced because /proc/sys/net/core/somaxconn is set to the lower value of 128.\n1:M 18 Jun 12:16:04.071 # Server started, Redis version 3.2.12\n1:M 18 Jun 12:16:04.071 # WARNING you have Transparent Huge Pages (THP) support enabled in your kernel. This will create latency and memory usage issues with Redis. To fix this issue run the command 'echo never > /sys/kernel/mm/transparent_hugepage/enabled' as root, and add it to your /etc/rc.local in order to retain the setting after a reboot. Redis must be restarted after THP is disabled.\n1:M 18 Jun 12:16:04.071 * The server is now ready to accept connections on port 6379\n"
+[AfterEach] [k8s.io] Kubectl logs
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1196
+STEP: using delete to clean up resources
+Jun 18 12:16:08.244: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 delete --grace-period=0 --force -f - --namespace=kubectl-4243'
+Jun 18 12:16:08.404: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n"
+Jun 18 12:16:08.404: INFO: stdout: "replicationcontroller \"redis-master\" force deleted\n"
+Jun 18 12:16:08.405: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 get rc,svc -l name=nginx --no-headers --namespace=kubectl-4243'
+Jun 18 12:16:08.577: INFO: stderr: "No resources found.\n"
+Jun 18 12:16:08.577: INFO: stdout: ""
+Jun 18 12:16:08.577: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 get pods -l name=nginx --namespace=kubectl-4243 -o go-template={{ range .items }}{{ if not .metadata.deletionTimestamp }}{{ .metadata.name }}{{ "\n" }}{{ end }}{{ end }}'
+Jun 18 12:16:08.712: INFO: stderr: ""
+Jun 18 12:16:08.712: INFO: stdout: ""
+[AfterEach] [sig-cli] Kubectl client
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 12:16:08.712: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "kubectl-4243" for this suite.
+Jun 18 12:16:14.757: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 12:16:15.207: INFO: namespace kubectl-4243 deletion completed in 6.477929291s
 
-• [SLOW TEST:9.043 seconds]
-[sig-storage] ConfigMap
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:32
-  should be consumable from pods in volume with mappings [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+• [SLOW TEST:13.523 seconds]
+[sig-cli] Kubectl client
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23
+  [k8s.io] Kubectl logs
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
+    should be able to retrieve and filter logs  [Conformance]
+    /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-SS
+SSSSSSSSSSSSSS
 ------------------------------
-[k8s.io] Probing container 
-  should be restarted with a /healthz http liveness probe [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [k8s.io] Probing container
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+[sig-node] Downward API 
+  should provide pod name, namespace and IP address as env vars [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-node] Downward API
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 19:11:51.263: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-STEP: Building a namespace api object, basename container-probe
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in container-probe-3447
+Jun 18 12:16:15.207: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename downward-api
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in downward-api-9672
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [k8s.io] Probing container
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/container_probe.go:51
-[It] should be restarted with a /healthz http liveness probe [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: Creating pod liveness-http in namespace container-probe-3447
-Jun  4 19:11:55.526: INFO: Started pod liveness-http in namespace container-probe-3447
-STEP: checking the pod's current state and verifying that restartCount is present
-Jun  4 19:11:55.538: INFO: Initial restart count of pod liveness-http is 0
-Jun  4 19:12:17.757: INFO: Restart count of pod container-probe-3447/liveness-http is now 1 (22.219240175s elapsed)
-STEP: deleting the pod
-[AfterEach] [k8s.io] Probing container
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 19:12:17.793: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "container-probe-3447" for this suite.
-Jun  4 19:12:23.910: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 19:12:24.425: INFO: namespace container-probe-3447 deletion completed in 6.616422562s
+[It] should provide pod name, namespace and IP address as env vars [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating a pod to test downward api env vars
+Jun 18 12:16:15.441: INFO: Waiting up to 5m0s for pod "downward-api-de802103-91c2-11e9-a25d-8608290c688a" in namespace "downward-api-9672" to be "success or failure"
+Jun 18 12:16:15.460: INFO: Pod "downward-api-de802103-91c2-11e9-a25d-8608290c688a": Phase="Pending", Reason="", readiness=false. Elapsed: 18.961555ms
+Jun 18 12:16:17.473: INFO: Pod "downward-api-de802103-91c2-11e9-a25d-8608290c688a": Phase="Running", Reason="", readiness=true. Elapsed: 2.032427636s
+Jun 18 12:16:19.487: INFO: Pod "downward-api-de802103-91c2-11e9-a25d-8608290c688a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.046275279s
+STEP: Saw pod success
+Jun 18 12:16:19.487: INFO: Pod "downward-api-de802103-91c2-11e9-a25d-8608290c688a" satisfied condition "success or failure"
+Jun 18 12:16:19.500: INFO: Trying to get logs from node 10.72.74.184 pod downward-api-de802103-91c2-11e9-a25d-8608290c688a container dapi-container: 
+STEP: delete the pod
+Jun 18 12:16:19.572: INFO: Waiting for pod downward-api-de802103-91c2-11e9-a25d-8608290c688a to disappear
+Jun 18 12:16:19.586: INFO: Pod downward-api-de802103-91c2-11e9-a25d-8608290c688a no longer exists
+[AfterEach] [sig-node] Downward API
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 12:16:19.586: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "downward-api-9672" for this suite.
+Jun 18 12:16:25.626: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 12:16:26.044: INFO: namespace downward-api-9672 deletion completed in 6.444276996s
 
-• [SLOW TEST:33.162 seconds]
-[k8s.io] Probing container
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
-  should be restarted with a /healthz http liveness probe [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-------------------------------
-SS
+• [SLOW TEST:10.837 seconds]
+[sig-node] Downward API
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downward_api.go:38
+  should provide pod name, namespace and IP address as env vars [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-[sig-api-machinery] Aggregator 
-  Should be able to support the 1.10 Sample API Server using the current Aggregator [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [sig-api-machinery] Aggregator
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+[k8s.io] Docker Containers 
+  should be able to override the image's default command and arguments [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [k8s.io] Docker Containers
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 19:12:24.425: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-STEP: Building a namespace api object, basename aggregator
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in aggregator-6804
+Jun 18 12:16:26.045: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename containers
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in containers-2109
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-api-machinery] Aggregator
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/aggregator.go:69
-[It] Should be able to support the 1.10 Sample API Server using the current Aggregator [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: Registering the sample API server.
-Jun  4 19:12:25.512: INFO: deployment "sample-apiserver-deployment" doesn't have the required revision set
-Jun  4 19:12:27.617: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63695272345, loc:(*time.Location)(0x8a140e0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63695272345, loc:(*time.Location)(0x8a140e0)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63695272345, loc:(*time.Location)(0x8a140e0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63695272345, loc:(*time.Location)(0x8a140e0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-65db6755fc\" is progressing."}}, CollisionCount:(*int32)(nil)}
-Jun  4 19:12:29.625: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63695272345, loc:(*time.Location)(0x8a140e0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63695272345, loc:(*time.Location)(0x8a140e0)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63695272345, loc:(*time.Location)(0x8a140e0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63695272345, loc:(*time.Location)(0x8a140e0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-65db6755fc\" is progressing."}}, CollisionCount:(*int32)(nil)}
-Jun  4 19:12:31.627: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63695272345, loc:(*time.Location)(0x8a140e0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63695272345, loc:(*time.Location)(0x8a140e0)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63695272345, loc:(*time.Location)(0x8a140e0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63695272345, loc:(*time.Location)(0x8a140e0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-65db6755fc\" is progressing."}}, CollisionCount:(*int32)(nil)}
-Jun  4 19:12:33.627: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63695272345, loc:(*time.Location)(0x8a140e0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63695272345, loc:(*time.Location)(0x8a140e0)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63695272345, loc:(*time.Location)(0x8a140e0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63695272345, loc:(*time.Location)(0x8a140e0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-65db6755fc\" is progressing."}}, CollisionCount:(*int32)(nil)}
-Jun  4 19:12:35.627: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63695272345, loc:(*time.Location)(0x8a140e0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63695272345, loc:(*time.Location)(0x8a140e0)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63695272345, loc:(*time.Location)(0x8a140e0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63695272345, loc:(*time.Location)(0x8a140e0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-65db6755fc\" is progressing."}}, CollisionCount:(*int32)(nil)}
-Jun  4 19:12:41.229: INFO: Waited 3.561172369s for the sample-apiserver to be ready to handle requests.
-[AfterEach] [sig-api-machinery] Aggregator
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/aggregator.go:60
-[AfterEach] [sig-api-machinery] Aggregator
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 19:12:41.729: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "aggregator-6804" for this suite.
-Jun  4 19:12:47.809: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 19:12:48.183: INFO: namespace aggregator-6804 deletion completed in 6.441587803s
+[It] should be able to override the image's default command and arguments [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating a pod to test override all
+Jun 18 12:16:26.267: INFO: Waiting up to 5m0s for pod "client-containers-e4f4311d-91c2-11e9-a25d-8608290c688a" in namespace "containers-2109" to be "success or failure"
+Jun 18 12:16:26.300: INFO: Pod "client-containers-e4f4311d-91c2-11e9-a25d-8608290c688a": Phase="Pending", Reason="", readiness=false. Elapsed: 33.224321ms
+Jun 18 12:16:28.314: INFO: Pod "client-containers-e4f4311d-91c2-11e9-a25d-8608290c688a": Phase="Pending", Reason="", readiness=false. Elapsed: 2.047100207s
+Jun 18 12:16:30.332: INFO: Pod "client-containers-e4f4311d-91c2-11e9-a25d-8608290c688a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.065214199s
+STEP: Saw pod success
+Jun 18 12:16:30.332: INFO: Pod "client-containers-e4f4311d-91c2-11e9-a25d-8608290c688a" satisfied condition "success or failure"
+Jun 18 12:16:30.349: INFO: Trying to get logs from node 10.72.74.138 pod client-containers-e4f4311d-91c2-11e9-a25d-8608290c688a container test-container: 
+STEP: delete the pod
+Jun 18 12:16:30.432: INFO: Waiting for pod client-containers-e4f4311d-91c2-11e9-a25d-8608290c688a to disappear
+Jun 18 12:16:30.447: INFO: Pod client-containers-e4f4311d-91c2-11e9-a25d-8608290c688a no longer exists
+[AfterEach] [k8s.io] Docker Containers
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 12:16:30.447: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "containers-2109" for this suite.
+Jun 18 12:16:38.489: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 12:16:39.281: INFO: namespace containers-2109 deletion completed in 8.820019195s
 
-• [SLOW TEST:23.758 seconds]
-[sig-api-machinery] Aggregator
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:22
-  Should be able to support the 1.10 Sample API Server using the current Aggregator [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+• [SLOW TEST:13.237 seconds]
+[k8s.io] Docker Containers
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
+  should be able to override the image's default command and arguments [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+SSSSSSSSSSSSSSSS
 ------------------------------
-[sig-cli] Kubectl client [k8s.io] Kubectl describe 
-  should check if kubectl describe prints relevant information for rc and pods  [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+[sig-storage] EmptyDir volumes 
+  should support (root,0644,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-storage] EmptyDir volumes
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 19:12:48.184: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-STEP: Building a namespace api object, basename kubectl
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-519
+Jun 18 12:16:39.282: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename emptydir
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in emptydir-9084
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:213
-[It] should check if kubectl describe prints relevant information for rc and pods  [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-Jun  4 19:12:48.488: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 version --client'
-Jun  4 19:12:48.529: INFO: stderr: ""
-Jun  4 19:12:48.529: INFO: stdout: "Client Version: version.Info{Major:\"1\", Minor:\"14\", GitVersion:\"v1.14.2\", GitCommit:\"66049e3b21efe110454d67df4fa62b08ea79a19b\", GitTreeState:\"clean\", BuildDate:\"2019-05-16T16:23:09Z\", GoVersion:\"go1.12.5\", Compiler:\"gc\", Platform:\"linux/amd64\"}\n"
-Jun  4 19:12:48.533: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 create -f - --namespace=kubectl-519'
-Jun  4 19:12:48.812: INFO: stderr: ""
-Jun  4 19:12:48.812: INFO: stdout: "replicationcontroller/redis-master created\n"
-Jun  4 19:12:48.812: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 create -f - --namespace=kubectl-519'
-Jun  4 19:12:49.012: INFO: stderr: ""
-Jun  4 19:12:49.012: INFO: stdout: "service/redis-master created\n"
-STEP: Waiting for Redis master to start.
-Jun  4 19:12:50.025: INFO: Selector matched 1 pods for map[app:redis]
-Jun  4 19:12:50.025: INFO: Found 0 / 1
-Jun  4 19:12:51.025: INFO: Selector matched 1 pods for map[app:redis]
-Jun  4 19:12:51.025: INFO: Found 1 / 1
-Jun  4 19:12:51.025: INFO: WaitFor completed with timeout 5m0s.  Pods found = 1 out of 1
-Jun  4 19:12:51.054: INFO: Selector matched 1 pods for map[app:redis]
-Jun  4 19:12:51.054: INFO: ForEach: Found 1 pods from the filter.  Now looping through them.
-Jun  4 19:12:51.054: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 describe pod redis-master-fwbpj --namespace=kubectl-519'
-Jun  4 19:12:51.272: INFO: stderr: ""
-Jun  4 19:12:51.272: INFO: stdout: "Name:               redis-master-fwbpj\nNamespace:          kubectl-519\nPriority:           0\nPriorityClassName:  \nNode:               10.212.23.189/10.212.23.189\nStart Time:         Tue, 04 Jun 2019 19:12:48 +0000\nLabels:             app=redis\n                    role=master\nAnnotations:        kubernetes.io/psp: e2e-test-privileged-psp\nStatus:             Running\nIP:                 172.30.193.9\nControlled By:      ReplicationController/redis-master\nContainers:\n  redis-master:\n    Container ID:   containerd://0b87a53d106567874a13e82a34171693fed03b78ec8a47efa8cf88f8e8090df7\n    Image:          gcr.io/kubernetes-e2e-test-images/redis:1.0\n    Image ID:       gcr.io/kubernetes-e2e-test-images/redis@sha256:af4748d1655c08dc54d4be5182135395db9ce87aba2d4699b26b14ae197c5830\n    Port:           6379/TCP\n    Host Port:      0/TCP\n    State:          Running\n      Started:      Tue, 04 Jun 2019 19:12:49 +0000\n    Ready:          True\n    Restart Count:  0\n    Environment:    \n    Mounts:\n      /var/run/secrets/kubernetes.io/serviceaccount from default-token-7dvvz (ro)\nConditions:\n  Type              Status\n  Initialized       True \n  Ready             True \n  ContainersReady   True \n  PodScheduled      True \nVolumes:\n  default-token-7dvvz:\n    Type:        Secret (a volume populated by a Secret)\n    SecretName:  default-token-7dvvz\n    Optional:    false\nQoS Class:       BestEffort\nNode-Selectors:  \nTolerations:     node.kubernetes.io/not-ready:NoExecute for 300s\n                 node.kubernetes.io/unreachable:NoExecute for 300s\nEvents:\n  Type    Reason     Age   From                    Message\n  ----    ------     ----  ----                    -------\n  Normal  Scheduled  3s    default-scheduler       Successfully assigned kubectl-519/redis-master-fwbpj to 10.212.23.189\n  Normal  Pulled     2s    kubelet, 10.212.23.189  Container image \"gcr.io/kubernetes-e2e-test-images/redis:1.0\" already present on machine\n  Normal  Created    2s    kubelet, 10.212.23.189  Created container redis-master\n  Normal  Started    2s    kubelet, 10.212.23.189  Started container redis-master\n"
-Jun  4 19:12:51.272: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 describe rc redis-master --namespace=kubectl-519'
-Jun  4 19:12:51.409: INFO: stderr: ""
-Jun  4 19:12:51.409: INFO: stdout: "Name:         redis-master\nNamespace:    kubectl-519\nSelector:     app=redis,role=master\nLabels:       app=redis\n              role=master\nAnnotations:  \nReplicas:     1 current / 1 desired\nPods Status:  1 Running / 0 Waiting / 0 Succeeded / 0 Failed\nPod Template:\n  Labels:  app=redis\n           role=master\n  Containers:\n   redis-master:\n    Image:        gcr.io/kubernetes-e2e-test-images/redis:1.0\n    Port:         6379/TCP\n    Host Port:    0/TCP\n    Environment:  \n    Mounts:       \n  Volumes:        \nEvents:\n  Type    Reason            Age   From                    Message\n  ----    ------            ----  ----                    -------\n  Normal  SuccessfulCreate  3s    replication-controller  Created pod: redis-master-fwbpj\n"
-Jun  4 19:12:51.409: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 describe service redis-master --namespace=kubectl-519'
-Jun  4 19:12:51.542: INFO: stderr: ""
-Jun  4 19:12:51.542: INFO: stdout: "Name:              redis-master\nNamespace:         kubectl-519\nLabels:            app=redis\n                   role=master\nAnnotations:       \nSelector:          app=redis,role=master\nType:              ClusterIP\nIP:                172.21.2.39\nPort:                6379/TCP\nTargetPort:        redis-server/TCP\nEndpoints:         172.30.193.9:6379\nSession Affinity:  None\nEvents:            \n"
-Jun  4 19:12:51.559: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 describe node 10.212.23.161'
-Jun  4 19:12:51.707: INFO: stderr: ""
-Jun  4 19:12:51.707: INFO: stdout: "Name:               10.212.23.161\nRoles:              \nLabels:             arch=amd64\n                    beta.kubernetes.io/arch=amd64\n                    beta.kubernetes.io/instance-type=b3c.4x16.encrypted\n                    beta.kubernetes.io/os=linux\n                    failure-domain.beta.kubernetes.io/region=jp-tok\n                    failure-domain.beta.kubernetes.io/zone=tok02\n                    ibm-cloud.kubernetes.io/encrypted-docker-data=true\n                    ibm-cloud.kubernetes.io/external-ip=162.133.73.155\n                    ibm-cloud.kubernetes.io/ha-worker=true\n                    ibm-cloud.kubernetes.io/iaas-provider=softlayer\n                    ibm-cloud.kubernetes.io/internal-ip=10.212.23.161\n                    ibm-cloud.kubernetes.io/machine-type=b3c.4x16.encrypted\n                    ibm-cloud.kubernetes.io/os=UBUNTU_18_64\n                    ibm-cloud.kubernetes.io/region=jp-tok\n                    ibm-cloud.kubernetes.io/sgx-enabled=false\n                    ibm-cloud.kubernetes.io/worker-id=kube-tok02-cra6d87a4394f440e3becc50f237563475-w2\n                    ibm-cloud.kubernetes.io/worker-pool-id=a6d87a4394f440e3becc50f237563475-76069cf\n                    ibm-cloud.kubernetes.io/worker-version=1.14.2_1521\n                    ibm-cloud.kubernetes.io/zone=tok02\n                    kubernetes.io/arch=amd64\n                    kubernetes.io/hostname=10.212.23.161\n                    kubernetes.io/os=linux\n                    privateVLAN=2633737\n                    publicVLAN=2633735\nAnnotations:        node.alpha.kubernetes.io/ttl: 0\n                    volumes.kubernetes.io/controller-managed-attach-detach: true\nCreationTimestamp:  Tue, 04 Jun 2019 17:44:38 +0000\nTaints:             \nUnschedulable:      false\nConditions:\n  Type             Status  LastHeartbeatTime                 LastTransitionTime                Reason                       Message\n  ----             ------  -----------------                 ------------------                ------                       -------\n  MemoryPressure   False   Tue, 04 Jun 2019 19:12:15 +0000   Tue, 04 Jun 2019 17:44:38 +0000   KubeletHasSufficientMemory   kubelet has sufficient memory available\n  DiskPressure     False   Tue, 04 Jun 2019 19:12:15 +0000   Tue, 04 Jun 2019 17:44:38 +0000   KubeletHasNoDiskPressure     kubelet has no disk pressure\n  PIDPressure      False   Tue, 04 Jun 2019 19:12:15 +0000   Tue, 04 Jun 2019 17:44:38 +0000   KubeletHasSufficientPID      kubelet has sufficient PID available\n  Ready            True    Tue, 04 Jun 2019 19:12:15 +0000   Tue, 04 Jun 2019 17:44:48 +0000   KubeletReady                 kubelet is posting ready status. AppArmor enabled\nAddresses:\n  InternalIP:  10.212.23.161\n  ExternalIP:  162.133.73.155\n  Hostname:    10.212.23.161\nCapacity:\n cpu:                4\n ephemeral-storage:  102685624Ki\n hugepages-1Gi:      0\n hugepages-2Mi:      0\n memory:             16419912Ki\n pods:               110\nAllocatable:\n cpu:                3910m\n ephemeral-storage:  99892574949\n hugepages-1Gi:      0\n hugepages-2Mi:      0\n memory:             13627464Ki\n pods:               110\nSystem Info:\n Machine ID:                 2da5f13f22d94415a0a6a0a126b35a58\n System UUID:                6850DEB9-2314-F9FF-5790-7BBED769079E\n Boot ID:                    576911d2-6f90-4570-a7b4-d983e7916d38\n Kernel Version:             4.15.0-50-generic\n OS Image:                   Ubuntu 18.04.2 LTS\n Operating System:           linux\n Architecture:               amd64\n Container Runtime Version:  containerd://1.2.6\n Kubelet Version:            v1.14.2+IKS\n Kube-Proxy Version:         v1.14.2+IKS\nProviderID:                  ibm://d18c889395112a40d2f4e3065f237a7d///a6d87a4394f440e3becc50f237563475/kube-tok02-cra6d87a4394f440e3becc50f237563475-w2\nNon-terminated Pods:         (13 in total)\n  Namespace                  Name                                                               CPU Requests  CPU Limits  Memory Requests  Memory Limits  AGE\n  ---------                  ----                                                               ------------  ----------  ---------------  -------------  ---\n  heptio-sonobuoy            sonobuoy-systemd-logs-daemon-set-d63a0136eef44275-l8n68            0 (0%)        0 (0%)      0 (0%)           0 (0%)         45m\n  kube-system                calico-kube-controllers-64ccf5d5d4-k2qnd                           10m (0%)      0 (0%)      25Mi (0%)        0 (0%)         92m\n  kube-system                calico-node-wtt96                                                  250m (6%)     0 (0%)      80Mi (0%)        0 (0%)         88m\n  kube-system                coredns-autoscaler-5846b97547-t7mrd                                20m (0%)      0 (0%)      10Mi (0%)        0 (0%)         91m\n  kube-system                coredns-cc5d484bb-9nhk2                                            100m (2%)     0 (0%)      70Mi (0%)        400Mi (3%)     91m\n  kube-system                ibm-file-plugin-69d4bc9c8f-jwhqt                                   50m (1%)      200m (5%)   100Mi (0%)       0 (0%)         90m\n  kube-system                ibm-keepalived-watcher-vl7mx                                       5m (0%)       0 (0%)      10Mi (0%)        0 (0%)         88m\n  kube-system                ibm-kube-fluentd-cj2l2                                             25m (0%)      300m (7%)   150Mi (1%)       800M (5%)      87m\n  kube-system                ibm-master-proxy-static-10.212.23.161                              25m (0%)      300m (7%)   32M (0%)         512M (3%)      88m\n  kube-system                ibm-storage-watcher-f74547b86-jhnt2                                50m (1%)      200m (5%)   100Mi (0%)       0 (0%)         90m\n  kube-system                kubernetes-dashboard-b4f5d97fc-rrmfx                               50m (1%)      0 (0%)      100Mi (0%)       0 (0%)         88m\n  kube-system                public-cra6d87a4394f440e3becc50f237563475-alb1-79855b4ffb-mqcml    0 (0%)        0 (0%)      0 (0%)           0 (0%)         83m\n  kube-system                vpn-7bbfb5bc49-4mjxm                                               5m (0%)       0 (0%)      5Mi (0%)         0 (0%)         88m\nAllocated resources:\n  (Total limits may be over 100 percent, i.e., overcommitted.)\n  Resource           Requests       Limits\n  --------           --------       ------\n  cpu                590m (15%)     1 (25%)\n  memory             696850Ki (5%)  1690850Ki (12%)\n  ephemeral-storage  0 (0%)         0 (0%)\nEvents:              \n"
-Jun  4 19:12:51.708: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 describe namespace kubectl-519'
-Jun  4 19:12:51.822: INFO: stderr: ""
-Jun  4 19:12:51.822: INFO: stdout: "Name:         kubectl-519\nLabels:       e2e-framework=kubectl\n              e2e-run=79e879d1-86f6-11e9-8d1b-467ee19922ac\nAnnotations:  \nStatus:       Active\n\nNo resource quota.\n\nNo resource limits.\n"
-[AfterEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 19:12:51.822: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "kubectl-519" for this suite.
-Jun  4 19:13:15.879: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 19:13:16.303: INFO: namespace kubectl-519 deletion completed in 24.46491202s
+[It] should support (root,0644,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating a pod to test emptydir 0644 on tmpfs
+Jun 18 12:16:40.103: INFO: Waiting up to 5m0s for pod "pod-ed3347f7-91c2-11e9-a25d-8608290c688a" in namespace "emptydir-9084" to be "success or failure"
+Jun 18 12:16:40.116: INFO: Pod "pod-ed3347f7-91c2-11e9-a25d-8608290c688a": Phase="Pending", Reason="", readiness=false. Elapsed: 12.568727ms
+Jun 18 12:16:42.130: INFO: Pod "pod-ed3347f7-91c2-11e9-a25d-8608290c688a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.026523541s
+STEP: Saw pod success
+Jun 18 12:16:42.130: INFO: Pod "pod-ed3347f7-91c2-11e9-a25d-8608290c688a" satisfied condition "success or failure"
+Jun 18 12:16:42.143: INFO: Trying to get logs from node 10.72.74.184 pod pod-ed3347f7-91c2-11e9-a25d-8608290c688a container test-container: 
+STEP: delete the pod
+Jun 18 12:16:42.218: INFO: Waiting for pod pod-ed3347f7-91c2-11e9-a25d-8608290c688a to disappear
+Jun 18 12:16:42.230: INFO: Pod pod-ed3347f7-91c2-11e9-a25d-8608290c688a no longer exists
+[AfterEach] [sig-storage] EmptyDir volumes
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 12:16:42.230: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "emptydir-9084" for this suite.
+Jun 18 12:16:50.270: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 12:16:51.154: INFO: namespace emptydir-9084 deletion completed in 8.909164296s
 
-• [SLOW TEST:28.119 seconds]
-[sig-cli] Kubectl client
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23
-  [k8s.io] Kubectl describe
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
-    should check if kubectl describe prints relevant information for rc and pods  [Conformance]
-    /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+• [SLOW TEST:11.872 seconds]
+[sig-storage] EmptyDir volumes
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:41
+  should support (root,0644,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-SSSSSSSSSS
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-cli] Kubectl client [k8s.io] Proxy server 
-  should support proxy with --port 0  [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+[sig-storage] Downward API volume 
+  should provide container's memory request [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-storage] Downward API volume
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 19:13:16.303: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-STEP: Building a namespace api object, basename kubectl
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-2722
+Jun 18 12:16:51.158: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename downward-api
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in downward-api-4052
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:213
-[It] should support proxy with --port 0  [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: starting the proxy server
-Jun  4 19:13:16.524: INFO: Asynchronously running '/usr/local/bin/kubectl kubectl --kubeconfig=/tmp/kubeconfig-229234504 proxy -p 0 --disable-filter'
-STEP: curling proxy /api/ output
-[AfterEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 19:13:16.597: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "kubectl-2722" for this suite.
-Jun  4 19:13:22.654: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 19:13:23.085: INFO: namespace kubectl-2722 deletion completed in 6.473702252s
+[BeforeEach] [sig-storage] Downward API volume
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:39
+[It] should provide container's memory request [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating a pod to test downward API volume plugin
+Jun 18 12:16:51.381: INFO: Waiting up to 5m0s for pod "downwardapi-volume-f3ec6b8b-91c2-11e9-a25d-8608290c688a" in namespace "downward-api-4052" to be "success or failure"
+Jun 18 12:16:51.404: INFO: Pod "downwardapi-volume-f3ec6b8b-91c2-11e9-a25d-8608290c688a": Phase="Pending", Reason="", readiness=false. Elapsed: 22.767361ms
+Jun 18 12:16:53.418: INFO: Pod "downwardapi-volume-f3ec6b8b-91c2-11e9-a25d-8608290c688a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.036811483s
+STEP: Saw pod success
+Jun 18 12:16:53.418: INFO: Pod "downwardapi-volume-f3ec6b8b-91c2-11e9-a25d-8608290c688a" satisfied condition "success or failure"
+Jun 18 12:16:53.432: INFO: Trying to get logs from node 10.72.74.138 pod downwardapi-volume-f3ec6b8b-91c2-11e9-a25d-8608290c688a container client-container: 
+STEP: delete the pod
+Jun 18 12:16:53.507: INFO: Waiting for pod downwardapi-volume-f3ec6b8b-91c2-11e9-a25d-8608290c688a to disappear
+Jun 18 12:16:53.520: INFO: Pod downwardapi-volume-f3ec6b8b-91c2-11e9-a25d-8608290c688a no longer exists
+[AfterEach] [sig-storage] Downward API volume
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 12:16:53.520: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "downward-api-4052" for this suite.
+Jun 18 12:17:01.567: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 12:17:03.013: INFO: namespace downward-api-4052 deletion completed in 9.478086285s
 
-• [SLOW TEST:6.782 seconds]
-[sig-cli] Kubectl client
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23
-  [k8s.io] Proxy server
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
-    should support proxy with --port 0  [Conformance]
-    /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+• [SLOW TEST:11.855 seconds]
+[sig-storage] Downward API volume
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:34
+  should provide container's memory request [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-SSSSSSSSSSSSSSSSSSSS
+SSS
 ------------------------------
-[sig-cli] Kubectl client [k8s.io] Guestbook application 
-  should create and stop a working application  [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+[sig-storage] Secrets 
+  should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-storage] Secrets
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 19:13:23.087: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-STEP: Building a namespace api object, basename kubectl
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-375
+Jun 18 12:17:03.013: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename secrets
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in secrets-5210
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:213
-[It] should create and stop a working application  [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: creating all guestbook components
-Jun  4 19:13:23.300: INFO: apiVersion: v1
-kind: Service
-metadata:
-  name: redis-slave
-  labels:
-    app: redis
-    role: slave
-    tier: backend
-spec:
-  ports:
-  - port: 6379
-  selector:
-    app: redis
-    role: slave
-    tier: backend
-
-Jun  4 19:13:23.300: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 create -f - --namespace=kubectl-375'
-Jun  4 19:13:23.508: INFO: stderr: ""
-Jun  4 19:13:23.508: INFO: stdout: "service/redis-slave created\n"
-Jun  4 19:13:23.508: INFO: apiVersion: v1
-kind: Service
-metadata:
-  name: redis-master
-  labels:
-    app: redis
-    role: master
-    tier: backend
-spec:
-  ports:
-  - port: 6379
-    targetPort: 6379
-  selector:
-    app: redis
-    role: master
-    tier: backend
-
-Jun  4 19:13:23.508: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 create -f - --namespace=kubectl-375'
-Jun  4 19:13:23.700: INFO: stderr: ""
-Jun  4 19:13:23.700: INFO: stdout: "service/redis-master created\n"
-Jun  4 19:13:23.700: INFO: apiVersion: v1
-kind: Service
-metadata:
-  name: frontend
-  labels:
-    app: guestbook
-    tier: frontend
-spec:
-  # if your cluster supports it, uncomment the following to automatically create
-  # an external load-balanced IP for the frontend service.
-  # type: LoadBalancer
-  ports:
-  - port: 80
-  selector:
-    app: guestbook
-    tier: frontend
+[It] should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating secret with name secret-test-faff3743-91c2-11e9-a25d-8608290c688a
+STEP: Creating a pod to test consume secrets
+Jun 18 12:17:03.263: INFO: Waiting up to 5m0s for pod "pod-secrets-fb015912-91c2-11e9-a25d-8608290c688a" in namespace "secrets-5210" to be "success or failure"
+Jun 18 12:17:03.282: INFO: Pod "pod-secrets-fb015912-91c2-11e9-a25d-8608290c688a": Phase="Pending", Reason="", readiness=false. Elapsed: 19.749267ms
+Jun 18 12:17:05.296: INFO: Pod "pod-secrets-fb015912-91c2-11e9-a25d-8608290c688a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.033271178s
+STEP: Saw pod success
+Jun 18 12:17:05.296: INFO: Pod "pod-secrets-fb015912-91c2-11e9-a25d-8608290c688a" satisfied condition "success or failure"
+Jun 18 12:17:05.319: INFO: Trying to get logs from node 10.72.74.184 pod pod-secrets-fb015912-91c2-11e9-a25d-8608290c688a container secret-volume-test: 
+STEP: delete the pod
+Jun 18 12:17:05.470: INFO: Waiting for pod pod-secrets-fb015912-91c2-11e9-a25d-8608290c688a to disappear
+Jun 18 12:17:05.487: INFO: Pod pod-secrets-fb015912-91c2-11e9-a25d-8608290c688a no longer exists
+[AfterEach] [sig-storage] Secrets
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 12:17:05.487: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "secrets-5210" for this suite.
+Jun 18 12:17:11.535: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 12:17:11.935: INFO: namespace secrets-5210 deletion completed in 6.434969246s
 
-Jun  4 19:13:23.700: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 create -f - --namespace=kubectl-375'
-Jun  4 19:13:23.955: INFO: stderr: ""
-Jun  4 19:13:23.955: INFO: stdout: "service/frontend created\n"
-Jun  4 19:13:23.955: INFO: apiVersion: apps/v1
-kind: Deployment
-metadata:
-  name: frontend
-spec:
-  replicas: 3
-  selector:
-    matchLabels:
-      app: guestbook
-      tier: frontend
-  template:
-    metadata:
-      labels:
-        app: guestbook
-        tier: frontend
-    spec:
-      containers:
-      - name: php-redis
-        image: gcr.io/google-samples/gb-frontend:v6
-        resources:
-          requests:
-            cpu: 100m
-            memory: 100Mi
-        env:
-        - name: GET_HOSTS_FROM
-          value: dns
-          # If your cluster config does not include a dns service, then to
-          # instead access environment variables to find service host
-          # info, comment out the 'value: dns' line above, and uncomment the
-          # line below:
-          # value: env
-        ports:
-        - containerPort: 80
+• [SLOW TEST:8.923 seconds]
+[sig-storage] Secrets
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/secrets_volume.go:33
+  should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSS
+------------------------------
+[k8s.io] Container Runtime blackbox test when starting a container that exits 
+  should run with the expected status [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [k8s.io] Container Runtime
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Jun 18 12:17:11.936: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename container-runtime
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in container-runtime-3359
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should run with the expected status [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Container 'terminate-cmd-rpa': should get the expected 'RestartCount'
+STEP: Container 'terminate-cmd-rpa': should get the expected 'Phase'
+STEP: Container 'terminate-cmd-rpa': should get the expected 'Ready' condition
+STEP: Container 'terminate-cmd-rpa': should get the expected 'State'
+STEP: Container 'terminate-cmd-rpa': should be possible to delete [NodeConformance]
+STEP: Container 'terminate-cmd-rpof': should get the expected 'RestartCount'
+STEP: Container 'terminate-cmd-rpof': should get the expected 'Phase'
+STEP: Container 'terminate-cmd-rpof': should get the expected 'Ready' condition
+STEP: Container 'terminate-cmd-rpof': should get the expected 'State'
+STEP: Container 'terminate-cmd-rpof': should be possible to delete [NodeConformance]
+STEP: Container 'terminate-cmd-rpn': should get the expected 'RestartCount'
+STEP: Container 'terminate-cmd-rpn': should get the expected 'Phase'
+STEP: Container 'terminate-cmd-rpn': should get the expected 'Ready' condition
+STEP: Container 'terminate-cmd-rpn': should get the expected 'State'
+STEP: Container 'terminate-cmd-rpn': should be possible to delete [NodeConformance]
+[AfterEach] [k8s.io] Container Runtime
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 12:17:37.633: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "container-runtime-3359" for this suite.
+Jun 18 12:17:45.852: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 12:17:46.234: INFO: namespace container-runtime-3359 deletion completed in 8.544948426s
 
-Jun  4 19:13:23.955: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 create -f - --namespace=kubectl-375'
-Jun  4 19:13:24.189: INFO: stderr: ""
-Jun  4 19:13:24.189: INFO: stdout: "deployment.apps/frontend created\n"
-Jun  4 19:13:24.190: INFO: apiVersion: apps/v1
-kind: Deployment
-metadata:
-  name: redis-master
-spec:
-  replicas: 1
-  selector:
-    matchLabels:
-      app: redis
-      role: master
-      tier: backend
-  template:
-    metadata:
-      labels:
-        app: redis
-        role: master
-        tier: backend
-    spec:
-      containers:
-      - name: master
-        image: gcr.io/kubernetes-e2e-test-images/redis:1.0
-        resources:
-          requests:
-            cpu: 100m
-            memory: 100Mi
-        ports:
-        - containerPort: 6379
+• [SLOW TEST:34.299 seconds]
+[k8s.io] Container Runtime
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
+  blackbox test
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/runtime.go:37
+    when starting a container that exits
+    /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/runtime.go:38
+      should run with the expected status [NodeConformance] [Conformance]
+      /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSSSSSSSSSS
+------------------------------
+[sig-node] Downward API 
+  should provide pod UID as env vars [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-node] Downward API
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Jun 18 12:17:46.238: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename downward-api
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in downward-api-5809
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should provide pod UID as env vars [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating a pod to test downward api env vars
+Jun 18 12:17:46.488: INFO: Waiting up to 5m0s for pod "downward-api-14c4e721-91c3-11e9-a25d-8608290c688a" in namespace "downward-api-5809" to be "success or failure"
+Jun 18 12:17:46.501: INFO: Pod "downward-api-14c4e721-91c3-11e9-a25d-8608290c688a": Phase="Pending", Reason="", readiness=false. Elapsed: 13.39034ms
+Jun 18 12:17:48.515: INFO: Pod "downward-api-14c4e721-91c3-11e9-a25d-8608290c688a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.02743636s
+STEP: Saw pod success
+Jun 18 12:17:48.515: INFO: Pod "downward-api-14c4e721-91c3-11e9-a25d-8608290c688a" satisfied condition "success or failure"
+Jun 18 12:17:48.531: INFO: Trying to get logs from node 10.72.74.184 pod downward-api-14c4e721-91c3-11e9-a25d-8608290c688a container dapi-container: 
+STEP: delete the pod
+Jun 18 12:17:48.613: INFO: Waiting for pod downward-api-14c4e721-91c3-11e9-a25d-8608290c688a to disappear
+Jun 18 12:17:48.625: INFO: Pod downward-api-14c4e721-91c3-11e9-a25d-8608290c688a no longer exists
+[AfterEach] [sig-node] Downward API
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 12:17:48.625: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "downward-api-5809" for this suite.
+Jun 18 12:17:54.665: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 12:17:55.326: INFO: namespace downward-api-5809 deletion completed in 6.688373577s
 
-Jun  4 19:13:24.190: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 create -f - --namespace=kubectl-375'
-Jun  4 19:13:24.470: INFO: stderr: ""
-Jun  4 19:13:24.470: INFO: stdout: "deployment.apps/redis-master created\n"
-Jun  4 19:13:24.470: INFO: apiVersion: apps/v1
-kind: Deployment
-metadata:
-  name: redis-slave
-spec:
-  replicas: 2
-  selector:
-    matchLabels:
-      app: redis
-      role: slave
-      tier: backend
-  template:
-    metadata:
-      labels:
-        app: redis
-        role: slave
-        tier: backend
-    spec:
-      containers:
-      - name: slave
-        image: gcr.io/google-samples/gb-redisslave:v3
-        resources:
-          requests:
-            cpu: 100m
-            memory: 100Mi
-        env:
-        - name: GET_HOSTS_FROM
-          value: dns
-          # If your cluster config does not include a dns service, then to
-          # instead access an environment variable to find the master
-          # service's host, comment out the 'value: dns' line above, and
-          # uncomment the line below:
-          # value: env
-        ports:
-        - containerPort: 6379
+• [SLOW TEST:9.089 seconds]
+[sig-node] Downward API
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downward_api.go:38
+  should provide pod UID as env vars [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+------------------------------
+[sig-storage] Secrets 
+  optional updates should be reflected in volume [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-storage] Secrets
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Jun 18 12:17:55.330: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename secrets
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in secrets-1293
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] optional updates should be reflected in volume [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating secret with name s-test-opt-del-1a2eea19-91c3-11e9-a25d-8608290c688a
+STEP: Creating secret with name s-test-opt-upd-1a2eea7a-91c3-11e9-a25d-8608290c688a
+STEP: Creating the pod
+STEP: Deleting secret s-test-opt-del-1a2eea19-91c3-11e9-a25d-8608290c688a
+STEP: Updating secret s-test-opt-upd-1a2eea7a-91c3-11e9-a25d-8608290c688a
+STEP: Creating secret with name s-test-opt-create-1a2eeaa5-91c3-11e9-a25d-8608290c688a
+STEP: waiting to observe update in volume
+[AfterEach] [sig-storage] Secrets
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 12:18:00.163: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "secrets-1293" for this suite.
+Jun 18 12:18:24.209: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 12:18:25.182: INFO: namespace secrets-1293 deletion completed in 25.00521693s
 
-Jun  4 19:13:24.470: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 create -f - --namespace=kubectl-375'
-Jun  4 19:13:24.694: INFO: stderr: ""
-Jun  4 19:13:24.694: INFO: stdout: "deployment.apps/redis-slave created\n"
-STEP: validating guestbook app
-Jun  4 19:13:24.694: INFO: Waiting for all frontend pods to be Running.
-Jun  4 19:13:39.745: INFO: Waiting for frontend to serve content.
-Jun  4 19:13:44.785: INFO: Failed to get response from guestbook. err: , response: 
-Fatal error: Uncaught exception 'Predis\Connection\ConnectionException' with message 'Connection timed out [tcp://redis-slave:6379]' in /usr/local/lib/php/Predis/Connection/AbstractConnection.php:155 -Stack trace: -#0 /usr/local/lib/php/Predis/Connection/StreamConnection.php(128): Predis\Connection\AbstractConnection->onConnectionError('Connection time...', 110) -#1 /usr/local/lib/php/Predis/Connection/StreamConnection.php(178): Predis\Connection\StreamConnection->createStreamSocket(Object(Predis\Connection\Parameters), 'tcp://redis-sla...', 4) -#2 /usr/local/lib/php/Predis/Connection/StreamConnection.php(100): Predis\Connection\StreamConnection->tcpStreamInitializer(Object(Predis\Connection\Parameters)) -#3 /usr/local/lib/php/Predis/Connection/AbstractConnection.php(81): Predis\Connection\StreamConnection->createResource() -#4 /usr/local/lib/php/Predis/Connection/StreamConnection.php(258): Predis\Connection\AbstractConnection->connect() -#5 /usr/local/lib/php/Predis/Connection/AbstractConnection.php(180): Predis\Connection\Stre in /usr/local/lib/php/Predis/Connection/AbstractConnection.php on line 155
+• [SLOW TEST:29.852 seconds] +[sig-storage] Secrets +/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/secrets_volume.go:33 + optional updates should be reflected in volume [NodeConformance] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +------------------------------ +[k8s.io] KubeletManagedEtcHosts + should test kubelet managed /etc/hosts file [LinuxOnly] [NodeConformance] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [k8s.io] KubeletManagedEtcHosts + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +STEP: Creating a kubernetes client +Jun 18 12:18:25.183: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150 +STEP: Building a namespace api object, basename e2e-kubelet-etc-hosts +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-kubelet-etc-hosts-7610 +STEP: Waiting for a default service account to be provisioned in namespace +[It] should test kubelet managed /etc/hosts file [LinuxOnly] [NodeConformance] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: Setting up the test +STEP: Creating hostNetwork=false pod +STEP: Creating hostNetwork=true pod +STEP: Running the test +STEP: Verifying /etc/hosts of container is kubelet-managed for pod with hostNetwork=false +Jun 18 12:18:29.499: INFO: ExecWithOptions {Command:[cat /etc/hosts] Namespace:e2e-kubelet-etc-hosts-7610 PodName:test-pod ContainerName:busybox-1 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} +Jun 18 12:18:29.499: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150 +Jun 18 12:18:29.729: INFO: Exec stderr: "" +Jun 18 12:18:29.729: INFO: ExecWithOptions {Command:[cat /etc/hosts-original] Namespace:e2e-kubelet-etc-hosts-7610 PodName:test-pod ContainerName:busybox-1 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} +Jun 18 12:18:29.729: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150 +Jun 18 12:18:29.940: INFO: Exec stderr: "" +Jun 18 12:18:29.940: INFO: ExecWithOptions {Command:[cat /etc/hosts] Namespace:e2e-kubelet-etc-hosts-7610 PodName:test-pod ContainerName:busybox-2 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} +Jun 18 12:18:29.940: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150 +Jun 18 12:18:30.191: INFO: Exec stderr: "" +Jun 18 12:18:30.191: INFO: ExecWithOptions {Command:[cat /etc/hosts-original] Namespace:e2e-kubelet-etc-hosts-7610 PodName:test-pod ContainerName:busybox-2 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} +Jun 18 12:18:30.191: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150 +Jun 18 12:18:30.453: INFO: Exec stderr: "" +STEP: Verifying /etc/hosts of container is not kubelet-managed since container specifies /etc/hosts mount +Jun 18 12:18:30.453: INFO: ExecWithOptions {Command:[cat /etc/hosts] Namespace:e2e-kubelet-etc-hosts-7610 PodName:test-pod ContainerName:busybox-3 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} +Jun 18 12:18:30.453: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150 +Jun 18 12:18:30.725: INFO: Exec stderr: "" +Jun 18 12:18:30.725: INFO: ExecWithOptions {Command:[cat /etc/hosts-original] Namespace:e2e-kubelet-etc-hosts-7610 PodName:test-pod ContainerName:busybox-3 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} +Jun 18 12:18:30.725: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150 +Jun 18 12:18:30.950: INFO: Exec stderr: "" +STEP: Verifying /etc/hosts content of container is not kubelet-managed for pod with hostNetwork=true +Jun 18 12:18:30.950: INFO: ExecWithOptions {Command:[cat /etc/hosts] Namespace:e2e-kubelet-etc-hosts-7610 PodName:test-host-network-pod ContainerName:busybox-1 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} +Jun 18 12:18:30.950: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150 +Jun 18 12:18:31.298: INFO: Exec stderr: "" +Jun 18 12:18:31.298: INFO: ExecWithOptions {Command:[cat /etc/hosts-original] Namespace:e2e-kubelet-etc-hosts-7610 PodName:test-host-network-pod ContainerName:busybox-1 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} +Jun 18 12:18:31.299: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150 +Jun 18 12:18:31.925: INFO: Exec stderr: "" +Jun 18 12:18:31.925: INFO: ExecWithOptions {Command:[cat /etc/hosts] Namespace:e2e-kubelet-etc-hosts-7610 PodName:test-host-network-pod ContainerName:busybox-2 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} +Jun 18 12:18:31.925: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150 +Jun 18 12:18:33.096: INFO: Exec stderr: "" +Jun 18 12:18:33.096: INFO: ExecWithOptions {Command:[cat /etc/hosts-original] Namespace:e2e-kubelet-etc-hosts-7610 PodName:test-host-network-pod ContainerName:busybox-2 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} +Jun 18 12:18:33.096: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150 +Jun 18 12:18:33.364: INFO: Exec stderr: "" +[AfterEach] [k8s.io] KubeletManagedEtcHosts + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Jun 18 12:18:33.364: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "e2e-kubelet-etc-hosts-7610" for this suite. +Jun 18 12:19:11.411: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:19:12.142: INFO: namespace e2e-kubelet-etc-hosts-7610 deletion completed in 38.763124923s -Jun 4 19:13:49.820: INFO: Trying to add a new entry to the guestbook. -Jun 4 19:13:49.847: INFO: Verifying that added entry can be retrieved. -STEP: using delete to clean up resources -Jun 4 19:13:49.880: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 delete --grace-period=0 --force -f - --namespace=kubectl-375' -Jun 4 19:13:50.042: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" -Jun 4 19:13:50.042: INFO: stdout: "service \"redis-slave\" force deleted\n" -STEP: using delete to clean up resources -Jun 4 19:13:50.042: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 delete --grace-period=0 --force -f - --namespace=kubectl-375' -Jun 4 19:13:50.181: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" -Jun 4 19:13:50.181: INFO: stdout: "service \"redis-master\" force deleted\n" -STEP: using delete to clean up resources -Jun 4 19:13:50.181: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 delete --grace-period=0 --force -f - --namespace=kubectl-375' -Jun 4 19:13:50.325: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" -Jun 4 19:13:50.325: INFO: stdout: "service \"frontend\" force deleted\n" -STEP: using delete to clean up resources -Jun 4 19:13:50.325: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 delete --grace-period=0 --force -f - --namespace=kubectl-375' -Jun 4 19:13:50.435: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" -Jun 4 19:13:50.435: INFO: stdout: "deployment.apps \"frontend\" force deleted\n" -STEP: using delete to clean up resources -Jun 4 19:13:50.435: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 delete --grace-period=0 --force -f - --namespace=kubectl-375' -Jun 4 19:13:50.552: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" -Jun 4 19:13:50.552: INFO: stdout: "deployment.apps \"redis-master\" force deleted\n" -STEP: using delete to clean up resources -Jun 4 19:13:50.552: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 delete --grace-period=0 --force -f - --namespace=kubectl-375' -Jun 4 19:13:50.647: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" -Jun 4 19:13:50.647: INFO: stdout: "deployment.apps \"redis-slave\" force deleted\n" +• [SLOW TEST:46.959 seconds] +[k8s.io] KubeletManagedEtcHosts +/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687 + should test kubelet managed /etc/hosts file [LinuxOnly] [NodeConformance] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +------------------------------ +SS +------------------------------ +[sig-cli] Kubectl client [k8s.io] Kubectl version + should check is all data is printed [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [sig-cli] Kubectl client + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +STEP: Creating a kubernetes client +Jun 18 12:19:12.142: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150 +STEP: Building a namespace api object, basename kubectl +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-615 +STEP: Waiting for a default service account to be provisioned in namespace +[BeforeEach] [sig-cli] Kubectl client + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:213 +[It] should check is all data is printed [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +Jun 18 12:19:12.350: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 version' +Jun 18 12:19:12.459: INFO: stderr: "" +Jun 18 12:19:12.459: INFO: stdout: "Client Version: version.Info{Major:\"1\", Minor:\"14\", GitVersion:\"v1.14.3\", GitCommit:\"5e53fd6bc17c0dec8434817e69b04a25d8ae0ff0\", GitTreeState:\"clean\", BuildDate:\"2019-06-06T01:44:30Z\", GoVersion:\"go1.12.5\", Compiler:\"gc\", Platform:\"linux/amd64\"}\nServer Version: version.Info{Major:\"1\", Minor:\"14\", GitVersion:\"v1.14.3+IKS\", GitCommit:\"1c6446568f0a7d9d62dfe2f6c6a8388b3d4b1a6b\", GitTreeState:\"clean\", BuildDate:\"2019-06-13T12:53:41Z\", GoVersion:\"go1.12.5\", Compiler:\"gc\", Platform:\"linux/amd64\"}\n" [AfterEach] [sig-cli] Kubectl client - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 -Jun 4 19:13:50.647: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "kubectl-375" for this suite. -Jun 4 19:14:30.707: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -Jun 4 19:14:31.269: INFO: namespace kubectl-375 deletion completed in 40.602577051s + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Jun 18 12:19:12.459: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "kubectl-615" for this suite. +Jun 18 12:19:18.522: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:19:18.916: INFO: namespace kubectl-615 deletion completed in 6.442072789s -• [SLOW TEST:68.182 seconds] +• [SLOW TEST:6.774 seconds] [sig-cli] Kubectl client -/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 - [k8s.io] Guestbook application - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687 - should create and stop a working application [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 + [k8s.io] Kubectl version + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687 + should check is all data is printed [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 ------------------------------ -SSSSSSSSS +S ------------------------------ [sig-storage] Downward API volume - should update labels on modification [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 + should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 [BeforeEach] [sig-storage] Downward API volume - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 STEP: Creating a kubernetes client -Jun 4 19:14:31.269: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504 +Jun 18 12:19:18.917: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150 STEP: Building a namespace api object, basename downward-api -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in downward-api-8925 +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in downward-api-7807 STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Downward API volume - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:39 -[It] should update labels on modification [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -STEP: Creating the pod -Jun 4 19:14:34.112: INFO: Successfully updated pod "labelsupdatefb264543-86fc-11e9-8d1b-467ee19922ac" + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:39 +[It] should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: Creating a pod to test downward API volume plugin +Jun 18 12:19:19.158: INFO: Waiting up to 5m0s for pod "downwardapi-volume-4c0062b1-91c3-11e9-a25d-8608290c688a" in namespace "downward-api-7807" to be "success or failure" +Jun 18 12:19:19.174: INFO: Pod "downwardapi-volume-4c0062b1-91c3-11e9-a25d-8608290c688a": Phase="Pending", Reason="", readiness=false. Elapsed: 15.3725ms +Jun 18 12:19:21.188: INFO: Pod "downwardapi-volume-4c0062b1-91c3-11e9-a25d-8608290c688a": Phase="Pending", Reason="", readiness=false. Elapsed: 2.030362364s +Jun 18 12:19:23.204: INFO: Pod "downwardapi-volume-4c0062b1-91c3-11e9-a25d-8608290c688a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.045526432s +STEP: Saw pod success +Jun 18 12:19:23.204: INFO: Pod "downwardapi-volume-4c0062b1-91c3-11e9-a25d-8608290c688a" satisfied condition "success or failure" +Jun 18 12:19:23.217: INFO: Trying to get logs from node 10.72.74.184 pod downwardapi-volume-4c0062b1-91c3-11e9-a25d-8608290c688a container client-container: +STEP: delete the pod +Jun 18 12:19:23.300: INFO: Waiting for pod downwardapi-volume-4c0062b1-91c3-11e9-a25d-8608290c688a to disappear +Jun 18 12:19:23.313: INFO: Pod downwardapi-volume-4c0062b1-91c3-11e9-a25d-8608290c688a no longer exists [AfterEach] [sig-storage] Downward API volume - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 -Jun 4 19:14:36.209: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "downward-api-8925" for this suite. -Jun 4 19:15:00.275: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -Jun 4 19:15:00.726: INFO: namespace downward-api-8925 deletion completed in 24.500802974s + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Jun 18 12:19:23.313: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "downward-api-7807" for this suite. +Jun 18 12:19:29.353: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:19:29.746: INFO: namespace downward-api-7807 deletion completed in 6.418946536s -• [SLOW TEST:29.457 seconds] +• [SLOW TEST:10.829 seconds] [sig-storage] Downward API volume -/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:34 - should update labels on modification [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 ------------------------------- -SSSSSSSSSSSSS +/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:34 + should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 ------------------------------ -[sig-storage] Projected downwardAPI - should provide container's memory limit [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -[BeforeEach] [sig-storage] Projected downwardAPI - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +[sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] + Scaling should happen in predictable order and halt if any stateful pod is unhealthy [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [sig-apps] StatefulSet + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 STEP: Creating a kubernetes client -Jun 4 19:15:00.727: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504 -STEP: Building a namespace api object, basename projected -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-6172 +Jun 18 12:19:29.746: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150 +STEP: Building a namespace api object, basename statefulset +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in statefulset-1300 STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [sig-storage] Projected downwardAPI - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:39 -[It] should provide container's memory limit [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -STEP: Creating a pod to test downward API volume plugin -Jun 4 19:15:01.010: INFO: Waiting up to 5m0s for pod "downwardapi-volume-0cb35950-86fd-11e9-8d1b-467ee19922ac" in namespace "projected-6172" to be "success or failure" -Jun 4 19:15:01.022: INFO: Pod "downwardapi-volume-0cb35950-86fd-11e9-8d1b-467ee19922ac": Phase="Pending", Reason="", readiness=false. Elapsed: 12.228849ms -Jun 4 19:15:03.034: INFO: Pod "downwardapi-volume-0cb35950-86fd-11e9-8d1b-467ee19922ac": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.023564955s -STEP: Saw pod success -Jun 4 19:15:03.034: INFO: Pod "downwardapi-volume-0cb35950-86fd-11e9-8d1b-467ee19922ac" satisfied condition "success or failure" -Jun 4 19:15:03.045: INFO: Trying to get logs from node 10.212.23.164 pod downwardapi-volume-0cb35950-86fd-11e9-8d1b-467ee19922ac container client-container: -STEP: delete the pod -Jun 4 19:15:03.112: INFO: Waiting for pod downwardapi-volume-0cb35950-86fd-11e9-8d1b-467ee19922ac to disappear -Jun 4 19:15:03.122: INFO: Pod downwardapi-volume-0cb35950-86fd-11e9-8d1b-467ee19922ac no longer exists -[AfterEach] [sig-storage] Projected downwardAPI - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 -Jun 4 19:15:03.122: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "projected-6172" for this suite. -Jun 4 19:15:09.195: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -Jun 4 19:15:09.588: INFO: namespace projected-6172 deletion completed in 6.45011402s +[BeforeEach] [sig-apps] StatefulSet + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:59 +[BeforeEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:74 +STEP: Creating service test in namespace statefulset-1300 +[It] Scaling should happen in predictable order and halt if any stateful pod is unhealthy [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: Initializing watcher for selector baz=blah,foo=bar +STEP: Creating stateful set ss in namespace statefulset-1300 +STEP: Waiting until all stateful set ss replicas will be running in namespace statefulset-1300 +Jun 18 12:19:29.991: INFO: Found 0 stateful pods, waiting for 1 +Jun 18 12:19:40.007: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true +STEP: Confirming that stateful set scale up will halt with unhealthy stateful pod +Jun 18 12:19:40.027: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-1300 ss-0 -- /bin/sh -x -c mv -v /usr/share/nginx/html/index.html /tmp/ || true' +Jun 18 12:19:40.448: INFO: stderr: "+ mv -v /usr/share/nginx/html/index.html /tmp/\n" +Jun 18 12:19:40.448: INFO: stdout: "'/usr/share/nginx/html/index.html' -> '/tmp/index.html'\n" +Jun 18 12:19:40.448: INFO: stdout of mv -v /usr/share/nginx/html/index.html /tmp/ || true on ss-0: '/usr/share/nginx/html/index.html' -> '/tmp/index.html' + +Jun 18 12:19:40.469: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=true +Jun 18 12:19:50.484: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=false +Jun 18 12:19:50.484: INFO: Waiting for statefulset status.replicas updated to 0 +Jun 18 12:19:50.539: INFO: Verifying statefulset ss doesn't scale past 1 for another 9.99999815s +Jun 18 12:19:51.554: INFO: Verifying statefulset ss doesn't scale past 1 for another 8.987542643s +Jun 18 12:19:52.568: INFO: Verifying statefulset ss doesn't scale past 1 for another 7.97265331s +Jun 18 12:19:53.585: INFO: Verifying statefulset ss doesn't scale past 1 for another 6.95873704s +Jun 18 12:19:54.599: INFO: Verifying statefulset ss doesn't scale past 1 for another 5.941376581s +Jun 18 12:19:55.613: INFO: Verifying statefulset ss doesn't scale past 1 for another 4.927226841s +Jun 18 12:19:56.629: INFO: Verifying statefulset ss doesn't scale past 1 for another 3.913675328s +Jun 18 12:19:57.643: INFO: Verifying statefulset ss doesn't scale past 1 for another 2.897190147s +Jun 18 12:19:58.656: INFO: Verifying statefulset ss doesn't scale past 1 for another 1.883756876s +Jun 18 12:19:59.670: INFO: Verifying statefulset ss doesn't scale past 1 for another 870.762652ms +STEP: Scaling up stateful set ss to 3 replicas and waiting until all of them will be running in namespace statefulset-1300 +Jun 18 12:20:00.684: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-1300 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true' +Jun 18 12:20:01.111: INFO: stderr: "+ mv -v /tmp/index.html /usr/share/nginx/html/\n" +Jun 18 12:20:01.111: INFO: stdout: "'/tmp/index.html' -> '/usr/share/nginx/html/index.html'\n" +Jun 18 12:20:01.111: INFO: stdout of mv -v /tmp/index.html /usr/share/nginx/html/ || true on ss-0: '/tmp/index.html' -> '/usr/share/nginx/html/index.html' + +Jun 18 12:20:01.125: INFO: Found 1 stateful pods, waiting for 3 +Jun 18 12:20:11.139: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true +Jun 18 12:20:11.139: INFO: Waiting for pod ss-1 to enter Running - Ready=true, currently Running - Ready=true +Jun 18 12:20:11.139: INFO: Waiting for pod ss-2 to enter Running - Ready=true, currently Running - Ready=true +STEP: Verifying that stateful set ss was scaled up in order +STEP: Scale down will halt with unhealthy stateful pod +Jun 18 12:20:11.160: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-1300 ss-0 -- /bin/sh -x -c mv -v /usr/share/nginx/html/index.html /tmp/ || true' +Jun 18 12:20:11.542: INFO: stderr: "+ mv -v /usr/share/nginx/html/index.html /tmp/\n" +Jun 18 12:20:11.542: INFO: stdout: "'/usr/share/nginx/html/index.html' -> '/tmp/index.html'\n" +Jun 18 12:20:11.542: INFO: stdout of mv -v /usr/share/nginx/html/index.html /tmp/ || true on ss-0: '/usr/share/nginx/html/index.html' -> '/tmp/index.html' + +Jun 18 12:20:11.543: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-1300 ss-1 -- /bin/sh -x -c mv -v /usr/share/nginx/html/index.html /tmp/ || true' +Jun 18 12:20:11.943: INFO: stderr: "+ mv -v /usr/share/nginx/html/index.html /tmp/\n" +Jun 18 12:20:11.943: INFO: stdout: "'/usr/share/nginx/html/index.html' -> '/tmp/index.html'\n" +Jun 18 12:20:11.943: INFO: stdout of mv -v /usr/share/nginx/html/index.html /tmp/ || true on ss-1: '/usr/share/nginx/html/index.html' -> '/tmp/index.html' + +Jun 18 12:20:11.943: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-1300 ss-2 -- /bin/sh -x -c mv -v /usr/share/nginx/html/index.html /tmp/ || true' +Jun 18 12:20:12.385: INFO: stderr: "+ mv -v /usr/share/nginx/html/index.html /tmp/\n" +Jun 18 12:20:12.385: INFO: stdout: "'/usr/share/nginx/html/index.html' -> '/tmp/index.html'\n" +Jun 18 12:20:12.385: INFO: stdout of mv -v /usr/share/nginx/html/index.html /tmp/ || true on ss-2: '/usr/share/nginx/html/index.html' -> '/tmp/index.html' + +Jun 18 12:20:12.385: INFO: Waiting for statefulset status.replicas updated to 0 +Jun 18 12:20:12.399: INFO: Waiting for stateful set status.readyReplicas to become 0, currently 1 +Jun 18 12:20:22.424: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=false +Jun 18 12:20:22.424: INFO: Waiting for pod ss-1 to enter Running - Ready=false, currently Running - Ready=false +Jun 18 12:20:22.424: INFO: Waiting for pod ss-2 to enter Running - Ready=false, currently Running - Ready=false +Jun 18 12:20:22.479: INFO: Verifying statefulset ss doesn't scale past 3 for another 9.99999817s +Jun 18 12:20:23.494: INFO: Verifying statefulset ss doesn't scale past 3 for another 8.970987997s +Jun 18 12:20:24.527: INFO: Verifying statefulset ss doesn't scale past 3 for another 7.956590512s +Jun 18 12:20:25.541: INFO: Verifying statefulset ss doesn't scale past 3 for another 6.92354235s +Jun 18 12:20:26.555: INFO: Verifying statefulset ss doesn't scale past 3 for another 5.909426024s +Jun 18 12:20:27.569: INFO: Verifying statefulset ss doesn't scale past 3 for another 4.895508083s +Jun 18 12:20:28.582: INFO: Verifying statefulset ss doesn't scale past 3 for another 3.88189114s +Jun 18 12:20:29.596: INFO: Verifying statefulset ss doesn't scale past 3 for another 2.867996693s +Jun 18 12:20:30.612: INFO: Verifying statefulset ss doesn't scale past 3 for another 1.85414232s +Jun 18 12:20:31.628: INFO: Verifying statefulset ss doesn't scale past 3 for another 838.629601ms +STEP: Scaling down stateful set ss to 0 replicas and waiting until none of pods will run in namespacestatefulset-1300 +Jun 18 12:20:32.644: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-1300 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true' +Jun 18 12:20:33.043: INFO: stderr: "+ mv -v /tmp/index.html /usr/share/nginx/html/\n" +Jun 18 12:20:33.043: INFO: stdout: "'/tmp/index.html' -> '/usr/share/nginx/html/index.html'\n" +Jun 18 12:20:33.043: INFO: stdout of mv -v /tmp/index.html /usr/share/nginx/html/ || true on ss-0: '/tmp/index.html' -> '/usr/share/nginx/html/index.html' + +Jun 18 12:20:33.043: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-1300 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true' +Jun 18 12:20:33.914: INFO: stderr: "+ mv -v /tmp/index.html /usr/share/nginx/html/\n" +Jun 18 12:20:33.915: INFO: stdout: "'/tmp/index.html' -> '/usr/share/nginx/html/index.html'\n" +Jun 18 12:20:33.915: INFO: stdout of mv -v /tmp/index.html /usr/share/nginx/html/ || true on ss-1: '/tmp/index.html' -> '/usr/share/nginx/html/index.html' + +Jun 18 12:20:33.915: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-1300 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true' +Jun 18 12:20:34.271: INFO: rc: 1 +Jun 18 12:20:34.271: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-1300 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] [] error: Internal error occurred: error executing command in container: failed to exec in container: failed to load task: no running task found: not found + [] 0xc0027822d0 exit status 1 true [0xc00018d9e0 0xc00018db48 0xc00018dba8] [0xc00018d9e0 0xc00018db48 0xc00018dba8] [0xc00018daa0 0xc00018db90] [0x9c00a0 0x9c00a0] 0xc000abbb60 }: +Command stdout: + +stderr: +error: Internal error occurred: error executing command in container: failed to exec in container: failed to load task: no running task found: not found + +error: +exit status 1 + +Jun 18 12:20:44.272: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-1300 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true' +Jun 18 12:20:44.543: INFO: rc: 1 +Jun 18 12:20:44.543: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-1300 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] [] error: unable to upgrade connection: container not found ("nginx") + [] 0xc003132cc0 exit status 1 true [0xc0019a8368 0xc0019a83b0 0xc0019a83e0] [0xc0019a8368 0xc0019a83b0 0xc0019a83e0] [0xc0019a83a8 0xc0019a83d0] [0x9c00a0 0x9c00a0] 0xc0031d8a20 }: +Command stdout: + +stderr: +error: unable to upgrade connection: container not found ("nginx") + +error: +exit status 1 + +Jun 18 12:20:54.544: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-1300 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true' +Jun 18 12:20:54.679: INFO: rc: 1 +Jun 18 12:20:54.679: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-1300 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] [] Error from server (NotFound): pods "ss-2" not found + [] 0xc003133020 exit status 1 true [0xc0019a83e8 0xc0019a8410 0xc0019a8430] [0xc0019a83e8 0xc0019a8410 0xc0019a8430] [0xc0019a8400 0xc0019a8428] [0x9c00a0 0x9c00a0] 0xc0031d8d80 }: +Command stdout: + +stderr: +Error from server (NotFound): pods "ss-2" not found + +error: +exit status 1 + +Jun 18 12:21:04.679: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-1300 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true' +Jun 18 12:21:04.786: INFO: rc: 1 +Jun 18 12:21:04.786: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-1300 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] [] Error from server (NotFound): pods "ss-2" not found + [] 0xc003133470 exit status 1 true [0xc0019a8438 0xc0019a8488 0xc0019a84e8] [0xc0019a8438 0xc0019a8488 0xc0019a84e8] [0xc0019a8470 0xc0019a84b0] [0x9c00a0 0x9c00a0] 0xc0031d90e0 }: +Command stdout: + +stderr: +Error from server (NotFound): pods "ss-2" not found + +error: +exit status 1 + +Jun 18 12:21:14.786: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-1300 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true' +Jun 18 12:21:14.917: INFO: rc: 1 +Jun 18 12:21:14.917: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-1300 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] [] Error from server (NotFound): pods "ss-2" not found + [] 0xc003133920 exit status 1 true [0xc0019a8508 0xc0019a8528 0xc0019a8578] [0xc0019a8508 0xc0019a8528 0xc0019a8578] [0xc0019a8520 0xc0019a8548] [0x9c00a0 0x9c00a0] 0xc0031d95c0 }: +Command stdout: + +stderr: +Error from server (NotFound): pods "ss-2" not found + +error: +exit status 1 + +Jun 18 12:21:24.917: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-1300 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true' +Jun 18 12:21:25.032: INFO: rc: 1 +Jun 18 12:21:25.032: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-1300 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] [] Error from server (NotFound): pods "ss-2" not found + [] 0xc002782690 exit status 1 true [0xc00018dbc0 0xc00018dc48 0xc00018dca0] [0xc00018dbc0 0xc00018dc48 0xc00018dca0] [0xc00018dc00 0xc00018dc88] [0x9c00a0 0x9c00a0] 0xc00294e240 }: +Command stdout: + +stderr: +Error from server (NotFound): pods "ss-2" not found + +error: +exit status 1 + +Jun 18 12:21:35.032: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-1300 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true' +Jun 18 12:21:35.152: INFO: rc: 1 +Jun 18 12:21:35.153: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-1300 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] [] Error from server (NotFound): pods "ss-2" not found + [] 0xc002939e90 exit status 1 true [0xc000011ee0 0xc000011f38 0xc000011fb8] [0xc000011ee0 0xc000011f38 0xc000011fb8] [0xc000011f10 0xc000011fb0] [0x9c00a0 0x9c00a0] 0xc002f25200 }: +Command stdout: + +stderr: +Error from server (NotFound): pods "ss-2" not found + +error: +exit status 1 + +Jun 18 12:21:45.153: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-1300 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true' +Jun 18 12:21:45.282: INFO: rc: 1 +Jun 18 12:21:45.282: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-1300 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] [] Error from server (NotFound): pods "ss-2" not found + [] 0xc002b92300 exit status 1 true [0xc00088a018 0xc00088a2a8 0xc00088a3b8] [0xc00088a018 0xc00088a2a8 0xc00088a3b8] [0xc00088a220 0xc00088a328] [0x9c00a0 0x9c00a0] 0xc002bd22a0 }: +Command stdout: + +stderr: +Error from server (NotFound): pods "ss-2" not found -• [SLOW TEST:8.861 seconds] -[sig-storage] Projected downwardAPI -/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:33 - should provide container's memory limit [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +error: +exit status 1 + +Jun 18 12:21:55.282: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-1300 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true' +Jun 18 12:21:55.391: INFO: rc: 1 +Jun 18 12:21:55.391: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-1300 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] [] Error from server (NotFound): pods "ss-2" not found + [] 0xc002782b40 exit status 1 true [0xc00018dcc0 0xc00018dd78 0xc00018ddc0] [0xc00018dcc0 0xc00018dd78 0xc00018ddc0] [0xc00018dd58 0xc00018dda8] [0x9c00a0 0x9c00a0] 0xc00294e660 }: +Command stdout: + +stderr: +Error from server (NotFound): pods "ss-2" not found + +error: +exit status 1 + +Jun 18 12:22:05.392: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-1300 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true' +Jun 18 12:22:05.513: INFO: rc: 1 +Jun 18 12:22:05.513: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-1300 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] [] Error from server (NotFound): pods "ss-2" not found + [] 0xc002782ea0 exit status 1 true [0xc00018ddf8 0xc00018de90 0xc00018df00] [0xc00018ddf8 0xc00018de90 0xc00018df00] [0xc00018de60 0xc00018deb0] [0x9c00a0 0x9c00a0] 0xc00294ea80 }: +Command stdout: + +stderr: +Error from server (NotFound): pods "ss-2" not found + +error: +exit status 1 + +Jun 18 12:22:15.514: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-1300 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true' +Jun 18 12:22:15.659: INFO: rc: 1 +Jun 18 12:22:15.659: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-1300 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] [] Error from server (NotFound): pods "ss-2" not found + [] 0xc002783200 exit status 1 true [0xc00018df48 0xc00018dfe0 0xc001a3e020] [0xc00018df48 0xc00018dfe0 0xc001a3e020] [0xc00018dfd8 0xc001a3e000] [0x9c00a0 0x9c00a0] 0xc00294ede0 }: +Command stdout: + +stderr: +Error from server (NotFound): pods "ss-2" not found + +error: +exit status 1 + +Jun 18 12:22:25.660: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-1300 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true' +Jun 18 12:22:25.772: INFO: rc: 1 +Jun 18 12:22:25.773: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-1300 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] [] Error from server (NotFound): pods "ss-2" not found + [] 0xc001b32300 exit status 1 true [0xc00018d588 0xc00018d688 0xc00018d770] [0xc00018d588 0xc00018d688 0xc00018d770] [0xc00018d678 0xc00018d730] [0x9c00a0 0x9c00a0] 0xc000abab40 }: +Command stdout: + +stderr: +Error from server (NotFound): pods "ss-2" not found + +error: +exit status 1 + +Jun 18 12:22:35.773: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-1300 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true' +Jun 18 12:22:35.890: INFO: rc: 1 +Jun 18 12:22:35.890: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-1300 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] [] Error from server (NotFound): pods "ss-2" not found + [] 0xc001e5e300 exit status 1 true [0xc0000cc020 0xc0000cc4a8 0xc00088a220] [0xc0000cc020 0xc0000cc4a8 0xc00088a220] [0xc0000cc3e8 0xc00088a170] [0x9c00a0 0x9c00a0] 0xc002478ea0 }: +Command stdout: + +stderr: +Error from server (NotFound): pods "ss-2" not found + +error: +exit status 1 + +Jun 18 12:22:45.890: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-1300 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true' +Jun 18 12:22:46.011: INFO: rc: 1 +Jun 18 12:22:46.011: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-1300 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] [] Error from server (NotFound): pods "ss-2" not found + [] 0xc002f86300 exit status 1 true [0xc001a3e000 0xc001a3e040 0xc001a3e078] [0xc001a3e000 0xc001a3e040 0xc001a3e078] [0xc001a3e038 0xc001a3e058] [0x9c00a0 0x9c00a0] 0xc002bd22a0 }: +Command stdout: + +stderr: +Error from server (NotFound): pods "ss-2" not found + +error: +exit status 1 + +Jun 18 12:22:56.011: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-1300 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true' +Jun 18 12:22:56.952: INFO: rc: 1 +Jun 18 12:22:56.952: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-1300 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] [] Error from server (NotFound): pods "ss-2" not found + [] 0xc002f86660 exit status 1 true [0xc001a3e088 0xc001a3e0f8 0xc001a3e158] [0xc001a3e088 0xc001a3e0f8 0xc001a3e158] [0xc001a3e0e8 0xc001a3e148] [0x9c00a0 0x9c00a0] 0xc002bd2600 }: +Command stdout: + +stderr: +Error from server (NotFound): pods "ss-2" not found + +error: +exit status 1 + +Jun 18 12:23:06.952: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-1300 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true' +Jun 18 12:23:07.068: INFO: rc: 1 +Jun 18 12:23:07.068: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-1300 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] [] Error from server (NotFound): pods "ss-2" not found + [] 0xc001e5e660 exit status 1 true [0xc00088a2a8 0xc00088a3b8 0xc00088a570] [0xc00088a2a8 0xc00088a3b8 0xc00088a570] [0xc00088a328 0xc00088a4e8] [0x9c00a0 0x9c00a0] 0xc002479740 }: +Command stdout: + +stderr: +Error from server (NotFound): pods "ss-2" not found + +error: +exit status 1 + +Jun 18 12:23:17.068: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-1300 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true' +Jun 18 12:23:17.193: INFO: rc: 1 +Jun 18 12:23:17.193: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-1300 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] [] Error from server (NotFound): pods "ss-2" not found + [] 0xc001b32660 exit status 1 true [0xc00018d7a0 0xc00018d800 0xc00018d9c0] [0xc00018d7a0 0xc00018d800 0xc00018d9c0] [0xc00018d7d8 0xc00018d930] [0x9c00a0 0x9c00a0] 0xc000abb620 }: +Command stdout: + +stderr: +Error from server (NotFound): pods "ss-2" not found + +error: +exit status 1 + +Jun 18 12:23:27.193: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-1300 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true' +Jun 18 12:23:27.309: INFO: rc: 1 +Jun 18 12:23:27.309: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-1300 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] [] Error from server (NotFound): pods "ss-2" not found + [] 0xc001b329c0 exit status 1 true [0xc00018d9e0 0xc00018db48 0xc00018dba8] [0xc00018d9e0 0xc00018db48 0xc00018dba8] [0xc00018daa0 0xc00018db90] [0x9c00a0 0x9c00a0] 0xc000abbc80 }: +Command stdout: + +stderr: +Error from server (NotFound): pods "ss-2" not found + +error: +exit status 1 + +Jun 18 12:23:37.309: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-1300 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true' +Jun 18 12:23:37.428: INFO: rc: 1 +Jun 18 12:23:37.428: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-1300 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] [] Error from server (NotFound): pods "ss-2" not found + [] 0xc001b33350 exit status 1 true [0xc00018dbc0 0xc00018dc48 0xc00018dca0] [0xc00018dbc0 0xc00018dc48 0xc00018dca0] [0xc00018dc00 0xc00018dc88] [0x9c00a0 0x9c00a0] 0xc00294e300 }: +Command stdout: + +stderr: +Error from server (NotFound): pods "ss-2" not found + +error: +exit status 1 + +Jun 18 12:23:47.428: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-1300 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true' +Jun 18 12:23:48.112: INFO: rc: 1 +Jun 18 12:23:48.112: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-1300 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] [] Error from server (NotFound): pods "ss-2" not found + [] 0xc002f869f0 exit status 1 true [0xc001a3e180 0xc001a3e1e0 0xc001a3e220] [0xc001a3e180 0xc001a3e1e0 0xc001a3e220] [0xc001a3e1c8 0xc001a3e1f0] [0x9c00a0 0x9c00a0] 0xc002bd29c0 }: +Command stdout: + +stderr: +Error from server (NotFound): pods "ss-2" not found + +error: +exit status 1 + +Jun 18 12:23:58.112: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-1300 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true' +Jun 18 12:23:58.257: INFO: rc: 1 +Jun 18 12:23:58.257: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-1300 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] [] Error from server (NotFound): pods "ss-2" not found + [] 0xc002f86d20 exit status 1 true [0xc001a3e248 0xc001a3e288 0xc001a3e2f0] [0xc001a3e248 0xc001a3e288 0xc001a3e2f0] [0xc001a3e278 0xc001a3e2b8] [0x9c00a0 0x9c00a0] 0xc002bd2de0 }: +Command stdout: + +stderr: +Error from server (NotFound): pods "ss-2" not found + +error: +exit status 1 + +Jun 18 12:24:08.257: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-1300 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true' +Jun 18 12:24:08.399: INFO: rc: 1 +Jun 18 12:24:08.399: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-1300 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] [] Error from server (NotFound): pods "ss-2" not found + [] 0xc002f87050 exit status 1 true [0xc001a3e308 0xc001a3e338 0xc001a3e368] [0xc001a3e308 0xc001a3e338 0xc001a3e368] [0xc001a3e330 0xc001a3e358] [0x9c00a0 0x9c00a0] 0xc002bd3200 }: +Command stdout: + +stderr: +Error from server (NotFound): pods "ss-2" not found + +error: +exit status 1 + +Jun 18 12:24:18.400: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-1300 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true' +Jun 18 12:24:18.532: INFO: rc: 1 +Jun 18 12:24:18.532: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-1300 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] [] Error from server (NotFound): pods "ss-2" not found + [] 0xc001e5e9f0 exit status 1 true [0xc00088a580 0xc00088a658 0xc00088a748] [0xc00088a580 0xc00088a658 0xc00088a748] [0xc00088a640 0xc00088a6f8] [0x9c00a0 0x9c00a0] 0xc002479b60 }: +Command stdout: + +stderr: +Error from server (NotFound): pods "ss-2" not found + +error: +exit status 1 + +Jun 18 12:24:28.532: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-1300 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true' +Jun 18 12:24:28.644: INFO: rc: 1 +Jun 18 12:24:28.644: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-1300 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] [] Error from server (NotFound): pods "ss-2" not found + [] 0xc001b32330 exit status 1 true [0xc0000cc0b0 0xc00018d538 0xc00018d678] [0xc0000cc0b0 0xc00018d538 0xc00018d678] [0xc0000cc4a8 0xc00018d660] [0x9c00a0 0x9c00a0] 0xc000abab40 }: +Command stdout: + +stderr: +Error from server (NotFound): pods "ss-2" not found + +error: +exit status 1 + +Jun 18 12:24:38.645: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-1300 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true' +Jun 18 12:24:38.780: INFO: rc: 1 +Jun 18 12:24:38.781: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-1300 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] [] Error from server (NotFound): pods "ss-2" not found + [] 0xc001e5e360 exit status 1 true [0xc00088a018 0xc00088a2a8 0xc00088a3b8] [0xc00088a018 0xc00088a2a8 0xc00088a3b8] [0xc00088a220 0xc00088a328] [0x9c00a0 0x9c00a0] 0xc00294e480 }: +Command stdout: + +stderr: +Error from server (NotFound): pods "ss-2" not found + +error: +exit status 1 + +Jun 18 12:24:48.781: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-1300 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true' +Jun 18 12:24:48.899: INFO: rc: 1 +Jun 18 12:24:48.899: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-1300 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] [] Error from server (NotFound): pods "ss-2" not found + [] 0xc002b92300 exit status 1 true [0xc001a3e000 0xc001a3e040 0xc001a3e078] [0xc001a3e000 0xc001a3e040 0xc001a3e078] [0xc001a3e038 0xc001a3e058] [0x9c00a0 0x9c00a0] 0xc002478ea0 }: +Command stdout: + +stderr: +Error from server (NotFound): pods "ss-2" not found + +error: +exit status 1 + +Jun 18 12:24:58.899: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-1300 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true' +Jun 18 12:24:59.010: INFO: rc: 1 +Jun 18 12:24:59.010: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-1300 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] [] Error from server (NotFound): pods "ss-2" not found + [] 0xc001e5e6f0 exit status 1 true [0xc00088a438 0xc00088a580 0xc00088a658] [0xc00088a438 0xc00088a580 0xc00088a658] [0xc00088a570 0xc00088a640] [0x9c00a0 0x9c00a0] 0xc00294e840 }: +Command stdout: + +stderr: +Error from server (NotFound): pods "ss-2" not found + +error: +exit status 1 + +Jun 18 12:25:09.010: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-1300 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true' +Jun 18 12:25:09.126: INFO: rc: 1 +Jun 18 12:25:09.126: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-1300 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] [] Error from server (NotFound): pods "ss-2" not found + [] 0xc001b326c0 exit status 1 true [0xc00018d688 0xc00018d770 0xc00018d7d8] [0xc00018d688 0xc00018d770 0xc00018d7d8] [0xc00018d730 0xc00018d7c8] [0x9c00a0 0x9c00a0] 0xc000abb620 }: +Command stdout: + +stderr: +Error from server (NotFound): pods "ss-2" not found + +error: +exit status 1 + +Jun 18 12:25:19.127: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-1300 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true' +Jun 18 12:25:19.247: INFO: rc: 1 +Jun 18 12:25:19.248: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-1300 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] [] Error from server (NotFound): pods "ss-2" not found + [] 0xc001e5ea80 exit status 1 true [0xc00088a6b8 0xc00088a7b8 0xc00088a818] [0xc00088a6b8 0xc00088a7b8 0xc00088a818] [0xc00088a748 0xc00088a808] [0x9c00a0 0x9c00a0] 0xc00294ec00 }: +Command stdout: + +stderr: +Error from server (NotFound): pods "ss-2" not found + +error: +exit status 1 + +Jun 18 12:25:29.248: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-1300 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true' +Jun 18 12:25:29.373: INFO: rc: 1 +Jun 18 12:25:29.373: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-1300 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] [] Error from server (NotFound): pods "ss-2" not found + [] 0xc002b92630 exit status 1 true [0xc001a3e088 0xc001a3e0f8 0xc001a3e158] [0xc001a3e088 0xc001a3e0f8 0xc001a3e158] [0xc001a3e0e8 0xc001a3e148] [0x9c00a0 0x9c00a0] 0xc002479740 }: +Command stdout: + +stderr: +Error from server (NotFound): pods "ss-2" not found + +error: +exit status 1 + +Jun 18 12:25:39.373: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-1300 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true' +Jun 18 12:25:39.507: INFO: rc: 1 +Jun 18 12:25:39.507: INFO: stdout of mv -v /tmp/index.html /usr/share/nginx/html/ || true on ss-2: +Jun 18 12:25:39.507: INFO: Scaling statefulset ss to 0 +STEP: Verifying that stateful set ss was scaled down in reverse order +[AfterEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:85 +Jun 18 12:25:39.546: INFO: Deleting all statefulset in ns statefulset-1300 +Jun 18 12:25:39.558: INFO: Scaling statefulset ss to 0 +Jun 18 12:25:39.602: INFO: Waiting for statefulset status.replicas updated to 0 +Jun 18 12:25:39.613: INFO: Deleting statefulset ss +[AfterEach] [sig-apps] StatefulSet + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Jun 18 12:25:39.664: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "statefulset-1300" for this suite. +Jun 18 12:25:47.706: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:25:48.150: INFO: namespace statefulset-1300 deletion completed in 8.469690612s + +• [SLOW TEST:378.404 seconds] +[sig-apps] StatefulSet +/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22 + [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687 + Scaling should happen in predictable order and halt if any stateful pod is unhealthy [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 ------------------------------ -SSSSSSSSSSSSSSSSSSSSS +SSSSSSS ------------------------------ -[sig-storage] HostPath - should give a volume the correct mode [LinuxOnly] [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -[BeforeEach] [sig-storage] HostPath - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +[sig-network] DNS + should provide DNS for the cluster [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [sig-network] DNS + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 STEP: Creating a kubernetes client -Jun 4 19:15:09.588: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504 -STEP: Building a namespace api object, basename hostpath -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in hostpath-542 +Jun 18 12:25:48.151: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150 +STEP: Building a namespace api object, basename dns +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in dns-9235 STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [sig-storage] HostPath - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/host_path.go:37 -[It] should give a volume the correct mode [LinuxOnly] [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -STEP: Creating a pod to test hostPath mode -Jun 4 19:15:09.823: INFO: Waiting up to 5m0s for pod "pod-host-path-test" in namespace "hostpath-542" to be "success or failure" -Jun 4 19:15:09.839: INFO: Pod "pod-host-path-test": Phase="Pending", Reason="", readiness=false. Elapsed: 16.15424ms -Jun 4 19:15:11.854: INFO: Pod "pod-host-path-test": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.031049472s -STEP: Saw pod success -Jun 4 19:15:11.854: INFO: Pod "pod-host-path-test" satisfied condition "success or failure" -Jun 4 19:15:11.866: INFO: Trying to get logs from node 10.212.23.189 pod pod-host-path-test container test-container-1: -STEP: delete the pod -Jun 4 19:15:11.928: INFO: Waiting for pod pod-host-path-test to disappear -Jun 4 19:15:11.941: INFO: Pod pod-host-path-test no longer exists -[AfterEach] [sig-storage] HostPath - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 -Jun 4 19:15:11.941: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "hostpath-542" for this suite. -Jun 4 19:15:18.004: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -Jun 4 19:15:18.386: INFO: namespace hostpath-542 deletion completed in 6.426474297s +[It] should provide DNS for the cluster [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: Running these commands on wheezy: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search kubernetes.default.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_udp@kubernetes.default.svc.cluster.local;check="$$(dig +tcp +noall +answer +search kubernetes.default.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@kubernetes.default.svc.cluster.local;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-9235.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@PodARecord;sleep 1; done -• [SLOW TEST:8.798 seconds] -[sig-storage] HostPath -/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/host_path.go:34 - should give a volume the correct mode [LinuxOnly] [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: Running these commands on jessie: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search kubernetes.default.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_udp@kubernetes.default.svc.cluster.local;check="$$(dig +tcp +noall +answer +search kubernetes.default.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_tcp@kubernetes.default.svc.cluster.local;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-9235.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_tcp@PodARecord;sleep 1; done + +STEP: creating a pod to probe DNS +STEP: submitting the pod to kubernetes +STEP: retrieving the pod +STEP: looking for the results for each expected name from probers +Jun 18 12:26:04.590: INFO: DNS probes using dns-9235/dns-test-33fec811-91c4-11e9-a25d-8608290c688a succeeded + +STEP: deleting the pod +[AfterEach] [sig-network] DNS + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Jun 18 12:26:04.647: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "dns-9235" for this suite. +Jun 18 12:26:13.120: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:26:13.518: INFO: namespace dns-9235 deletion completed in 8.423762596s + +• [SLOW TEST:25.367 seconds] +[sig-network] DNS +/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:22 + should provide DNS for the cluster [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSS +SSSSSSSSS ------------------------------ -[sig-storage] Subpath Atomic writer volumes - should support subpaths with projected pod [LinuxOnly] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -[BeforeEach] [sig-storage] Subpath - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +[k8s.io] Docker Containers + should be able to override the image's default command (docker entrypoint) [NodeConformance] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [k8s.io] Docker Containers + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 STEP: Creating a kubernetes client -Jun 4 19:15:18.386: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504 -STEP: Building a namespace api object, basename subpath -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in subpath-4808 +Jun 18 12:26:13.518: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150 +STEP: Building a namespace api object, basename containers +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in containers-445 STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] Atomic writer volumes - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:38 -STEP: Setting up data -[It] should support subpaths with projected pod [LinuxOnly] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -STEP: Creating pod pod-subpath-test-projected-gzfx -STEP: Creating a pod to test atomic-volume-subpath -Jun 4 19:15:18.656: INFO: Waiting up to 5m0s for pod "pod-subpath-test-projected-gzfx" in namespace "subpath-4808" to be "success or failure" -Jun 4 19:15:18.667: INFO: Pod "pod-subpath-test-projected-gzfx": Phase="Pending", Reason="", readiness=false. Elapsed: 11.025599ms -Jun 4 19:15:20.679: INFO: Pod "pod-subpath-test-projected-gzfx": Phase="Running", Reason="", readiness=true. Elapsed: 2.022962001s -Jun 4 19:15:22.691: INFO: Pod "pod-subpath-test-projected-gzfx": Phase="Running", Reason="", readiness=true. Elapsed: 4.035053693s -Jun 4 19:15:24.706: INFO: Pod "pod-subpath-test-projected-gzfx": Phase="Running", Reason="", readiness=true. Elapsed: 6.050492237s -Jun 4 19:15:26.719: INFO: Pod "pod-subpath-test-projected-gzfx": Phase="Running", Reason="", readiness=true. Elapsed: 8.062911524s -Jun 4 19:15:28.731: INFO: Pod "pod-subpath-test-projected-gzfx": Phase="Running", Reason="", readiness=true. Elapsed: 10.075373238s -Jun 4 19:15:30.745: INFO: Pod "pod-subpath-test-projected-gzfx": Phase="Running", Reason="", readiness=true. Elapsed: 12.089425007s -Jun 4 19:15:32.757: INFO: Pod "pod-subpath-test-projected-gzfx": Phase="Running", Reason="", readiness=true. Elapsed: 14.101333443s -Jun 4 19:15:34.770: INFO: Pod "pod-subpath-test-projected-gzfx": Phase="Running", Reason="", readiness=true. Elapsed: 16.113779568s -Jun 4 19:15:36.785: INFO: Pod "pod-subpath-test-projected-gzfx": Phase="Running", Reason="", readiness=true. Elapsed: 18.12938965s -Jun 4 19:15:38.798: INFO: Pod "pod-subpath-test-projected-gzfx": Phase="Running", Reason="", readiness=true. Elapsed: 20.142262681s -Jun 4 19:15:40.812: INFO: Pod "pod-subpath-test-projected-gzfx": Phase="Succeeded", Reason="", readiness=false. Elapsed: 22.156284075s +[It] should be able to override the image's default command (docker entrypoint) [NodeConformance] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: Creating a pod to test override command +Jun 18 12:26:13.737: INFO: Waiting up to 5m0s for pod "client-containers-431d6d16-91c4-11e9-a25d-8608290c688a" in namespace "containers-445" to be "success or failure" +Jun 18 12:26:13.753: INFO: Pod "client-containers-431d6d16-91c4-11e9-a25d-8608290c688a": Phase="Pending", Reason="", readiness=false. Elapsed: 15.505708ms +Jun 18 12:26:15.766: INFO: Pod "client-containers-431d6d16-91c4-11e9-a25d-8608290c688a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.029075774s STEP: Saw pod success -Jun 4 19:15:40.812: INFO: Pod "pod-subpath-test-projected-gzfx" satisfied condition "success or failure" -Jun 4 19:15:40.823: INFO: Trying to get logs from node 10.212.23.189 pod pod-subpath-test-projected-gzfx container test-container-subpath-projected-gzfx: +Jun 18 12:26:15.766: INFO: Pod "client-containers-431d6d16-91c4-11e9-a25d-8608290c688a" satisfied condition "success or failure" +Jun 18 12:26:15.778: INFO: Trying to get logs from node 10.72.74.138 pod client-containers-431d6d16-91c4-11e9-a25d-8608290c688a container test-container: STEP: delete the pod -Jun 4 19:15:40.886: INFO: Waiting for pod pod-subpath-test-projected-gzfx to disappear -Jun 4 19:15:40.897: INFO: Pod pod-subpath-test-projected-gzfx no longer exists -STEP: Deleting pod pod-subpath-test-projected-gzfx -Jun 4 19:15:40.897: INFO: Deleting pod "pod-subpath-test-projected-gzfx" in namespace "subpath-4808" -[AfterEach] [sig-storage] Subpath - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 -Jun 4 19:15:40.908: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "subpath-4808" for this suite. -Jun 4 19:15:46.969: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -Jun 4 19:15:47.470: INFO: namespace subpath-4808 deletion completed in 6.54628627s +Jun 18 12:26:16.010: INFO: Waiting for pod client-containers-431d6d16-91c4-11e9-a25d-8608290c688a to disappear +Jun 18 12:26:16.022: INFO: Pod client-containers-431d6d16-91c4-11e9-a25d-8608290c688a no longer exists +[AfterEach] [k8s.io] Docker Containers + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Jun 18 12:26:16.022: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "containers-445" for this suite. +Jun 18 12:26:24.069: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:26:24.486: INFO: namespace containers-445 deletion completed in 8.449220258s -• [SLOW TEST:29.085 seconds] -[sig-storage] Subpath -/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:22 - Atomic writer volumes - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:34 - should support subpaths with projected pod [LinuxOnly] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +• [SLOW TEST:10.968 seconds] +[k8s.io] Docker Containers +/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687 + should be able to override the image's default command (docker entrypoint) [NodeConformance] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 ------------------------------ -SSSSSSSSSSSSSSSSS +SSSSSS ------------------------------ -[sig-storage] ConfigMap - should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -[BeforeEach] [sig-storage] ConfigMap - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +[sig-node] ConfigMap + should be consumable via the environment [NodeConformance] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [sig-node] ConfigMap + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 STEP: Creating a kubernetes client -Jun 4 19:15:47.470: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504 +Jun 18 12:26:24.486: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150 STEP: Building a namespace api object, basename configmap -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in configmap-5947 +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in configmap-6138 STEP: Waiting for a default service account to be provisioned in namespace -[It] should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -STEP: Creating configMap with name configmap-test-volume-288fc425-86fd-11e9-8d1b-467ee19922ac +[It] should be consumable via the environment [NodeConformance] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: Creating configMap configmap-6138/configmap-test-49a8115f-91c4-11e9-a25d-8608290c688a STEP: Creating a pod to test consume configMaps -Jun 4 19:15:47.720: INFO: Waiting up to 5m0s for pod "pod-configmaps-2891980e-86fd-11e9-8d1b-467ee19922ac" in namespace "configmap-5947" to be "success or failure" -Jun 4 19:15:47.755: INFO: Pod "pod-configmaps-2891980e-86fd-11e9-8d1b-467ee19922ac": Phase="Pending", Reason="", readiness=false. Elapsed: 35.186391ms -Jun 4 19:15:49.770: INFO: Pod "pod-configmaps-2891980e-86fd-11e9-8d1b-467ee19922ac": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.049749767s +Jun 18 12:26:24.727: INFO: Waiting up to 5m0s for pod "pod-configmaps-49aa315e-91c4-11e9-a25d-8608290c688a" in namespace "configmap-6138" to be "success or failure" +Jun 18 12:26:24.741: INFO: Pod "pod-configmaps-49aa315e-91c4-11e9-a25d-8608290c688a": Phase="Pending", Reason="", readiness=false. Elapsed: 13.217915ms +Jun 18 12:26:26.757: INFO: Pod "pod-configmaps-49aa315e-91c4-11e9-a25d-8608290c688a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.029343175s STEP: Saw pod success -Jun 4 19:15:49.770: INFO: Pod "pod-configmaps-2891980e-86fd-11e9-8d1b-467ee19922ac" satisfied condition "success or failure" -Jun 4 19:15:49.780: INFO: Trying to get logs from node 10.212.23.164 pod pod-configmaps-2891980e-86fd-11e9-8d1b-467ee19922ac container configmap-volume-test: +Jun 18 12:26:26.757: INFO: Pod "pod-configmaps-49aa315e-91c4-11e9-a25d-8608290c688a" satisfied condition "success or failure" +Jun 18 12:26:26.770: INFO: Trying to get logs from node 10.72.74.184 pod pod-configmaps-49aa315e-91c4-11e9-a25d-8608290c688a container env-test: STEP: delete the pod -Jun 4 19:15:49.842: INFO: Waiting for pod pod-configmaps-2891980e-86fd-11e9-8d1b-467ee19922ac to disappear -Jun 4 19:15:49.853: INFO: Pod pod-configmaps-2891980e-86fd-11e9-8d1b-467ee19922ac no longer exists -[AfterEach] [sig-storage] ConfigMap - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 -Jun 4 19:15:49.853: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "configmap-5947" for this suite. -Jun 4 19:15:55.909: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -Jun 4 19:15:56.257: INFO: namespace configmap-5947 deletion completed in 6.390343178s +Jun 18 12:26:27.267: INFO: Waiting for pod pod-configmaps-49aa315e-91c4-11e9-a25d-8608290c688a to disappear +Jun 18 12:26:27.284: INFO: Pod pod-configmaps-49aa315e-91c4-11e9-a25d-8608290c688a no longer exists +[AfterEach] [sig-node] ConfigMap + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Jun 18 12:26:27.284: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "configmap-6138" for this suite. +Jun 18 12:26:35.326: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:26:35.739: INFO: namespace configmap-6138 deletion completed in 8.440580061s -• [SLOW TEST:8.787 seconds] -[sig-storage] ConfigMap -/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:32 - should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +• [SLOW TEST:11.253 seconds] +[sig-node] ConfigMap +/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/configmap.go:32 + should be consumable via the environment [NodeConformance] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +SSSSSS ------------------------------ -[k8s.io] Variable Expansion - should allow composing env vars into new env vars [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -[BeforeEach] [k8s.io] Variable Expansion - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +[sig-storage] Subpath Atomic writer volumes + should support subpaths with configmap pod with mountPath of existing file [LinuxOnly] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [sig-storage] Subpath + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 STEP: Creating a kubernetes client -Jun 4 19:15:56.257: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504 -STEP: Building a namespace api object, basename var-expansion -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in var-expansion-3519 +Jun 18 12:26:35.739: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150 +STEP: Building a namespace api object, basename subpath +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in subpath-7191 STEP: Waiting for a default service account to be provisioned in namespace -[It] should allow composing env vars into new env vars [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -STEP: Creating a pod to test env composition -Jun 4 19:15:56.499: INFO: Waiting up to 5m0s for pod "var-expansion-2dcd3db9-86fd-11e9-8d1b-467ee19922ac" in namespace "var-expansion-3519" to be "success or failure" -Jun 4 19:15:56.510: INFO: Pod "var-expansion-2dcd3db9-86fd-11e9-8d1b-467ee19922ac": Phase="Pending", Reason="", readiness=false. Elapsed: 11.140209ms -Jun 4 19:15:58.523: INFO: Pod "var-expansion-2dcd3db9-86fd-11e9-8d1b-467ee19922ac": Phase="Running", Reason="", readiness=true. Elapsed: 2.023516374s -Jun 4 19:16:00.534: INFO: Pod "var-expansion-2dcd3db9-86fd-11e9-8d1b-467ee19922ac": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.035336595s +[BeforeEach] Atomic writer volumes + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:38 +STEP: Setting up data +[It] should support subpaths with configmap pod with mountPath of existing file [LinuxOnly] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: Creating pod pod-subpath-test-configmap-75hm +STEP: Creating a pod to test atomic-volume-subpath +Jun 18 12:26:35.986: INFO: Waiting up to 5m0s for pod "pod-subpath-test-configmap-75hm" in namespace "subpath-7191" to be "success or failure" +Jun 18 12:26:35.999: INFO: Pod "pod-subpath-test-configmap-75hm": Phase="Pending", Reason="", readiness=false. Elapsed: 13.213461ms +Jun 18 12:26:38.014: INFO: Pod "pod-subpath-test-configmap-75hm": Phase="Running", Reason="", readiness=true. Elapsed: 2.027956263s +Jun 18 12:26:40.029: INFO: Pod "pod-subpath-test-configmap-75hm": Phase="Running", Reason="", readiness=true. Elapsed: 4.04319998s +Jun 18 12:26:42.042: INFO: Pod "pod-subpath-test-configmap-75hm": Phase="Running", Reason="", readiness=true. Elapsed: 6.056375656s +Jun 18 12:26:44.056: INFO: Pod "pod-subpath-test-configmap-75hm": Phase="Running", Reason="", readiness=true. Elapsed: 8.07010449s +Jun 18 12:26:46.070: INFO: Pod "pod-subpath-test-configmap-75hm": Phase="Running", Reason="", readiness=true. Elapsed: 10.083944926s +Jun 18 12:26:48.084: INFO: Pod "pod-subpath-test-configmap-75hm": Phase="Running", Reason="", readiness=true. Elapsed: 12.097938824s +Jun 18 12:26:50.097: INFO: Pod "pod-subpath-test-configmap-75hm": Phase="Running", Reason="", readiness=true. Elapsed: 14.111546131s +Jun 18 12:26:52.111: INFO: Pod "pod-subpath-test-configmap-75hm": Phase="Running", Reason="", readiness=true. Elapsed: 16.125524104s +Jun 18 12:26:54.125: INFO: Pod "pod-subpath-test-configmap-75hm": Phase="Running", Reason="", readiness=true. Elapsed: 18.138755313s +Jun 18 12:26:56.140: INFO: Pod "pod-subpath-test-configmap-75hm": Phase="Running", Reason="", readiness=true. Elapsed: 20.154157138s +Jun 18 12:26:58.272: INFO: Pod "pod-subpath-test-configmap-75hm": Phase="Running", Reason="", readiness=true. Elapsed: 22.286564364s +Jun 18 12:27:00.287: INFO: Pod "pod-subpath-test-configmap-75hm": Phase="Succeeded", Reason="", readiness=false. Elapsed: 24.30082145s STEP: Saw pod success -Jun 4 19:16:00.534: INFO: Pod "var-expansion-2dcd3db9-86fd-11e9-8d1b-467ee19922ac" satisfied condition "success or failure" -Jun 4 19:16:00.545: INFO: Trying to get logs from node 10.212.23.189 pod var-expansion-2dcd3db9-86fd-11e9-8d1b-467ee19922ac container dapi-container: +Jun 18 12:27:00.287: INFO: Pod "pod-subpath-test-configmap-75hm" satisfied condition "success or failure" +Jun 18 12:27:00.299: INFO: Trying to get logs from node 10.72.74.138 pod pod-subpath-test-configmap-75hm container test-container-subpath-configmap-75hm: STEP: delete the pod -Jun 4 19:16:00.655: INFO: Waiting for pod var-expansion-2dcd3db9-86fd-11e9-8d1b-467ee19922ac to disappear -Jun 4 19:16:00.669: INFO: Pod var-expansion-2dcd3db9-86fd-11e9-8d1b-467ee19922ac no longer exists -[AfterEach] [k8s.io] Variable Expansion - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 -Jun 4 19:16:00.669: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "var-expansion-3519" for this suite. -Jun 4 19:16:06.728: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -Jun 4 19:16:07.082: INFO: namespace var-expansion-3519 deletion completed in 6.396686406s +Jun 18 12:27:00.374: INFO: Waiting for pod pod-subpath-test-configmap-75hm to disappear +Jun 18 12:27:00.389: INFO: Pod pod-subpath-test-configmap-75hm no longer exists +STEP: Deleting pod pod-subpath-test-configmap-75hm +Jun 18 12:27:00.389: INFO: Deleting pod "pod-subpath-test-configmap-75hm" in namespace "subpath-7191" +[AfterEach] [sig-storage] Subpath + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Jun 18 12:27:00.405: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "subpath-7191" for this suite. +Jun 18 12:27:08.651: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:27:09.047: INFO: namespace subpath-7191 deletion completed in 8.42174543s -• [SLOW TEST:10.825 seconds] -[k8s.io] Variable Expansion -/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687 - should allow composing env vars into new env vars [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +• [SLOW TEST:33.308 seconds] +[sig-storage] Subpath +/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:22 + Atomic writer volumes + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:34 + should support subpaths with configmap pod with mountPath of existing file [LinuxOnly] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Projected downwardAPI should update labels on modification [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 [BeforeEach] [sig-storage] Projected downwardAPI - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 STEP: Creating a kubernetes client -Jun 4 19:16:07.083: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504 +Jun 18 12:27:09.048: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150 STEP: Building a namespace api object, basename projected -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-2809 +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-8828 STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Projected downwardAPI - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:39 + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:39 [It] should update labels on modification [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 STEP: Creating the pod -Jun 4 19:16:09.988: INFO: Successfully updated pod "labelsupdate3440d7df-86fd-11e9-8d1b-467ee19922ac" +Jun 18 12:27:13.920: INFO: Successfully updated pod "labelsupdate64379ab5-91c4-11e9-a25d-8608290c688a" [AfterEach] [sig-storage] Projected downwardAPI - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 -Jun 4 19:16:14.083: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "projected-2809" for this suite. -Jun 4 19:16:36.141: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -Jun 4 19:16:36.470: INFO: namespace projected-2809 deletion completed in 22.372664963s + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Jun 18 12:27:15.987: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "projected-8828" for this suite. +Jun 18 12:27:40.042: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:27:40.544: INFO: namespace projected-8828 deletion completed in 24.536766698s -• [SLOW TEST:29.388 seconds] +• [SLOW TEST:31.496 seconds] [sig-storage] Projected downwardAPI -/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:33 +/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:33 should update labels on modification [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 ------------------------------- -SSSSSSSSSSSSSSSSSS ------------------------------- -[sig-api-machinery] Watchers - should observe an object deletion if it stops meeting the requirements of the selector [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -[BeforeEach] [sig-api-machinery] Watchers - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 -STEP: Creating a kubernetes client -Jun 4 19:16:36.471: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504 -STEP: Building a namespace api object, basename watch -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in watch-7660 -STEP: Waiting for a default service account to be provisioned in namespace -[It] should observe an object deletion if it stops meeting the requirements of the selector [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -STEP: creating a watch on configmaps with a certain label -STEP: creating a new configmap -STEP: modifying the configmap once -STEP: changing the label value of the configmap -STEP: Expecting to observe a delete notification for the watched object -Jun 4 19:16:36.896: INFO: Got : ADDED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-label-changed,GenerateName:,Namespace:watch-7660,SelfLink:/api/v1/namespaces/watch-7660/configmaps/e2e-watch-test-label-changed,UID:45dc2fcc-86fd-11e9-8318-1e5386706511,ResourceVersion:20120,Generation:0,CreationTimestamp:2019-06-04 19:16:36 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: label-changed-and-restored,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{},BinaryData:map[string][]byte{},} -Jun 4 19:16:36.896: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-label-changed,GenerateName:,Namespace:watch-7660,SelfLink:/api/v1/namespaces/watch-7660/configmaps/e2e-watch-test-label-changed,UID:45dc2fcc-86fd-11e9-8318-1e5386706511,ResourceVersion:20121,Generation:0,CreationTimestamp:2019-06-04 19:16:36 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: label-changed-and-restored,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{mutation: 1,},BinaryData:map[string][]byte{},} -Jun 4 19:16:36.896: INFO: Got : DELETED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-label-changed,GenerateName:,Namespace:watch-7660,SelfLink:/api/v1/namespaces/watch-7660/configmaps/e2e-watch-test-label-changed,UID:45dc2fcc-86fd-11e9-8318-1e5386706511,ResourceVersion:20122,Generation:0,CreationTimestamp:2019-06-04 19:16:36 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: label-changed-and-restored,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{mutation: 1,},BinaryData:map[string][]byte{},} -STEP: modifying the configmap a second time -STEP: Expecting not to observe a notification because the object no longer meets the selector's requirements -STEP: changing the label value of the configmap back -STEP: modifying the configmap a third time -STEP: deleting the configmap -STEP: Expecting to observe an add notification for the watched object when the label value was restored -Jun 4 19:16:47.080: INFO: Got : ADDED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-label-changed,GenerateName:,Namespace:watch-7660,SelfLink:/api/v1/namespaces/watch-7660/configmaps/e2e-watch-test-label-changed,UID:45dc2fcc-86fd-11e9-8318-1e5386706511,ResourceVersion:20141,Generation:0,CreationTimestamp:2019-06-04 19:16:36 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: label-changed-and-restored,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},} -Jun 4 19:16:47.080: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-label-changed,GenerateName:,Namespace:watch-7660,SelfLink:/api/v1/namespaces/watch-7660/configmaps/e2e-watch-test-label-changed,UID:45dc2fcc-86fd-11e9-8318-1e5386706511,ResourceVersion:20142,Generation:0,CreationTimestamp:2019-06-04 19:16:36 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: label-changed-and-restored,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{mutation: 3,},BinaryData:map[string][]byte{},} -Jun 4 19:16:47.080: INFO: Got : DELETED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-label-changed,GenerateName:,Namespace:watch-7660,SelfLink:/api/v1/namespaces/watch-7660/configmaps/e2e-watch-test-label-changed,UID:45dc2fcc-86fd-11e9-8318-1e5386706511,ResourceVersion:20143,Generation:0,CreationTimestamp:2019-06-04 19:16:36 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: label-changed-and-restored,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{mutation: 3,},BinaryData:map[string][]byte{},} -[AfterEach] [sig-api-machinery] Watchers - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 -Jun 4 19:16:47.080: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "watch-7660" for this suite. -Jun 4 19:16:53.142: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -Jun 4 19:16:53.496: INFO: namespace watch-7660 deletion completed in 6.399611132s - -• [SLOW TEST:17.026 seconds] -[sig-api-machinery] Watchers -/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:22 - should observe an object deletion if it stops meeting the requirements of the selector [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 ------------------------------ -SSSSSSSSSSSSSSSSSS +SSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Secrets - should be consumable from pods in volume with mappings and Item Mode set [LinuxOnly] [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 + should be consumable from pods in volume [NodeConformance] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 [BeforeEach] [sig-storage] Secrets - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 STEP: Creating a kubernetes client -Jun 4 19:16:53.498: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504 +Jun 18 12:27:40.547: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150 STEP: Building a namespace api object, basename secrets -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in secrets-7035 +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in secrets-645 STEP: Waiting for a default service account to be provisioned in namespace -[It] should be consumable from pods in volume with mappings and Item Mode set [LinuxOnly] [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -STEP: Creating secret with name secret-test-map-4feb2470-86fd-11e9-8d1b-467ee19922ac +[It] should be consumable from pods in volume [NodeConformance] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: Creating secret with name secret-test-76fd9dfb-91c4-11e9-a25d-8608290c688a STEP: Creating a pod to test consume secrets -Jun 4 19:16:53.744: INFO: Waiting up to 5m0s for pod "pod-secrets-4fec86ca-86fd-11e9-8d1b-467ee19922ac" in namespace "secrets-7035" to be "success or failure" -Jun 4 19:16:53.754: INFO: Pod "pod-secrets-4fec86ca-86fd-11e9-8d1b-467ee19922ac": Phase="Pending", Reason="", readiness=false. Elapsed: 10.396696ms -Jun 4 19:16:55.767: INFO: Pod "pod-secrets-4fec86ca-86fd-11e9-8d1b-467ee19922ac": Phase="Running", Reason="", readiness=true. Elapsed: 2.023577712s -Jun 4 19:16:57.788: INFO: Pod "pod-secrets-4fec86ca-86fd-11e9-8d1b-467ee19922ac": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.044182428s +Jun 18 12:27:40.782: INFO: Waiting up to 5m0s for pod "pod-secrets-76ff631b-91c4-11e9-a25d-8608290c688a" in namespace "secrets-645" to be "success or failure" +Jun 18 12:27:40.795: INFO: Pod "pod-secrets-76ff631b-91c4-11e9-a25d-8608290c688a": Phase="Pending", Reason="", readiness=false. Elapsed: 12.832439ms +Jun 18 12:27:42.810: INFO: Pod "pod-secrets-76ff631b-91c4-11e9-a25d-8608290c688a": Phase="Running", Reason="", readiness=true. Elapsed: 2.02744448s +Jun 18 12:27:44.823: INFO: Pod "pod-secrets-76ff631b-91c4-11e9-a25d-8608290c688a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.040595823s STEP: Saw pod success -Jun 4 19:16:57.788: INFO: Pod "pod-secrets-4fec86ca-86fd-11e9-8d1b-467ee19922ac" satisfied condition "success or failure" -Jun 4 19:16:57.806: INFO: Trying to get logs from node 10.212.23.164 pod pod-secrets-4fec86ca-86fd-11e9-8d1b-467ee19922ac container secret-volume-test: +Jun 18 12:27:44.823: INFO: Pod "pod-secrets-76ff631b-91c4-11e9-a25d-8608290c688a" satisfied condition "success or failure" +Jun 18 12:27:44.839: INFO: Trying to get logs from node 10.72.74.138 pod pod-secrets-76ff631b-91c4-11e9-a25d-8608290c688a container secret-volume-test: STEP: delete the pod -Jun 4 19:16:57.885: INFO: Waiting for pod pod-secrets-4fec86ca-86fd-11e9-8d1b-467ee19922ac to disappear -Jun 4 19:16:57.908: INFO: Pod pod-secrets-4fec86ca-86fd-11e9-8d1b-467ee19922ac no longer exists +Jun 18 12:27:44.916: INFO: Waiting for pod pod-secrets-76ff631b-91c4-11e9-a25d-8608290c688a to disappear +Jun 18 12:27:44.939: INFO: Pod pod-secrets-76ff631b-91c4-11e9-a25d-8608290c688a no longer exists [AfterEach] [sig-storage] Secrets - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 -Jun 4 19:16:57.908: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "secrets-7035" for this suite. -Jun 4 19:17:03.976: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -Jun 4 19:17:04.409: INFO: namespace secrets-7035 deletion completed in 6.475563568s + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Jun 18 12:27:44.939: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "secrets-645" for this suite. +Jun 18 12:27:52.980: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:27:53.799: INFO: namespace secrets-645 deletion completed in 8.846340805s -• [SLOW TEST:10.911 seconds] +• [SLOW TEST:13.253 seconds] [sig-storage] Secrets -/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/secrets_volume.go:33 - should be consumable from pods in volume with mappings and Item Mode set [LinuxOnly] [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/secrets_volume.go:33 + should be consumable from pods in volume [NodeConformance] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 ------------------------------ -SS +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-storage] EmptyDir volumes - volume on default medium should have the correct mode [LinuxOnly] [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -[BeforeEach] [sig-storage] EmptyDir volumes - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +[sig-api-machinery] Garbage collector + should keep the rc around until all its pods are deleted if the deleteOptions says so [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [sig-api-machinery] Garbage collector + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 STEP: Creating a kubernetes client -Jun 4 19:17:04.409: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504 -STEP: Building a namespace api object, basename emptydir -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in emptydir-7185 +Jun 18 12:27:53.800: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150 +STEP: Building a namespace api object, basename gc +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in gc-701 STEP: Waiting for a default service account to be provisioned in namespace -[It] volume on default medium should have the correct mode [LinuxOnly] [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -STEP: Creating a pod to test emptydir volume type on node default medium -Jun 4 19:17:04.647: INFO: Waiting up to 5m0s for pod "pod-566bb59a-86fd-11e9-8d1b-467ee19922ac" in namespace "emptydir-7185" to be "success or failure" -Jun 4 19:17:04.657: INFO: Pod "pod-566bb59a-86fd-11e9-8d1b-467ee19922ac": Phase="Pending", Reason="", readiness=false. Elapsed: 10.736432ms -Jun 4 19:17:06.670: INFO: Pod "pod-566bb59a-86fd-11e9-8d1b-467ee19922ac": Phase="Running", Reason="", readiness=true. Elapsed: 2.023331558s -Jun 4 19:17:08.685: INFO: Pod "pod-566bb59a-86fd-11e9-8d1b-467ee19922ac": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.038622708s -STEP: Saw pod success -Jun 4 19:17:08.685: INFO: Pod "pod-566bb59a-86fd-11e9-8d1b-467ee19922ac" satisfied condition "success or failure" -Jun 4 19:17:08.697: INFO: Trying to get logs from node 10.212.23.189 pod pod-566bb59a-86fd-11e9-8d1b-467ee19922ac container test-container: -STEP: delete the pod -Jun 4 19:17:08.763: INFO: Waiting for pod pod-566bb59a-86fd-11e9-8d1b-467ee19922ac to disappear -Jun 4 19:17:08.774: INFO: Pod pod-566bb59a-86fd-11e9-8d1b-467ee19922ac no longer exists -[AfterEach] [sig-storage] EmptyDir volumes - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 -Jun 4 19:17:08.774: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "emptydir-7185" for this suite. -Jun 4 19:17:14.833: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -Jun 4 19:17:15.167: INFO: namespace emptydir-7185 deletion completed in 6.375049217s +[It] should keep the rc around until all its pods are deleted if the deleteOptions says so [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: create the rc +STEP: delete the rc +STEP: wait for the rc to be deleted +STEP: Gathering metrics +W0618 12:28:00.114921 20 metrics_grabber.go:79] Master node is not registered. Grabbing metrics from Scheduler, ControllerManager and ClusterAutoscaler is disabled. +Jun 18 12:28:00.115: INFO: For apiserver_request_total: +For apiserver_request_latencies_summary: +For apiserver_init_events_total: +For garbage_collector_attempt_to_delete_queue_latency: +For garbage_collector_attempt_to_delete_work_duration: +For garbage_collector_attempt_to_orphan_queue_latency: +For garbage_collector_attempt_to_orphan_work_duration: +For garbage_collector_dirty_processing_latency_microseconds: +For garbage_collector_event_processing_latency_microseconds: +For garbage_collector_graph_changes_queue_latency: +For garbage_collector_graph_changes_work_duration: +For garbage_collector_orphan_processing_latency_microseconds: +For namespace_queue_latency: +For namespace_queue_latency_sum: +For namespace_queue_latency_count: +For namespace_retries: +For namespace_work_duration: +For namespace_work_duration_sum: +For namespace_work_duration_count: +For function_duration_seconds: +For errors_total: +For evicted_pods_total: -• [SLOW TEST:10.758 seconds] -[sig-storage] EmptyDir volumes -/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:41 - volume on default medium should have the correct mode [LinuxOnly] [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[AfterEach] [sig-api-machinery] Garbage collector + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Jun 18 12:28:00.115: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "gc-701" for this suite. +Jun 18 12:28:08.153: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:28:08.587: INFO: namespace gc-701 deletion completed in 8.460480078s + +• [SLOW TEST:14.787 seconds] +[sig-api-machinery] Garbage collector +/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:22 + should keep the rc around until all its pods are deleted if the deleteOptions says so [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 ------------------------------ -S +SSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[k8s.io] Variable Expansion - should allow substituting values in a container's command [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -[BeforeEach] [k8s.io] Variable Expansion - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +[sig-cli] Kubectl client [k8s.io] Kubectl run rc + should create an rc from an image [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [sig-cli] Kubectl client + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 STEP: Creating a kubernetes client -Jun 4 19:17:15.167: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504 -STEP: Building a namespace api object, basename var-expansion -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in var-expansion-7803 +Jun 18 12:28:08.590: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150 +STEP: Building a namespace api object, basename kubectl +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-529 STEP: Waiting for a default service account to be provisioned in namespace -[It] should allow substituting values in a container's command [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -STEP: Creating a pod to test substitution in container's command -Jun 4 19:17:15.411: INFO: Waiting up to 5m0s for pod "var-expansion-5cd52886-86fd-11e9-8d1b-467ee19922ac" in namespace "var-expansion-7803" to be "success or failure" -Jun 4 19:17:15.422: INFO: Pod "var-expansion-5cd52886-86fd-11e9-8d1b-467ee19922ac": Phase="Pending", Reason="", readiness=false. Elapsed: 10.767422ms -Jun 4 19:17:17.436: INFO: Pod "var-expansion-5cd52886-86fd-11e9-8d1b-467ee19922ac": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.024946374s -STEP: Saw pod success -Jun 4 19:17:17.436: INFO: Pod "var-expansion-5cd52886-86fd-11e9-8d1b-467ee19922ac" satisfied condition "success or failure" -Jun 4 19:17:17.447: INFO: Trying to get logs from node 10.212.23.161 pod var-expansion-5cd52886-86fd-11e9-8d1b-467ee19922ac container dapi-container: -STEP: delete the pod -Jun 4 19:17:17.520: INFO: Waiting for pod var-expansion-5cd52886-86fd-11e9-8d1b-467ee19922ac to disappear -Jun 4 19:17:17.531: INFO: Pod var-expansion-5cd52886-86fd-11e9-8d1b-467ee19922ac no longer exists -[AfterEach] [k8s.io] Variable Expansion - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 -Jun 4 19:17:17.531: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "var-expansion-7803" for this suite. -Jun 4 19:17:23.588: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -Jun 4 19:17:23.977: INFO: namespace var-expansion-7803 deletion completed in 6.430397186s +[BeforeEach] [sig-cli] Kubectl client + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:213 +[BeforeEach] [k8s.io] Kubectl run rc + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1354 +[It] should create an rc from an image [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: running the image docker.io/library/nginx:1.14-alpine +Jun 18 12:28:08.796: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 run e2e-test-nginx-rc --image=docker.io/library/nginx:1.14-alpine --generator=run/v1 --namespace=kubectl-529' +Jun 18 12:28:09.049: INFO: stderr: "kubectl run --generator=run/v1 is DEPRECATED and will be removed in a future version. Use kubectl run --generator=run-pod/v1 or kubectl create instead.\n" +Jun 18 12:28:09.049: INFO: stdout: "replicationcontroller/e2e-test-nginx-rc created\n" +STEP: verifying the rc e2e-test-nginx-rc was created +STEP: verifying the pod controlled by rc e2e-test-nginx-rc was created +STEP: confirm that you can get logs from an rc +Jun 18 12:28:09.087: INFO: Waiting up to 5m0s for 1 pods to be running and ready: [e2e-test-nginx-rc-f47w4] +Jun 18 12:28:09.087: INFO: Waiting up to 5m0s for pod "e2e-test-nginx-rc-f47w4" in namespace "kubectl-529" to be "running and ready" +Jun 18 12:28:09.101: INFO: Pod "e2e-test-nginx-rc-f47w4": Phase="Pending", Reason="", readiness=false. Elapsed: 14.434911ms +Jun 18 12:28:11.115: INFO: Pod "e2e-test-nginx-rc-f47w4": Phase="Pending", Reason="", readiness=false. Elapsed: 2.028632869s +Jun 18 12:28:13.128: INFO: Pod "e2e-test-nginx-rc-f47w4": Phase="Running", Reason="", readiness=true. Elapsed: 4.04162889s +Jun 18 12:28:13.128: INFO: Pod "e2e-test-nginx-rc-f47w4" satisfied condition "running and ready" +Jun 18 12:28:13.129: INFO: Wanted all 1 pods to be running and ready. Result: true. Pods: [e2e-test-nginx-rc-f47w4] +Jun 18 12:28:13.129: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 logs rc/e2e-test-nginx-rc --namespace=kubectl-529' +Jun 18 12:28:13.519: INFO: stderr: "" +Jun 18 12:28:13.519: INFO: stdout: "" +[AfterEach] [k8s.io] Kubectl run rc + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1359 +Jun 18 12:28:13.519: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 delete rc e2e-test-nginx-rc --namespace=kubectl-529' +Jun 18 12:28:13.703: INFO: stderr: "" +Jun 18 12:28:13.703: INFO: stdout: "replicationcontroller \"e2e-test-nginx-rc\" deleted\n" +[AfterEach] [sig-cli] Kubectl client + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Jun 18 12:28:13.703: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "kubectl-529" for this suite. +Jun 18 12:28:37.987: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:28:38.393: INFO: namespace kubectl-529 deletion completed in 24.443082836s -• [SLOW TEST:8.810 seconds] -[k8s.io] Variable Expansion -/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687 - should allow substituting values in a container's command [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +• [SLOW TEST:29.804 seconds] +[sig-cli] Kubectl client +/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 + [k8s.io] Kubectl run rc + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687 + should create an rc from an image [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 ------------------------------ SSSS ------------------------------ -[sig-storage] EmptyDir volumes - should support (non-root,0777,default) [LinuxOnly] [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -[BeforeEach] [sig-storage] EmptyDir volumes - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +[sig-cli] Kubectl client [k8s.io] Kubectl run deployment + should create a deployment from an image [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [sig-cli] Kubectl client + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 STEP: Creating a kubernetes client -Jun 4 19:17:23.977: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504 -STEP: Building a namespace api object, basename emptydir -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in emptydir-4177 +Jun 18 12:28:38.394: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150 +STEP: Building a namespace api object, basename kubectl +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-9432 STEP: Waiting for a default service account to be provisioned in namespace -[It] should support (non-root,0777,default) [LinuxOnly] [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -STEP: Creating a pod to test emptydir 0777 on node default medium -Jun 4 19:17:24.212: INFO: Waiting up to 5m0s for pod "pod-6214f45f-86fd-11e9-8d1b-467ee19922ac" in namespace "emptydir-4177" to be "success or failure" -Jun 4 19:17:24.228: INFO: Pod "pod-6214f45f-86fd-11e9-8d1b-467ee19922ac": Phase="Pending", Reason="", readiness=false. Elapsed: 16.023177ms -Jun 4 19:17:26.255: INFO: Pod "pod-6214f45f-86fd-11e9-8d1b-467ee19922ac": Phase="Pending", Reason="", readiness=false. Elapsed: 2.043342119s -Jun 4 19:17:28.268: INFO: Pod "pod-6214f45f-86fd-11e9-8d1b-467ee19922ac": Phase="Pending", Reason="", readiness=false. Elapsed: 4.055958605s -Jun 4 19:17:30.282: INFO: Pod "pod-6214f45f-86fd-11e9-8d1b-467ee19922ac": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.069806145s +[BeforeEach] [sig-cli] Kubectl client + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:213 +[BeforeEach] [k8s.io] Kubectl run deployment + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1455 +[It] should create a deployment from an image [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: running the image docker.io/library/nginx:1.14-alpine +Jun 18 12:28:38.589: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 run e2e-test-nginx-deployment --image=docker.io/library/nginx:1.14-alpine --generator=deployment/v1beta1 --namespace=kubectl-9432' +Jun 18 12:28:38.723: INFO: stderr: "kubectl run --generator=deployment/v1beta1 is DEPRECATED and will be removed in a future version. Use kubectl run --generator=run-pod/v1 or kubectl create instead.\n" +Jun 18 12:28:38.723: INFO: stdout: "deployment.extensions/e2e-test-nginx-deployment created\n" +STEP: verifying the deployment e2e-test-nginx-deployment was created +STEP: verifying the pod controlled by deployment e2e-test-nginx-deployment was created +[AfterEach] [k8s.io] Kubectl run deployment + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1460 +Jun 18 12:28:42.756: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 delete deployment e2e-test-nginx-deployment --namespace=kubectl-9432' +Jun 18 12:28:42.931: INFO: stderr: "" +Jun 18 12:28:42.931: INFO: stdout: "deployment.extensions \"e2e-test-nginx-deployment\" deleted\n" +[AfterEach] [sig-cli] Kubectl client + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Jun 18 12:28:42.931: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "kubectl-9432" for this suite. +Jun 18 12:29:06.972: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:29:07.372: INFO: namespace kubectl-9432 deletion completed in 24.425693349s + +• [SLOW TEST:28.978 seconds] +[sig-cli] Kubectl client +/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 + [k8s.io] Kubectl run deployment + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687 + should create a deployment from an image [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +------------------------------ +SSSSSSSSSSS +------------------------------ +[sig-node] ConfigMap + should be consumable via environment variable [NodeConformance] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [sig-node] ConfigMap + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +STEP: Creating a kubernetes client +Jun 18 12:29:07.372: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150 +STEP: Building a namespace api object, basename configmap +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in configmap-6741 +STEP: Waiting for a default service account to be provisioned in namespace +[It] should be consumable via environment variable [NodeConformance] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: Creating configMap configmap-6741/configmap-test-aabdd51d-91c4-11e9-a25d-8608290c688a +STEP: Creating a pod to test consume configMaps +Jun 18 12:29:07.610: INFO: Waiting up to 5m0s for pod "pod-configmaps-aac00012-91c4-11e9-a25d-8608290c688a" in namespace "configmap-6741" to be "success or failure" +Jun 18 12:29:07.631: INFO: Pod "pod-configmaps-aac00012-91c4-11e9-a25d-8608290c688a": Phase="Pending", Reason="", readiness=false. Elapsed: 21.700654ms +Jun 18 12:29:09.645: INFO: Pod "pod-configmaps-aac00012-91c4-11e9-a25d-8608290c688a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.035413686s STEP: Saw pod success -Jun 4 19:17:30.282: INFO: Pod "pod-6214f45f-86fd-11e9-8d1b-467ee19922ac" satisfied condition "success or failure" -Jun 4 19:17:30.293: INFO: Trying to get logs from node 10.212.23.164 pod pod-6214f45f-86fd-11e9-8d1b-467ee19922ac container test-container: +Jun 18 12:29:09.645: INFO: Pod "pod-configmaps-aac00012-91c4-11e9-a25d-8608290c688a" satisfied condition "success or failure" +Jun 18 12:29:09.658: INFO: Trying to get logs from node 10.72.74.184 pod pod-configmaps-aac00012-91c4-11e9-a25d-8608290c688a container env-test: STEP: delete the pod -Jun 4 19:17:30.353: INFO: Waiting for pod pod-6214f45f-86fd-11e9-8d1b-467ee19922ac to disappear -Jun 4 19:17:30.364: INFO: Pod pod-6214f45f-86fd-11e9-8d1b-467ee19922ac no longer exists -[AfterEach] [sig-storage] EmptyDir volumes - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 -Jun 4 19:17:30.364: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "emptydir-4177" for this suite. -Jun 4 19:17:36.425: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -Jun 4 19:17:36.796: INFO: namespace emptydir-4177 deletion completed in 6.413691116s +Jun 18 12:29:10.083: INFO: Waiting for pod pod-configmaps-aac00012-91c4-11e9-a25d-8608290c688a to disappear +Jun 18 12:29:10.099: INFO: Pod pod-configmaps-aac00012-91c4-11e9-a25d-8608290c688a no longer exists +[AfterEach] [sig-node] ConfigMap + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Jun 18 12:29:10.099: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "configmap-6741" for this suite. +Jun 18 12:29:16.145: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:29:17.567: INFO: namespace configmap-6741 deletion completed in 7.451134758s -• [SLOW TEST:12.819 seconds] -[sig-storage] EmptyDir volumes -/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:41 - should support (non-root,0777,default) [LinuxOnly] [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +• [SLOW TEST:10.196 seconds] +[sig-node] ConfigMap +/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/configmap.go:32 + should be consumable via environment variable [NodeConformance] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 ------------------------------ -SSSSSSSSSSSS +SSS ------------------------------ -[sig-apps] Daemon set [Serial] - should run and stop complex daemon [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -[BeforeEach] [sig-apps] Daemon set [Serial] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +[sig-api-machinery] Secrets + should be consumable via the environment [NodeConformance] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [sig-api-machinery] Secrets + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 STEP: Creating a kubernetes client -Jun 4 19:17:36.797: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504 -STEP: Building a namespace api object, basename daemonsets -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in daemonsets-6397 +Jun 18 12:29:17.568: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150 +STEP: Building a namespace api object, basename secrets +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in secrets-1732 STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [sig-apps] Daemon set [Serial] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:102 -[It] should run and stop complex daemon [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -Jun 4 19:17:37.084: INFO: Creating daemon "daemon-set" with a node selector -STEP: Initially, daemon pods should not be running on any nodes. -Jun 4 19:17:37.108: INFO: Number of nodes with available pods: 0 -Jun 4 19:17:37.108: INFO: Number of running nodes: 0, number of available pods: 0 -STEP: Change node label to blue, check that daemon pod is launched. -Jun 4 19:17:37.163: INFO: Number of nodes with available pods: 0 -Jun 4 19:17:37.163: INFO: Node 10.212.23.161 is running more than one daemon pod -Jun 4 19:17:38.177: INFO: Number of nodes with available pods: 0 -Jun 4 19:17:38.177: INFO: Node 10.212.23.161 is running more than one daemon pod -Jun 4 19:17:39.176: INFO: Number of nodes with available pods: 1 -Jun 4 19:17:39.176: INFO: Number of running nodes: 1, number of available pods: 1 -STEP: Update the node label to green, and wait for daemons to be unscheduled -Jun 4 19:17:39.268: INFO: Number of nodes with available pods: 0 -Jun 4 19:17:39.268: INFO: Number of running nodes: 0, number of available pods: 0 -STEP: Update DaemonSet node selector to green, and change its update strategy to RollingUpdate -Jun 4 19:17:39.299: INFO: Number of nodes with available pods: 0 -Jun 4 19:17:39.299: INFO: Node 10.212.23.161 is running more than one daemon pod -Jun 4 19:17:40.313: INFO: Number of nodes with available pods: 0 -Jun 4 19:17:40.313: INFO: Node 10.212.23.161 is running more than one daemon pod -Jun 4 19:17:41.311: INFO: Number of nodes with available pods: 0 -Jun 4 19:17:41.311: INFO: Node 10.212.23.161 is running more than one daemon pod -Jun 4 19:17:42.312: INFO: Number of nodes with available pods: 0 -Jun 4 19:17:42.312: INFO: Node 10.212.23.161 is running more than one daemon pod -Jun 4 19:17:43.312: INFO: Number of nodes with available pods: 0 -Jun 4 19:17:43.312: INFO: Node 10.212.23.161 is running more than one daemon pod -Jun 4 19:17:44.312: INFO: Number of nodes with available pods: 0 -Jun 4 19:17:44.312: INFO: Node 10.212.23.161 is running more than one daemon pod -Jun 4 19:17:45.310: INFO: Number of nodes with available pods: 1 -Jun 4 19:17:45.310: INFO: Number of running nodes: 1, number of available pods: 1 -[AfterEach] [sig-apps] Daemon set [Serial] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:68 -STEP: Deleting DaemonSet "daemon-set" -STEP: deleting DaemonSet.extensions daemon-set in namespace daemonsets-6397, will wait for the garbage collector to delete the pods -Jun 4 19:17:45.420: INFO: Deleting DaemonSet.extensions daemon-set took: 27.506078ms -Jun 4 19:17:45.522: INFO: Terminating DaemonSet.extensions daemon-set pods took: 101.102471ms -Jun 4 19:17:49.334: INFO: Number of nodes with available pods: 0 -Jun 4 19:17:49.334: INFO: Number of running nodes: 0, number of available pods: 0 -Jun 4 19:17:49.345: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"selfLink":"/apis/apps/v1/namespaces/daemonsets-6397/daemonsets","resourceVersion":"20468"},"items":null} +[It] should be consumable via the environment [NodeConformance] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: creating secret secrets-1732/secret-test-b0d37b4f-91c4-11e9-a25d-8608290c688a +STEP: Creating a pod to test consume secrets +Jun 18 12:29:17.820: INFO: Waiting up to 5m0s for pod "pod-configmaps-b0d55af0-91c4-11e9-a25d-8608290c688a" in namespace "secrets-1732" to be "success or failure" +Jun 18 12:29:17.836: INFO: Pod "pod-configmaps-b0d55af0-91c4-11e9-a25d-8608290c688a": Phase="Pending", Reason="", readiness=false. Elapsed: 15.802943ms +Jun 18 12:29:19.849: INFO: Pod "pod-configmaps-b0d55af0-91c4-11e9-a25d-8608290c688a": Phase="Running", Reason="", readiness=true. Elapsed: 2.02943893s +Jun 18 12:29:21.864: INFO: Pod "pod-configmaps-b0d55af0-91c4-11e9-a25d-8608290c688a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.043832397s +STEP: Saw pod success +Jun 18 12:29:21.864: INFO: Pod "pod-configmaps-b0d55af0-91c4-11e9-a25d-8608290c688a" satisfied condition "success or failure" +Jun 18 12:29:21.889: INFO: Trying to get logs from node 10.72.74.138 pod pod-configmaps-b0d55af0-91c4-11e9-a25d-8608290c688a container env-test: +STEP: delete the pod +Jun 18 12:29:21.987: INFO: Waiting for pod pod-configmaps-b0d55af0-91c4-11e9-a25d-8608290c688a to disappear +Jun 18 12:29:22.002: INFO: Pod pod-configmaps-b0d55af0-91c4-11e9-a25d-8608290c688a no longer exists +[AfterEach] [sig-api-machinery] Secrets + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Jun 18 12:29:22.002: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "secrets-1732" for this suite. +Jun 18 12:29:30.044: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:29:30.461: INFO: namespace secrets-1732 deletion completed in 8.445131986s -Jun 4 19:17:49.356: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"selfLink":"/api/v1/namespaces/daemonsets-6397/pods","resourceVersion":"20468"},"items":null} +• [SLOW TEST:12.893 seconds] +[sig-api-machinery] Secrets +/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/secrets.go:32 + should be consumable via the environment [NodeConformance] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +------------------------------ +SSSS +------------------------------ +[sig-cli] Kubectl client [k8s.io] Kubectl patch + should add annotations for pods in rc [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [sig-cli] Kubectl client + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +STEP: Creating a kubernetes client +Jun 18 12:29:30.461: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150 +STEP: Building a namespace api object, basename kubectl +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-3180 +STEP: Waiting for a default service account to be provisioned in namespace +[BeforeEach] [sig-cli] Kubectl client + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:213 +[It] should add annotations for pods in rc [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: creating Redis RC +Jun 18 12:29:30.654: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 create -f - --namespace=kubectl-3180' +Jun 18 12:29:30.875: INFO: stderr: "" +Jun 18 12:29:30.875: INFO: stdout: "replicationcontroller/redis-master created\n" +STEP: Waiting for Redis master to start. +Jun 18 12:29:31.889: INFO: Selector matched 1 pods for map[app:redis] +Jun 18 12:29:31.889: INFO: Found 0 / 1 +Jun 18 12:29:32.890: INFO: Selector matched 1 pods for map[app:redis] +Jun 18 12:29:32.891: INFO: Found 1 / 1 +Jun 18 12:29:32.891: INFO: WaitFor completed with timeout 5m0s. Pods found = 1 out of 1 +STEP: patching all pods +Jun 18 12:29:32.905: INFO: Selector matched 1 pods for map[app:redis] +Jun 18 12:29:32.905: INFO: ForEach: Found 1 pods from the filter. Now looping through them. +Jun 18 12:29:32.905: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 patch pod redis-master-q4k6k --namespace=kubectl-3180 -p {"metadata":{"annotations":{"x":"y"}}}' +Jun 18 12:29:33.064: INFO: stderr: "" +Jun 18 12:29:33.064: INFO: stdout: "pod/redis-master-q4k6k patched\n" +STEP: checking annotations +Jun 18 12:29:33.089: INFO: Selector matched 1 pods for map[app:redis] +Jun 18 12:29:33.089: INFO: ForEach: Found 1 pods from the filter. Now looping through them. +[AfterEach] [sig-cli] Kubectl client + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Jun 18 12:29:33.089: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "kubectl-3180" for this suite. +Jun 18 12:29:59.134: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:29:59.549: INFO: namespace kubectl-3180 deletion completed in 26.446082442s -[AfterEach] [sig-apps] Daemon set [Serial] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 -Jun 4 19:17:49.426: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "daemonsets-6397" for this suite. -Jun 4 19:17:55.479: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -Jun 4 19:17:55.817: INFO: namespace daemonsets-6397 deletion completed in 6.379467663s +• [SLOW TEST:29.088 seconds] +[sig-cli] Kubectl client +/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 + [k8s.io] Kubectl patch + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687 + should add annotations for pods in rc [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +------------------------------ +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +------------------------------ +[sig-auth] ServiceAccounts + should allow opting out of API token automount [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [sig-auth] ServiceAccounts + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +STEP: Creating a kubernetes client +Jun 18 12:29:59.550: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150 +STEP: Building a namespace api object, basename svcaccounts +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in svcaccounts-8580 +STEP: Waiting for a default service account to be provisioned in namespace +[It] should allow opting out of API token automount [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: getting the auto-created API token +Jun 18 12:30:00.304: INFO: created pod pod-service-account-defaultsa +Jun 18 12:30:00.304: INFO: pod pod-service-account-defaultsa service account token volume mount: true +Jun 18 12:30:00.327: INFO: created pod pod-service-account-mountsa +Jun 18 12:30:00.327: INFO: pod pod-service-account-mountsa service account token volume mount: true +Jun 18 12:30:00.342: INFO: created pod pod-service-account-nomountsa +Jun 18 12:30:00.342: INFO: pod pod-service-account-nomountsa service account token volume mount: false +Jun 18 12:30:00.363: INFO: created pod pod-service-account-defaultsa-mountspec +Jun 18 12:30:00.363: INFO: pod pod-service-account-defaultsa-mountspec service account token volume mount: true +Jun 18 12:30:00.381: INFO: created pod pod-service-account-mountsa-mountspec +Jun 18 12:30:00.381: INFO: pod pod-service-account-mountsa-mountspec service account token volume mount: true +Jun 18 12:30:00.396: INFO: created pod pod-service-account-nomountsa-mountspec +Jun 18 12:30:00.396: INFO: pod pod-service-account-nomountsa-mountspec service account token volume mount: true +Jun 18 12:30:00.411: INFO: created pod pod-service-account-defaultsa-nomountspec +Jun 18 12:30:00.411: INFO: pod pod-service-account-defaultsa-nomountspec service account token volume mount: false +Jun 18 12:30:00.425: INFO: created pod pod-service-account-mountsa-nomountspec +Jun 18 12:30:00.425: INFO: pod pod-service-account-mountsa-nomountspec service account token volume mount: false +Jun 18 12:30:00.438: INFO: created pod pod-service-account-nomountsa-nomountspec +Jun 18 12:30:00.438: INFO: pod pod-service-account-nomountsa-nomountspec service account token volume mount: false +[AfterEach] [sig-auth] ServiceAccounts + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Jun 18 12:30:00.438: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "svcaccounts-8580" for this suite. +Jun 18 12:30:08.478: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:30:08.879: INFO: namespace svcaccounts-8580 deletion completed in 8.426017718s -• [SLOW TEST:19.020 seconds] -[sig-apps] Daemon set [Serial] -/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22 - should run and stop complex daemon [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +• [SLOW TEST:9.329 seconds] +[sig-auth] ServiceAccounts +/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/auth/framework.go:22 + should allow opting out of API token automount [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 ------------------------------ -SSSSS +SSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-storage] Subpath Atomic writer volumes - should support subpaths with configmap pod [LinuxOnly] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -[BeforeEach] [sig-storage] Subpath - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +[sig-node] Downward API + should provide default limits.cpu/memory from node allocatable [NodeConformance] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [sig-node] Downward API + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 STEP: Creating a kubernetes client -Jun 4 19:17:55.817: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504 -STEP: Building a namespace api object, basename subpath -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in subpath-5023 +Jun 18 12:30:08.879: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150 +STEP: Building a namespace api object, basename downward-api +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in downward-api-5442 STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] Atomic writer volumes - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:38 -STEP: Setting up data -[It] should support subpaths with configmap pod [LinuxOnly] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -STEP: Creating pod pod-subpath-test-configmap-rc9b -STEP: Creating a pod to test atomic-volume-subpath -Jun 4 19:17:56.106: INFO: Waiting up to 5m0s for pod "pod-subpath-test-configmap-rc9b" in namespace "subpath-5023" to be "success or failure" -Jun 4 19:17:56.118: INFO: Pod "pod-subpath-test-configmap-rc9b": Phase="Pending", Reason="", readiness=false. Elapsed: 11.674357ms -Jun 4 19:17:58.128: INFO: Pod "pod-subpath-test-configmap-rc9b": Phase="Running", Reason="", readiness=true. Elapsed: 2.022330037s -Jun 4 19:18:00.142: INFO: Pod "pod-subpath-test-configmap-rc9b": Phase="Running", Reason="", readiness=true. Elapsed: 4.035742352s -Jun 4 19:18:02.154: INFO: Pod "pod-subpath-test-configmap-rc9b": Phase="Running", Reason="", readiness=true. Elapsed: 6.047809454s -Jun 4 19:18:04.167: INFO: Pod "pod-subpath-test-configmap-rc9b": Phase="Running", Reason="", readiness=true. Elapsed: 8.061052624s -Jun 4 19:18:06.181: INFO: Pod "pod-subpath-test-configmap-rc9b": Phase="Running", Reason="", readiness=true. Elapsed: 10.07462533s -Jun 4 19:18:08.193: INFO: Pod "pod-subpath-test-configmap-rc9b": Phase="Running", Reason="", readiness=true. Elapsed: 12.087007425s -Jun 4 19:18:10.205: INFO: Pod "pod-subpath-test-configmap-rc9b": Phase="Running", Reason="", readiness=true. Elapsed: 14.099068766s -Jun 4 19:18:12.218: INFO: Pod "pod-subpath-test-configmap-rc9b": Phase="Running", Reason="", readiness=true. Elapsed: 16.112220586s -Jun 4 19:18:14.235: INFO: Pod "pod-subpath-test-configmap-rc9b": Phase="Running", Reason="", readiness=true. Elapsed: 18.128955293s -Jun 4 19:18:16.248: INFO: Pod "pod-subpath-test-configmap-rc9b": Phase="Running", Reason="", readiness=true. Elapsed: 20.142401181s -Jun 4 19:18:18.261: INFO: Pod "pod-subpath-test-configmap-rc9b": Phase="Succeeded", Reason="", readiness=false. Elapsed: 22.154993051s +[It] should provide default limits.cpu/memory from node allocatable [NodeConformance] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: Creating a pod to test downward api env vars +Jun 18 12:30:09.106: INFO: Waiting up to 5m0s for pod "downward-api-cf674cec-91c4-11e9-a25d-8608290c688a" in namespace "downward-api-5442" to be "success or failure" +Jun 18 12:30:09.119: INFO: Pod "downward-api-cf674cec-91c4-11e9-a25d-8608290c688a": Phase="Pending", Reason="", readiness=false. Elapsed: 12.253306ms +Jun 18 12:30:11.132: INFO: Pod "downward-api-cf674cec-91c4-11e9-a25d-8608290c688a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.025730153s STEP: Saw pod success -Jun 4 19:18:18.261: INFO: Pod "pod-subpath-test-configmap-rc9b" satisfied condition "success or failure" -Jun 4 19:18:18.273: INFO: Trying to get logs from node 10.212.23.189 pod pod-subpath-test-configmap-rc9b container test-container-subpath-configmap-rc9b: +Jun 18 12:30:11.132: INFO: Pod "downward-api-cf674cec-91c4-11e9-a25d-8608290c688a" satisfied condition "success or failure" +Jun 18 12:30:11.147: INFO: Trying to get logs from node 10.72.74.184 pod downward-api-cf674cec-91c4-11e9-a25d-8608290c688a container dapi-container: STEP: delete the pod -Jun 4 19:18:18.355: INFO: Waiting for pod pod-subpath-test-configmap-rc9b to disappear -Jun 4 19:18:18.367: INFO: Pod pod-subpath-test-configmap-rc9b no longer exists -STEP: Deleting pod pod-subpath-test-configmap-rc9b -Jun 4 19:18:18.367: INFO: Deleting pod "pod-subpath-test-configmap-rc9b" in namespace "subpath-5023" -[AfterEach] [sig-storage] Subpath - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 -Jun 4 19:18:18.378: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "subpath-5023" for this suite. -Jun 4 19:18:24.442: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -Jun 4 19:18:24.783: INFO: namespace subpath-5023 deletion completed in 6.389240718s +Jun 18 12:30:11.247: INFO: Waiting for pod downward-api-cf674cec-91c4-11e9-a25d-8608290c688a to disappear +Jun 18 12:30:11.260: INFO: Pod downward-api-cf674cec-91c4-11e9-a25d-8608290c688a no longer exists +[AfterEach] [sig-node] Downward API + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Jun 18 12:30:11.260: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "downward-api-5442" for this suite. +Jun 18 12:30:17.317: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:30:17.728: INFO: namespace downward-api-5442 deletion completed in 6.453240989s -• [SLOW TEST:28.966 seconds] -[sig-storage] Subpath -/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:22 - Atomic writer volumes - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:34 - should support subpaths with configmap pod [LinuxOnly] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +• [SLOW TEST:8.849 seconds] +[sig-node] Downward API +/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downward_api.go:38 + should provide default limits.cpu/memory from node allocatable [NodeConformance] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +------------------------------ +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +------------------------------ +[sig-storage] ConfigMap + should be consumable from pods in volume with mappings as non-root [LinuxOnly] [NodeConformance] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [sig-storage] ConfigMap + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +STEP: Creating a kubernetes client +Jun 18 12:30:17.729: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150 +STEP: Building a namespace api object, basename configmap +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in configmap-3250 +STEP: Waiting for a default service account to be provisioned in namespace +[It] should be consumable from pods in volume with mappings as non-root [LinuxOnly] [NodeConformance] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: Creating configMap with name configmap-test-volume-map-d4acff47-91c4-11e9-a25d-8608290c688a +STEP: Creating a pod to test consume configMaps +Jun 18 12:30:17.964: INFO: Waiting up to 5m0s for pod "pod-configmaps-d4af35c3-91c4-11e9-a25d-8608290c688a" in namespace "configmap-3250" to be "success or failure" +Jun 18 12:30:17.979: INFO: Pod "pod-configmaps-d4af35c3-91c4-11e9-a25d-8608290c688a": Phase="Pending", Reason="", readiness=false. Elapsed: 15.084091ms +Jun 18 12:30:19.995: INFO: Pod "pod-configmaps-d4af35c3-91c4-11e9-a25d-8608290c688a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.031319151s +STEP: Saw pod success +Jun 18 12:30:19.995: INFO: Pod "pod-configmaps-d4af35c3-91c4-11e9-a25d-8608290c688a" satisfied condition "success or failure" +Jun 18 12:30:20.022: INFO: Trying to get logs from node 10.72.74.138 pod pod-configmaps-d4af35c3-91c4-11e9-a25d-8608290c688a container configmap-volume-test: +STEP: delete the pod +Jun 18 12:30:20.102: INFO: Waiting for pod pod-configmaps-d4af35c3-91c4-11e9-a25d-8608290c688a to disappear +Jun 18 12:30:20.114: INFO: Pod pod-configmaps-d4af35c3-91c4-11e9-a25d-8608290c688a no longer exists +[AfterEach] [sig-storage] ConfigMap + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Jun 18 12:30:20.114: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "configmap-3250" for this suite. +Jun 18 12:30:28.158: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:30:28.541: INFO: namespace configmap-3250 deletion completed in 8.412449447s + +• [SLOW TEST:10.812 seconds] +[sig-storage] ConfigMap +/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:32 + should be consumable from pods in volume with mappings as non-root [LinuxOnly] [NodeConformance] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 ------------------------------ -SSSSSSSSSSSSSSSSSS +SSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-storage] Downward API volume - should provide container's memory limit [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -[BeforeEach] [sig-storage] Downward API volume - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +[sig-network] Networking Granular Checks: Pods + should function for node-pod communication: http [LinuxOnly] [NodeConformance] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [sig-network] Networking + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 STEP: Creating a kubernetes client -Jun 4 19:18:24.783: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504 -STEP: Building a namespace api object, basename downward-api -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in downward-api-7801 +Jun 18 12:30:28.541: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150 +STEP: Building a namespace api object, basename pod-network-test +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in pod-network-test-2560 STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [sig-storage] Downward API volume - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:39 -[It] should provide container's memory limit [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -STEP: Creating a pod to test downward API volume plugin -Jun 4 19:18:25.022: INFO: Waiting up to 5m0s for pod "downwardapi-volume-865413c0-86fd-11e9-8d1b-467ee19922ac" in namespace "downward-api-7801" to be "success or failure" -Jun 4 19:18:25.039: INFO: Pod "downwardapi-volume-865413c0-86fd-11e9-8d1b-467ee19922ac": Phase="Pending", Reason="", readiness=false. Elapsed: 17.125919ms -Jun 4 19:18:27.052: INFO: Pod "downwardapi-volume-865413c0-86fd-11e9-8d1b-467ee19922ac": Phase="Pending", Reason="", readiness=false. Elapsed: 2.030382954s -Jun 4 19:18:29.065: INFO: Pod "downwardapi-volume-865413c0-86fd-11e9-8d1b-467ee19922ac": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.043077722s -STEP: Saw pod success -Jun 4 19:18:29.065: INFO: Pod "downwardapi-volume-865413c0-86fd-11e9-8d1b-467ee19922ac" satisfied condition "success or failure" -Jun 4 19:18:29.076: INFO: Trying to get logs from node 10.212.23.164 pod downwardapi-volume-865413c0-86fd-11e9-8d1b-467ee19922ac container client-container: -STEP: delete the pod -Jun 4 19:18:29.140: INFO: Waiting for pod downwardapi-volume-865413c0-86fd-11e9-8d1b-467ee19922ac to disappear -Jun 4 19:18:29.150: INFO: Pod downwardapi-volume-865413c0-86fd-11e9-8d1b-467ee19922ac no longer exists -[AfterEach] [sig-storage] Downward API volume - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 -Jun 4 19:18:29.150: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "downward-api-7801" for this suite. -Jun 4 19:18:35.208: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -Jun 4 19:18:35.551: INFO: namespace downward-api-7801 deletion completed in 6.385381632s +[It] should function for node-pod communication: http [LinuxOnly] [NodeConformance] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: Performing setup for networking test in namespace pod-network-test-2560 +STEP: creating a selector +STEP: Creating the service pods in kubernetes +Jun 18 12:30:28.738: INFO: Waiting up to 10m0s for all (but 0) nodes to be schedulable +STEP: Creating test pods +Jun 18 12:30:51.021: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s --max-time 15 --connect-timeout 1 http://172.30.142.31:8080/hostName | grep -v '^\s*$'] Namespace:pod-network-test-2560 PodName:host-test-container-pod ContainerName:hostexec Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} +Jun 18 12:30:51.021: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150 +Jun 18 12:30:51.868: INFO: Found all expected endpoints: [netserver-0] +Jun 18 12:30:51.882: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s --max-time 15 --connect-timeout 1 http://172.30.206.148:8080/hostName | grep -v '^\s*$'] Namespace:pod-network-test-2560 PodName:host-test-container-pod ContainerName:hostexec Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} +Jun 18 12:30:51.882: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150 +Jun 18 12:30:52.157: INFO: Found all expected endpoints: [netserver-1] +Jun 18 12:30:52.171: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s --max-time 15 --connect-timeout 1 http://172.30.80.151:8080/hostName | grep -v '^\s*$'] Namespace:pod-network-test-2560 PodName:host-test-container-pod ContainerName:hostexec Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} +Jun 18 12:30:52.171: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150 +Jun 18 12:30:52.420: INFO: Found all expected endpoints: [netserver-2] +[AfterEach] [sig-network] Networking + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Jun 18 12:30:52.420: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "pod-network-test-2560" for this suite. +Jun 18 12:31:16.462: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:31:16.866: INFO: namespace pod-network-test-2560 deletion completed in 24.43117234s -• [SLOW TEST:10.768 seconds] -[sig-storage] Downward API volume -/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:34 - should provide container's memory limit [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +• [SLOW TEST:48.325 seconds] +[sig-network] Networking +/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:25 + Granular Checks: Pods + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:28 + should function for node-pod communication: http [LinuxOnly] [NodeConformance] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 ------------------------------ -SSSS +[sig-apps] Deployment + deployment should support rollover [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [sig-apps] Deployment + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +STEP: Creating a kubernetes client +Jun 18 12:31:16.867: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150 +STEP: Building a namespace api object, basename deployment +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in deployment-7214 +STEP: Waiting for a default service account to be provisioned in namespace +[BeforeEach] [sig-apps] Deployment + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:65 +[It] deployment should support rollover [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +Jun 18 12:31:17.178: INFO: Pod name rollover-pod: Found 0 pods out of 1 +Jun 18 12:31:22.192: INFO: Pod name rollover-pod: Found 1 pods out of 1 +STEP: ensuring each pod is running +Jun 18 12:31:22.192: INFO: Waiting for pods owned by replica set "test-rollover-controller" to become ready +Jun 18 12:31:24.204: INFO: Creating deployment "test-rollover-deployment" +Jun 18 12:31:24.687: INFO: Make sure deployment "test-rollover-deployment" performs scaling operations +Jun 18 12:31:26.703: INFO: Check revision of new replica set for deployment "test-rollover-deployment" +Jun 18 12:31:26.839: INFO: Ensure that both replica sets have 1 created replica +Jun 18 12:31:26.860: INFO: Rollover old replica sets for deployment "test-rollover-deployment" with new image update +Jun 18 12:31:26.876: INFO: Updating deployment test-rollover-deployment +Jun 18 12:31:26.876: INFO: Wait deployment "test-rollover-deployment" to be observed by the deployment controller +Jun 18 12:31:28.890: INFO: Wait for revision update of deployment "test-rollover-deployment" to 2 +Jun 18 12:31:28.909: INFO: Make sure deployment "test-rollover-deployment" is complete +Jun 18 12:31:28.938: INFO: all replica sets need to contain the pod-template-hash label +Jun 18 12:31:28.938: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:2, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63696457884, loc:(*time.Location)(0x8a1a0e0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63696457884, loc:(*time.Location)(0x8a1a0e0)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63696457888, loc:(*time.Location)(0x8a1a0e0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63696457884, loc:(*time.Location)(0x8a1a0e0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-766b4d6c9d\" is progressing."}}, CollisionCount:(*int32)(nil)} +Jun 18 12:31:30.957: INFO: all replica sets need to contain the pod-template-hash label +Jun 18 12:31:30.958: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:2, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63696457884, loc:(*time.Location)(0x8a1a0e0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63696457884, loc:(*time.Location)(0x8a1a0e0)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63696457888, loc:(*time.Location)(0x8a1a0e0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63696457884, loc:(*time.Location)(0x8a1a0e0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-766b4d6c9d\" is progressing."}}, CollisionCount:(*int32)(nil)} +Jun 18 12:31:32.957: INFO: all replica sets need to contain the pod-template-hash label +Jun 18 12:31:32.957: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:2, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63696457884, loc:(*time.Location)(0x8a1a0e0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63696457884, loc:(*time.Location)(0x8a1a0e0)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63696457888, loc:(*time.Location)(0x8a1a0e0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63696457884, loc:(*time.Location)(0x8a1a0e0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-766b4d6c9d\" is progressing."}}, CollisionCount:(*int32)(nil)} +Jun 18 12:31:34.958: INFO: all replica sets need to contain the pod-template-hash label +Jun 18 12:31:34.958: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:2, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63696457884, loc:(*time.Location)(0x8a1a0e0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63696457884, loc:(*time.Location)(0x8a1a0e0)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63696457888, loc:(*time.Location)(0x8a1a0e0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63696457884, loc:(*time.Location)(0x8a1a0e0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-766b4d6c9d\" is progressing."}}, CollisionCount:(*int32)(nil)} +Jun 18 12:31:36.959: INFO: all replica sets need to contain the pod-template-hash label +Jun 18 12:31:36.959: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:2, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63696457884, loc:(*time.Location)(0x8a1a0e0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63696457884, loc:(*time.Location)(0x8a1a0e0)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63696457888, loc:(*time.Location)(0x8a1a0e0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63696457884, loc:(*time.Location)(0x8a1a0e0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-766b4d6c9d\" is progressing."}}, CollisionCount:(*int32)(nil)} +Jun 18 12:31:38.959: INFO: +Jun 18 12:31:38.959: INFO: Ensure that both old replica sets have no replicas +[AfterEach] [sig-apps] Deployment + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:59 +Jun 18 12:31:38.991: INFO: Deployment "test-rollover-deployment": +&Deployment{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-rollover-deployment,GenerateName:,Namespace:deployment-7214,SelfLink:/apis/apps/v1/namespaces/deployment-7214/deployments/test-rollover-deployment,UID:fc2f24a5-91c4-11e9-a08a-ee7a14707756,ResourceVersion:98071,Generation:2,CreationTimestamp:2019-06-18 12:31:24 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: rollover-pod,},Annotations:map[string]string{deployment.kubernetes.io/revision: 2,},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:DeploymentSpec{Replicas:*1,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: rollover-pod,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: rollover-pod,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[],Containers:[{redis gcr.io/kubernetes-e2e-test-images/redis:1.0 [] [] [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,},},Strategy:DeploymentStrategy{Type:RollingUpdate,RollingUpdate:&RollingUpdateDeployment{MaxUnavailable:0,MaxSurge:1,},},MinReadySeconds:10,RevisionHistoryLimit:*10,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:2,Replicas:1,UpdatedReplicas:1,AvailableReplicas:1,UnavailableReplicas:0,Conditions:[{Available True 2019-06-18 12:31:24 +0000 UTC 2019-06-18 12:31:24 +0000 UTC MinimumReplicasAvailable Deployment has minimum availability.} {Progressing True 2019-06-18 12:31:38 +0000 UTC 2019-06-18 12:31:24 +0000 UTC NewReplicaSetAvailable ReplicaSet "test-rollover-deployment-766b4d6c9d" has successfully progressed.}],ReadyReplicas:1,CollisionCount:nil,},} + +Jun 18 12:31:39.003: INFO: New ReplicaSet "test-rollover-deployment-766b4d6c9d" of Deployment "test-rollover-deployment": +&ReplicaSet{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-rollover-deployment-766b4d6c9d,GenerateName:,Namespace:deployment-7214,SelfLink:/apis/apps/v1/namespaces/deployment-7214/replicasets/test-rollover-deployment-766b4d6c9d,UID:fdc68f3b-91c4-11e9-a08a-ee7a14707756,ResourceVersion:98061,Generation:2,CreationTimestamp:2019-06-18 12:31:26 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: rollover-pod,pod-template-hash: 766b4d6c9d,},Annotations:map[string]string{deployment.kubernetes.io/desired-replicas: 1,deployment.kubernetes.io/max-replicas: 2,deployment.kubernetes.io/revision: 2,},OwnerReferences:[{apps/v1 Deployment test-rollover-deployment fc2f24a5-91c4-11e9-a08a-ee7a14707756 0xc0016ecfa7 0xc0016ecfa8}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:ReplicaSetSpec{Replicas:*1,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: rollover-pod,pod-template-hash: 766b4d6c9d,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: rollover-pod,pod-template-hash: 766b4d6c9d,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[],Containers:[{redis gcr.io/kubernetes-e2e-test-images/redis:1.0 [] [] [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,},},MinReadySeconds:10,},Status:ReplicaSetStatus{Replicas:1,FullyLabeledReplicas:1,ObservedGeneration:2,ReadyReplicas:1,AvailableReplicas:1,Conditions:[],},} +Jun 18 12:31:39.003: INFO: All old ReplicaSets of Deployment "test-rollover-deployment": +Jun 18 12:31:39.004: INFO: &ReplicaSet{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-rollover-controller,GenerateName:,Namespace:deployment-7214,SelfLink:/apis/apps/v1/namespaces/deployment-7214/replicasets/test-rollover-controller,UID:f7edc43c-91c4-11e9-a08a-ee7a14707756,ResourceVersion:98070,Generation:2,CreationTimestamp:2019-06-18 12:31:17 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: rollover-pod,pod: nginx,},Annotations:map[string]string{deployment.kubernetes.io/desired-replicas: 1,deployment.kubernetes.io/max-replicas: 2,},OwnerReferences:[{apps/v1 Deployment test-rollover-deployment fc2f24a5-91c4-11e9-a08a-ee7a14707756 0xc0016ecdc7 0xc0016ecdc8}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:ReplicaSetSpec{Replicas:*0,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: rollover-pod,pod: nginx,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: rollover-pod,pod: nginx,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,},},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:0,FullyLabeledReplicas:0,ObservedGeneration:2,ReadyReplicas:0,AvailableReplicas:0,Conditions:[],},} +Jun 18 12:31:39.004: INFO: &ReplicaSet{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-rollover-deployment-6455657675,GenerateName:,Namespace:deployment-7214,SelfLink:/apis/apps/v1/namespaces/deployment-7214/replicasets/test-rollover-deployment-6455657675,UID:fc79ac33-91c4-11e9-a08a-ee7a14707756,ResourceVersion:98025,Generation:2,CreationTimestamp:2019-06-18 12:31:24 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: rollover-pod,pod-template-hash: 6455657675,},Annotations:map[string]string{deployment.kubernetes.io/desired-replicas: 1,deployment.kubernetes.io/max-replicas: 2,deployment.kubernetes.io/revision: 1,},OwnerReferences:[{apps/v1 Deployment test-rollover-deployment fc2f24a5-91c4-11e9-a08a-ee7a14707756 0xc0016eceb7 0xc0016eceb8}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:ReplicaSetSpec{Replicas:*0,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: rollover-pod,pod-template-hash: 6455657675,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: rollover-pod,pod-template-hash: 6455657675,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[],Containers:[{redis-slave gcr.io/google_samples/gb-redisslave:nonexistent [] [] [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,},},MinReadySeconds:10,},Status:ReplicaSetStatus{Replicas:0,FullyLabeledReplicas:0,ObservedGeneration:2,ReadyReplicas:0,AvailableReplicas:0,Conditions:[],},} +Jun 18 12:31:39.017: INFO: Pod "test-rollover-deployment-766b4d6c9d-6s6ql" is available: +&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-rollover-deployment-766b4d6c9d-6s6ql,GenerateName:test-rollover-deployment-766b4d6c9d-,Namespace:deployment-7214,SelfLink:/api/v1/namespaces/deployment-7214/pods/test-rollover-deployment-766b4d6c9d-6s6ql,UID:fdcd27ba-91c4-11e9-a08a-ee7a14707756,ResourceVersion:98042,Generation:0,CreationTimestamp:2019-06-18 12:31:26 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: rollover-pod,pod-template-hash: 766b4d6c9d,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet test-rollover-deployment-766b4d6c9d fdc68f3b-91c4-11e9-a08a-ee7a14707756 0xc0016edb07 0xc0016edb08}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-dqspq {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-dqspq,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{redis gcr.io/kubernetes-e2e-test-images/redis:1.0 [] [] [] [] [] {map[] map[]} [{default-token-dqspq true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.72.74.184,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc0016edb80} {node.kubernetes.io/unreachable Exists NoExecute 0xc0016edba0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Running,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:31:26 +0000 UTC } {Ready True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:31:28 +0000 UTC } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:31:28 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 12:31:26 +0000 UTC }],Message:,Reason:,HostIP:10.72.74.184,PodIP:172.30.142.34,StartTime:2019-06-18 12:31:26 +0000 UTC,ContainerStatuses:[{redis {nil ContainerStateRunning{StartedAt:2019-06-18 12:31:28 +0000 UTC,} nil} {nil nil nil} true 0 gcr.io/kubernetes-e2e-test-images/redis:1.0 gcr.io/kubernetes-e2e-test-images/redis@sha256:af4748d1655c08dc54d4be5182135395db9ce87aba2d4699b26b14ae197c5830 containerd://0929048e8c502ee47ddb87fc69aad2d1d0d355a0666d0f20cd2701f284b03a88}],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} +[AfterEach] [sig-apps] Deployment + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Jun 18 12:31:39.017: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "deployment-7214" for this suite. +Jun 18 12:31:47.059: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:31:47.594: INFO: namespace deployment-7214 deletion completed in 8.563577329s + +• [SLOW TEST:30.728 seconds] +[sig-apps] Deployment +/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22 + deployment should support rollover [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 ------------------------------ -[sig-cli] Kubectl client [k8s.io] Kubectl label - should update the label on a resource [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[sig-cli] Kubectl client [k8s.io] Kubectl replace + should update a single-container pod's image [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 [BeforeEach] [sig-cli] Kubectl client - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 STEP: Creating a kubernetes client -Jun 4 19:18:35.551: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504 +Jun 18 12:31:47.594: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150 STEP: Building a namespace api object, basename kubectl -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-5446 +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-2630 STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:213 -[BeforeEach] [k8s.io] Kubectl label - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1108 -STEP: creating the pod -Jun 4 19:18:35.770: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 create -f - --namespace=kubectl-5446' -Jun 4 19:18:36.068: INFO: stderr: "" -Jun 4 19:18:36.068: INFO: stdout: "pod/pause created\n" -Jun 4 19:18:36.068: INFO: Waiting up to 5m0s for 1 pods to be running and ready: [pause] -Jun 4 19:18:36.068: INFO: Waiting up to 5m0s for pod "pause" in namespace "kubectl-5446" to be "running and ready" -Jun 4 19:18:36.083: INFO: Pod "pause": Phase="Pending", Reason="", readiness=false. Elapsed: 15.212473ms -Jun 4 19:18:38.095: INFO: Pod "pause": Phase="Running", Reason="", readiness=true. Elapsed: 2.027393136s -Jun 4 19:18:38.095: INFO: Pod "pause" satisfied condition "running and ready" -Jun 4 19:18:38.095: INFO: Wanted all 1 pods to be running and ready. Result: true. Pods: [pause] -[It] should update the label on a resource [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -STEP: adding the label testing-label with value testing-label-value to a pod -Jun 4 19:18:38.095: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 label pods pause testing-label=testing-label-value --namespace=kubectl-5446' -Jun 4 19:18:38.207: INFO: stderr: "" -Jun 4 19:18:38.207: INFO: stdout: "pod/pause labeled\n" -STEP: verifying the pod has the label testing-label with the value testing-label-value -Jun 4 19:18:38.207: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 get pod pause -L testing-label --namespace=kubectl-5446' -Jun 4 19:18:38.294: INFO: stderr: "" -Jun 4 19:18:38.294: INFO: stdout: "NAME READY STATUS RESTARTS AGE TESTING-LABEL\npause 1/1 Running 0 2s testing-label-value\n" -STEP: removing the label testing-label of a pod -Jun 4 19:18:38.294: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 label pods pause testing-label- --namespace=kubectl-5446' -Jun 4 19:18:38.391: INFO: stderr: "" -Jun 4 19:18:38.391: INFO: stdout: "pod/pause labeled\n" -STEP: verifying the pod doesn't have the label testing-label -Jun 4 19:18:38.391: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 get pod pause -L testing-label --namespace=kubectl-5446' -Jun 4 19:18:38.477: INFO: stderr: "" -Jun 4 19:18:38.477: INFO: stdout: "NAME READY STATUS RESTARTS AGE TESTING-LABEL\npause 1/1 Running 0 2s \n" -[AfterEach] [k8s.io] Kubectl label - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1115 -STEP: using delete to clean up resources -Jun 4 19:18:38.478: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 delete --grace-period=0 --force -f - --namespace=kubectl-5446' -Jun 4 19:18:38.612: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" -Jun 4 19:18:38.612: INFO: stdout: "pod \"pause\" force deleted\n" -Jun 4 19:18:38.612: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 get rc,svc -l name=pause --no-headers --namespace=kubectl-5446' -Jun 4 19:18:38.714: INFO: stderr: "No resources found.\n" -Jun 4 19:18:38.714: INFO: stdout: "" -Jun 4 19:18:38.714: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 get pods -l name=pause --namespace=kubectl-5446 -o go-template={{ range .items }}{{ if not .metadata.deletionTimestamp }}{{ .metadata.name }}{{ "\n" }}{{ end }}{{ end }}' -Jun 4 19:18:38.901: INFO: stderr: "" -Jun 4 19:18:38.901: INFO: stdout: "" + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:213 +[BeforeEach] [k8s.io] Kubectl replace + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1619 +[It] should update a single-container pod's image [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: running the image docker.io/library/nginx:1.14-alpine +Jun 18 12:31:47.795: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 run e2e-test-nginx-pod --generator=run-pod/v1 --image=docker.io/library/nginx:1.14-alpine --labels=run=e2e-test-nginx-pod --namespace=kubectl-2630' +Jun 18 12:31:47.942: INFO: stderr: "" +Jun 18 12:31:47.942: INFO: stdout: "pod/e2e-test-nginx-pod created\n" +STEP: verifying the pod e2e-test-nginx-pod is running +STEP: verifying the pod e2e-test-nginx-pod was created +Jun 18 12:31:52.993: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 get pod e2e-test-nginx-pod --namespace=kubectl-2630 -o json' +Jun 18 12:31:53.120: INFO: stderr: "" +Jun 18 12:31:53.120: INFO: stdout: "{\n \"apiVersion\": \"v1\",\n \"kind\": \"Pod\",\n \"metadata\": {\n \"annotations\": {\n \"kubernetes.io/psp\": \"e2e-test-privileged-psp\"\n },\n \"creationTimestamp\": \"2019-06-18T12:31:47Z\",\n \"labels\": {\n \"run\": \"e2e-test-nginx-pod\"\n },\n \"name\": \"e2e-test-nginx-pod\",\n \"namespace\": \"kubectl-2630\",\n \"resourceVersion\": \"98154\",\n \"selfLink\": \"/api/v1/namespaces/kubectl-2630/pods/e2e-test-nginx-pod\",\n \"uid\": \"0a50f92f-91c5-11e9-aff5-3a82590dc32b\"\n },\n \"spec\": {\n \"containers\": [\n {\n \"image\": \"docker.io/library/nginx:1.14-alpine\",\n \"imagePullPolicy\": \"IfNotPresent\",\n \"name\": \"e2e-test-nginx-pod\",\n \"resources\": {},\n \"terminationMessagePath\": \"/dev/termination-log\",\n \"terminationMessagePolicy\": \"File\",\n \"volumeMounts\": [\n {\n \"mountPath\": \"/var/run/secrets/kubernetes.io/serviceaccount\",\n \"name\": \"default-token-fjkhk\",\n \"readOnly\": true\n }\n ]\n }\n ],\n \"dnsPolicy\": \"ClusterFirst\",\n \"enableServiceLinks\": true,\n \"nodeName\": \"10.72.74.184\",\n \"priority\": 0,\n \"restartPolicy\": \"Always\",\n \"schedulerName\": \"default-scheduler\",\n \"securityContext\": {},\n \"serviceAccount\": \"default\",\n \"serviceAccountName\": \"default\",\n \"terminationGracePeriodSeconds\": 30,\n \"tolerations\": [\n {\n \"effect\": \"NoExecute\",\n \"key\": \"node.kubernetes.io/not-ready\",\n \"operator\": \"Exists\",\n \"tolerationSeconds\": 300\n },\n {\n \"effect\": \"NoExecute\",\n \"key\": \"node.kubernetes.io/unreachable\",\n \"operator\": \"Exists\",\n \"tolerationSeconds\": 300\n }\n ],\n \"volumes\": [\n {\n \"name\": \"default-token-fjkhk\",\n \"secret\": {\n \"defaultMode\": 420,\n \"secretName\": \"default-token-fjkhk\"\n }\n }\n ]\n },\n \"status\": {\n \"conditions\": [\n {\n \"lastProbeTime\": null,\n \"lastTransitionTime\": \"2019-06-18T12:31:47Z\",\n \"status\": \"True\",\n \"type\": \"Initialized\"\n },\n {\n \"lastProbeTime\": null,\n \"lastTransitionTime\": \"2019-06-18T12:31:50Z\",\n \"status\": \"True\",\n \"type\": \"Ready\"\n },\n {\n \"lastProbeTime\": null,\n \"lastTransitionTime\": \"2019-06-18T12:31:50Z\",\n \"status\": \"True\",\n \"type\": \"ContainersReady\"\n },\n {\n \"lastProbeTime\": null,\n \"lastTransitionTime\": \"2019-06-18T12:31:47Z\",\n \"status\": \"True\",\n \"type\": \"PodScheduled\"\n }\n ],\n \"containerStatuses\": [\n {\n \"containerID\": \"containerd://1782e322d35674a10d26489fb0e8a0271107eda8098578549b21d22df4691a27\",\n \"image\": \"docker.io/library/nginx:1.14-alpine\",\n \"imageID\": \"docker.io/library/nginx@sha256:485b610fefec7ff6c463ced9623314a04ed67e3945b9c08d7e53a47f6d108dc7\",\n \"lastState\": {},\n \"name\": \"e2e-test-nginx-pod\",\n \"ready\": true,\n \"restartCount\": 0,\n \"state\": {\n \"running\": {\n \"startedAt\": \"2019-06-18T12:31:49Z\"\n }\n }\n }\n ],\n \"hostIP\": \"10.72.74.184\",\n \"phase\": \"Running\",\n \"podIP\": \"172.30.142.35\",\n \"qosClass\": \"BestEffort\",\n \"startTime\": \"2019-06-18T12:31:47Z\"\n }\n}\n" +STEP: replace the image in the pod +Jun 18 12:31:53.120: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 replace -f - --namespace=kubectl-2630' +Jun 18 12:31:53.462: INFO: stderr: "" +Jun 18 12:31:53.462: INFO: stdout: "pod/e2e-test-nginx-pod replaced\n" +STEP: verifying the pod e2e-test-nginx-pod has the right image docker.io/library/busybox:1.29 +[AfterEach] [k8s.io] Kubectl replace + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1624 +Jun 18 12:31:53.484: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 delete pods e2e-test-nginx-pod --namespace=kubectl-2630' +Jun 18 12:31:55.452: INFO: stderr: "" +Jun 18 12:31:55.452: INFO: stdout: "pod \"e2e-test-nginx-pod\" deleted\n" [AfterEach] [sig-cli] Kubectl client - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 -Jun 4 19:18:38.901: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "kubectl-5446" for this suite. -Jun 4 19:18:44.963: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -Jun 4 19:18:45.524: INFO: namespace kubectl-5446 deletion completed in 6.604748266s + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Jun 18 12:31:55.452: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "kubectl-2630" for this suite. +Jun 18 12:32:01.492: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:32:01.926: INFO: namespace kubectl-2630 deletion completed in 6.460014009s -• [SLOW TEST:9.973 seconds] +• [SLOW TEST:14.331 seconds] [sig-cli] Kubectl client -/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 - [k8s.io] Kubectl label - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687 - should update the label on a resource [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 + [k8s.io] Kubectl replace + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687 + should update a single-container pod's image [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 ------------------------------ -SSSSSSSSSSSSSSSSSS +SSSSSSSSSSSSSSS ------------------------------ -[k8s.io] Pods - should support retrieving logs from the container over websockets [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -[BeforeEach] [k8s.io] Pods - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +[sig-apps] Daemon set [Serial] + should rollback without unnecessary restarts [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [sig-apps] Daemon set [Serial] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 STEP: Creating a kubernetes client -Jun 4 19:18:45.525: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504 -STEP: Building a namespace api object, basename pods -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in pods-9601 +Jun 18 12:32:01.930: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150 +STEP: Building a namespace api object, basename daemonsets +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in daemonsets-1025 STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [k8s.io] Pods - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:135 -[It] should support retrieving logs from the container over websockets [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -Jun 4 19:18:45.759: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504 -STEP: creating the pod -STEP: submitting the pod to kubernetes -[AfterEach] [k8s.io] Pods - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 -Jun 4 19:18:47.953: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "pods-9601" for this suite. -Jun 4 19:19:28.016: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -Jun 4 19:19:28.470: INFO: namespace pods-9601 deletion completed in 40.498287049s +[BeforeEach] [sig-apps] Daemon set [Serial] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:102 +[It] should rollback without unnecessary restarts [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +Jun 18 12:32:02.312: INFO: Create a RollingUpdate DaemonSet +Jun 18 12:32:02.330: INFO: Check that daemon pods launch on every node of the cluster +Jun 18 12:32:02.355: INFO: Number of nodes with available pods: 0 +Jun 18 12:32:02.355: INFO: Node 10.72.74.138 is running more than one daemon pod +Jun 18 12:32:03.404: INFO: Number of nodes with available pods: 0 +Jun 18 12:32:03.404: INFO: Node 10.72.74.138 is running more than one daemon pod +Jun 18 12:32:04.384: INFO: Number of nodes with available pods: 2 +Jun 18 12:32:04.384: INFO: Node 10.72.74.184 is running more than one daemon pod +Jun 18 12:32:05.385: INFO: Number of nodes with available pods: 3 +Jun 18 12:32:05.385: INFO: Number of running nodes: 3, number of available pods: 3 +Jun 18 12:32:05.385: INFO: Update the DaemonSet to trigger a rollout +Jun 18 12:32:05.413: INFO: Updating DaemonSet daemon-set +Jun 18 12:32:17.453: INFO: Roll back the DaemonSet before rollout is complete +Jun 18 12:32:17.484: INFO: Updating DaemonSet daemon-set +Jun 18 12:32:17.484: INFO: Make sure DaemonSet rollback is complete +Jun 18 12:32:17.497: INFO: Wrong image for pod: daemon-set-njqjq. Expected: docker.io/library/nginx:1.14-alpine, got: foo:non-existent. +Jun 18 12:32:17.497: INFO: Pod daemon-set-njqjq is not available +Jun 18 12:32:18.525: INFO: Wrong image for pod: daemon-set-njqjq. Expected: docker.io/library/nginx:1.14-alpine, got: foo:non-existent. +Jun 18 12:32:18.525: INFO: Pod daemon-set-njqjq is not available +Jun 18 12:32:19.525: INFO: Wrong image for pod: daemon-set-njqjq. Expected: docker.io/library/nginx:1.14-alpine, got: foo:non-existent. +Jun 18 12:32:19.525: INFO: Pod daemon-set-njqjq is not available +Jun 18 12:32:20.527: INFO: Pod daemon-set-ph48k is not available +[AfterEach] [sig-apps] Daemon set [Serial] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:68 +STEP: Deleting DaemonSet "daemon-set" +STEP: deleting DaemonSet.extensions daemon-set in namespace daemonsets-1025, will wait for the garbage collector to delete the pods +Jun 18 12:32:20.661: INFO: Deleting DaemonSet.extensions daemon-set took: 32.792078ms +Jun 18 12:32:20.861: INFO: Terminating DaemonSet.extensions daemon-set pods took: 200.273932ms +Jun 18 12:32:33.974: INFO: Number of nodes with available pods: 0 +Jun 18 12:32:33.974: INFO: Number of running nodes: 0, number of available pods: 0 +Jun 18 12:32:33.986: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"selfLink":"/apis/apps/v1/namespaces/daemonsets-1025/daemonsets","resourceVersion":"98375"},"items":null} -• [SLOW TEST:42.946 seconds] -[k8s.io] Pods -/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687 - should support retrieving logs from the container over websockets [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +Jun 18 12:32:33.998: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"selfLink":"/api/v1/namespaces/daemonsets-1025/pods","resourceVersion":"98375"},"items":null} + +[AfterEach] [sig-apps] Daemon set [Serial] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Jun 18 12:32:34.045: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "daemonsets-1025" for this suite. +Jun 18 12:32:42.083: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:32:42.940: INFO: namespace daemonsets-1025 deletion completed in 8.883051314s + +• [SLOW TEST:41.011 seconds] +[sig-apps] Daemon set [Serial] +/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22 + should rollback without unnecessary restarts [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 ------------------------------ -SSSSSSSSS +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] ConfigMap - should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 + optional updates should be reflected in volume [NodeConformance] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 [BeforeEach] [sig-storage] ConfigMap - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 STEP: Creating a kubernetes client -Jun 4 19:19:28.470: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504 +Jun 18 12:32:42.941: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150 STEP: Building a namespace api object, basename configmap -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in configmap-3805 +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in configmap-6245 STEP: Waiting for a default service account to be provisioned in namespace -[It] should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -STEP: Creating configMap with name configmap-test-volume-ac4a29c3-86fd-11e9-8d1b-467ee19922ac -STEP: Creating a pod to test consume configMaps -Jun 4 19:19:28.725: INFO: Waiting up to 5m0s for pod "pod-configmaps-ac4c7904-86fd-11e9-8d1b-467ee19922ac" in namespace "configmap-3805" to be "success or failure" -Jun 4 19:19:28.742: INFO: Pod "pod-configmaps-ac4c7904-86fd-11e9-8d1b-467ee19922ac": Phase="Pending", Reason="", readiness=false. Elapsed: 16.795501ms -Jun 4 19:19:30.756: INFO: Pod "pod-configmaps-ac4c7904-86fd-11e9-8d1b-467ee19922ac": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.030967016s -STEP: Saw pod success -Jun 4 19:19:30.756: INFO: Pod "pod-configmaps-ac4c7904-86fd-11e9-8d1b-467ee19922ac" satisfied condition "success or failure" -Jun 4 19:19:30.768: INFO: Trying to get logs from node 10.212.23.161 pod pod-configmaps-ac4c7904-86fd-11e9-8d1b-467ee19922ac container configmap-volume-test: -STEP: delete the pod -Jun 4 19:19:30.893: INFO: Waiting for pod pod-configmaps-ac4c7904-86fd-11e9-8d1b-467ee19922ac to disappear -Jun 4 19:19:30.914: INFO: Pod pod-configmaps-ac4c7904-86fd-11e9-8d1b-467ee19922ac no longer exists +[It] optional updates should be reflected in volume [NodeConformance] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: Creating configMap with name cm-test-opt-del-2b3d1ab5-91c5-11e9-a25d-8608290c688a +STEP: Creating configMap with name cm-test-opt-upd-2b3d1b06-91c5-11e9-a25d-8608290c688a +STEP: Creating the pod +STEP: Deleting configmap cm-test-opt-del-2b3d1ab5-91c5-11e9-a25d-8608290c688a +STEP: Updating configmap cm-test-opt-upd-2b3d1b06-91c5-11e9-a25d-8608290c688a +STEP: Creating configMap with name cm-test-opt-create-2b3d1b30-91c5-11e9-a25d-8608290c688a +STEP: waiting to observe update in volume [AfterEach] [sig-storage] ConfigMap - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 -Jun 4 19:19:30.914: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "configmap-3805" for this suite. -Jun 4 19:19:36.972: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -Jun 4 19:19:37.577: INFO: namespace configmap-3805 deletion completed in 6.648696726s + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Jun 18 12:34:03.846: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "configmap-6245" for this suite. +Jun 18 12:34:27.887: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:34:28.342: INFO: namespace configmap-6245 deletion completed in 24.481309421s -• [SLOW TEST:9.107 seconds] +• [SLOW TEST:105.401 seconds] [sig-storage] ConfigMap -/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:32 - should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:32 + optional updates should be reflected in volume [NodeConformance] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 ------------------------------ -[sig-storage] EmptyDir volumes - should support (non-root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -[BeforeEach] [sig-storage] EmptyDir volumes - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +SSSSS +------------------------------ +[sig-storage] EmptyDir wrapper volumes + should not conflict [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [sig-storage] EmptyDir wrapper volumes + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 STEP: Creating a kubernetes client -Jun 4 19:19:37.578: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504 -STEP: Building a namespace api object, basename emptydir -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in emptydir-2678 +Jun 18 12:34:28.343: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150 +STEP: Building a namespace api object, basename emptydir-wrapper +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in emptydir-wrapper-7799 STEP: Waiting for a default service account to be provisioned in namespace -[It] should support (non-root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -STEP: Creating a pod to test emptydir 0777 on tmpfs -Jun 4 19:19:37.955: INFO: Waiting up to 5m0s for pod "pod-b1c75267-86fd-11e9-8d1b-467ee19922ac" in namespace "emptydir-2678" to be "success or failure" -Jun 4 19:19:37.967: INFO: Pod "pod-b1c75267-86fd-11e9-8d1b-467ee19922ac": Phase="Pending", Reason="", readiness=false. Elapsed: 11.770319ms -Jun 4 19:19:39.979: INFO: Pod "pod-b1c75267-86fd-11e9-8d1b-467ee19922ac": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.023389805s -STEP: Saw pod success -Jun 4 19:19:39.979: INFO: Pod "pod-b1c75267-86fd-11e9-8d1b-467ee19922ac" satisfied condition "success or failure" -Jun 4 19:19:40.078: INFO: Trying to get logs from node 10.212.23.164 pod pod-b1c75267-86fd-11e9-8d1b-467ee19922ac container test-container: -STEP: delete the pod -Jun 4 19:19:40.143: INFO: Waiting for pod pod-b1c75267-86fd-11e9-8d1b-467ee19922ac to disappear -Jun 4 19:19:40.154: INFO: Pod pod-b1c75267-86fd-11e9-8d1b-467ee19922ac no longer exists -[AfterEach] [sig-storage] EmptyDir volumes - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 -Jun 4 19:19:40.154: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "emptydir-2678" for this suite. -Jun 4 19:19:46.214: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -Jun 4 19:19:46.597: INFO: namespace emptydir-2678 deletion completed in 6.425108731s +[It] should not conflict [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: Cleaning up the secret +STEP: Cleaning up the configmap +STEP: Cleaning up the pod +[AfterEach] [sig-storage] EmptyDir wrapper volumes + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Jun 18 12:34:30.742: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "emptydir-wrapper-7799" for this suite. +Jun 18 12:34:36.781: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:34:37.184: INFO: namespace emptydir-wrapper-7799 deletion completed in 6.428443989s -• [SLOW TEST:9.019 seconds] -[sig-storage] EmptyDir volumes -/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:41 - should support (non-root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +• [SLOW TEST:8.841 seconds] +[sig-storage] EmptyDir wrapper volumes +/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:22 + should not conflict [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 ------------------------------ -SSSSSS +SSSS ------------------------------ -[k8s.io] Kubelet when scheduling a busybox command in a pod - should print the output to logs [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -[BeforeEach] [k8s.io] Kubelet - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +[sig-cli] Kubectl client [k8s.io] Kubectl expose + should create services for rc [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [sig-cli] Kubectl client + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 STEP: Creating a kubernetes client -Jun 4 19:19:46.597: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504 -STEP: Building a namespace api object, basename kubelet-test -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubelet-test-2310 +Jun 18 12:34:37.185: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150 +STEP: Building a namespace api object, basename kubectl +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-3191 STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [k8s.io] Kubelet - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:37 -[It] should print the output to logs [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -[AfterEach] [k8s.io] Kubelet - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 -Jun 4 19:19:49.038: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "kubelet-test-2310" for this suite. -Jun 4 19:20:39.098: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -Jun 4 19:20:39.489: INFO: namespace kubelet-test-2310 deletion completed in 50.435848967s +[BeforeEach] [sig-cli] Kubectl client + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:213 +[It] should create services for rc [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: creating Redis RC +Jun 18 12:34:37.386: INFO: namespace kubectl-3191 +Jun 18 12:34:37.387: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 create -f - --namespace=kubectl-3191' +Jun 18 12:34:37.619: INFO: stderr: "" +Jun 18 12:34:37.619: INFO: stdout: "replicationcontroller/redis-master created\n" +STEP: Waiting for Redis master to start. +Jun 18 12:34:38.635: INFO: Selector matched 1 pods for map[app:redis] +Jun 18 12:34:38.635: INFO: Found 0 / 1 +Jun 18 12:34:39.635: INFO: Selector matched 1 pods for map[app:redis] +Jun 18 12:34:39.635: INFO: Found 0 / 1 +Jun 18 12:34:40.633: INFO: Selector matched 1 pods for map[app:redis] +Jun 18 12:34:40.633: INFO: Found 1 / 1 +Jun 18 12:34:40.633: INFO: WaitFor completed with timeout 5m0s. Pods found = 1 out of 1 +Jun 18 12:34:40.649: INFO: Selector matched 1 pods for map[app:redis] +Jun 18 12:34:40.649: INFO: ForEach: Found 1 pods from the filter. Now looping through them. +Jun 18 12:34:40.649: INFO: wait on redis-master startup in kubectl-3191 +Jun 18 12:34:40.649: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 logs redis-master-9gjtb redis-master --namespace=kubectl-3191' +Jun 18 12:34:40.804: INFO: stderr: "" +Jun 18 12:34:40.804: INFO: stdout: " _._ \n _.-``__ ''-._ \n _.-`` `. `_. ''-._ Redis 3.2.12 (35a5711f/0) 64 bit\n .-`` .-```. ```\\/ _.,_ ''-._ \n ( ' , .-` | `, ) Running in standalone mode\n |`-._`-...-` __...-.``-._|'` _.-'| Port: 6379\n | `-._ `._ / _.-' | PID: 1\n `-._ `-._ `-./ _.-' _.-' \n |`-._`-._ `-.__.-' _.-'_.-'| \n | `-._`-._ _.-'_.-' | http://redis.io \n `-._ `-._`-.__.-'_.-' _.-' \n |`-._`-._ `-.__.-' _.-'_.-'| \n | `-._`-._ _.-'_.-' | \n `-._ `-._`-.__.-'_.-' _.-' \n `-._ `-.__.-' _.-' \n `-._ _.-' \n `-.__.-' \n\n1:M 18 Jun 12:34:38.966 # WARNING: The TCP backlog setting of 511 cannot be enforced because /proc/sys/net/core/somaxconn is set to the lower value of 128.\n1:M 18 Jun 12:34:38.966 # Server started, Redis version 3.2.12\n1:M 18 Jun 12:34:38.966 # WARNING you have Transparent Huge Pages (THP) support enabled in your kernel. This will create latency and memory usage issues with Redis. To fix this issue run the command 'echo never > /sys/kernel/mm/transparent_hugepage/enabled' as root, and add it to your /etc/rc.local in order to retain the setting after a reboot. Redis must be restarted after THP is disabled.\n1:M 18 Jun 12:34:38.966 * The server is now ready to accept connections on port 6379\n" +STEP: exposing RC +Jun 18 12:34:40.804: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 expose rc redis-master --name=rm2 --port=1234 --target-port=6379 --namespace=kubectl-3191' +Jun 18 12:34:41.003: INFO: stderr: "" +Jun 18 12:34:41.003: INFO: stdout: "service/rm2 exposed\n" +Jun 18 12:34:41.010: INFO: Service rm2 in namespace kubectl-3191 found. +STEP: exposing service +Jun 18 12:34:43.029: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 expose service rm2 --name=rm3 --port=2345 --target-port=6379 --namespace=kubectl-3191' +Jun 18 12:34:43.504: INFO: stderr: "" +Jun 18 12:34:43.504: INFO: stdout: "service/rm3 exposed\n" +Jun 18 12:34:43.513: INFO: Service rm3 in namespace kubectl-3191 found. +[AfterEach] [sig-cli] Kubectl client + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Jun 18 12:34:45.528: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "kubectl-3191" for this suite. +Jun 18 12:35:09.574: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:35:09.974: INFO: namespace kubectl-3191 deletion completed in 24.431040164s -• [SLOW TEST:52.893 seconds] -[k8s.io] Kubelet -/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687 - when scheduling a busybox command in a pod - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:40 - should print the output to logs [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +• [SLOW TEST:32.790 seconds] +[sig-cli] Kubectl client +/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 + [k8s.io] Kubectl expose + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687 + should create services for rc [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 ------------------------------ -SSSSSSSSSSSS +SSSSSSSSSSSSSSSSS ------------------------------ -[k8s.io] Variable Expansion - should allow substituting values in a container's args [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -[BeforeEach] [k8s.io] Variable Expansion - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +[k8s.io] Probing container + should have monotonically increasing restart count [NodeConformance] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [k8s.io] Probing container + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 STEP: Creating a kubernetes client -Jun 4 19:20:39.490: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504 -STEP: Building a namespace api object, basename var-expansion -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in var-expansion-3087 +Jun 18 12:35:09.977: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150 +STEP: Building a namespace api object, basename container-probe +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in container-probe-4624 STEP: Waiting for a default service account to be provisioned in namespace -[It] should allow substituting values in a container's args [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -STEP: Creating a pod to test substitution in container's args -Jun 4 19:20:39.779: INFO: Waiting up to 5m0s for pod "var-expansion-d6a67eb3-86fd-11e9-8d1b-467ee19922ac" in namespace "var-expansion-3087" to be "success or failure" -Jun 4 19:20:39.790: INFO: Pod "var-expansion-d6a67eb3-86fd-11e9-8d1b-467ee19922ac": Phase="Pending", Reason="", readiness=false. Elapsed: 11.466693ms -Jun 4 19:20:41.803: INFO: Pod "var-expansion-d6a67eb3-86fd-11e9-8d1b-467ee19922ac": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.024026449s -STEP: Saw pod success -Jun 4 19:20:41.803: INFO: Pod "var-expansion-d6a67eb3-86fd-11e9-8d1b-467ee19922ac" satisfied condition "success or failure" -Jun 4 19:20:41.815: INFO: Trying to get logs from node 10.212.23.189 pod var-expansion-d6a67eb3-86fd-11e9-8d1b-467ee19922ac container dapi-container: -STEP: delete the pod -Jun 4 19:20:41.893: INFO: Waiting for pod var-expansion-d6a67eb3-86fd-11e9-8d1b-467ee19922ac to disappear -Jun 4 19:20:41.905: INFO: Pod var-expansion-d6a67eb3-86fd-11e9-8d1b-467ee19922ac no longer exists -[AfterEach] [k8s.io] Variable Expansion - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 -Jun 4 19:20:41.905: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "var-expansion-3087" for this suite. -Jun 4 19:20:47.965: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -Jun 4 19:20:48.400: INFO: namespace var-expansion-3087 deletion completed in 6.479666956s +[BeforeEach] [k8s.io] Probing container + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/container_probe.go:51 +[It] should have monotonically increasing restart count [NodeConformance] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: Creating pod liveness-http in namespace container-probe-4624 +Jun 18 12:35:16.318: INFO: Started pod liveness-http in namespace container-probe-4624 +STEP: checking the pod's current state and verifying that restartCount is present +Jun 18 12:35:16.331: INFO: Initial restart count of pod liveness-http is 0 +Jun 18 12:35:33.434: INFO: Restart count of pod container-probe-4624/liveness-http is now 1 (17.103003078s elapsed) +Jun 18 12:35:53.861: INFO: Restart count of pod container-probe-4624/liveness-http is now 2 (37.530436606s elapsed) +Jun 18 12:36:12.830: INFO: Restart count of pod container-probe-4624/liveness-http is now 3 (56.498612908s elapsed) +Jun 18 12:36:32.983: INFO: Restart count of pod container-probe-4624/liveness-http is now 4 (1m16.652049452s elapsed) +Jun 18 12:37:35.567: INFO: Restart count of pod container-probe-4624/liveness-http is now 5 (2m19.236179964s elapsed) +STEP: deleting the pod +[AfterEach] [k8s.io] Probing container + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Jun 18 12:37:35.620: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "container-probe-4624" for this suite. +Jun 18 12:37:41.667: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:37:42.056: INFO: namespace container-probe-4624 deletion completed in 6.415080222s -• [SLOW TEST:8.910 seconds] -[k8s.io] Variable Expansion -/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687 - should allow substituting values in a container's args [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +• [SLOW TEST:152.080 seconds] +[k8s.io] Probing container +/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687 + should have monotonically increasing restart count [NodeConformance] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSS +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-storage] Projected downwardAPI - should provide container's cpu request [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -[BeforeEach] [sig-storage] Projected downwardAPI - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +[sig-network] Services + should serve multiport endpoints from pods [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [sig-network] Services + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 STEP: Creating a kubernetes client -Jun 4 19:20:48.400: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504 -STEP: Building a namespace api object, basename projected -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-5784 +Jun 18 12:37:42.060: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150 +STEP: Building a namespace api object, basename services +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in services-7024 STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [sig-storage] Projected downwardAPI - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:39 -[It] should provide container's cpu request [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -STEP: Creating a pod to test downward API volume plugin -Jun 4 19:20:48.891: INFO: Waiting up to 5m0s for pod "downwardapi-volume-dbf49029-86fd-11e9-8d1b-467ee19922ac" in namespace "projected-5784" to be "success or failure" -Jun 4 19:20:48.902: INFO: Pod "downwardapi-volume-dbf49029-86fd-11e9-8d1b-467ee19922ac": Phase="Pending", Reason="", readiness=false. Elapsed: 11.616651ms -Jun 4 19:20:50.917: INFO: Pod "downwardapi-volume-dbf49029-86fd-11e9-8d1b-467ee19922ac": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.026276289s -STEP: Saw pod success -Jun 4 19:20:50.917: INFO: Pod "downwardapi-volume-dbf49029-86fd-11e9-8d1b-467ee19922ac" satisfied condition "success or failure" -Jun 4 19:20:50.929: INFO: Trying to get logs from node 10.212.23.164 pod downwardapi-volume-dbf49029-86fd-11e9-8d1b-467ee19922ac container client-container: -STEP: delete the pod -Jun 4 19:20:50.998: INFO: Waiting for pod downwardapi-volume-dbf49029-86fd-11e9-8d1b-467ee19922ac to disappear -Jun 4 19:20:51.008: INFO: Pod downwardapi-volume-dbf49029-86fd-11e9-8d1b-467ee19922ac no longer exists -[AfterEach] [sig-storage] Projected downwardAPI - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 -Jun 4 19:20:51.008: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "projected-5784" for this suite. -Jun 4 19:20:57.064: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -Jun 4 19:20:57.401: INFO: namespace projected-5784 deletion completed in 6.378382134s +[BeforeEach] [sig-network] Services + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:86 +[It] should serve multiport endpoints from pods [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: creating service multi-endpoint-test in namespace services-7024 +STEP: waiting up to 3m0s for service multi-endpoint-test in namespace services-7024 to expose endpoints map[] +Jun 18 12:37:42.302: INFO: successfully validated that service multi-endpoint-test in namespace services-7024 exposes endpoints map[] (8.120249ms elapsed) +STEP: Creating pod pod1 in namespace services-7024 +STEP: waiting up to 3m0s for service multi-endpoint-test in namespace services-7024 to expose endpoints map[pod1:[100]] +Jun 18 12:37:45.422: INFO: successfully validated that service multi-endpoint-test in namespace services-7024 exposes endpoints map[pod1:[100]] (3.094186904s elapsed) +STEP: Creating pod pod2 in namespace services-7024 +STEP: waiting up to 3m0s for service multi-endpoint-test in namespace services-7024 to expose endpoints map[pod1:[100] pod2:[101]] +Jun 18 12:37:47.553: INFO: successfully validated that service multi-endpoint-test in namespace services-7024 exposes endpoints map[pod1:[100] pod2:[101]] (2.11600532s elapsed) +STEP: Deleting pod pod1 in namespace services-7024 +STEP: waiting up to 3m0s for service multi-endpoint-test in namespace services-7024 to expose endpoints map[pod2:[101]] +Jun 18 12:37:47.606: INFO: successfully validated that service multi-endpoint-test in namespace services-7024 exposes endpoints map[pod2:[101]] (27.960653ms elapsed) +STEP: Deleting pod pod2 in namespace services-7024 +STEP: waiting up to 3m0s for service multi-endpoint-test in namespace services-7024 to expose endpoints map[] +Jun 18 12:37:47.639: INFO: successfully validated that service multi-endpoint-test in namespace services-7024 exposes endpoints map[] (6.129941ms elapsed) +[AfterEach] [sig-network] Services + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Jun 18 12:37:47.685: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "services-7024" for this suite. +Jun 18 12:38:11.727: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:38:12.197: INFO: namespace services-7024 deletion completed in 24.496730823s +[AfterEach] [sig-network] Services + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:91 -• [SLOW TEST:9.001 seconds] -[sig-storage] Projected downwardAPI -/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:33 - should provide container's cpu request [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +• [SLOW TEST:30.137 seconds] +[sig-network] Services +/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:22 + should serve multiport endpoints from pods [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +SSSSSSSSSS ------------------------------ [k8s.io] Pods - should contain environment variables for services [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 + should support remote command execution over websockets [NodeConformance] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 [BeforeEach] [k8s.io] Pods - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 STEP: Creating a kubernetes client -Jun 4 19:20:57.402: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504 +Jun 18 12:38:12.197: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150 STEP: Building a namespace api object, basename pods -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in pods-610 +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in pods-790 STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Pods - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:135 -[It] should contain environment variables for services [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -Jun 4 19:20:59.756: INFO: Waiting up to 5m0s for pod "client-envvars-e28be41b-86fd-11e9-8d1b-467ee19922ac" in namespace "pods-610" to be "success or failure" -Jun 4 19:20:59.766: INFO: Pod "client-envvars-e28be41b-86fd-11e9-8d1b-467ee19922ac": Phase="Pending", Reason="", readiness=false. Elapsed: 10.598358ms -Jun 4 19:21:01.787: INFO: Pod "client-envvars-e28be41b-86fd-11e9-8d1b-467ee19922ac": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.03129987s -STEP: Saw pod success -Jun 4 19:21:01.787: INFO: Pod "client-envvars-e28be41b-86fd-11e9-8d1b-467ee19922ac" satisfied condition "success or failure" -Jun 4 19:21:01.805: INFO: Trying to get logs from node 10.212.23.189 pod client-envvars-e28be41b-86fd-11e9-8d1b-467ee19922ac container env3cont: -STEP: delete the pod -Jun 4 19:21:01.884: INFO: Waiting for pod client-envvars-e28be41b-86fd-11e9-8d1b-467ee19922ac to disappear -Jun 4 19:21:01.894: INFO: Pod client-envvars-e28be41b-86fd-11e9-8d1b-467ee19922ac no longer exists + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:135 +[It] should support remote command execution over websockets [NodeConformance] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +Jun 18 12:38:12.398: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150 +STEP: creating the pod +STEP: submitting the pod to kubernetes [AfterEach] [k8s.io] Pods - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 -Jun 4 19:21:01.894: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "pods-610" for this suite. -Jun 4 19:21:46.155: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -Jun 4 19:21:46.594: INFO: namespace pods-610 deletion completed in 44.681637549s + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Jun 18 12:38:16.710: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "pods-790" for this suite. +Jun 18 12:39:02.763: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:39:03.264: INFO: namespace pods-790 deletion completed in 46.52980902s -• [SLOW TEST:49.193 seconds] +• [SLOW TEST:51.067 seconds] [k8s.io] Pods -/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687 - should contain environment variables for services [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687 + should support remote command execution over websockets [NodeConformance] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 ------------------------------ -SSSSS +SSSS ------------------------------ -[k8s.io] Probing container - with readiness probe should not be ready before initial delay and never restart [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -[BeforeEach] [k8s.io] Probing container - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +[sig-storage] Projected configMap + should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [sig-storage] Projected configMap + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 STEP: Creating a kubernetes client -Jun 4 19:21:46.594: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504 -STEP: Building a namespace api object, basename container-probe -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in container-probe-6671 +Jun 18 12:39:03.264: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150 +STEP: Building a namespace api object, basename projected +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-9866 STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [k8s.io] Probing container - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/container_probe.go:51 -[It] with readiness probe should not be ready before initial delay and never restart [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -Jun 4 19:22:06.873: INFO: Container started at 2019-06-04 19:21:47 +0000 UTC, pod became ready at 2019-06-04 19:22:05 +0000 UTC -[AfterEach] [k8s.io] Probing container - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 -Jun 4 19:22:06.873: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "container-probe-6671" for this suite. -Jun 4 19:22:31.002: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -Jun 4 19:22:31.517: INFO: namespace container-probe-6671 deletion completed in 24.627740226s +[It] should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: Creating configMap with name projected-configmap-test-volume-map-0ded1975-91c6-11e9-a25d-8608290c688a +STEP: Creating a pod to test consume configMaps +Jun 18 12:39:03.515: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-0defdcb0-91c6-11e9-a25d-8608290c688a" in namespace "projected-9866" to be "success or failure" +Jun 18 12:39:03.528: INFO: Pod "pod-projected-configmaps-0defdcb0-91c6-11e9-a25d-8608290c688a": Phase="Pending", Reason="", readiness=false. Elapsed: 13.385576ms +Jun 18 12:39:05.547: INFO: Pod "pod-projected-configmaps-0defdcb0-91c6-11e9-a25d-8608290c688a": Phase="Running", Reason="", readiness=true. Elapsed: 2.031943012s +Jun 18 12:39:07.561: INFO: Pod "pod-projected-configmaps-0defdcb0-91c6-11e9-a25d-8608290c688a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.045881975s +STEP: Saw pod success +Jun 18 12:39:07.561: INFO: Pod "pod-projected-configmaps-0defdcb0-91c6-11e9-a25d-8608290c688a" satisfied condition "success or failure" +Jun 18 12:39:07.574: INFO: Trying to get logs from node 10.72.74.138 pod pod-projected-configmaps-0defdcb0-91c6-11e9-a25d-8608290c688a container projected-configmap-volume-test: +STEP: delete the pod +Jun 18 12:39:07.653: INFO: Waiting for pod pod-projected-configmaps-0defdcb0-91c6-11e9-a25d-8608290c688a to disappear +Jun 18 12:39:07.665: INFO: Pod pod-projected-configmaps-0defdcb0-91c6-11e9-a25d-8608290c688a no longer exists +[AfterEach] [sig-storage] Projected configMap + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Jun 18 12:39:07.665: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "projected-9866" for this suite. +Jun 18 12:39:13.721: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:39:14.808: INFO: namespace projected-9866 deletion completed in 7.12842628s -• [SLOW TEST:44.922 seconds] -[k8s.io] Probing container -/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687 - with readiness probe should not be ready before initial delay and never restart [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +• [SLOW TEST:11.544 seconds] +[sig-storage] Projected configMap +/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_configmap.go:33 + should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 ------------------------------ -SSS +SSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-storage] Subpath Atomic writer volumes - should support subpaths with secret pod [LinuxOnly] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -[BeforeEach] [sig-storage] Subpath - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +[sig-storage] HostPath + should give a volume the correct mode [LinuxOnly] [NodeConformance] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [sig-storage] HostPath + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 STEP: Creating a kubernetes client -Jun 4 19:22:31.517: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504 -STEP: Building a namespace api object, basename subpath -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in subpath-7318 +Jun 18 12:39:14.810: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150 +STEP: Building a namespace api object, basename hostpath +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in hostpath-4185 STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] Atomic writer volumes - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:38 -STEP: Setting up data -[It] should support subpaths with secret pod [LinuxOnly] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -STEP: Creating pod pod-subpath-test-secret-2dbd -STEP: Creating a pod to test atomic-volume-subpath -Jun 4 19:22:31.774: INFO: Waiting up to 5m0s for pod "pod-subpath-test-secret-2dbd" in namespace "subpath-7318" to be "success or failure" -Jun 4 19:22:31.788: INFO: Pod "pod-subpath-test-secret-2dbd": Phase="Pending", Reason="", readiness=false. Elapsed: 14.264008ms -Jun 4 19:22:33.800: INFO: Pod "pod-subpath-test-secret-2dbd": Phase="Running", Reason="", readiness=true. Elapsed: 2.026208572s -Jun 4 19:22:35.813: INFO: Pod "pod-subpath-test-secret-2dbd": Phase="Running", Reason="", readiness=true. Elapsed: 4.039242292s -Jun 4 19:22:37.825: INFO: Pod "pod-subpath-test-secret-2dbd": Phase="Running", Reason="", readiness=true. Elapsed: 6.05069228s -Jun 4 19:22:39.839: INFO: Pod "pod-subpath-test-secret-2dbd": Phase="Running", Reason="", readiness=true. Elapsed: 8.064985756s -Jun 4 19:22:41.853: INFO: Pod "pod-subpath-test-secret-2dbd": Phase="Running", Reason="", readiness=true. Elapsed: 10.078630962s -Jun 4 19:22:43.866: INFO: Pod "pod-subpath-test-secret-2dbd": Phase="Running", Reason="", readiness=true. Elapsed: 12.092368658s -Jun 4 19:22:45.878: INFO: Pod "pod-subpath-test-secret-2dbd": Phase="Running", Reason="", readiness=true. Elapsed: 14.104391263s -Jun 4 19:22:47.914: INFO: Pod "pod-subpath-test-secret-2dbd": Phase="Running", Reason="", readiness=true. Elapsed: 16.140364181s -Jun 4 19:22:49.928: INFO: Pod "pod-subpath-test-secret-2dbd": Phase="Running", Reason="", readiness=true. Elapsed: 18.154339632s -Jun 4 19:22:51.942: INFO: Pod "pod-subpath-test-secret-2dbd": Phase="Running", Reason="", readiness=true. Elapsed: 20.168340852s -Jun 4 19:22:53.955: INFO: Pod "pod-subpath-test-secret-2dbd": Phase="Succeeded", Reason="", readiness=false. Elapsed: 22.181077708s +[BeforeEach] [sig-storage] HostPath + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/host_path.go:37 +[It] should give a volume the correct mode [LinuxOnly] [NodeConformance] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: Creating a pod to test hostPath mode +Jun 18 12:39:15.032: INFO: Waiting up to 5m0s for pod "pod-host-path-test" in namespace "hostpath-4185" to be "success or failure" +Jun 18 12:39:15.046: INFO: Pod "pod-host-path-test": Phase="Pending", Reason="", readiness=false. Elapsed: 13.638317ms +Jun 18 12:39:17.060: INFO: Pod "pod-host-path-test": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.027216513s STEP: Saw pod success -Jun 4 19:22:53.955: INFO: Pod "pod-subpath-test-secret-2dbd" satisfied condition "success or failure" -Jun 4 19:22:53.966: INFO: Trying to get logs from node 10.212.23.189 pod pod-subpath-test-secret-2dbd container test-container-subpath-secret-2dbd: +Jun 18 12:39:17.060: INFO: Pod "pod-host-path-test" satisfied condition "success or failure" +Jun 18 12:39:17.073: INFO: Trying to get logs from node 10.72.74.184 pod pod-host-path-test container test-container-1: STEP: delete the pod -Jun 4 19:22:54.093: INFO: Waiting for pod pod-subpath-test-secret-2dbd to disappear -Jun 4 19:22:54.105: INFO: Pod pod-subpath-test-secret-2dbd no longer exists -STEP: Deleting pod pod-subpath-test-secret-2dbd -Jun 4 19:22:54.105: INFO: Deleting pod "pod-subpath-test-secret-2dbd" in namespace "subpath-7318" -[AfterEach] [sig-storage] Subpath - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 -Jun 4 19:22:54.116: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "subpath-7318" for this suite. -Jun 4 19:23:00.175: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -Jun 4 19:23:00.585: INFO: namespace subpath-7318 deletion completed in 6.454269602s +Jun 18 12:39:17.148: INFO: Waiting for pod pod-host-path-test to disappear +Jun 18 12:39:17.163: INFO: Pod pod-host-path-test no longer exists +[AfterEach] [sig-storage] HostPath + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Jun 18 12:39:17.164: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "hostpath-4185" for this suite. +Jun 18 12:39:23.878: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:39:24.298: INFO: namespace hostpath-4185 deletion completed in 7.120247408s -• [SLOW TEST:29.068 seconds] -[sig-storage] Subpath -/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:22 - Atomic writer volumes - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:34 - should support subpaths with secret pod [LinuxOnly] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +• [SLOW TEST:9.489 seconds] +[sig-storage] HostPath +/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/host_path.go:34 + should give a volume the correct mode [LinuxOnly] [NodeConformance] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +------------------------------ +S +------------------------------ +[sig-network] Networking Granular Checks: Pods + should function for node-pod communication: udp [LinuxOnly] [NodeConformance] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [sig-network] Networking + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +STEP: Creating a kubernetes client +Jun 18 12:39:24.298: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150 +STEP: Building a namespace api object, basename pod-network-test +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in pod-network-test-1015 +STEP: Waiting for a default service account to be provisioned in namespace +[It] should function for node-pod communication: udp [LinuxOnly] [NodeConformance] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: Performing setup for networking test in namespace pod-network-test-1015 +STEP: creating a selector +STEP: Creating the service pods in kubernetes +Jun 18 12:39:24.510: INFO: Waiting up to 10m0s for all (but 0) nodes to be schedulable +STEP: Creating test pods +Jun 18 12:39:44.806: INFO: ExecWithOptions {Command:[/bin/sh -c echo hostName | nc -w 1 -u 172.30.142.42 8081 | grep -v '^\s*$'] Namespace:pod-network-test-1015 PodName:host-test-container-pod ContainerName:hostexec Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} +Jun 18 12:39:44.806: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150 +Jun 18 12:39:46.034: INFO: Found all expected endpoints: [netserver-0] +Jun 18 12:39:46.046: INFO: ExecWithOptions {Command:[/bin/sh -c echo hostName | nc -w 1 -u 172.30.206.155 8081 | grep -v '^\s*$'] Namespace:pod-network-test-1015 PodName:host-test-container-pod ContainerName:hostexec Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} +Jun 18 12:39:46.046: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150 +Jun 18 12:39:48.527: INFO: Found all expected endpoints: [netserver-1] +Jun 18 12:39:48.540: INFO: ExecWithOptions {Command:[/bin/sh -c echo hostName | nc -w 1 -u 172.30.80.155 8081 | grep -v '^\s*$'] Namespace:pod-network-test-1015 PodName:host-test-container-pod ContainerName:hostexec Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} +Jun 18 12:39:48.540: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150 +Jun 18 12:39:49.841: INFO: Found all expected endpoints: [netserver-2] +[AfterEach] [sig-network] Networking + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Jun 18 12:39:49.841: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "pod-network-test-1015" for this suite. +Jun 18 12:40:13.888: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:40:14.334: INFO: namespace pod-network-test-1015 deletion completed in 24.473289895s + +• [SLOW TEST:50.036 seconds] +[sig-network] Networking +/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:25 + Granular Checks: Pods + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:28 + should function for node-pod communication: udp [LinuxOnly] [NodeConformance] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +------------------------------ +SSSSSSS ------------------------------ [sig-storage] EmptyDir volumes volume on tmpfs should have the correct mode [LinuxOnly] [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 [BeforeEach] [sig-storage] EmptyDir volumes - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 STEP: Creating a kubernetes client -Jun 4 19:23:00.585: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504 +Jun 18 12:40:14.336: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150 STEP: Building a namespace api object, basename emptydir -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in emptydir-1366 +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in emptydir-3101 STEP: Waiting for a default service account to be provisioned in namespace [It] volume on tmpfs should have the correct mode [LinuxOnly] [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 STEP: Creating a pod to test emptydir volume type on tmpfs -Jun 4 19:23:00.825: INFO: Waiting up to 5m0s for pod "pod-2ab85f99-86fe-11e9-8d1b-467ee19922ac" in namespace "emptydir-1366" to be "success or failure" -Jun 4 19:23:00.837: INFO: Pod "pod-2ab85f99-86fe-11e9-8d1b-467ee19922ac": Phase="Pending", Reason="", readiness=false. Elapsed: 11.680388ms -Jun 4 19:23:02.850: INFO: Pod "pod-2ab85f99-86fe-11e9-8d1b-467ee19922ac": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.024219294s +Jun 18 12:40:14.566: INFO: Waiting up to 5m0s for pod "pod-3847db3b-91c6-11e9-a25d-8608290c688a" in namespace "emptydir-3101" to be "success or failure" +Jun 18 12:40:14.588: INFO: Pod "pod-3847db3b-91c6-11e9-a25d-8608290c688a": Phase="Pending", Reason="", readiness=false. Elapsed: 21.462068ms +Jun 18 12:40:16.607: INFO: Pod "pod-3847db3b-91c6-11e9-a25d-8608290c688a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.040565275s STEP: Saw pod success -Jun 4 19:23:02.850: INFO: Pod "pod-2ab85f99-86fe-11e9-8d1b-467ee19922ac" satisfied condition "success or failure" -Jun 4 19:23:02.861: INFO: Trying to get logs from node 10.212.23.161 pod pod-2ab85f99-86fe-11e9-8d1b-467ee19922ac container test-container: +Jun 18 12:40:16.607: INFO: Pod "pod-3847db3b-91c6-11e9-a25d-8608290c688a" satisfied condition "success or failure" +Jun 18 12:40:16.620: INFO: Trying to get logs from node 10.72.74.138 pod pod-3847db3b-91c6-11e9-a25d-8608290c688a container test-container: STEP: delete the pod -Jun 4 19:23:02.955: INFO: Waiting for pod pod-2ab85f99-86fe-11e9-8d1b-467ee19922ac to disappear -Jun 4 19:23:02.966: INFO: Pod pod-2ab85f99-86fe-11e9-8d1b-467ee19922ac no longer exists +Jun 18 12:40:16.695: INFO: Waiting for pod pod-3847db3b-91c6-11e9-a25d-8608290c688a to disappear +Jun 18 12:40:16.708: INFO: Pod pod-3847db3b-91c6-11e9-a25d-8608290c688a no longer exists [AfterEach] [sig-storage] EmptyDir volumes - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 -Jun 4 19:23:02.966: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "emptydir-1366" for this suite. -Jun 4 19:23:09.038: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -Jun 4 19:23:09.392: INFO: namespace emptydir-1366 deletion completed in 6.411470414s + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Jun 18 12:40:16.708: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "emptydir-3101" for this suite. +Jun 18 12:40:24.747: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:40:25.982: INFO: namespace emptydir-3101 deletion completed in 9.259895751s -• [SLOW TEST:8.807 seconds] +• [SLOW TEST:11.646 seconds] [sig-storage] EmptyDir volumes -/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:41 +/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:41 volume on tmpfs should have the correct mode [LinuxOnly] [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 ------------------------------- -SSSSSSSSSSSSSSSSSSS ------------------------------- -[sig-storage] Downward API volume - should update annotations on modification [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -[BeforeEach] [sig-storage] Downward API volume - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 -STEP: Creating a kubernetes client -Jun 4 19:23:09.392: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504 -STEP: Building a namespace api object, basename downward-api -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in downward-api-7859 -STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [sig-storage] Downward API volume - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:39 -[It] should update annotations on modification [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -STEP: Creating the pod -Jun 4 19:23:12.282: INFO: Successfully updated pod "annotationupdate2fff55c3-86fe-11e9-8d1b-467ee19922ac" -[AfterEach] [sig-storage] Downward API volume - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 -Jun 4 19:23:14.337: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "downward-api-7859" for this suite. -Jun 4 19:23:38.395: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -Jun 4 19:23:38.790: INFO: namespace downward-api-7859 deletion completed in 24.437686762s - -• [SLOW TEST:29.398 seconds] -[sig-storage] Downward API volume -/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:34 - should update annotations on modification [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSSS +SSS ------------------------------ -[sig-storage] Projected secret - should be consumable in multiple volumes in a pod [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -[BeforeEach] [sig-storage] Projected secret - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +[sig-storage] Projected configMap + should be consumable from pods in volume with mappings [NodeConformance] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [sig-storage] Projected configMap + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 STEP: Creating a kubernetes client -Jun 4 19:23:38.792: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504 +Jun 18 12:40:25.982: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150 STEP: Building a namespace api object, basename projected -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-3672 +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-12 STEP: Waiting for a default service account to be provisioned in namespace -[It] should be consumable in multiple volumes in a pod [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -STEP: Creating secret with name projected-secret-test-417e0c6f-86fe-11e9-8d1b-467ee19922ac -STEP: Creating a pod to test consume secrets -Jun 4 19:23:39.041: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-417f70ad-86fe-11e9-8d1b-467ee19922ac" in namespace "projected-3672" to be "success or failure" -Jun 4 19:23:39.053: INFO: Pod "pod-projected-secrets-417f70ad-86fe-11e9-8d1b-467ee19922ac": Phase="Pending", Reason="", readiness=false. Elapsed: 12.076877ms -Jun 4 19:23:41.066: INFO: Pod "pod-projected-secrets-417f70ad-86fe-11e9-8d1b-467ee19922ac": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.024812572s +[It] should be consumable from pods in volume with mappings [NodeConformance] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: Creating configMap with name projected-configmap-test-volume-map-3f38cd50-91c6-11e9-a25d-8608290c688a +STEP: Creating a pod to test consume configMaps +Jun 18 12:40:26.215: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-3f3b09ae-91c6-11e9-a25d-8608290c688a" in namespace "projected-12" to be "success or failure" +Jun 18 12:40:26.229: INFO: Pod "pod-projected-configmaps-3f3b09ae-91c6-11e9-a25d-8608290c688a": Phase="Pending", Reason="", readiness=false. Elapsed: 13.640317ms +Jun 18 12:40:28.243: INFO: Pod "pod-projected-configmaps-3f3b09ae-91c6-11e9-a25d-8608290c688a": Phase="Pending", Reason="", readiness=false. Elapsed: 2.028218669s +Jun 18 12:40:30.257: INFO: Pod "pod-projected-configmaps-3f3b09ae-91c6-11e9-a25d-8608290c688a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.042108987s STEP: Saw pod success -Jun 4 19:23:41.066: INFO: Pod "pod-projected-secrets-417f70ad-86fe-11e9-8d1b-467ee19922ac" satisfied condition "success or failure" -Jun 4 19:23:41.078: INFO: Trying to get logs from node 10.212.23.161 pod pod-projected-secrets-417f70ad-86fe-11e9-8d1b-467ee19922ac container secret-volume-test: +Jun 18 12:40:30.257: INFO: Pod "pod-projected-configmaps-3f3b09ae-91c6-11e9-a25d-8608290c688a" satisfied condition "success or failure" +Jun 18 12:40:30.270: INFO: Trying to get logs from node 10.72.74.184 pod pod-projected-configmaps-3f3b09ae-91c6-11e9-a25d-8608290c688a container projected-configmap-volume-test: STEP: delete the pod -Jun 4 19:23:41.139: INFO: Waiting for pod pod-projected-secrets-417f70ad-86fe-11e9-8d1b-467ee19922ac to disappear -Jun 4 19:23:41.149: INFO: Pod pod-projected-secrets-417f70ad-86fe-11e9-8d1b-467ee19922ac no longer exists -[AfterEach] [sig-storage] Projected secret - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 -Jun 4 19:23:41.149: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "projected-3672" for this suite. -Jun 4 19:23:47.210: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -Jun 4 19:23:47.589: INFO: namespace projected-3672 deletion completed in 6.424843801s - -• [SLOW TEST:8.797 seconds] -[sig-storage] Projected secret -/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_secret.go:33 - should be consumable in multiple volumes in a pod [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 ------------------------------- -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------- -[k8s.io] Kubelet when scheduling a read only busybox container - should not write to root filesystem [LinuxOnly] [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -[BeforeEach] [k8s.io] Kubelet - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 -STEP: Creating a kubernetes client -Jun 4 19:23:47.593: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504 -STEP: Building a namespace api object, basename kubelet-test -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubelet-test-6456 -STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [k8s.io] Kubelet - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:37 -[It] should not write to root filesystem [LinuxOnly] [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -[AfterEach] [k8s.io] Kubelet - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 -Jun 4 19:23:49.899: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "kubelet-test-6456" for this suite. -Jun 4 19:24:29.961: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -Jun 4 19:24:30.379: INFO: namespace kubelet-test-6456 deletion completed in 40.464711477s +Jun 18 12:40:30.346: INFO: Waiting for pod pod-projected-configmaps-3f3b09ae-91c6-11e9-a25d-8608290c688a to disappear +Jun 18 12:40:30.359: INFO: Pod pod-projected-configmaps-3f3b09ae-91c6-11e9-a25d-8608290c688a no longer exists +[AfterEach] [sig-storage] Projected configMap + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Jun 18 12:40:30.359: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "projected-12" for this suite. +Jun 18 12:40:36.397: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:40:36.835: INFO: namespace projected-12 deletion completed in 6.463651944s -• [SLOW TEST:42.786 seconds] -[k8s.io] Kubelet -/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687 - when scheduling a read only busybox container - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:187 - should not write to root filesystem [LinuxOnly] [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +• [SLOW TEST:10.853 seconds] +[sig-storage] Projected configMap +/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_configmap.go:33 + should be consumable from pods in volume with mappings [NodeConformance] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +SSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-cli] Kubectl client [k8s.io] Kubectl run job - should create a job from an image when restart is OnFailure [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -[BeforeEach] [sig-cli] Kubectl client - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +[sig-storage] EmptyDir volumes + should support (non-root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [sig-storage] EmptyDir volumes + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 STEP: Creating a kubernetes client -Jun 4 19:24:30.379: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504 -STEP: Building a namespace api object, basename kubectl -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-7464 +Jun 18 12:40:36.836: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150 +STEP: Building a namespace api object, basename emptydir +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in emptydir-5162 STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [sig-cli] Kubectl client - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:213 -[BeforeEach] [k8s.io] Kubectl run job - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1510 -[It] should create a job from an image when restart is OnFailure [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -STEP: running the image docker.io/library/nginx:1.14-alpine -Jun 4 19:24:30.595: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 run e2e-test-nginx-job --restart=OnFailure --generator=job/v1 --image=docker.io/library/nginx:1.14-alpine --namespace=kubectl-7464' -Jun 4 19:24:30.759: INFO: stderr: "kubectl run --generator=job/v1 is DEPRECATED and will be removed in a future version. Use kubectl run --generator=run-pod/v1 or kubectl create instead.\n" -Jun 4 19:24:30.759: INFO: stdout: "job.batch/e2e-test-nginx-job created\n" -STEP: verifying the job e2e-test-nginx-job was created -[AfterEach] [k8s.io] Kubectl run job - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1515 -Jun 4 19:24:30.772: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 delete jobs e2e-test-nginx-job --namespace=kubectl-7464' -Jun 4 19:24:30.892: INFO: stderr: "" -Jun 4 19:24:30.892: INFO: stdout: "job.batch \"e2e-test-nginx-job\" deleted\n" -[AfterEach] [sig-cli] Kubectl client - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 -Jun 4 19:24:30.892: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "kubectl-7464" for this suite. -Jun 4 19:24:54.950: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -Jun 4 19:24:55.312: INFO: namespace kubectl-7464 deletion completed in 24.404040464s +[It] should support (non-root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: Creating a pod to test emptydir 0666 on tmpfs +Jun 18 12:40:37.793: INFO: Waiting up to 5m0s for pod "pod-4621abf6-91c6-11e9-a25d-8608290c688a" in namespace "emptydir-5162" to be "success or failure" +Jun 18 12:40:37.807: INFO: Pod "pod-4621abf6-91c6-11e9-a25d-8608290c688a": Phase="Pending", Reason="", readiness=false. Elapsed: 13.294138ms +Jun 18 12:40:39.822: INFO: Pod "pod-4621abf6-91c6-11e9-a25d-8608290c688a": Phase="Pending", Reason="", readiness=false. Elapsed: 2.028785639s +Jun 18 12:40:41.836: INFO: Pod "pod-4621abf6-91c6-11e9-a25d-8608290c688a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.042628886s +STEP: Saw pod success +Jun 18 12:40:41.836: INFO: Pod "pod-4621abf6-91c6-11e9-a25d-8608290c688a" satisfied condition "success or failure" +Jun 18 12:40:41.849: INFO: Trying to get logs from node 10.72.74.138 pod pod-4621abf6-91c6-11e9-a25d-8608290c688a container test-container: +STEP: delete the pod +Jun 18 12:40:41.938: INFO: Waiting for pod pod-4621abf6-91c6-11e9-a25d-8608290c688a to disappear +Jun 18 12:40:41.950: INFO: Pod pod-4621abf6-91c6-11e9-a25d-8608290c688a no longer exists +[AfterEach] [sig-storage] EmptyDir volumes + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Jun 18 12:40:41.950: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "emptydir-5162" for this suite. +Jun 18 12:40:47.993: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:40:48.428: INFO: namespace emptydir-5162 deletion completed in 6.461088514s -• [SLOW TEST:24.933 seconds] -[sig-cli] Kubectl client -/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 - [k8s.io] Kubectl run job - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687 - should create a job from an image when restart is OnFailure [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +• [SLOW TEST:11.592 seconds] +[sig-storage] EmptyDir volumes +/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:41 + should support (non-root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 ------------------------------ -SSSSSSSSSSSSSSSS +SSSSSS ------------------------------ -[sig-storage] Projected combined - should project all components that make up the projection API [Projection][NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -[BeforeEach] [sig-storage] Projected combined - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +[k8s.io] Container Lifecycle Hook when create a pod with lifecycle hook + should execute poststart http hook properly [NodeConformance] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [k8s.io] Container Lifecycle Hook + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 STEP: Creating a kubernetes client -Jun 4 19:24:55.312: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504 -STEP: Building a namespace api object, basename projected -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-7399 +Jun 18 12:40:48.428: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150 +STEP: Building a namespace api object, basename container-lifecycle-hook +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in container-lifecycle-hook-5092 STEP: Waiting for a default service account to be provisioned in namespace -[It] should project all components that make up the projection API [Projection][NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -STEP: Creating configMap with name configmap-projected-all-test-volume-6f1b5009-86fe-11e9-8d1b-467ee19922ac -STEP: Creating secret with name secret-projected-all-test-volume-6f1b4ff2-86fe-11e9-8d1b-467ee19922ac -STEP: Creating a pod to test Check all projections for projected volume plugin -Jun 4 19:24:55.581: INFO: Waiting up to 5m0s for pod "projected-volume-6f1b4fc4-86fe-11e9-8d1b-467ee19922ac" in namespace "projected-7399" to be "success or failure" -Jun 4 19:24:55.590: INFO: Pod "projected-volume-6f1b4fc4-86fe-11e9-8d1b-467ee19922ac": Phase="Pending", Reason="", readiness=false. Elapsed: 9.675522ms -Jun 4 19:24:57.606: INFO: Pod "projected-volume-6f1b4fc4-86fe-11e9-8d1b-467ee19922ac": Phase="Pending", Reason="", readiness=false. Elapsed: 2.024903222s -Jun 4 19:24:59.621: INFO: Pod "projected-volume-6f1b4fc4-86fe-11e9-8d1b-467ee19922ac": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.040453421s -STEP: Saw pod success -Jun 4 19:24:59.621: INFO: Pod "projected-volume-6f1b4fc4-86fe-11e9-8d1b-467ee19922ac" satisfied condition "success or failure" -Jun 4 19:24:59.635: INFO: Trying to get logs from node 10.212.23.189 pod projected-volume-6f1b4fc4-86fe-11e9-8d1b-467ee19922ac container projected-all-volume-test: -STEP: delete the pod -Jun 4 19:24:59.695: INFO: Waiting for pod projected-volume-6f1b4fc4-86fe-11e9-8d1b-467ee19922ac to disappear -Jun 4 19:24:59.766: INFO: Pod projected-volume-6f1b4fc4-86fe-11e9-8d1b-467ee19922ac no longer exists -[AfterEach] [sig-storage] Projected combined - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 -Jun 4 19:24:59.766: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "projected-7399" for this suite. -Jun 4 19:25:05.825: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -Jun 4 19:25:06.159: INFO: namespace projected-7399 deletion completed in 6.375177281s +[BeforeEach] when create a pod with lifecycle hook + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:61 +STEP: create the container to handle the HTTPGet hook request. +[It] should execute poststart http hook properly [NodeConformance] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: create the pod with lifecycle hook +STEP: check poststart hook +STEP: delete the pod with lifecycle hook +Jun 18 12:40:54.812: INFO: Waiting for pod pod-with-poststart-http-hook to disappear +Jun 18 12:40:54.825: INFO: Pod pod-with-poststart-http-hook still exists +Jun 18 12:40:56.825: INFO: Waiting for pod pod-with-poststart-http-hook to disappear +Jun 18 12:40:56.838: INFO: Pod pod-with-poststart-http-hook still exists +Jun 18 12:40:58.825: INFO: Waiting for pod pod-with-poststart-http-hook to disappear +Jun 18 12:40:58.838: INFO: Pod pod-with-poststart-http-hook no longer exists +[AfterEach] [k8s.io] Container Lifecycle Hook + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Jun 18 12:40:58.838: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "container-lifecycle-hook-5092" for this suite. +Jun 18 12:41:22.878: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:41:23.879: INFO: namespace container-lifecycle-hook-5092 deletion completed in 25.027342335s -• [SLOW TEST:10.847 seconds] -[sig-storage] Projected combined -/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_combined.go:31 - should project all components that make up the projection API [Projection][NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +• [SLOW TEST:35.451 seconds] +[k8s.io] Container Lifecycle Hook +/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687 + when create a pod with lifecycle hook + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:40 + should execute poststart http hook properly [NodeConformance] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 ------------------------------ -SSSSSSSSSSSSSSSSSSSS +SSSSSSSSSSSSSS ------------------------------ [sig-cli] Kubectl client [k8s.io] Update Demo - should do a rolling update of a replication controller [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 + should create and stop a replication controller [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 [BeforeEach] [sig-cli] Kubectl client - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 STEP: Creating a kubernetes client -Jun 4 19:25:06.161: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504 +Jun 18 12:41:23.880: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150 STEP: Building a namespace api object, basename kubectl -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-4405 +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-9004 STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:213 + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:213 [BeforeEach] [k8s.io] Update Demo - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:265 -[It] should do a rolling update of a replication controller [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -STEP: creating the initial replication controller -Jun 4 19:25:06.455: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 create -f - --namespace=kubectl-4405' -Jun 4 19:25:06.638: INFO: stderr: "" -Jun 4 19:25:06.638: INFO: stdout: "replicationcontroller/update-demo-nautilus created\n" + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:265 +[It] should create and stop a replication controller [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: creating a replication controller +Jun 18 12:41:24.079: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 create -f - --namespace=kubectl-9004' +Jun 18 12:41:24.430: INFO: stderr: "" +Jun 18 12:41:24.430: INFO: stdout: "replicationcontroller/update-demo-nautilus created\n" STEP: waiting for all containers in name=update-demo pods to come up. -Jun 4 19:25:06.638: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-4405' -Jun 4 19:25:06.750: INFO: stderr: "" -Jun 4 19:25:06.750: INFO: stdout: "update-demo-nautilus-k7xdj update-demo-nautilus-lzbc9 " -Jun 4 19:25:06.750: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 get pods update-demo-nautilus-k7xdj -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-4405' -Jun 4 19:25:06.838: INFO: stderr: "" -Jun 4 19:25:06.838: INFO: stdout: "" -Jun 4 19:25:06.838: INFO: update-demo-nautilus-k7xdj is created but not running -Jun 4 19:25:11.838: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-4405' -Jun 4 19:25:11.958: INFO: stderr: "" -Jun 4 19:25:11.958: INFO: stdout: "update-demo-nautilus-k7xdj update-demo-nautilus-lzbc9 " -Jun 4 19:25:11.958: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 get pods update-demo-nautilus-k7xdj -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-4405' -Jun 4 19:25:12.052: INFO: stderr: "" -Jun 4 19:25:12.052: INFO: stdout: "true" -Jun 4 19:25:12.052: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 get pods update-demo-nautilus-k7xdj -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-4405' -Jun 4 19:25:12.146: INFO: stderr: "" -Jun 4 19:25:12.146: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0" -Jun 4 19:25:12.146: INFO: validating pod update-demo-nautilus-k7xdj -Jun 4 19:25:12.166: INFO: got data: { +Jun 18 12:41:24.430: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-9004' +Jun 18 12:41:24.560: INFO: stderr: "" +Jun 18 12:41:24.560: INFO: stdout: "update-demo-nautilus-lh8sm update-demo-nautilus-xbfbv " +Jun 18 12:41:24.560: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 get pods update-demo-nautilus-lh8sm -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-9004' +Jun 18 12:41:24.685: INFO: stderr: "" +Jun 18 12:41:24.685: INFO: stdout: "" +Jun 18 12:41:24.685: INFO: update-demo-nautilus-lh8sm is created but not running +Jun 18 12:41:29.685: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-9004' +Jun 18 12:41:29.834: INFO: stderr: "" +Jun 18 12:41:29.834: INFO: stdout: "update-demo-nautilus-lh8sm update-demo-nautilus-xbfbv " +Jun 18 12:41:29.834: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 get pods update-demo-nautilus-lh8sm -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-9004' +Jun 18 12:41:29.973: INFO: stderr: "" +Jun 18 12:41:29.973: INFO: stdout: "true" +Jun 18 12:41:29.973: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 get pods update-demo-nautilus-lh8sm -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-9004' +Jun 18 12:41:30.107: INFO: stderr: "" +Jun 18 12:41:30.107: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0" +Jun 18 12:41:30.107: INFO: validating pod update-demo-nautilus-lh8sm +Jun 18 12:41:30.811: INFO: got data: { "image": "nautilus.jpg" } -Jun 4 19:25:12.166: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . -Jun 4 19:25:12.166: INFO: update-demo-nautilus-k7xdj is verified up and running -Jun 4 19:25:12.166: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 get pods update-demo-nautilus-lzbc9 -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-4405' -Jun 4 19:25:12.266: INFO: stderr: "" -Jun 4 19:25:12.266: INFO: stdout: "true" -Jun 4 19:25:12.266: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 get pods update-demo-nautilus-lzbc9 -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-4405' -Jun 4 19:25:12.353: INFO: stderr: "" -Jun 4 19:25:12.353: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0" -Jun 4 19:25:12.353: INFO: validating pod update-demo-nautilus-lzbc9 -Jun 4 19:25:12.373: INFO: got data: { +Jun 18 12:41:30.811: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . +Jun 18 12:41:30.811: INFO: update-demo-nautilus-lh8sm is verified up and running +Jun 18 12:41:30.811: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 get pods update-demo-nautilus-xbfbv -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-9004' +Jun 18 12:41:30.944: INFO: stderr: "" +Jun 18 12:41:30.944: INFO: stdout: "true" +Jun 18 12:41:30.944: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 get pods update-demo-nautilus-xbfbv -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-9004' +Jun 18 12:41:31.069: INFO: stderr: "" +Jun 18 12:41:31.069: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0" +Jun 18 12:41:31.069: INFO: validating pod update-demo-nautilus-xbfbv +Jun 18 12:41:31.097: INFO: got data: { "image": "nautilus.jpg" } -Jun 4 19:25:12.373: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . -Jun 4 19:25:12.373: INFO: update-demo-nautilus-lzbc9 is verified up and running -STEP: rolling-update to new replication controller -Jun 4 19:25:12.374: INFO: scanned /root for discovery docs: -Jun 4 19:25:12.374: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 rolling-update update-demo-nautilus --update-period=1s -f - --namespace=kubectl-4405' -Jun 4 19:25:36.202: INFO: stderr: "Command \"rolling-update\" is deprecated, use \"rollout\" instead\n" -Jun 4 19:25:36.202: INFO: stdout: "Created update-demo-kitten\nScaling up update-demo-kitten from 0 to 2, scaling down update-demo-nautilus from 2 to 0 (keep 2 pods available, don't exceed 3 pods)\nScaling update-demo-kitten up to 1\nScaling update-demo-nautilus down to 1\nScaling update-demo-kitten up to 2\nScaling update-demo-nautilus down to 0\nUpdate succeeded. Deleting old controller: update-demo-nautilus\nRenaming update-demo-kitten to update-demo-nautilus\nreplicationcontroller/update-demo-nautilus rolling updated\n" -STEP: waiting for all containers in name=update-demo pods to come up. -Jun 4 19:25:36.202: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-4405' -Jun 4 19:25:36.299: INFO: stderr: "" -Jun 4 19:25:36.299: INFO: stdout: "update-demo-kitten-96w2r update-demo-kitten-r65hv " -Jun 4 19:25:36.299: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 get pods update-demo-kitten-96w2r -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-4405' -Jun 4 19:25:36.395: INFO: stderr: "" -Jun 4 19:25:36.395: INFO: stdout: "true" -Jun 4 19:25:36.395: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 get pods update-demo-kitten-96w2r -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-4405' -Jun 4 19:25:36.491: INFO: stderr: "" -Jun 4 19:25:36.491: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/kitten:1.0" -Jun 4 19:25:36.491: INFO: validating pod update-demo-kitten-96w2r -Jun 4 19:25:36.512: INFO: got data: { - "image": "kitten.jpg" -} - -Jun 4 19:25:36.512: INFO: Unmarshalled json jpg/img => {kitten.jpg} , expecting kitten.jpg . -Jun 4 19:25:36.512: INFO: update-demo-kitten-96w2r is verified up and running -Jun 4 19:25:36.512: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 get pods update-demo-kitten-r65hv -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-4405' -Jun 4 19:25:36.604: INFO: stderr: "" -Jun 4 19:25:36.604: INFO: stdout: "true" -Jun 4 19:25:36.604: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 get pods update-demo-kitten-r65hv -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-4405' -Jun 4 19:25:36.683: INFO: stderr: "" -Jun 4 19:25:36.683: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/kitten:1.0" -Jun 4 19:25:36.683: INFO: validating pod update-demo-kitten-r65hv -Jun 4 19:25:36.705: INFO: got data: { - "image": "kitten.jpg" -} - -Jun 4 19:25:36.705: INFO: Unmarshalled json jpg/img => {kitten.jpg} , expecting kitten.jpg . -Jun 4 19:25:36.705: INFO: update-demo-kitten-r65hv is verified up and running +Jun 18 12:41:31.097: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . +Jun 18 12:41:31.097: INFO: update-demo-nautilus-xbfbv is verified up and running +STEP: using delete to clean up resources +Jun 18 12:41:31.097: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 delete --grace-period=0 --force -f - --namespace=kubectl-9004' +Jun 18 12:41:31.245: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" +Jun 18 12:41:31.245: INFO: stdout: "replicationcontroller \"update-demo-nautilus\" force deleted\n" +Jun 18 12:41:31.245: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 get rc,svc -l name=update-demo --no-headers --namespace=kubectl-9004' +Jun 18 12:41:31.394: INFO: stderr: "No resources found.\n" +Jun 18 12:41:31.394: INFO: stdout: "" +Jun 18 12:41:31.394: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 get pods -l name=update-demo --namespace=kubectl-9004 -o go-template={{ range .items }}{{ if not .metadata.deletionTimestamp }}{{ .metadata.name }}{{ "\n" }}{{ end }}{{ end }}' +Jun 18 12:41:31.521: INFO: stderr: "" +Jun 18 12:41:31.521: INFO: stdout: "" [AfterEach] [sig-cli] Kubectl client - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 -Jun 4 19:25:36.705: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "kubectl-4405" for this suite. -Jun 4 19:26:00.768: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -Jun 4 19:26:01.187: INFO: namespace kubectl-4405 deletion completed in 24.46517278s + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Jun 18 12:41:31.522: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "kubectl-9004" for this suite. +Jun 18 12:41:55.561: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:41:55.970: INFO: namespace kubectl-9004 deletion completed in 24.434237507s -• [SLOW TEST:55.026 seconds] +• [SLOW TEST:32.090 seconds] [sig-cli] Kubectl client -/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 +/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 [k8s.io] Update Demo - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687 - should do a rolling update of a replication controller [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 ------------------------------- -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------- -[sig-api-machinery] Secrets - should be consumable via the environment [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -[BeforeEach] [sig-api-machinery] Secrets - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 -STEP: Creating a kubernetes client -Jun 4 19:26:01.187: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504 -STEP: Building a namespace api object, basename secrets -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in secrets-3194 -STEP: Waiting for a default service account to be provisioned in namespace -[It] should be consumable via the environment [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -STEP: creating secret secrets-3194/secret-test-9663900a-86fe-11e9-8d1b-467ee19922ac -STEP: Creating a pod to test consume secrets -Jun 4 19:26:01.466: INFO: Waiting up to 5m0s for pod "pod-configmaps-9664c876-86fe-11e9-8d1b-467ee19922ac" in namespace "secrets-3194" to be "success or failure" -Jun 4 19:26:01.476: INFO: Pod "pod-configmaps-9664c876-86fe-11e9-8d1b-467ee19922ac": Phase="Pending", Reason="", readiness=false. Elapsed: 9.629892ms -Jun 4 19:26:03.488: INFO: Pod "pod-configmaps-9664c876-86fe-11e9-8d1b-467ee19922ac": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.021107121s -STEP: Saw pod success -Jun 4 19:26:03.488: INFO: Pod "pod-configmaps-9664c876-86fe-11e9-8d1b-467ee19922ac" satisfied condition "success or failure" -Jun 4 19:26:03.498: INFO: Trying to get logs from node 10.212.23.189 pod pod-configmaps-9664c876-86fe-11e9-8d1b-467ee19922ac container env-test: -STEP: delete the pod -Jun 4 19:26:03.554: INFO: Waiting for pod pod-configmaps-9664c876-86fe-11e9-8d1b-467ee19922ac to disappear -Jun 4 19:26:03.565: INFO: Pod pod-configmaps-9664c876-86fe-11e9-8d1b-467ee19922ac no longer exists -[AfterEach] [sig-api-machinery] Secrets - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 -Jun 4 19:26:03.565: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "secrets-3194" for this suite. -Jun 4 19:26:09.624: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -Jun 4 19:26:09.963: INFO: namespace secrets-3194 deletion completed in 6.38302856s - -• [SLOW TEST:8.775 seconds] -[sig-api-machinery] Secrets -/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/secrets.go:32 - should be consumable via the environment [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 ------------------------------- -SSSSSSSSSSSSSS ------------------------------- -[sig-cli] Kubectl client [k8s.io] Kubectl version - should check is all data is printed [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -[BeforeEach] [sig-cli] Kubectl client - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 -STEP: Creating a kubernetes client -Jun 4 19:26:09.963: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504 -STEP: Building a namespace api object, basename kubectl -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-8412 -STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [sig-cli] Kubectl client - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:213 -[It] should check is all data is printed [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -Jun 4 19:26:10.179: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 version' -Jun 4 19:26:10.274: INFO: stderr: "" -Jun 4 19:26:10.274: INFO: stdout: "Client Version: version.Info{Major:\"1\", Minor:\"14\", GitVersion:\"v1.14.2\", GitCommit:\"66049e3b21efe110454d67df4fa62b08ea79a19b\", GitTreeState:\"clean\", BuildDate:\"2019-05-16T16:23:09Z\", GoVersion:\"go1.12.5\", Compiler:\"gc\", Platform:\"linux/amd64\"}\nServer Version: version.Info{Major:\"1\", Minor:\"14\", GitVersion:\"v1.14.2+IKS\", GitCommit:\"b3e16abec4c3166d8bb7524d9db770d265f6e470\", GitTreeState:\"clean\", BuildDate:\"2019-05-31T08:07:55Z\", GoVersion:\"go1.12.5\", Compiler:\"gc\", Platform:\"linux/amd64\"}\n" -[AfterEach] [sig-cli] Kubectl client - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 -Jun 4 19:26:10.275: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "kubectl-8412" for this suite. -Jun 4 19:26:16.334: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -Jun 4 19:26:16.679: INFO: namespace kubectl-8412 deletion completed in 6.387317667s - -• [SLOW TEST:6.716 seconds] -[sig-cli] Kubectl client -/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 - [k8s.io] Kubectl version - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687 - should check is all data is printed [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 ------------------------------- -SSSSSSSSSSSSSS ------------------------------- -[sig-network] Service endpoints latency - should not be very high [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -[BeforeEach] [sig-network] Service endpoints latency - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 -STEP: Creating a kubernetes client -Jun 4 19:26:16.679: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504 -STEP: Building a namespace api object, basename svc-latency -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in svc-latency-9600 -STEP: Waiting for a default service account to be provisioned in namespace -[It] should not be very high [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -STEP: creating replication controller svc-latency-rc in namespace svc-latency-9600 -I0604 19:26:16.908885 18 runners.go:184] Created replication controller with name: svc-latency-rc, namespace: svc-latency-9600, replica count: 1 -I0604 19:26:17.959446 18 runners.go:184] svc-latency-rc Pods: 1 out of 1 created, 0 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady -I0604 19:26:18.959659 18 runners.go:184] svc-latency-rc Pods: 1 out of 1 created, 1 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady -Jun 4 19:26:19.086: INFO: Created: latency-svc-t56z8 -Jun 4 19:26:19.097: INFO: Got endpoints: latency-svc-t56z8 [37.238595ms] -Jun 4 19:26:19.182: INFO: Created: latency-svc-887f9 -Jun 4 19:26:19.192: INFO: Got endpoints: latency-svc-887f9 [94.838171ms] -Jun 4 19:26:19.212: INFO: Created: latency-svc-7j6rw -Jun 4 19:26:19.212: INFO: Got endpoints: latency-svc-7j6rw [115.579039ms] -Jun 4 19:26:19.213: INFO: Created: latency-svc-w9tpf -Jun 4 19:26:19.221: INFO: Got endpoints: latency-svc-w9tpf [124.551224ms] -Jun 4 19:26:19.231: INFO: Created: latency-svc-kcqcp -Jun 4 19:26:19.240: INFO: Got endpoints: latency-svc-kcqcp [143.360159ms] -Jun 4 19:26:19.247: INFO: Created: latency-svc-4hnz9 -Jun 4 19:26:19.255: INFO: Got endpoints: latency-svc-4hnz9 [158.037564ms] -Jun 4 19:26:19.262: INFO: Created: latency-svc-rmwkq -Jun 4 19:26:19.270: INFO: Got endpoints: latency-svc-rmwkq [172.91278ms] -Jun 4 19:26:19.285: INFO: Created: latency-svc-wls5p -Jun 4 19:26:19.293: INFO: Got endpoints: latency-svc-wls5p [196.132435ms] -Jun 4 19:26:19.302: INFO: Created: latency-svc-f2fz7 -Jun 4 19:26:19.309: INFO: Got endpoints: latency-svc-f2fz7 [211.917412ms] -Jun 4 19:26:19.318: INFO: Created: latency-svc-bvvqk -Jun 4 19:26:19.326: INFO: Got endpoints: latency-svc-bvvqk [228.677877ms] -Jun 4 19:26:19.332: INFO: Created: latency-svc-k4mlt -Jun 4 19:26:19.340: INFO: Got endpoints: latency-svc-k4mlt [242.669947ms] -Jun 4 19:26:19.348: INFO: Created: latency-svc-7xjt5 -Jun 4 19:26:19.357: INFO: Got endpoints: latency-svc-7xjt5 [260.034823ms] -Jun 4 19:26:19.364: INFO: Created: latency-svc-pf7js -Jun 4 19:26:19.372: INFO: Got endpoints: latency-svc-pf7js [274.558732ms] -Jun 4 19:26:19.379: INFO: Created: latency-svc-62lf8 -Jun 4 19:26:19.388: INFO: Got endpoints: latency-svc-62lf8 [290.875109ms] -Jun 4 19:26:19.401: INFO: Created: latency-svc-t28sq -Jun 4 19:26:19.409: INFO: Got endpoints: latency-svc-t28sq [311.841661ms] -Jun 4 19:26:19.420: INFO: Created: latency-svc-wnq2n -Jun 4 19:26:19.432: INFO: Got endpoints: latency-svc-wnq2n [335.138389ms] -Jun 4 19:26:19.440: INFO: Created: latency-svc-2dft2 -Jun 4 19:26:19.448: INFO: Got endpoints: latency-svc-2dft2 [256.47726ms] -Jun 4 19:26:19.455: INFO: Created: latency-svc-786cl -Jun 4 19:26:19.464: INFO: Got endpoints: latency-svc-786cl [251.116341ms] -Jun 4 19:26:19.470: INFO: Created: latency-svc-cmm2j -Jun 4 19:26:19.477: INFO: Got endpoints: latency-svc-cmm2j [255.478673ms] -Jun 4 19:26:19.485: INFO: Created: latency-svc-jtjh4 -Jun 4 19:26:19.498: INFO: Got endpoints: latency-svc-jtjh4 [257.19528ms] -Jun 4 19:26:19.501: INFO: Created: latency-svc-6psn2 -Jun 4 19:26:19.509: INFO: Got endpoints: latency-svc-6psn2 [253.762432ms] -Jun 4 19:26:19.517: INFO: Created: latency-svc-chkw6 -Jun 4 19:26:19.525: INFO: Got endpoints: latency-svc-chkw6 [254.754267ms] -Jun 4 19:26:19.531: INFO: Created: latency-svc-4p4hx -Jun 4 19:26:19.538: INFO: Got endpoints: latency-svc-4p4hx [244.886572ms] -Jun 4 19:26:19.544: INFO: Created: latency-svc-wdbsr -Jun 4 19:26:19.553: INFO: Got endpoints: latency-svc-wdbsr [244.468587ms] -Jun 4 19:26:19.561: INFO: Created: latency-svc-gx4df -Jun 4 19:26:19.568: INFO: Got endpoints: latency-svc-gx4df [242.446175ms] -Jun 4 19:26:19.578: INFO: Created: latency-svc-dh2bw -Jun 4 19:26:19.585: INFO: Got endpoints: latency-svc-dh2bw [245.049301ms] -Jun 4 19:26:19.593: INFO: Created: latency-svc-wwp8h -Jun 4 19:26:19.608: INFO: Got endpoints: latency-svc-wwp8h [250.493819ms] -Jun 4 19:26:19.610: INFO: Created: latency-svc-ldw22 -Jun 4 19:26:19.615: INFO: Got endpoints: latency-svc-ldw22 [243.615876ms] -Jun 4 19:26:19.624: INFO: Created: latency-svc-khx7r -Jun 4 19:26:19.632: INFO: Got endpoints: latency-svc-khx7r [244.353961ms] -Jun 4 19:26:19.639: INFO: Created: latency-svc-4vnmd -Jun 4 19:26:19.648: INFO: Got endpoints: latency-svc-4vnmd [239.052778ms] -Jun 4 19:26:19.655: INFO: Created: latency-svc-xp7ch -Jun 4 19:26:19.664: INFO: Got endpoints: latency-svc-xp7ch [232.020984ms] -Jun 4 19:26:19.673: INFO: Created: latency-svc-9gjs4 -Jun 4 19:26:19.680: INFO: Got endpoints: latency-svc-9gjs4 [232.090754ms] -Jun 4 19:26:19.689: INFO: Created: latency-svc-kqx22 -Jun 4 19:26:19.697: INFO: Got endpoints: latency-svc-kqx22 [233.743323ms] -Jun 4 19:26:19.705: INFO: Created: latency-svc-l6565 -Jun 4 19:26:19.712: INFO: Got endpoints: latency-svc-l6565 [234.826591ms] -Jun 4 19:26:19.721: INFO: Created: latency-svc-m6827 -Jun 4 19:26:19.728: INFO: Got endpoints: latency-svc-m6827 [230.814682ms] -Jun 4 19:26:19.735: INFO: Created: latency-svc-wklxl -Jun 4 19:26:19.743: INFO: Got endpoints: latency-svc-wklxl [234.601356ms] -Jun 4 19:26:19.752: INFO: Created: latency-svc-r99r9 -Jun 4 19:26:19.760: INFO: Got endpoints: latency-svc-r99r9 [234.734525ms] -Jun 4 19:26:19.765: INFO: Created: latency-svc-4lq67 -Jun 4 19:26:19.774: INFO: Got endpoints: latency-svc-4lq67 [236.3418ms] -Jun 4 19:26:19.782: INFO: Created: latency-svc-bh99s -Jun 4 19:26:19.790: INFO: Got endpoints: latency-svc-bh99s [236.74393ms] -Jun 4 19:26:19.798: INFO: Created: latency-svc-hhpfk -Jun 4 19:26:19.805: INFO: Got endpoints: latency-svc-hhpfk [236.758339ms] -Jun 4 19:26:19.812: INFO: Created: latency-svc-sg2r8 -Jun 4 19:26:19.821: INFO: Got endpoints: latency-svc-sg2r8 [236.057045ms] -Jun 4 19:26:19.826: INFO: Created: latency-svc-xc5pd -Jun 4 19:26:19.835: INFO: Got endpoints: latency-svc-xc5pd [227.119888ms] -Jun 4 19:26:19.843: INFO: Created: latency-svc-c9j22 -Jun 4 19:26:19.850: INFO: Got endpoints: latency-svc-c9j22 [234.624926ms] -Jun 4 19:26:19.857: INFO: Created: latency-svc-4f9t2 -Jun 4 19:26:19.865: INFO: Got endpoints: latency-svc-4f9t2 [232.344729ms] -Jun 4 19:26:19.872: INFO: Created: latency-svc-p5tqp -Jun 4 19:26:19.879: INFO: Got endpoints: latency-svc-p5tqp [231.08744ms] -Jun 4 19:26:19.889: INFO: Created: latency-svc-zhrvv -Jun 4 19:26:19.898: INFO: Got endpoints: latency-svc-zhrvv [233.522445ms] -Jun 4 19:26:19.906: INFO: Created: latency-svc-xbvlg -Jun 4 19:26:19.914: INFO: Got endpoints: latency-svc-xbvlg [233.242337ms] -Jun 4 19:26:19.921: INFO: Created: latency-svc-bmk26 -Jun 4 19:26:19.935: INFO: Got endpoints: latency-svc-bmk26 [237.756003ms] -Jun 4 19:26:19.943: INFO: Created: latency-svc-c2m9m -Jun 4 19:26:19.951: INFO: Got endpoints: latency-svc-c2m9m [238.8081ms] -Jun 4 19:26:19.959: INFO: Created: latency-svc-zzhp5 -Jun 4 19:26:19.967: INFO: Got endpoints: latency-svc-zzhp5 [238.024564ms] -Jun 4 19:26:19.973: INFO: Created: latency-svc-dns46 -Jun 4 19:26:19.983: INFO: Got endpoints: latency-svc-dns46 [238.99403ms] -Jun 4 19:26:19.991: INFO: Created: latency-svc-xc4h9 -Jun 4 19:26:19.999: INFO: Got endpoints: latency-svc-xc4h9 [239.510071ms] -Jun 4 19:26:20.007: INFO: Created: latency-svc-twwlh -Jun 4 19:26:20.016: INFO: Got endpoints: latency-svc-twwlh [241.188486ms] -Jun 4 19:26:20.022: INFO: Created: latency-svc-trkrr -Jun 4 19:26:20.031: INFO: Got endpoints: latency-svc-trkrr [241.105214ms] -Jun 4 19:26:20.037: INFO: Created: latency-svc-fhvgg -Jun 4 19:26:20.046: INFO: Got endpoints: latency-svc-fhvgg [241.205522ms] -Jun 4 19:26:20.052: INFO: Created: latency-svc-k7kwc -Jun 4 19:26:20.061: INFO: Got endpoints: latency-svc-k7kwc [239.745753ms] -Jun 4 19:26:20.068: INFO: Created: latency-svc-8fbmn -Jun 4 19:26:20.076: INFO: Got endpoints: latency-svc-8fbmn [241.404952ms] -Jun 4 19:26:20.083: INFO: Created: latency-svc-qwjck -Jun 4 19:26:20.091: INFO: Got endpoints: latency-svc-qwjck [241.131855ms] -Jun 4 19:26:20.098: INFO: Created: latency-svc-t7d4k -Jun 4 19:26:20.106: INFO: Got endpoints: latency-svc-t7d4k [241.00784ms] -Jun 4 19:26:20.116: INFO: Created: latency-svc-hd4fz -Jun 4 19:26:20.120: INFO: Got endpoints: latency-svc-hd4fz [240.535233ms] -Jun 4 19:26:20.131: INFO: Created: latency-svc-wb4f9 -Jun 4 19:26:20.140: INFO: Got endpoints: latency-svc-wb4f9 [241.817434ms] -Jun 4 19:26:20.147: INFO: Created: latency-svc-hs652 -Jun 4 19:26:20.155: INFO: Got endpoints: latency-svc-hs652 [241.200231ms] -Jun 4 19:26:20.165: INFO: Created: latency-svc-z4mbr -Jun 4 19:26:20.172: INFO: Got endpoints: latency-svc-z4mbr [236.630127ms] -Jun 4 19:26:20.178: INFO: Created: latency-svc-8x5v4 -Jun 4 19:26:20.186: INFO: Got endpoints: latency-svc-8x5v4 [235.45258ms] -Jun 4 19:26:20.196: INFO: Created: latency-svc-nsdzj -Jun 4 19:26:20.206: INFO: Got endpoints: latency-svc-nsdzj [239.547037ms] -Jun 4 19:26:20.213: INFO: Created: latency-svc-rzsvp -Jun 4 19:26:20.221: INFO: Got endpoints: latency-svc-rzsvp [237.789968ms] -Jun 4 19:26:20.228: INFO: Created: latency-svc-5npdw -Jun 4 19:26:20.236: INFO: Got endpoints: latency-svc-5npdw [236.715522ms] -Jun 4 19:26:20.242: INFO: Created: latency-svc-d5vmv -Jun 4 19:26:20.251: INFO: Got endpoints: latency-svc-d5vmv [235.495642ms] -Jun 4 19:26:20.257: INFO: Created: latency-svc-wfhnb -Jun 4 19:26:20.266: INFO: Got endpoints: latency-svc-wfhnb [234.513841ms] -Jun 4 19:26:20.273: INFO: Created: latency-svc-v5cm2 -Jun 4 19:26:20.280: INFO: Got endpoints: latency-svc-v5cm2 [234.068268ms] -Jun 4 19:26:20.290: INFO: Created: latency-svc-btvsx -Jun 4 19:26:20.298: INFO: Got endpoints: latency-svc-btvsx [237.361084ms] -Jun 4 19:26:20.313: INFO: Created: latency-svc-2qwl4 -Jun 4 19:26:20.323: INFO: Got endpoints: latency-svc-2qwl4 [246.414021ms] -Jun 4 19:26:20.332: INFO: Created: latency-svc-dmp8n -Jun 4 19:26:20.332: INFO: Got endpoints: latency-svc-dmp8n [240.78399ms] -Jun 4 19:26:20.338: INFO: Created: latency-svc-679g4 -Jun 4 19:26:20.347: INFO: Got endpoints: latency-svc-679g4 [240.779753ms] -Jun 4 19:26:20.355: INFO: Created: latency-svc-x6mk7 -Jun 4 19:26:20.361: INFO: Got endpoints: latency-svc-x6mk7 [38.659696ms] -Jun 4 19:26:20.369: INFO: Created: latency-svc-xs5jn -Jun 4 19:26:20.383: INFO: Got endpoints: latency-svc-xs5jn [262.895617ms] -Jun 4 19:26:20.385: INFO: Created: latency-svc-klgks -Jun 4 19:26:20.393: INFO: Got endpoints: latency-svc-klgks [253.048372ms] -Jun 4 19:26:20.404: INFO: Created: latency-svc-tjr8z -Jun 4 19:26:20.413: INFO: Got endpoints: latency-svc-tjr8z [258.567851ms] -Jun 4 19:26:20.423: INFO: Created: latency-svc-hvn2w -Jun 4 19:26:20.431: INFO: Got endpoints: latency-svc-hvn2w [259.030979ms] -Jun 4 19:26:20.439: INFO: Created: latency-svc-frmbx -Jun 4 19:26:20.448: INFO: Got endpoints: latency-svc-frmbx [261.522799ms] -Jun 4 19:26:20.455: INFO: Created: latency-svc-4k8db -Jun 4 19:26:20.464: INFO: Got endpoints: latency-svc-4k8db [257.432063ms] -Jun 4 19:26:20.473: INFO: Created: latency-svc-zsbtx -Jun 4 19:26:20.481: INFO: Got endpoints: latency-svc-zsbtx [260.707759ms] -Jun 4 19:26:20.486: INFO: Created: latency-svc-nxxxd -Jun 4 19:26:20.494: INFO: Got endpoints: latency-svc-nxxxd [258.468479ms] -Jun 4 19:26:20.503: INFO: Created: latency-svc-xdt96 -Jun 4 19:26:20.511: INFO: Got endpoints: latency-svc-xdt96 [259.533973ms] -Jun 4 19:26:20.519: INFO: Created: latency-svc-v95st -Jun 4 19:26:20.526: INFO: Got endpoints: latency-svc-v95st [259.915244ms] -Jun 4 19:26:20.534: INFO: Created: latency-svc-9wqhq -Jun 4 19:26:20.542: INFO: Got endpoints: latency-svc-9wqhq [261.32782ms] -Jun 4 19:26:20.548: INFO: Created: latency-svc-tqktl -Jun 4 19:26:20.558: INFO: Got endpoints: latency-svc-tqktl [259.600212ms] -Jun 4 19:26:20.563: INFO: Created: latency-svc-qvgwb -Jun 4 19:26:20.572: INFO: Got endpoints: latency-svc-qvgwb [240.214675ms] -Jun 4 19:26:20.579: INFO: Created: latency-svc-4n2gs -Jun 4 19:26:20.594: INFO: Got endpoints: latency-svc-4n2gs [247.145301ms] -Jun 4 19:26:20.598: INFO: Created: latency-svc-897wg -Jun 4 19:26:20.605: INFO: Got endpoints: latency-svc-897wg [243.753153ms] -Jun 4 19:26:20.612: INFO: Created: latency-svc-bgr4v -Jun 4 19:26:20.620: INFO: Got endpoints: latency-svc-bgr4v [236.852557ms] -Jun 4 19:26:20.625: INFO: Created: latency-svc-pn4wp -Jun 4 19:26:20.634: INFO: Got endpoints: latency-svc-pn4wp [241.140536ms] -Jun 4 19:26:20.641: INFO: Created: latency-svc-dthqx -Jun 4 19:26:20.649: INFO: Got endpoints: latency-svc-dthqx [235.594931ms] -Jun 4 19:26:20.656: INFO: Created: latency-svc-72hnq -Jun 4 19:26:20.664: INFO: Got endpoints: latency-svc-72hnq [233.38584ms] -Jun 4 19:26:20.672: INFO: Created: latency-svc-2xtpv -Jun 4 19:26:20.679: INFO: Got endpoints: latency-svc-2xtpv [231.248681ms] -Jun 4 19:26:20.689: INFO: Created: latency-svc-thglr -Jun 4 19:26:20.696: INFO: Got endpoints: latency-svc-thglr [231.897495ms] -Jun 4 19:26:20.704: INFO: Created: latency-svc-nb2jn -Jun 4 19:26:20.721: INFO: Got endpoints: latency-svc-nb2jn [239.2904ms] -Jun 4 19:26:20.726: INFO: Created: latency-svc-p4khj -Jun 4 19:26:20.733: INFO: Got endpoints: latency-svc-p4khj [238.675244ms] -Jun 4 19:26:20.744: INFO: Created: latency-svc-75w59 -Jun 4 19:26:20.749: INFO: Got endpoints: latency-svc-75w59 [238.608432ms] -Jun 4 19:26:20.757: INFO: Created: latency-svc-85jnx -Jun 4 19:26:20.769: INFO: Got endpoints: latency-svc-85jnx [242.954464ms] -Jun 4 19:26:20.774: INFO: Created: latency-svc-sntbq -Jun 4 19:26:20.781: INFO: Got endpoints: latency-svc-sntbq [239.541915ms] -Jun 4 19:26:20.789: INFO: Created: latency-svc-wntx9 -Jun 4 19:26:20.797: INFO: Got endpoints: latency-svc-wntx9 [239.138462ms] -Jun 4 19:26:20.804: INFO: Created: latency-svc-w2rb6 -Jun 4 19:26:20.813: INFO: Got endpoints: latency-svc-w2rb6 [241.243986ms] -Jun 4 19:26:20.820: INFO: Created: latency-svc-jw5sc -Jun 4 19:26:20.828: INFO: Got endpoints: latency-svc-jw5sc [234.146786ms] -Jun 4 19:26:20.836: INFO: Created: latency-svc-v4lxc -Jun 4 19:26:20.843: INFO: Got endpoints: latency-svc-v4lxc [237.988004ms] -Jun 4 19:26:20.851: INFO: Created: latency-svc-sbb98 -Jun 4 19:26:20.860: INFO: Got endpoints: latency-svc-sbb98 [240.681004ms] -Jun 4 19:26:20.867: INFO: Created: latency-svc-wcmh9 -Jun 4 19:26:20.875: INFO: Got endpoints: latency-svc-wcmh9 [240.772025ms] -Jun 4 19:26:20.882: INFO: Created: latency-svc-bztx2 -Jun 4 19:26:20.891: INFO: Got endpoints: latency-svc-bztx2 [241.550934ms] -Jun 4 19:26:20.898: INFO: Created: latency-svc-tx647 -Jun 4 19:26:20.905: INFO: Got endpoints: latency-svc-tx647 [240.911574ms] -Jun 4 19:26:20.917: INFO: Created: latency-svc-gk7mj -Jun 4 19:26:20.922: INFO: Got endpoints: latency-svc-gk7mj [242.918768ms] -Jun 4 19:26:20.929: INFO: Created: latency-svc-2lpwt -Jun 4 19:26:20.938: INFO: Got endpoints: latency-svc-2lpwt [242.533617ms] -Jun 4 19:26:20.945: INFO: Created: latency-svc-lp7fd -Jun 4 19:26:20.953: INFO: Got endpoints: latency-svc-lp7fd [232.508941ms] -Jun 4 19:26:20.960: INFO: Created: latency-svc-jcfvf -Jun 4 19:26:20.967: INFO: Got endpoints: latency-svc-jcfvf [233.760024ms] -Jun 4 19:26:20.976: INFO: Created: latency-svc-xmqvz -Jun 4 19:26:20.986: INFO: Got endpoints: latency-svc-xmqvz [236.460661ms] -Jun 4 19:26:20.994: INFO: Created: latency-svc-b8n54 -Jun 4 19:26:21.002: INFO: Got endpoints: latency-svc-b8n54 [233.573503ms] -Jun 4 19:26:21.008: INFO: Created: latency-svc-wfs8k -Jun 4 19:26:21.018: INFO: Got endpoints: latency-svc-wfs8k [236.496406ms] -Jun 4 19:26:21.025: INFO: Created: latency-svc-cvrtr -Jun 4 19:26:21.032: INFO: Got endpoints: latency-svc-cvrtr [235.442362ms] -Jun 4 19:26:21.041: INFO: Created: latency-svc-pqx6m -Jun 4 19:26:21.049: INFO: Got endpoints: latency-svc-pqx6m [235.267309ms] -Jun 4 19:26:21.055: INFO: Created: latency-svc-mcdxr -Jun 4 19:26:21.063: INFO: Got endpoints: latency-svc-mcdxr [235.100366ms] -Jun 4 19:26:21.070: INFO: Created: latency-svc-k72jm -Jun 4 19:26:21.079: INFO: Got endpoints: latency-svc-k72jm [235.524745ms] -Jun 4 19:26:21.086: INFO: Created: latency-svc-qq5td -Jun 4 19:26:21.094: INFO: Got endpoints: latency-svc-qq5td [233.523075ms] -Jun 4 19:26:21.101: INFO: Created: latency-svc-2l2hp -Jun 4 19:26:21.108: INFO: Got endpoints: latency-svc-2l2hp [233.153527ms] -Jun 4 19:26:21.115: INFO: Created: latency-svc-fr8vw -Jun 4 19:26:21.124: INFO: Got endpoints: latency-svc-fr8vw [232.878474ms] -Jun 4 19:26:21.132: INFO: Created: latency-svc-szmkx -Jun 4 19:26:21.139: INFO: Got endpoints: latency-svc-szmkx [233.446789ms] -Jun 4 19:26:21.146: INFO: Created: latency-svc-qrxrt -Jun 4 19:26:21.154: INFO: Got endpoints: latency-svc-qrxrt [232.075325ms] -Jun 4 19:26:21.161: INFO: Created: latency-svc-8v84j -Jun 4 19:26:21.169: INFO: Got endpoints: latency-svc-8v84j [230.815417ms] -Jun 4 19:26:21.177: INFO: Created: latency-svc-vf8mh -Jun 4 19:26:21.185: INFO: Got endpoints: latency-svc-vf8mh [231.690879ms] -Jun 4 19:26:21.192: INFO: Created: latency-svc-2q6w7 -Jun 4 19:26:21.199: INFO: Got endpoints: latency-svc-2q6w7 [231.874912ms] -Jun 4 19:26:21.208: INFO: Created: latency-svc-vr8ph -Jun 4 19:26:21.219: INFO: Got endpoints: latency-svc-vr8ph [232.706997ms] -Jun 4 19:26:21.225: INFO: Created: latency-svc-4m7k5 -Jun 4 19:26:21.233: INFO: Got endpoints: latency-svc-4m7k5 [230.61579ms] -Jun 4 19:26:21.241: INFO: Created: latency-svc-gkjvm -Jun 4 19:26:21.248: INFO: Got endpoints: latency-svc-gkjvm [230.443672ms] -Jun 4 19:26:21.259: INFO: Created: latency-svc-726nh -Jun 4 19:26:21.266: INFO: Got endpoints: latency-svc-726nh [233.194166ms] -Jun 4 19:26:21.273: INFO: Created: latency-svc-6tkll -Jun 4 19:26:21.281: INFO: Got endpoints: latency-svc-6tkll [231.976975ms] -Jun 4 19:26:21.289: INFO: Created: latency-svc-pj5sq -Jun 4 19:26:21.297: INFO: Got endpoints: latency-svc-pj5sq [234.121928ms] -Jun 4 19:26:21.302: INFO: Created: latency-svc-j7z6x -Jun 4 19:26:21.310: INFO: Got endpoints: latency-svc-j7z6x [231.191698ms] -Jun 4 19:26:21.317: INFO: Created: latency-svc-dbkck -Jun 4 19:26:21.325: INFO: Got endpoints: latency-svc-dbkck [231.28169ms] -Jun 4 19:26:21.332: INFO: Created: latency-svc-xt4nz -Jun 4 19:26:21.341: INFO: Got endpoints: latency-svc-xt4nz [233.221703ms] -Jun 4 19:26:21.348: INFO: Created: latency-svc-bghdn -Jun 4 19:26:21.356: INFO: Got endpoints: latency-svc-bghdn [232.761262ms] -Jun 4 19:26:21.364: INFO: Created: latency-svc-z2vff -Jun 4 19:26:21.372: INFO: Got endpoints: latency-svc-z2vff [232.777174ms] -Jun 4 19:26:21.379: INFO: Created: latency-svc-b7hw7 -Jun 4 19:26:21.386: INFO: Got endpoints: latency-svc-b7hw7 [231.565323ms] -Jun 4 19:26:21.395: INFO: Created: latency-svc-j9m5s -Jun 4 19:26:21.403: INFO: Got endpoints: latency-svc-j9m5s [233.773298ms] -Jun 4 19:26:21.410: INFO: Created: latency-svc-rk2b4 -Jun 4 19:26:21.417: INFO: Got endpoints: latency-svc-rk2b4 [232.658511ms] -Jun 4 19:26:21.425: INFO: Created: latency-svc-jwbgc -Jun 4 19:26:21.433: INFO: Got endpoints: latency-svc-jwbgc [233.992645ms] -Jun 4 19:26:21.440: INFO: Created: latency-svc-md4t2 -Jun 4 19:26:21.448: INFO: Got endpoints: latency-svc-md4t2 [229.75267ms] -Jun 4 19:26:21.455: INFO: Created: latency-svc-hnkqd -Jun 4 19:26:21.464: INFO: Got endpoints: latency-svc-hnkqd [230.833303ms] -Jun 4 19:26:21.471: INFO: Created: latency-svc-4kv9l -Jun 4 19:26:21.480: INFO: Got endpoints: latency-svc-4kv9l [231.52112ms] -Jun 4 19:26:21.487: INFO: Created: latency-svc-fq489 -Jun 4 19:26:21.494: INFO: Got endpoints: latency-svc-fq489 [228.440585ms] -Jun 4 19:26:21.500: INFO: Created: latency-svc-fvs6j -Jun 4 19:26:21.509: INFO: Got endpoints: latency-svc-fvs6j [227.881408ms] -Jun 4 19:26:21.517: INFO: Created: latency-svc-n69s6 -Jun 4 19:26:21.525: INFO: Got endpoints: latency-svc-n69s6 [227.912325ms] -Jun 4 19:26:21.534: INFO: Created: latency-svc-t8f4s -Jun 4 19:26:21.542: INFO: Got endpoints: latency-svc-t8f4s [232.298929ms] -Jun 4 19:26:21.551: INFO: Created: latency-svc-hnxf2 -Jun 4 19:26:21.560: INFO: Got endpoints: latency-svc-hnxf2 [234.515191ms] -Jun 4 19:26:21.566: INFO: Created: latency-svc-64hsb -Jun 4 19:26:21.574: INFO: Got endpoints: latency-svc-64hsb [232.751054ms] -Jun 4 19:26:21.581: INFO: Created: latency-svc-4gwm6 -Jun 4 19:26:21.589: INFO: Got endpoints: latency-svc-4gwm6 [232.255983ms] -Jun 4 19:26:21.595: INFO: Created: latency-svc-zw7r2 -Jun 4 19:26:21.603: INFO: Got endpoints: latency-svc-zw7r2 [231.566828ms] -Jun 4 19:26:21.609: INFO: Created: latency-svc-ng75m -Jun 4 19:26:21.617: INFO: Got endpoints: latency-svc-ng75m [231.546282ms] -Jun 4 19:26:21.626: INFO: Created: latency-svc-f5mhx -Jun 4 19:26:21.634: INFO: Got endpoints: latency-svc-f5mhx [231.001825ms] -Jun 4 19:26:21.641: INFO: Created: latency-svc-krc58 -Jun 4 19:26:21.649: INFO: Got endpoints: latency-svc-krc58 [231.257422ms] -Jun 4 19:26:21.657: INFO: Created: latency-svc-gv5cn -Jun 4 19:26:21.666: INFO: Got endpoints: latency-svc-gv5cn [232.870815ms] -Jun 4 19:26:21.673: INFO: Created: latency-svc-fchsl -Jun 4 19:26:21.681: INFO: Got endpoints: latency-svc-fchsl [232.673866ms] -Jun 4 19:26:21.688: INFO: Created: latency-svc-d8n5f -Jun 4 19:26:21.697: INFO: Got endpoints: latency-svc-d8n5f [232.695717ms] -Jun 4 19:26:21.705: INFO: Created: latency-svc-m6gqh -Jun 4 19:26:21.713: INFO: Got endpoints: latency-svc-m6gqh [233.21085ms] -Jun 4 19:26:21.720: INFO: Created: latency-svc-m8zw4 -Jun 4 19:26:21.729: INFO: Got endpoints: latency-svc-m8zw4 [235.304957ms] -Jun 4 19:26:21.737: INFO: Created: latency-svc-cqnvx -Jun 4 19:26:21.753: INFO: Got endpoints: latency-svc-cqnvx [243.780018ms] -Jun 4 19:26:21.758: INFO: Created: latency-svc-btl89 -Jun 4 19:26:21.767: INFO: Got endpoints: latency-svc-btl89 [241.513682ms] -Jun 4 19:26:21.773: INFO: Created: latency-svc-nkxd8 -Jun 4 19:26:21.782: INFO: Got endpoints: latency-svc-nkxd8 [239.68161ms] -Jun 4 19:26:21.789: INFO: Created: latency-svc-czz8h -Jun 4 19:26:21.798: INFO: Got endpoints: latency-svc-czz8h [238.051105ms] -Jun 4 19:26:21.805: INFO: Created: latency-svc-7tnsg -Jun 4 19:26:21.813: INFO: Got endpoints: latency-svc-7tnsg [238.818445ms] -Jun 4 19:26:21.821: INFO: Created: latency-svc-x6hpr -Jun 4 19:26:21.828: INFO: Got endpoints: latency-svc-x6hpr [238.97533ms] -Jun 4 19:26:21.835: INFO: Created: latency-svc-ds664 -Jun 4 19:26:21.842: INFO: Got endpoints: latency-svc-ds664 [239.310052ms] -Jun 4 19:26:21.911: INFO: Created: latency-svc-scpv6 -Jun 4 19:26:21.915: INFO: Got endpoints: latency-svc-scpv6 [298.035746ms] -Jun 4 19:26:21.926: INFO: Created: latency-svc-9rwhw -Jun 4 19:26:21.934: INFO: Got endpoints: latency-svc-9rwhw [299.65863ms] -Jun 4 19:26:21.940: INFO: Created: latency-svc-mvs75 -Jun 4 19:26:21.949: INFO: Got endpoints: latency-svc-mvs75 [299.904649ms] -Jun 4 19:26:21.956: INFO: Created: latency-svc-5df4p -Jun 4 19:26:21.965: INFO: Got endpoints: latency-svc-5df4p [298.857944ms] -Jun 4 19:26:21.972: INFO: Created: latency-svc-wn7rx -Jun 4 19:26:21.979: INFO: Got endpoints: latency-svc-wn7rx [298.134961ms] -Jun 4 19:26:21.988: INFO: Created: latency-svc-44pdt -Jun 4 19:26:21.997: INFO: Got endpoints: latency-svc-44pdt [300.249911ms] -Jun 4 19:26:22.005: INFO: Created: latency-svc-2rccq -Jun 4 19:26:22.013: INFO: Got endpoints: latency-svc-2rccq [300.061302ms] -Jun 4 19:26:22.022: INFO: Created: latency-svc-mb6qw -Jun 4 19:26:22.030: INFO: Got endpoints: latency-svc-mb6qw [300.580208ms] -Jun 4 19:26:22.038: INFO: Created: latency-svc-qj5ht -Jun 4 19:26:22.045: INFO: Got endpoints: latency-svc-qj5ht [292.695884ms] -Jun 4 19:26:22.052: INFO: Created: latency-svc-9npt6 -Jun 4 19:26:22.062: INFO: Got endpoints: latency-svc-9npt6 [294.847693ms] -Jun 4 19:26:22.069: INFO: Created: latency-svc-l5w92 -Jun 4 19:26:22.077: INFO: Got endpoints: latency-svc-l5w92 [294.701211ms] -Jun 4 19:26:22.084: INFO: Created: latency-svc-s6lsc -Jun 4 19:26:22.094: INFO: Got endpoints: latency-svc-s6lsc [296.428111ms] -Jun 4 19:26:22.101: INFO: Created: latency-svc-jd5hc -Jun 4 19:26:22.109: INFO: Got endpoints: latency-svc-jd5hc [296.037611ms] -Jun 4 19:26:22.116: INFO: Created: latency-svc-zd6qg -Jun 4 19:26:22.124: INFO: Got endpoints: latency-svc-zd6qg [296.398208ms] -Jun 4 19:26:22.132: INFO: Created: latency-svc-4xcvc -Jun 4 19:26:22.139: INFO: Got endpoints: latency-svc-4xcvc [296.172239ms] -Jun 4 19:26:22.146: INFO: Created: latency-svc-9w7z2 -Jun 4 19:26:22.154: INFO: Got endpoints: latency-svc-9w7z2 [238.372425ms] -Jun 4 19:26:22.162: INFO: Created: latency-svc-7znw5 -Jun 4 19:26:22.170: INFO: Got endpoints: latency-svc-7znw5 [236.670162ms] -Jun 4 19:26:22.178: INFO: Created: latency-svc-qzmzn -Jun 4 19:26:22.187: INFO: Got endpoints: latency-svc-qzmzn [238.4903ms] -Jun 4 19:26:22.194: INFO: Created: latency-svc-kkmk6 -Jun 4 19:26:22.202: INFO: Got endpoints: latency-svc-kkmk6 [237.43408ms] -Jun 4 19:26:22.210: INFO: Created: latency-svc-fpnxn -Jun 4 19:26:22.218: INFO: Got endpoints: latency-svc-fpnxn [238.782921ms] -Jun 4 19:26:22.225: INFO: Created: latency-svc-5tftz -Jun 4 19:26:22.233: INFO: Got endpoints: latency-svc-5tftz [235.816764ms] -Jun 4 19:26:22.241: INFO: Created: latency-svc-pnxwp -Jun 4 19:26:22.250: INFO: Got endpoints: latency-svc-pnxwp [236.525656ms] -Jun 4 19:26:22.258: INFO: Created: latency-svc-9kndb -Jun 4 19:26:22.266: INFO: Got endpoints: latency-svc-9kndb [235.846542ms] -Jun 4 19:26:22.273: INFO: Created: latency-svc-sqlqq -Jun 4 19:26:22.281: INFO: Got endpoints: latency-svc-sqlqq [235.341964ms] -Jun 4 19:26:22.287: INFO: Created: latency-svc-wgq9p -Jun 4 19:26:22.295: INFO: Got endpoints: latency-svc-wgq9p [232.866113ms] -Jun 4 19:26:22.302: INFO: Created: latency-svc-cr4kl -Jun 4 19:26:22.310: INFO: Got endpoints: latency-svc-cr4kl [232.802015ms] -Jun 4 19:26:22.318: INFO: Created: latency-svc-wjhlr -Jun 4 19:26:22.326: INFO: Got endpoints: latency-svc-wjhlr [231.834031ms] -Jun 4 19:26:22.334: INFO: Created: latency-svc-9bqzz -Jun 4 19:26:22.342: INFO: Got endpoints: latency-svc-9bqzz [232.506815ms] -Jun 4 19:26:22.350: INFO: Created: latency-svc-pw2l4 -Jun 4 19:26:22.358: INFO: Got endpoints: latency-svc-pw2l4 [233.483791ms] -Jun 4 19:26:22.367: INFO: Created: latency-svc-gcqqk -Jun 4 19:26:22.373: INFO: Got endpoints: latency-svc-gcqqk [234.178026ms] -Jun 4 19:26:22.388: INFO: Created: latency-svc-s5dxf -Jun 4 19:26:22.389: INFO: Got endpoints: latency-svc-s5dxf [235.227035ms] -Jun 4 19:26:22.398: INFO: Created: latency-svc-kz6n6 -Jun 4 19:26:22.404: INFO: Got endpoints: latency-svc-kz6n6 [233.565858ms] -Jun 4 19:26:22.404: INFO: Latencies: [38.659696ms 94.838171ms 115.579039ms 124.551224ms 143.360159ms 158.037564ms 172.91278ms 196.132435ms 211.917412ms 227.119888ms 227.881408ms 227.912325ms 228.440585ms 228.677877ms 229.75267ms 230.443672ms 230.61579ms 230.814682ms 230.815417ms 230.833303ms 231.001825ms 231.08744ms 231.191698ms 231.248681ms 231.257422ms 231.28169ms 231.52112ms 231.546282ms 231.565323ms 231.566828ms 231.690879ms 231.834031ms 231.874912ms 231.897495ms 231.976975ms 232.020984ms 232.075325ms 232.090754ms 232.255983ms 232.298929ms 232.344729ms 232.506815ms 232.508941ms 232.658511ms 232.673866ms 232.695717ms 232.706997ms 232.751054ms 232.761262ms 232.777174ms 232.802015ms 232.866113ms 232.870815ms 232.878474ms 233.153527ms 233.194166ms 233.21085ms 233.221703ms 233.242337ms 233.38584ms 233.446789ms 233.483791ms 233.522445ms 233.523075ms 233.565858ms 233.573503ms 233.743323ms 233.760024ms 233.773298ms 233.992645ms 234.068268ms 234.121928ms 234.146786ms 234.178026ms 234.513841ms 234.515191ms 234.601356ms 234.624926ms 234.734525ms 234.826591ms 235.100366ms 235.227035ms 235.267309ms 235.304957ms 235.341964ms 235.442362ms 235.45258ms 235.495642ms 235.524745ms 235.594931ms 235.816764ms 235.846542ms 236.057045ms 236.3418ms 236.460661ms 236.496406ms 236.525656ms 236.630127ms 236.670162ms 236.715522ms 236.74393ms 236.758339ms 236.852557ms 237.361084ms 237.43408ms 237.756003ms 237.789968ms 237.988004ms 238.024564ms 238.051105ms 238.372425ms 238.4903ms 238.608432ms 238.675244ms 238.782921ms 238.8081ms 238.818445ms 238.97533ms 238.99403ms 239.052778ms 239.138462ms 239.2904ms 239.310052ms 239.510071ms 239.541915ms 239.547037ms 239.68161ms 239.745753ms 240.214675ms 240.535233ms 240.681004ms 240.772025ms 240.779753ms 240.78399ms 240.911574ms 241.00784ms 241.105214ms 241.131855ms 241.140536ms 241.188486ms 241.200231ms 241.205522ms 241.243986ms 241.404952ms 241.513682ms 241.550934ms 241.817434ms 242.446175ms 242.533617ms 242.669947ms 242.918768ms 242.954464ms 243.615876ms 243.753153ms 243.780018ms 244.353961ms 244.468587ms 244.886572ms 245.049301ms 246.414021ms 247.145301ms 250.493819ms 251.116341ms 253.048372ms 253.762432ms 254.754267ms 255.478673ms 256.47726ms 257.19528ms 257.432063ms 258.468479ms 258.567851ms 259.030979ms 259.533973ms 259.600212ms 259.915244ms 260.034823ms 260.707759ms 261.32782ms 261.522799ms 262.895617ms 274.558732ms 290.875109ms 292.695884ms 294.701211ms 294.847693ms 296.037611ms 296.172239ms 296.398208ms 296.428111ms 298.035746ms 298.134961ms 298.857944ms 299.65863ms 299.904649ms 300.061302ms 300.249911ms 300.580208ms 311.841661ms 335.138389ms] -Jun 4 19:26:22.404: INFO: 50 %ile: 236.74393ms -Jun 4 19:26:22.404: INFO: 90 %ile: 262.895617ms -Jun 4 19:26:22.404: INFO: 99 %ile: 311.841661ms -Jun 4 19:26:22.404: INFO: Total sample count: 200 -[AfterEach] [sig-network] Service endpoints latency - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 -Jun 4 19:26:22.404: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "svc-latency-9600" for this suite. -Jun 4 19:26:40.466: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -Jun 4 19:26:41.067: INFO: namespace svc-latency-9600 deletion completed in 18.645403625s - -• [SLOW TEST:24.388 seconds] -[sig-network] Service endpoints latency -/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:22 - should not be very high [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 ------------------------------- -SSSSSSSSSSSS ------------------------------- -[sig-storage] Secrets - should be consumable from pods in volume with mappings [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -[BeforeEach] [sig-storage] Secrets - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 -STEP: Creating a kubernetes client -Jun 4 19:26:41.067: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504 -STEP: Building a namespace api object, basename secrets -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in secrets-1105 -STEP: Waiting for a default service account to be provisioned in namespace -[It] should be consumable from pods in volume with mappings [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -STEP: Creating secret with name secret-test-map-ae233c43-86fe-11e9-8d1b-467ee19922ac -STEP: Creating a pod to test consume secrets -Jun 4 19:26:41.326: INFO: Waiting up to 5m0s for pod "pod-secrets-ae258bea-86fe-11e9-8d1b-467ee19922ac" in namespace "secrets-1105" to be "success or failure" -Jun 4 19:26:41.337: INFO: Pod "pod-secrets-ae258bea-86fe-11e9-8d1b-467ee19922ac": Phase="Pending", Reason="", readiness=false. Elapsed: 10.550291ms -Jun 4 19:26:43.349: INFO: Pod "pod-secrets-ae258bea-86fe-11e9-8d1b-467ee19922ac": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.022968635s -STEP: Saw pod success -Jun 4 19:26:43.349: INFO: Pod "pod-secrets-ae258bea-86fe-11e9-8d1b-467ee19922ac" satisfied condition "success or failure" -Jun 4 19:26:43.367: INFO: Trying to get logs from node 10.212.23.161 pod pod-secrets-ae258bea-86fe-11e9-8d1b-467ee19922ac container secret-volume-test: -STEP: delete the pod -Jun 4 19:26:43.429: INFO: Waiting for pod pod-secrets-ae258bea-86fe-11e9-8d1b-467ee19922ac to disappear -Jun 4 19:26:43.439: INFO: Pod pod-secrets-ae258bea-86fe-11e9-8d1b-467ee19922ac no longer exists -[AfterEach] [sig-storage] Secrets - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 -Jun 4 19:26:43.440: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "secrets-1105" for this suite. -Jun 4 19:26:49.501: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -Jun 4 19:26:49.884: INFO: namespace secrets-1105 deletion completed in 6.426006716s - -• [SLOW TEST:8.817 seconds] -[sig-storage] Secrets -/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/secrets_volume.go:33 - should be consumable from pods in volume with mappings [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687 + should create and stop a replication controller [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 ------------------------------ -SSSSSSSS +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[k8s.io] Docker Containers - should be able to override the image's default command (docker entrypoint) [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -[BeforeEach] [k8s.io] Docker Containers - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +[sig-storage] EmptyDir volumes + should support (non-root,0644,default) [LinuxOnly] [NodeConformance] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [sig-storage] EmptyDir volumes + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 STEP: Creating a kubernetes client -Jun 4 19:26:49.885: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504 -STEP: Building a namespace api object, basename containers -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in containers-2720 +Jun 18 12:41:55.971: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150 +STEP: Building a namespace api object, basename emptydir +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in emptydir-1662 STEP: Waiting for a default service account to be provisioned in namespace -[It] should be able to override the image's default command (docker entrypoint) [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -STEP: Creating a pod to test override command -Jun 4 19:26:50.180: INFO: Waiting up to 5m0s for pod "client-containers-b36d32f2-86fe-11e9-8d1b-467ee19922ac" in namespace "containers-2720" to be "success or failure" -Jun 4 19:26:50.193: INFO: Pod "client-containers-b36d32f2-86fe-11e9-8d1b-467ee19922ac": Phase="Pending", Reason="", readiness=false. Elapsed: 12.563182ms -Jun 4 19:26:52.205: INFO: Pod "client-containers-b36d32f2-86fe-11e9-8d1b-467ee19922ac": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.024859457s +[It] should support (non-root,0644,default) [LinuxOnly] [NodeConformance] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: Creating a pod to test emptydir 0644 on node default medium +Jun 18 12:41:56.196: INFO: Waiting up to 5m0s for pod "pod-74dcee08-91c6-11e9-a25d-8608290c688a" in namespace "emptydir-1662" to be "success or failure" +Jun 18 12:41:56.212: INFO: Pod "pod-74dcee08-91c6-11e9-a25d-8608290c688a": Phase="Pending", Reason="", readiness=false. Elapsed: 15.968677ms +Jun 18 12:41:58.225: INFO: Pod "pod-74dcee08-91c6-11e9-a25d-8608290c688a": Phase="Pending", Reason="", readiness=false. Elapsed: 2.029226162s +Jun 18 12:42:00.239: INFO: Pod "pod-74dcee08-91c6-11e9-a25d-8608290c688a": Phase="Pending", Reason="", readiness=false. Elapsed: 4.042684329s +Jun 18 12:42:02.253: INFO: Pod "pod-74dcee08-91c6-11e9-a25d-8608290c688a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.0565609s STEP: Saw pod success -Jun 4 19:26:52.205: INFO: Pod "client-containers-b36d32f2-86fe-11e9-8d1b-467ee19922ac" satisfied condition "success or failure" -Jun 4 19:26:52.217: INFO: Trying to get logs from node 10.212.23.164 pod client-containers-b36d32f2-86fe-11e9-8d1b-467ee19922ac container test-container: +Jun 18 12:42:02.253: INFO: Pod "pod-74dcee08-91c6-11e9-a25d-8608290c688a" satisfied condition "success or failure" +Jun 18 12:42:02.267: INFO: Trying to get logs from node 10.72.74.184 pod pod-74dcee08-91c6-11e9-a25d-8608290c688a container test-container: STEP: delete the pod -Jun 4 19:26:52.295: INFO: Waiting for pod client-containers-b36d32f2-86fe-11e9-8d1b-467ee19922ac to disappear -Jun 4 19:26:52.307: INFO: Pod client-containers-b36d32f2-86fe-11e9-8d1b-467ee19922ac no longer exists -[AfterEach] [k8s.io] Docker Containers - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 -Jun 4 19:26:52.307: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "containers-2720" for this suite. -Jun 4 19:26:58.367: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -Jun 4 19:26:58.778: INFO: namespace containers-2720 deletion completed in 6.456212574s +Jun 18 12:42:02.358: INFO: Waiting for pod pod-74dcee08-91c6-11e9-a25d-8608290c688a to disappear +Jun 18 12:42:02.371: INFO: Pod pod-74dcee08-91c6-11e9-a25d-8608290c688a no longer exists +[AfterEach] [sig-storage] EmptyDir volumes + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Jun 18 12:42:02.371: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "emptydir-1662" for this suite. +Jun 18 12:42:08.413: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:42:09.080: INFO: namespace emptydir-1662 deletion completed in 6.693078329s -• [SLOW TEST:8.894 seconds] -[k8s.io] Docker Containers -/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687 - should be able to override the image's default command (docker entrypoint) [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +• [SLOW TEST:13.109 seconds] +[sig-storage] EmptyDir volumes +/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:41 + should support (non-root,0644,default) [LinuxOnly] [NodeConformance] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 ------------------------------ -SSSSSSSSSSSSSSS +SSS ------------------------------ -[sig-storage] Downward API volume - should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -[BeforeEach] [sig-storage] Downward API volume - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +[sig-apps] Daemon set [Serial] + should run and stop complex daemon [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [sig-apps] Daemon set [Serial] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 STEP: Creating a kubernetes client -Jun 4 19:26:58.779: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504 -STEP: Building a namespace api object, basename downward-api -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in downward-api-3865 +Jun 18 12:42:09.081: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150 +STEP: Building a namespace api object, basename daemonsets +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in daemonsets-6213 STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [sig-storage] Downward API volume - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:39 -[It] should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -STEP: Creating a pod to test downward API volume plugin -Jun 4 19:26:59.083: INFO: Waiting up to 5m0s for pod "downwardapi-volume-b8bb31a8-86fe-11e9-8d1b-467ee19922ac" in namespace "downward-api-3865" to be "success or failure" -Jun 4 19:26:59.099: INFO: Pod "downwardapi-volume-b8bb31a8-86fe-11e9-8d1b-467ee19922ac": Phase="Pending", Reason="", readiness=false. Elapsed: 15.974461ms -Jun 4 19:27:01.113: INFO: Pod "downwardapi-volume-b8bb31a8-86fe-11e9-8d1b-467ee19922ac": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.030001254s -STEP: Saw pod success -Jun 4 19:27:01.113: INFO: Pod "downwardapi-volume-b8bb31a8-86fe-11e9-8d1b-467ee19922ac" satisfied condition "success or failure" -Jun 4 19:27:01.124: INFO: Trying to get logs from node 10.212.23.189 pod downwardapi-volume-b8bb31a8-86fe-11e9-8d1b-467ee19922ac container client-container: -STEP: delete the pod -Jun 4 19:27:01.192: INFO: Waiting for pod downwardapi-volume-b8bb31a8-86fe-11e9-8d1b-467ee19922ac to disappear -Jun 4 19:27:01.203: INFO: Pod downwardapi-volume-b8bb31a8-86fe-11e9-8d1b-467ee19922ac no longer exists -[AfterEach] [sig-storage] Downward API volume - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 -Jun 4 19:27:01.203: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "downward-api-3865" for this suite. -Jun 4 19:27:07.263: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -Jun 4 19:27:07.592: INFO: namespace downward-api-3865 deletion completed in 6.374336153s +[BeforeEach] [sig-apps] Daemon set [Serial] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:102 +[It] should run and stop complex daemon [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +Jun 18 12:42:09.363: INFO: Creating daemon "daemon-set" with a node selector +STEP: Initially, daemon pods should not be running on any nodes. +Jun 18 12:42:09.391: INFO: Number of nodes with available pods: 0 +Jun 18 12:42:09.391: INFO: Number of running nodes: 0, number of available pods: 0 +STEP: Change node label to blue, check that daemon pod is launched. +Jun 18 12:42:09.468: INFO: Number of nodes with available pods: 0 +Jun 18 12:42:09.468: INFO: Node 10.72.74.138 is running more than one daemon pod +Jun 18 12:42:10.484: INFO: Number of nodes with available pods: 0 +Jun 18 12:42:10.484: INFO: Node 10.72.74.138 is running more than one daemon pod +Jun 18 12:42:11.992: INFO: Number of nodes with available pods: 1 +Jun 18 12:42:11.992: INFO: Number of running nodes: 1, number of available pods: 1 +STEP: Update the node label to green, and wait for daemons to be unscheduled +Jun 18 12:42:12.055: INFO: Number of nodes with available pods: 1 +Jun 18 12:42:12.056: INFO: Number of running nodes: 0, number of available pods: 1 +Jun 18 12:42:13.070: INFO: Number of nodes with available pods: 0 +Jun 18 12:42:13.070: INFO: Number of running nodes: 0, number of available pods: 0 +STEP: Update DaemonSet node selector to green, and change its update strategy to RollingUpdate +Jun 18 12:42:13.099: INFO: Number of nodes with available pods: 0 +Jun 18 12:42:13.099: INFO: Node 10.72.74.138 is running more than one daemon pod +Jun 18 12:42:14.114: INFO: Number of nodes with available pods: 0 +Jun 18 12:42:14.114: INFO: Node 10.72.74.138 is running more than one daemon pod +Jun 18 12:42:15.114: INFO: Number of nodes with available pods: 0 +Jun 18 12:42:15.114: INFO: Node 10.72.74.138 is running more than one daemon pod +Jun 18 12:42:16.114: INFO: Number of nodes with available pods: 0 +Jun 18 12:42:16.114: INFO: Node 10.72.74.138 is running more than one daemon pod +Jun 18 12:42:17.114: INFO: Number of nodes with available pods: 0 +Jun 18 12:42:17.115: INFO: Node 10.72.74.138 is running more than one daemon pod +Jun 18 12:42:18.114: INFO: Number of nodes with available pods: 0 +Jun 18 12:42:18.114: INFO: Node 10.72.74.138 is running more than one daemon pod +Jun 18 12:42:19.113: INFO: Number of nodes with available pods: 0 +Jun 18 12:42:19.113: INFO: Node 10.72.74.138 is running more than one daemon pod +Jun 18 12:42:20.114: INFO: Number of nodes with available pods: 0 +Jun 18 12:42:20.114: INFO: Node 10.72.74.138 is running more than one daemon pod +Jun 18 12:42:21.113: INFO: Number of nodes with available pods: 0 +Jun 18 12:42:21.113: INFO: Node 10.72.74.138 is running more than one daemon pod +Jun 18 12:42:22.113: INFO: Number of nodes with available pods: 0 +Jun 18 12:42:22.113: INFO: Node 10.72.74.138 is running more than one daemon pod +Jun 18 12:42:23.113: INFO: Number of nodes with available pods: 0 +Jun 18 12:42:23.113: INFO: Node 10.72.74.138 is running more than one daemon pod +Jun 18 12:42:24.114: INFO: Number of nodes with available pods: 0 +Jun 18 12:42:24.114: INFO: Node 10.72.74.138 is running more than one daemon pod +Jun 18 12:42:25.115: INFO: Number of nodes with available pods: 0 +Jun 18 12:42:25.115: INFO: Node 10.72.74.138 is running more than one daemon pod +Jun 18 12:42:26.114: INFO: Number of nodes with available pods: 1 +Jun 18 12:42:26.114: INFO: Number of running nodes: 1, number of available pods: 1 +[AfterEach] [sig-apps] Daemon set [Serial] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:68 +STEP: Deleting DaemonSet "daemon-set" +STEP: deleting DaemonSet.extensions daemon-set in namespace daemonsets-6213, will wait for the garbage collector to delete the pods +Jun 18 12:42:26.237: INFO: Deleting DaemonSet.extensions daemon-set took: 32.819547ms +Jun 18 12:42:26.437: INFO: Terminating DaemonSet.extensions daemon-set pods took: 200.234069ms +Jun 18 12:42:33.950: INFO: Number of nodes with available pods: 0 +Jun 18 12:42:33.950: INFO: Number of running nodes: 0, number of available pods: 0 +Jun 18 12:42:33.962: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"selfLink":"/apis/apps/v1/namespaces/daemonsets-6213/daemonsets","resourceVersion":"100325"},"items":null} -• [SLOW TEST:8.813 seconds] -[sig-storage] Downward API volume -/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:34 - should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +Jun 18 12:42:33.974: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"selfLink":"/api/v1/namespaces/daemonsets-6213/pods","resourceVersion":"100325"},"items":null} + +[AfterEach] [sig-apps] Daemon set [Serial] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Jun 18 12:42:34.048: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "daemonsets-6213" for this suite. +Jun 18 12:42:42.086: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:42:42.482: INFO: namespace daemonsets-6213 deletion completed in 8.42165216s + +• [SLOW TEST:33.402 seconds] +[sig-apps] Daemon set [Serial] +/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22 + should run and stop complex daemon [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +S ------------------------------ -[k8s.io] Probing container - should have monotonically increasing restart count [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -[BeforeEach] [k8s.io] Probing container - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +[k8s.io] Pods + should get a host IP [NodeConformance] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [k8s.io] Pods + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 STEP: Creating a kubernetes client -Jun 4 19:27:07.594: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504 -STEP: Building a namespace api object, basename container-probe -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in container-probe-3883 +Jun 18 12:42:42.482: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150 +STEP: Building a namespace api object, basename pods +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in pods-1953 STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [k8s.io] Probing container - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/container_probe.go:51 -[It] should have monotonically increasing restart count [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -STEP: Creating pod liveness-http in namespace container-probe-3883 -Jun 4 19:27:13.851: INFO: Started pod liveness-http in namespace container-probe-3883 -STEP: checking the pod's current state and verifying that restartCount is present -Jun 4 19:27:13.864: INFO: Initial restart count of pod liveness-http is 0 -Jun 4 19:27:25.966: INFO: Restart count of pod container-probe-3883/liveness-http is now 1 (12.102220139s elapsed) -Jun 4 19:27:44.172: INFO: Restart count of pod container-probe-3883/liveness-http is now 2 (30.308283577s elapsed) -Jun 4 19:28:04.334: INFO: Restart count of pod container-probe-3883/liveness-http is now 3 (50.470592274s elapsed) -Jun 4 19:28:24.468: INFO: Restart count of pod container-probe-3883/liveness-http is now 4 (1m10.604372443s elapsed) -Jun 4 19:29:36.962: INFO: Restart count of pod container-probe-3883/liveness-http is now 5 (2m23.098048593s elapsed) -STEP: deleting the pod -[AfterEach] [k8s.io] Probing container - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 -Jun 4 19:29:37.002: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "container-probe-3883" for this suite. -Jun 4 19:29:43.064: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -Jun 4 19:29:43.487: INFO: namespace container-probe-3883 deletion completed in 6.46791316s +[BeforeEach] [k8s.io] Pods + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:135 +[It] should get a host IP [NodeConformance] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: creating pod +Jun 18 12:42:44.758: INFO: Pod pod-hostip-90951756-91c6-11e9-a25d-8608290c688a has hostIP: 10.72.74.138 +[AfterEach] [k8s.io] Pods + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Jun 18 12:42:44.758: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "pods-1953" for this suite. +Jun 18 12:43:08.801: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:43:09.260: INFO: namespace pods-1953 deletion completed in 24.48686398s -• [SLOW TEST:155.893 seconds] -[k8s.io] Probing container -/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687 - should have monotonically increasing restart count [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +• [SLOW TEST:26.778 seconds] +[k8s.io] Pods +/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687 + should get a host IP [NodeConformance] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 ------------------------------ -SSSS +SSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-node] Downward API - should provide pod name, namespace and IP address as env vars [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -[BeforeEach] [sig-node] Downward API - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +[sig-api-machinery] Watchers + should observe an object deletion if it stops meeting the requirements of the selector [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [sig-api-machinery] Watchers + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 STEP: Creating a kubernetes client -Jun 4 19:29:43.487: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504 -STEP: Building a namespace api object, basename downward-api -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in downward-api-5732 +Jun 18 12:43:09.260: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150 +STEP: Building a namespace api object, basename watch +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in watch-9234 STEP: Waiting for a default service account to be provisioned in namespace -[It] should provide pod name, namespace and IP address as env vars [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -STEP: Creating a pod to test downward api env vars -Jun 4 19:29:43.780: INFO: Waiting up to 5m0s for pod "downward-api-1ae6615c-86ff-11e9-8d1b-467ee19922ac" in namespace "downward-api-5732" to be "success or failure" -Jun 4 19:29:43.792: INFO: Pod "downward-api-1ae6615c-86ff-11e9-8d1b-467ee19922ac": Phase="Pending", Reason="", readiness=false. Elapsed: 11.674318ms -Jun 4 19:29:45.805: INFO: Pod "downward-api-1ae6615c-86ff-11e9-8d1b-467ee19922ac": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.024706941s -STEP: Saw pod success -Jun 4 19:29:45.805: INFO: Pod "downward-api-1ae6615c-86ff-11e9-8d1b-467ee19922ac" satisfied condition "success or failure" -Jun 4 19:29:45.867: INFO: Trying to get logs from node 10.212.23.164 pod downward-api-1ae6615c-86ff-11e9-8d1b-467ee19922ac container dapi-container: -STEP: delete the pod -Jun 4 19:29:45.937: INFO: Waiting for pod downward-api-1ae6615c-86ff-11e9-8d1b-467ee19922ac to disappear -Jun 4 19:29:45.948: INFO: Pod downward-api-1ae6615c-86ff-11e9-8d1b-467ee19922ac no longer exists -[AfterEach] [sig-node] Downward API - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 -Jun 4 19:29:45.948: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "downward-api-5732" for this suite. -Jun 4 19:29:52.004: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -Jun 4 19:29:52.352: INFO: namespace downward-api-5732 deletion completed in 6.389195297s +[It] should observe an object deletion if it stops meeting the requirements of the selector [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: creating a watch on configmaps with a certain label +STEP: creating a new configmap +STEP: modifying the configmap once +STEP: changing the label value of the configmap +STEP: Expecting to observe a delete notification for the watched object +Jun 18 12:43:09.542: INFO: Got : ADDED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-label-changed,GenerateName:,Namespace:watch-9234,SelfLink:/api/v1/namespaces/watch-9234/configmaps/e2e-watch-test-label-changed,UID:a08d82c6-91c6-11e9-a08a-ee7a14707756,ResourceVersion:100453,Generation:0,CreationTimestamp:2019-06-18 12:43:09 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: label-changed-and-restored,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{},BinaryData:map[string][]byte{},} +Jun 18 12:43:09.542: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-label-changed,GenerateName:,Namespace:watch-9234,SelfLink:/api/v1/namespaces/watch-9234/configmaps/e2e-watch-test-label-changed,UID:a08d82c6-91c6-11e9-a08a-ee7a14707756,ResourceVersion:100454,Generation:0,CreationTimestamp:2019-06-18 12:43:09 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: label-changed-and-restored,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{mutation: 1,},BinaryData:map[string][]byte{},} +Jun 18 12:43:09.542: INFO: Got : DELETED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-label-changed,GenerateName:,Namespace:watch-9234,SelfLink:/api/v1/namespaces/watch-9234/configmaps/e2e-watch-test-label-changed,UID:a08d82c6-91c6-11e9-a08a-ee7a14707756,ResourceVersion:100455,Generation:0,CreationTimestamp:2019-06-18 12:43:09 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: label-changed-and-restored,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{mutation: 1,},BinaryData:map[string][]byte{},} +STEP: modifying the configmap a second time +STEP: Expecting not to observe a notification because the object no longer meets the selector's requirements +STEP: changing the label value of the configmap back +STEP: modifying the configmap a third time +STEP: deleting the configmap +STEP: Expecting to observe an add notification for the watched object when the label value was restored +Jun 18 12:43:19.656: INFO: Got : ADDED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-label-changed,GenerateName:,Namespace:watch-9234,SelfLink:/api/v1/namespaces/watch-9234/configmaps/e2e-watch-test-label-changed,UID:a08d82c6-91c6-11e9-a08a-ee7a14707756,ResourceVersion:100474,Generation:0,CreationTimestamp:2019-06-18 12:43:09 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: label-changed-and-restored,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},} +Jun 18 12:43:19.656: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-label-changed,GenerateName:,Namespace:watch-9234,SelfLink:/api/v1/namespaces/watch-9234/configmaps/e2e-watch-test-label-changed,UID:a08d82c6-91c6-11e9-a08a-ee7a14707756,ResourceVersion:100475,Generation:0,CreationTimestamp:2019-06-18 12:43:09 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: label-changed-and-restored,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{mutation: 3,},BinaryData:map[string][]byte{},} +Jun 18 12:43:19.656: INFO: Got : DELETED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-label-changed,GenerateName:,Namespace:watch-9234,SelfLink:/api/v1/namespaces/watch-9234/configmaps/e2e-watch-test-label-changed,UID:a08d82c6-91c6-11e9-a08a-ee7a14707756,ResourceVersion:100476,Generation:0,CreationTimestamp:2019-06-18 12:43:09 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: label-changed-and-restored,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{mutation: 3,},BinaryData:map[string][]byte{},} +[AfterEach] [sig-api-machinery] Watchers + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Jun 18 12:43:19.656: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "watch-9234" for this suite. +Jun 18 12:43:25.701: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:43:26.887: INFO: namespace watch-9234 deletion completed in 7.216292164s -• [SLOW TEST:8.865 seconds] -[sig-node] Downward API -/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downward_api.go:38 - should provide pod name, namespace and IP address as env vars [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +• [SLOW TEST:17.627 seconds] +[sig-api-machinery] Watchers +/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:22 + should observe an object deletion if it stops meeting the requirements of the selector [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSSSS +SS ------------------------------ -[sig-cli] Kubectl client [k8s.io] Kubectl run deployment - should create a deployment from an image [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[sig-cli] Kubectl client [k8s.io] Kubectl run --rm job + should create a job from an image, then delete the job [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 [BeforeEach] [sig-cli] Kubectl client - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 STEP: Creating a kubernetes client -Jun 4 19:29:52.352: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504 +Jun 18 12:43:26.889: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150 STEP: Building a namespace api object, basename kubectl -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-4248 +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-3152 STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:213 -[BeforeEach] [k8s.io] Kubectl run deployment - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1455 -[It] should create a deployment from an image [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -STEP: running the image docker.io/library/nginx:1.14-alpine -Jun 4 19:29:52.566: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 run e2e-test-nginx-deployment --image=docker.io/library/nginx:1.14-alpine --generator=deployment/v1beta1 --namespace=kubectl-4248' -Jun 4 19:29:52.657: INFO: stderr: "kubectl run --generator=deployment/v1beta1 is DEPRECATED and will be removed in a future version. Use kubectl run --generator=run-pod/v1 or kubectl create instead.\n" -Jun 4 19:29:52.657: INFO: stdout: "deployment.extensions/e2e-test-nginx-deployment created\n" -STEP: verifying the deployment e2e-test-nginx-deployment was created -STEP: verifying the pod controlled by deployment e2e-test-nginx-deployment was created -[AfterEach] [k8s.io] Kubectl run deployment - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1460 -Jun 4 19:29:56.690: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 delete deployment e2e-test-nginx-deployment --namespace=kubectl-4248' -Jun 4 19:29:56.804: INFO: stderr: "" -Jun 4 19:29:56.804: INFO: stdout: "deployment.extensions \"e2e-test-nginx-deployment\" deleted\n" + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:213 +[It] should create a job from an image, then delete the job [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: executing a command with run --rm and attach with stdin +Jun 18 12:43:27.098: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 --namespace=kubectl-3152 run e2e-test-rm-busybox-job --image=docker.io/library/busybox:1.29 --rm=true --generator=job/v1 --restart=OnFailure --attach=true --stdin -- sh -c cat && echo 'stdin closed'' +Jun 18 12:43:28.789: INFO: stderr: "kubectl run --generator=job/v1 is DEPRECATED and will be removed in a future version. Use kubectl run --generator=run-pod/v1 or kubectl create instead.\nIf you don't see a command prompt, try pressing enter.\n" +Jun 18 12:43:28.789: INFO: stdout: "abcd1234stdin closed\njob.batch \"e2e-test-rm-busybox-job\" deleted\n" +STEP: verifying the job e2e-test-rm-busybox-job was deleted [AfterEach] [sig-cli] Kubectl client - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 -Jun 4 19:29:56.804: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "kubectl-4248" for this suite. -Jun 4 19:30:20.916: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -Jun 4 19:30:21.254: INFO: namespace kubectl-4248 deletion completed in 24.38296244s + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Jun 18 12:43:30.815: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "kubectl-3152" for this suite. +Jun 18 12:43:42.856: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:43:43.266: INFO: namespace kubectl-3152 deletion completed in 12.436852825s -• [SLOW TEST:28.902 seconds] +• [SLOW TEST:16.377 seconds] [sig-cli] Kubectl client -/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 - [k8s.io] Kubectl run deployment - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687 - should create a deployment from an image [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 + [k8s.io] Kubectl run --rm job + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687 + should create a job from an image, then delete the job [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 ------------------------------ -SSSSSSSSSSSSS +S ------------------------------ -[k8s.io] InitContainer [NodeConformance] - should not start app containers if init containers fail on a RestartAlways pod [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -[BeforeEach] [k8s.io] InitContainer [NodeConformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +[sig-cli] Kubectl client [k8s.io] Guestbook application + should create and stop a working application [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [sig-cli] Kubectl client + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 STEP: Creating a kubernetes client -Jun 4 19:30:21.254: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504 -STEP: Building a namespace api object, basename init-container -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in init-container-9462 +Jun 18 12:43:43.266: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150 +STEP: Building a namespace api object, basename kubectl +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-9827 STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [k8s.io] InitContainer [NodeConformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/init_container.go:43 -[It] should not start app containers if init containers fail on a RestartAlways pod [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -STEP: creating the pod -Jun 4 19:30:21.469: INFO: PodSpec: initContainers in spec.initContainers -Jun 4 19:31:10.655: INFO: init container has failed twice: &v1.Pod{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"pod-init-31606390-86ff-11e9-8d1b-467ee19922ac", GenerateName:"", Namespace:"init-container-9462", SelfLink:"/api/v1/namespaces/init-container-9462/pods/pod-init-31606390-86ff-11e9-8d1b-467ee19922ac", UID:"3161d002-86ff-11e9-8318-1e5386706511", ResourceVersion:"24529", Generation:0, CreationTimestamp:v1.Time{Time:time.Time{wall:0x0, ext:63695273421, loc:(*time.Location)(0x8a140e0)}}, DeletionTimestamp:(*v1.Time)(nil), DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"name":"foo", "time":"469382196"}, Annotations:map[string]string{"kubernetes.io/psp":"e2e-test-privileged-psp"}, OwnerReferences:[]v1.OwnerReference(nil), Initializers:(*v1.Initializers)(nil), Finalizers:[]string(nil), ClusterName:"", ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v1.PodSpec{Volumes:[]v1.Volume{v1.Volume{Name:"default-token-nd7jx", VolumeSource:v1.VolumeSource{HostPath:(*v1.HostPathVolumeSource)(nil), EmptyDir:(*v1.EmptyDirVolumeSource)(nil), GCEPersistentDisk:(*v1.GCEPersistentDiskVolumeSource)(nil), AWSElasticBlockStore:(*v1.AWSElasticBlockStoreVolumeSource)(nil), GitRepo:(*v1.GitRepoVolumeSource)(nil), Secret:(*v1.SecretVolumeSource)(0xc002c96000), NFS:(*v1.NFSVolumeSource)(nil), ISCSI:(*v1.ISCSIVolumeSource)(nil), Glusterfs:(*v1.GlusterfsVolumeSource)(nil), PersistentVolumeClaim:(*v1.PersistentVolumeClaimVolumeSource)(nil), RBD:(*v1.RBDVolumeSource)(nil), FlexVolume:(*v1.FlexVolumeSource)(nil), Cinder:(*v1.CinderVolumeSource)(nil), CephFS:(*v1.CephFSVolumeSource)(nil), Flocker:(*v1.FlockerVolumeSource)(nil), DownwardAPI:(*v1.DownwardAPIVolumeSource)(nil), FC:(*v1.FCVolumeSource)(nil), AzureFile:(*v1.AzureFileVolumeSource)(nil), ConfigMap:(*v1.ConfigMapVolumeSource)(nil), VsphereVolume:(*v1.VsphereVirtualDiskVolumeSource)(nil), Quobyte:(*v1.QuobyteVolumeSource)(nil), AzureDisk:(*v1.AzureDiskVolumeSource)(nil), PhotonPersistentDisk:(*v1.PhotonPersistentDiskVolumeSource)(nil), Projected:(*v1.ProjectedVolumeSource)(nil), PortworxVolume:(*v1.PortworxVolumeSource)(nil), ScaleIO:(*v1.ScaleIOVolumeSource)(nil), StorageOS:(*v1.StorageOSVolumeSource)(nil), CSI:(*v1.CSIVolumeSource)(nil)}}}, InitContainers:[]v1.Container{v1.Container{Name:"init1", Image:"docker.io/library/busybox:1.29", Command:[]string{"/bin/false"}, Args:[]string(nil), WorkingDir:"", Ports:[]v1.ContainerPort(nil), EnvFrom:[]v1.EnvFromSource(nil), Env:[]v1.EnvVar(nil), Resources:v1.ResourceRequirements{Limits:v1.ResourceList(nil), Requests:v1.ResourceList(nil)}, VolumeMounts:[]v1.VolumeMount{v1.VolumeMount{Name:"default-token-nd7jx", ReadOnly:true, MountPath:"/var/run/secrets/kubernetes.io/serviceaccount", SubPath:"", MountPropagation:(*v1.MountPropagationMode)(nil), SubPathExpr:""}}, VolumeDevices:[]v1.VolumeDevice(nil), LivenessProbe:(*v1.Probe)(nil), ReadinessProbe:(*v1.Probe)(nil), Lifecycle:(*v1.Lifecycle)(nil), TerminationMessagePath:"/dev/termination-log", TerminationMessagePolicy:"File", ImagePullPolicy:"IfNotPresent", SecurityContext:(*v1.SecurityContext)(nil), Stdin:false, StdinOnce:false, TTY:false}, v1.Container{Name:"init2", Image:"docker.io/library/busybox:1.29", Command:[]string{"/bin/true"}, Args:[]string(nil), WorkingDir:"", Ports:[]v1.ContainerPort(nil), EnvFrom:[]v1.EnvFromSource(nil), Env:[]v1.EnvVar(nil), Resources:v1.ResourceRequirements{Limits:v1.ResourceList(nil), Requests:v1.ResourceList(nil)}, VolumeMounts:[]v1.VolumeMount{v1.VolumeMount{Name:"default-token-nd7jx", ReadOnly:true, MountPath:"/var/run/secrets/kubernetes.io/serviceaccount", SubPath:"", MountPropagation:(*v1.MountPropagationMode)(nil), SubPathExpr:""}}, VolumeDevices:[]v1.VolumeDevice(nil), LivenessProbe:(*v1.Probe)(nil), ReadinessProbe:(*v1.Probe)(nil), Lifecycle:(*v1.Lifecycle)(nil), TerminationMessagePath:"/dev/termination-log", TerminationMessagePolicy:"File", ImagePullPolicy:"IfNotPresent", SecurityContext:(*v1.SecurityContext)(nil), Stdin:false, StdinOnce:false, TTY:false}}, Containers:[]v1.Container{v1.Container{Name:"run1", Image:"k8s.gcr.io/pause:3.1", Command:[]string(nil), Args:[]string(nil), WorkingDir:"", Ports:[]v1.ContainerPort(nil), EnvFrom:[]v1.EnvFromSource(nil), Env:[]v1.EnvVar(nil), Resources:v1.ResourceRequirements{Limits:v1.ResourceList{"cpu":resource.Quantity{i:resource.int64Amount{value:100, scale:-3}, d:resource.infDecAmount{Dec:(*inf.Dec)(nil)}, s:"100m", Format:"DecimalSI"}, "memory":resource.Quantity{i:resource.int64Amount{value:52428800, scale:0}, d:resource.infDecAmount{Dec:(*inf.Dec)(nil)}, s:"52428800", Format:"DecimalSI"}}, Requests:v1.ResourceList{"cpu":resource.Quantity{i:resource.int64Amount{value:100, scale:-3}, d:resource.infDecAmount{Dec:(*inf.Dec)(nil)}, s:"100m", Format:"DecimalSI"}, "memory":resource.Quantity{i:resource.int64Amount{value:52428800, scale:0}, d:resource.infDecAmount{Dec:(*inf.Dec)(nil)}, s:"52428800", Format:"DecimalSI"}}}, VolumeMounts:[]v1.VolumeMount{v1.VolumeMount{Name:"default-token-nd7jx", ReadOnly:true, MountPath:"/var/run/secrets/kubernetes.io/serviceaccount", SubPath:"", MountPropagation:(*v1.MountPropagationMode)(nil), SubPathExpr:""}}, VolumeDevices:[]v1.VolumeDevice(nil), LivenessProbe:(*v1.Probe)(nil), ReadinessProbe:(*v1.Probe)(nil), Lifecycle:(*v1.Lifecycle)(nil), TerminationMessagePath:"/dev/termination-log", TerminationMessagePolicy:"File", ImagePullPolicy:"IfNotPresent", SecurityContext:(*v1.SecurityContext)(nil), Stdin:false, StdinOnce:false, TTY:false}}, RestartPolicy:"Always", TerminationGracePeriodSeconds:(*int64)(0xc00133a0b8), ActiveDeadlineSeconds:(*int64)(nil), DNSPolicy:"ClusterFirst", NodeSelector:map[string]string(nil), ServiceAccountName:"default", DeprecatedServiceAccount:"default", AutomountServiceAccountToken:(*bool)(nil), NodeName:"10.212.23.161", HostNetwork:false, HostPID:false, HostIPC:false, ShareProcessNamespace:(*bool)(nil), SecurityContext:(*v1.PodSecurityContext)(0xc002e8e000), ImagePullSecrets:[]v1.LocalObjectReference(nil), Hostname:"", Subdomain:"", Affinity:(*v1.Affinity)(nil), SchedulerName:"default-scheduler", Tolerations:[]v1.Toleration{v1.Toleration{Key:"node.kubernetes.io/not-ready", Operator:"Exists", Value:"", Effect:"NoExecute", TolerationSeconds:(*int64)(0xc00133a1e0)}, v1.Toleration{Key:"node.kubernetes.io/unreachable", Operator:"Exists", Value:"", Effect:"NoExecute", TolerationSeconds:(*int64)(0xc00133a210)}}, HostAliases:[]v1.HostAlias(nil), PriorityClassName:"", Priority:(*int32)(0xc00133a218), DNSConfig:(*v1.PodDNSConfig)(nil), ReadinessGates:[]v1.PodReadinessGate(nil), RuntimeClassName:(*string)(nil), EnableServiceLinks:(*bool)(0xc00133a21c)}, Status:v1.PodStatus{Phase:"Pending", Conditions:[]v1.PodCondition{v1.PodCondition{Type:"Initialized", Status:"False", LastProbeTime:v1.Time{Time:time.Time{wall:0x0, ext:0, loc:(*time.Location)(nil)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63695273421, loc:(*time.Location)(0x8a140e0)}}, Reason:"ContainersNotInitialized", Message:"containers with incomplete status: [init1 init2]"}, v1.PodCondition{Type:"Ready", Status:"False", LastProbeTime:v1.Time{Time:time.Time{wall:0x0, ext:0, loc:(*time.Location)(nil)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63695273421, loc:(*time.Location)(0x8a140e0)}}, Reason:"ContainersNotReady", Message:"containers with unready status: [run1]"}, v1.PodCondition{Type:"ContainersReady", Status:"False", LastProbeTime:v1.Time{Time:time.Time{wall:0x0, ext:0, loc:(*time.Location)(nil)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63695273421, loc:(*time.Location)(0x8a140e0)}}, Reason:"ContainersNotReady", Message:"containers with unready status: [run1]"}, v1.PodCondition{Type:"PodScheduled", Status:"True", LastProbeTime:v1.Time{Time:time.Time{wall:0x0, ext:0, loc:(*time.Location)(nil)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63695273421, loc:(*time.Location)(0x8a140e0)}}, Reason:"", Message:""}}, Message:"", Reason:"", NominatedNodeName:"", HostIP:"10.212.23.161", PodIP:"172.30.169.167", StartTime:(*v1.Time)(0xc0029a0060), InitContainerStatuses:[]v1.ContainerStatus{v1.ContainerStatus{Name:"init1", State:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(nil), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(0xc001c4b0a0)}, LastTerminationState:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(nil), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(0xc001c4b110)}, Ready:false, RestartCount:3, Image:"docker.io/library/busybox:1.29", ImageID:"docker.io/library/busybox@sha256:8ccbac733d19c0dd4d70b4f0c1e12245b5fa3ad24758a11035ee505c629c0796", ContainerID:"containerd://6ef8f52250f054ca4eee3654b207669fb85b68ba9273223e47c9c12a20401cf2"}, v1.ContainerStatus{Name:"init2", State:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(0xc0029a00a0), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(nil)}, LastTerminationState:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(nil), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(nil)}, Ready:false, RestartCount:0, Image:"docker.io/library/busybox:1.29", ImageID:"", ContainerID:""}}, ContainerStatuses:[]v1.ContainerStatus{v1.ContainerStatus{Name:"run1", State:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(0xc0029a0080), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(nil)}, LastTerminationState:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(nil), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(nil)}, Ready:false, RestartCount:0, Image:"k8s.gcr.io/pause:3.1", ImageID:"", ContainerID:""}}, QOSClass:"Guaranteed"}} -[AfterEach] [k8s.io] InitContainer [NodeConformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 -Jun 4 19:31:10.656: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "init-container-9462" for this suite. -Jun 4 19:31:34.716: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -Jun 4 19:31:35.104: INFO: namespace init-container-9462 deletion completed in 24.431071649s +[BeforeEach] [sig-cli] Kubectl client + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:213 +[It] should create and stop a working application [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: creating all guestbook components +Jun 18 12:43:43.459: INFO: apiVersion: v1 +kind: Service +metadata: + name: redis-slave + labels: + app: redis + role: slave + tier: backend +spec: + ports: + - port: 6379 + selector: + app: redis + role: slave + tier: backend + +Jun 18 12:43:43.459: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 create -f - --namespace=kubectl-9827' +Jun 18 12:43:43.710: INFO: stderr: "" +Jun 18 12:43:43.710: INFO: stdout: "service/redis-slave created\n" +Jun 18 12:43:43.710: INFO: apiVersion: v1 +kind: Service +metadata: + name: redis-master + labels: + app: redis + role: master + tier: backend +spec: + ports: + - port: 6379 + targetPort: 6379 + selector: + app: redis + role: master + tier: backend + +Jun 18 12:43:43.710: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 create -f - --namespace=kubectl-9827' +Jun 18 12:43:44.114: INFO: stderr: "" +Jun 18 12:43:44.114: INFO: stdout: "service/redis-master created\n" +Jun 18 12:43:44.114: INFO: apiVersion: v1 +kind: Service +metadata: + name: frontend + labels: + app: guestbook + tier: frontend +spec: + # if your cluster supports it, uncomment the following to automatically create + # an external load-balanced IP for the frontend service. + # type: LoadBalancer + ports: + - port: 80 + selector: + app: guestbook + tier: frontend + +Jun 18 12:43:44.114: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 create -f - --namespace=kubectl-9827' +Jun 18 12:43:44.409: INFO: stderr: "" +Jun 18 12:43:44.409: INFO: stdout: "service/frontend created\n" +Jun 18 12:43:44.410: INFO: apiVersion: apps/v1 +kind: Deployment +metadata: + name: frontend +spec: + replicas: 3 + selector: + matchLabels: + app: guestbook + tier: frontend + template: + metadata: + labels: + app: guestbook + tier: frontend + spec: + containers: + - name: php-redis + image: gcr.io/google-samples/gb-frontend:v6 + resources: + requests: + cpu: 100m + memory: 100Mi + env: + - name: GET_HOSTS_FROM + value: dns + # If your cluster config does not include a dns service, then to + # instead access environment variables to find service host + # info, comment out the 'value: dns' line above, and uncomment the + # line below: + # value: env + ports: + - containerPort: 80 + +Jun 18 12:43:44.410: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 create -f - --namespace=kubectl-9827' +Jun 18 12:43:44.738: INFO: stderr: "" +Jun 18 12:43:44.738: INFO: stdout: "deployment.apps/frontend created\n" +Jun 18 12:43:44.738: INFO: apiVersion: apps/v1 +kind: Deployment +metadata: + name: redis-master +spec: + replicas: 1 + selector: + matchLabels: + app: redis + role: master + tier: backend + template: + metadata: + labels: + app: redis + role: master + tier: backend + spec: + containers: + - name: master + image: gcr.io/kubernetes-e2e-test-images/redis:1.0 + resources: + requests: + cpu: 100m + memory: 100Mi + ports: + - containerPort: 6379 + +Jun 18 12:43:44.738: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 create -f - --namespace=kubectl-9827' +Jun 18 12:43:45.030: INFO: stderr: "" +Jun 18 12:43:45.030: INFO: stdout: "deployment.apps/redis-master created\n" +Jun 18 12:43:45.030: INFO: apiVersion: apps/v1 +kind: Deployment +metadata: + name: redis-slave +spec: + replicas: 2 + selector: + matchLabels: + app: redis + role: slave + tier: backend + template: + metadata: + labels: + app: redis + role: slave + tier: backend + spec: + containers: + - name: slave + image: gcr.io/google-samples/gb-redisslave:v3 + resources: + requests: + cpu: 100m + memory: 100Mi + env: + - name: GET_HOSTS_FROM + value: dns + # If your cluster config does not include a dns service, then to + # instead access an environment variable to find the master + # service's host, comment out the 'value: dns' line above, and + # uncomment the line below: + # value: env + ports: + - containerPort: 6379 + +Jun 18 12:43:45.031: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 create -f - --namespace=kubectl-9827' +Jun 18 12:43:45.353: INFO: stderr: "" +Jun 18 12:43:45.353: INFO: stdout: "deployment.apps/redis-slave created\n" +STEP: validating guestbook app +Jun 18 12:43:45.353: INFO: Waiting for all frontend pods to be Running. +Jun 18 12:44:05.404: INFO: Waiting for frontend to serve content. +Jun 18 12:44:10.450: INFO: Failed to get response from guestbook. err: , response:
+Fatal error: Uncaught exception 'Predis\Connection\ConnectionException' with message 'Connection timed out [tcp://redis-slave:6379]' in /usr/local/lib/php/Predis/Connection/AbstractConnection.php:155 +Stack trace: +#0 /usr/local/lib/php/Predis/Connection/StreamConnection.php(128): Predis\Connection\AbstractConnection->onConnectionError('Connection time...', 110) +#1 /usr/local/lib/php/Predis/Connection/StreamConnection.php(178): Predis\Connection\StreamConnection->createStreamSocket(Object(Predis\Connection\Parameters), 'tcp://redis-sla...', 4) +#2 /usr/local/lib/php/Predis/Connection/StreamConnection.php(100): Predis\Connection\StreamConnection->tcpStreamInitializer(Object(Predis\Connection\Parameters)) +#3 /usr/local/lib/php/Predis/Connection/AbstractConnection.php(81): Predis\Connection\StreamConnection->createResource() +#4 /usr/local/lib/php/Predis/Connection/StreamConnection.php(258): Predis\Connection\AbstractConnection->connect() +#5 /usr/local/lib/php/Predis/Connection/AbstractConnection.php(180): Predis\Connection\Stre in /usr/local/lib/php/Predis/Connection/AbstractConnection.php on line 155
+ +Jun 18 12:44:15.491: INFO: Trying to add a new entry to the guestbook. +Jun 18 12:44:15.522: INFO: Verifying that added entry can be retrieved. +STEP: using delete to clean up resources +Jun 18 12:44:15.551: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 delete --grace-period=0 --force -f - --namespace=kubectl-9827' +Jun 18 12:44:15.767: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" +Jun 18 12:44:15.767: INFO: stdout: "service \"redis-slave\" force deleted\n" +STEP: using delete to clean up resources +Jun 18 12:44:15.767: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 delete --grace-period=0 --force -f - --namespace=kubectl-9827' +Jun 18 12:44:15.976: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" +Jun 18 12:44:15.976: INFO: stdout: "service \"redis-master\" force deleted\n" +STEP: using delete to clean up resources +Jun 18 12:44:15.976: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 delete --grace-period=0 --force -f - --namespace=kubectl-9827' +Jun 18 12:44:16.172: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" +Jun 18 12:44:16.172: INFO: stdout: "service \"frontend\" force deleted\n" +STEP: using delete to clean up resources +Jun 18 12:44:16.172: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 delete --grace-period=0 --force -f - --namespace=kubectl-9827' +Jun 18 12:44:16.326: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" +Jun 18 12:44:16.326: INFO: stdout: "deployment.apps \"frontend\" force deleted\n" +STEP: using delete to clean up resources +Jun 18 12:44:16.326: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 delete --grace-period=0 --force -f - --namespace=kubectl-9827' +Jun 18 12:44:16.551: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" +Jun 18 12:44:16.551: INFO: stdout: "deployment.apps \"redis-master\" force deleted\n" +STEP: using delete to clean up resources +Jun 18 12:44:16.551: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 delete --grace-period=0 --force -f - --namespace=kubectl-9827' +Jun 18 12:44:16.702: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" +Jun 18 12:44:16.702: INFO: stdout: "deployment.apps \"redis-slave\" force deleted\n" +[AfterEach] [sig-cli] Kubectl client + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Jun 18 12:44:16.702: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "kubectl-9827" for this suite. +Jun 18 12:44:58.747: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:44:59.149: INFO: namespace kubectl-9827 deletion completed in 42.428876206s -• [SLOW TEST:73.849 seconds] -[k8s.io] InitContainer [NodeConformance] -/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687 - should not start app containers if init containers fail on a RestartAlways pod [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +• [SLOW TEST:75.884 seconds] +[sig-cli] Kubectl client +/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 + [k8s.io] Guestbook application + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687 + should create and stop a working application [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 ------------------------------ -SSSSSSSSSSSSSSSSS +SSSSSS ------------------------------ -[sig-storage] Projected downwardAPI +[sig-storage] Downward API volume should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -[BeforeEach] [sig-storage] Projected downwardAPI - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [sig-storage] Downward API volume + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 STEP: Creating a kubernetes client -Jun 4 19:31:35.104: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504 -STEP: Building a namespace api object, basename projected -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-7592 +Jun 18 12:44:59.150: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150 +STEP: Building a namespace api object, basename downward-api +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in downward-api-7392 STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [sig-storage] Projected downwardAPI - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:39 +[BeforeEach] [sig-storage] Downward API volume + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:39 [It] should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 STEP: Creating a pod to test downward API volume plugin -Jun 4 19:31:35.381: INFO: Waiting up to 5m0s for pod "downwardapi-volume-5d6b213b-86ff-11e9-8d1b-467ee19922ac" in namespace "projected-7592" to be "success or failure" -Jun 4 19:31:35.392: INFO: Pod "downwardapi-volume-5d6b213b-86ff-11e9-8d1b-467ee19922ac": Phase="Pending", Reason="", readiness=false. Elapsed: 11.135239ms -Jun 4 19:31:37.404: INFO: Pod "downwardapi-volume-5d6b213b-86ff-11e9-8d1b-467ee19922ac": Phase="Running", Reason="", readiness=true. Elapsed: 2.023520659s -Jun 4 19:31:39.418: INFO: Pod "downwardapi-volume-5d6b213b-86ff-11e9-8d1b-467ee19922ac": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.036992919s +Jun 18 12:44:59.381: INFO: Waiting up to 5m0s for pod "downwardapi-volume-e20ba4f9-91c6-11e9-a25d-8608290c688a" in namespace "downward-api-7392" to be "success or failure" +Jun 18 12:44:59.395: INFO: Pod "downwardapi-volume-e20ba4f9-91c6-11e9-a25d-8608290c688a": Phase="Pending", Reason="", readiness=false. Elapsed: 13.787396ms +Jun 18 12:45:01.409: INFO: Pod "downwardapi-volume-e20ba4f9-91c6-11e9-a25d-8608290c688a": Phase="Running", Reason="", readiness=true. Elapsed: 2.027580593s +Jun 18 12:45:03.425: INFO: Pod "downwardapi-volume-e20ba4f9-91c6-11e9-a25d-8608290c688a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.043529028s STEP: Saw pod success -Jun 4 19:31:39.418: INFO: Pod "downwardapi-volume-5d6b213b-86ff-11e9-8d1b-467ee19922ac" satisfied condition "success or failure" -Jun 4 19:31:39.430: INFO: Trying to get logs from node 10.212.23.164 pod downwardapi-volume-5d6b213b-86ff-11e9-8d1b-467ee19922ac container client-container: +Jun 18 12:45:03.425: INFO: Pod "downwardapi-volume-e20ba4f9-91c6-11e9-a25d-8608290c688a" satisfied condition "success or failure" +Jun 18 12:45:03.439: INFO: Trying to get logs from node 10.72.74.138 pod downwardapi-volume-e20ba4f9-91c6-11e9-a25d-8608290c688a container client-container: STEP: delete the pod -Jun 4 19:31:39.517: INFO: Waiting for pod downwardapi-volume-5d6b213b-86ff-11e9-8d1b-467ee19922ac to disappear -Jun 4 19:31:39.528: INFO: Pod downwardapi-volume-5d6b213b-86ff-11e9-8d1b-467ee19922ac no longer exists -[AfterEach] [sig-storage] Projected downwardAPI - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 -Jun 4 19:31:39.528: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "projected-7592" for this suite. -Jun 4 19:31:45.587: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -Jun 4 19:31:45.969: INFO: namespace projected-7592 deletion completed in 6.424619516s +Jun 18 12:45:03.518: INFO: Waiting for pod downwardapi-volume-e20ba4f9-91c6-11e9-a25d-8608290c688a to disappear +Jun 18 12:45:03.530: INFO: Pod downwardapi-volume-e20ba4f9-91c6-11e9-a25d-8608290c688a no longer exists +[AfterEach] [sig-storage] Downward API volume + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Jun 18 12:45:03.530: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "downward-api-7392" for this suite. +Jun 18 12:45:09.570: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:45:09.967: INFO: namespace downward-api-7392 deletion completed in 6.423440505s -• [SLOW TEST:10.865 seconds] -[sig-storage] Projected downwardAPI -/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:33 +• [SLOW TEST:10.818 seconds] +[sig-storage] Downward API volume +/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:34 should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 ------------------------------- -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------- -[k8s.io] InitContainer [NodeConformance] - should not start app containers and fail the pod if init containers fail on a RestartNever pod [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -[BeforeEach] [k8s.io] InitContainer [NodeConformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 -STEP: Creating a kubernetes client -Jun 4 19:31:45.970: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504 -STEP: Building a namespace api object, basename init-container -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in init-container-6884 -STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [k8s.io] InitContainer [NodeConformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/init_container.go:43 -[It] should not start app containers and fail the pod if init containers fail on a RestartNever pod [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -STEP: creating the pod -Jun 4 19:31:46.194: INFO: PodSpec: initContainers in spec.initContainers -[AfterEach] [k8s.io] InitContainer [NodeConformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 -Jun 4 19:31:49.505: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "init-container-6884" for this suite. -Jun 4 19:31:55.597: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -Jun 4 19:31:56.005: INFO: namespace init-container-6884 deletion completed in 6.482764094s - -• [SLOW TEST:10.035 seconds] -[k8s.io] InitContainer [NodeConformance] -/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687 - should not start app containers and fail the pod if init containers fail on a RestartNever pod [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 ------------------------------- -SSSSSSSSSS ------------------------------- -[k8s.io] Docker Containers - should be able to override the image's default arguments (docker cmd) [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -[BeforeEach] [k8s.io] Docker Containers - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 -STEP: Creating a kubernetes client -Jun 4 19:31:56.006: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504 -STEP: Building a namespace api object, basename containers -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in containers-2346 -STEP: Waiting for a default service account to be provisioned in namespace -[It] should be able to override the image's default arguments (docker cmd) [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -STEP: Creating a pod to test override arguments -Jun 4 19:31:56.245: INFO: Waiting up to 5m0s for pod "client-containers-69dac7b5-86ff-11e9-8d1b-467ee19922ac" in namespace "containers-2346" to be "success or failure" -Jun 4 19:31:56.256: INFO: Pod "client-containers-69dac7b5-86ff-11e9-8d1b-467ee19922ac": Phase="Pending", Reason="", readiness=false. Elapsed: 10.801242ms -Jun 4 19:31:58.272: INFO: Pod "client-containers-69dac7b5-86ff-11e9-8d1b-467ee19922ac": Phase="Pending", Reason="", readiness=false. Elapsed: 2.027387251s -Jun 4 19:32:00.286: INFO: Pod "client-containers-69dac7b5-86ff-11e9-8d1b-467ee19922ac": Phase="Pending", Reason="", readiness=false. Elapsed: 4.040729801s -Jun 4 19:32:02.307: INFO: Pod "client-containers-69dac7b5-86ff-11e9-8d1b-467ee19922ac": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.062024241s -STEP: Saw pod success -Jun 4 19:32:02.307: INFO: Pod "client-containers-69dac7b5-86ff-11e9-8d1b-467ee19922ac" satisfied condition "success or failure" -Jun 4 19:32:02.318: INFO: Trying to get logs from node 10.212.23.189 pod client-containers-69dac7b5-86ff-11e9-8d1b-467ee19922ac container test-container: -STEP: delete the pod -Jun 4 19:32:02.387: INFO: Waiting for pod client-containers-69dac7b5-86ff-11e9-8d1b-467ee19922ac to disappear -Jun 4 19:32:02.401: INFO: Pod client-containers-69dac7b5-86ff-11e9-8d1b-467ee19922ac no longer exists -[AfterEach] [k8s.io] Docker Containers - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 -Jun 4 19:32:02.401: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "containers-2346" for this suite. -Jun 4 19:32:08.464: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -Jun 4 19:32:08.823: INFO: namespace containers-2346 deletion completed in 6.405618028s - -• [SLOW TEST:12.817 seconds] -[k8s.io] Docker Containers -/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687 - should be able to override the image's default arguments (docker cmd) [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 ------------------------------ -SSSSSSSS +SSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Projected configMap - should be consumable from pods in volume as non-root [LinuxOnly] [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 + optional updates should be reflected in volume [NodeConformance] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 [BeforeEach] [sig-storage] Projected configMap - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 STEP: Creating a kubernetes client -Jun 4 19:32:08.823: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504 +Jun 18 12:45:09.968: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150 STEP: Building a namespace api object, basename projected -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-9800 +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-1723 STEP: Waiting for a default service account to be provisioned in namespace -[It] should be consumable from pods in volume as non-root [LinuxOnly] [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -STEP: Creating configMap with name projected-configmap-test-volume-717fabc2-86ff-11e9-8d1b-467ee19922ac -STEP: Creating a pod to test consume configMaps -Jun 4 19:32:09.082: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-7181bbee-86ff-11e9-8d1b-467ee19922ac" in namespace "projected-9800" to be "success or failure" -Jun 4 19:32:09.100: INFO: Pod "pod-projected-configmaps-7181bbee-86ff-11e9-8d1b-467ee19922ac": Phase="Pending", Reason="", readiness=false. Elapsed: 17.425267ms -Jun 4 19:32:11.113: INFO: Pod "pod-projected-configmaps-7181bbee-86ff-11e9-8d1b-467ee19922ac": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.030375441s -STEP: Saw pod success -Jun 4 19:32:11.113: INFO: Pod "pod-projected-configmaps-7181bbee-86ff-11e9-8d1b-467ee19922ac" satisfied condition "success or failure" -Jun 4 19:32:11.125: INFO: Trying to get logs from node 10.212.23.164 pod pod-projected-configmaps-7181bbee-86ff-11e9-8d1b-467ee19922ac container projected-configmap-volume-test: -STEP: delete the pod -Jun 4 19:32:11.195: INFO: Waiting for pod pod-projected-configmaps-7181bbee-86ff-11e9-8d1b-467ee19922ac to disappear -Jun 4 19:32:11.206: INFO: Pod pod-projected-configmaps-7181bbee-86ff-11e9-8d1b-467ee19922ac no longer exists +[It] optional updates should be reflected in volume [NodeConformance] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: Creating configMap with name cm-test-opt-del-e88146c5-91c6-11e9-a25d-8608290c688a +STEP: Creating configMap with name cm-test-opt-upd-e8814719-91c6-11e9-a25d-8608290c688a +STEP: Creating the pod +STEP: Deleting configmap cm-test-opt-del-e88146c5-91c6-11e9-a25d-8608290c688a +STEP: Updating configmap cm-test-opt-upd-e8814719-91c6-11e9-a25d-8608290c688a +STEP: Creating configMap with name cm-test-opt-create-e8814741-91c6-11e9-a25d-8608290c688a +STEP: waiting to observe update in volume [AfterEach] [sig-storage] Projected configMap - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 -Jun 4 19:32:11.206: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "projected-9800" for this suite. -Jun 4 19:32:19.262: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -Jun 4 19:32:19.612: INFO: namespace projected-9800 deletion completed in 8.390940613s + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Jun 18 12:46:36.985: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "projected-1723" for this suite. +Jun 18 12:47:01.025: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:47:01.440: INFO: namespace projected-1723 deletion completed in 24.44058393s -• [SLOW TEST:10.788 seconds] +• [SLOW TEST:111.472 seconds] [sig-storage] Projected configMap -/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_configmap.go:33 - should be consumable from pods in volume as non-root [LinuxOnly] [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_configmap.go:33 + optional updates should be reflected in volume [NodeConformance] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 ------------------------------ -SSSSSSSS +SSSSSSSSSSS ------------------------------ [sig-apps] Daemon set [Serial] - should rollback without unnecessary restarts [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 + should update pod when spec was updated and update strategy is RollingUpdate [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 [BeforeEach] [sig-apps] Daemon set [Serial] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 STEP: Creating a kubernetes client -Jun 4 19:32:19.612: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504 +Jun 18 12:47:01.441: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150 STEP: Building a namespace api object, basename daemonsets -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in daemonsets-1795 +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in daemonsets-255 STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] Daemon set [Serial] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:102 -[It] should rollback without unnecessary restarts [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -Jun 4 19:32:19.903: INFO: Create a RollingUpdate DaemonSet -Jun 4 19:32:19.915: INFO: Check that daemon pods launch on every node of the cluster -Jun 4 19:32:19.940: INFO: Number of nodes with available pods: 0 -Jun 4 19:32:19.940: INFO: Node 10.212.23.161 is running more than one daemon pod -Jun 4 19:32:20.983: INFO: Number of nodes with available pods: 0 -Jun 4 19:32:20.983: INFO: Node 10.212.23.161 is running more than one daemon pod -Jun 4 19:32:21.969: INFO: Number of nodes with available pods: 3 -Jun 4 19:32:21.969: INFO: Number of running nodes: 3, number of available pods: 3 -Jun 4 19:32:21.969: INFO: Update the DaemonSet to trigger a rollout -Jun 4 19:32:21.995: INFO: Updating DaemonSet daemon-set -Jun 4 19:32:37.135: INFO: Roll back the DaemonSet before rollout is complete -Jun 4 19:32:37.466: INFO: Updating DaemonSet daemon-set -Jun 4 19:32:37.466: INFO: Make sure DaemonSet rollback is complete -Jun 4 19:32:37.478: INFO: Wrong image for pod: daemon-set-6qkrj. Expected: docker.io/library/nginx:1.14-alpine, got: foo:non-existent. -Jun 4 19:32:37.478: INFO: Pod daemon-set-6qkrj is not available -Jun 4 19:32:38.570: INFO: Wrong image for pod: daemon-set-6qkrj. Expected: docker.io/library/nginx:1.14-alpine, got: foo:non-existent. -Jun 4 19:32:38.570: INFO: Pod daemon-set-6qkrj is not available -Jun 4 19:32:39.570: INFO: Wrong image for pod: daemon-set-6qkrj. Expected: docker.io/library/nginx:1.14-alpine, got: foo:non-existent. -Jun 4 19:32:39.570: INFO: Pod daemon-set-6qkrj is not available -Jun 4 19:32:40.569: INFO: Pod daemon-set-4zskj is not available + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:102 +[It] should update pod when spec was updated and update strategy is RollingUpdate [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +Jun 18 12:47:01.710: INFO: Creating simple daemon set daemon-set +STEP: Check that daemon pods launch on every node of the cluster. +Jun 18 12:47:01.748: INFO: Number of nodes with available pods: 0 +Jun 18 12:47:01.749: INFO: Node 10.72.74.138 is running more than one daemon pod +Jun 18 12:47:02.777: INFO: Number of nodes with available pods: 0 +Jun 18 12:47:02.777: INFO: Node 10.72.74.138 is running more than one daemon pod +Jun 18 12:47:03.778: INFO: Number of nodes with available pods: 1 +Jun 18 12:47:03.778: INFO: Node 10.72.74.138 is running more than one daemon pod +Jun 18 12:47:04.779: INFO: Number of nodes with available pods: 3 +Jun 18 12:47:04.779: INFO: Number of running nodes: 3, number of available pods: 3 +STEP: Update daemon pods image. +STEP: Check that daemon pods images are updated. +Jun 18 12:47:04.903: INFO: Wrong image for pod: daemon-set-jv7sw. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Jun 18 12:47:04.903: INFO: Wrong image for pod: daemon-set-lbnhb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Jun 18 12:47:04.903: INFO: Wrong image for pod: daemon-set-r88gk. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Jun 18 12:47:05.927: INFO: Wrong image for pod: daemon-set-jv7sw. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Jun 18 12:47:05.927: INFO: Wrong image for pod: daemon-set-lbnhb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Jun 18 12:47:05.927: INFO: Wrong image for pod: daemon-set-r88gk. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Jun 18 12:47:06.928: INFO: Wrong image for pod: daemon-set-jv7sw. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Jun 18 12:47:06.928: INFO: Wrong image for pod: daemon-set-lbnhb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Jun 18 12:47:06.928: INFO: Wrong image for pod: daemon-set-r88gk. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Jun 18 12:47:07.928: INFO: Wrong image for pod: daemon-set-jv7sw. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Jun 18 12:47:07.928: INFO: Wrong image for pod: daemon-set-lbnhb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Jun 18 12:47:07.928: INFO: Wrong image for pod: daemon-set-r88gk. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Jun 18 12:47:07.928: INFO: Pod daemon-set-r88gk is not available +Jun 18 12:47:08.928: INFO: Wrong image for pod: daemon-set-jv7sw. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Jun 18 12:47:08.928: INFO: Wrong image for pod: daemon-set-lbnhb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Jun 18 12:47:08.928: INFO: Wrong image for pod: daemon-set-r88gk. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Jun 18 12:47:08.928: INFO: Pod daemon-set-r88gk is not available +Jun 18 12:47:09.929: INFO: Wrong image for pod: daemon-set-jv7sw. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Jun 18 12:47:09.929: INFO: Wrong image for pod: daemon-set-lbnhb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Jun 18 12:47:09.929: INFO: Wrong image for pod: daemon-set-r88gk. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Jun 18 12:47:09.929: INFO: Pod daemon-set-r88gk is not available +Jun 18 12:47:10.928: INFO: Wrong image for pod: daemon-set-jv7sw. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Jun 18 12:47:10.929: INFO: Wrong image for pod: daemon-set-lbnhb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Jun 18 12:47:10.929: INFO: Wrong image for pod: daemon-set-r88gk. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Jun 18 12:47:10.929: INFO: Pod daemon-set-r88gk is not available +Jun 18 12:47:11.928: INFO: Wrong image for pod: daemon-set-jv7sw. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Jun 18 12:47:11.928: INFO: Wrong image for pod: daemon-set-lbnhb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Jun 18 12:47:11.928: INFO: Wrong image for pod: daemon-set-r88gk. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Jun 18 12:47:11.928: INFO: Pod daemon-set-r88gk is not available +Jun 18 12:47:12.928: INFO: Wrong image for pod: daemon-set-jv7sw. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Jun 18 12:47:12.928: INFO: Wrong image for pod: daemon-set-lbnhb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Jun 18 12:47:12.928: INFO: Wrong image for pod: daemon-set-r88gk. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Jun 18 12:47:12.928: INFO: Pod daemon-set-r88gk is not available +Jun 18 12:47:13.929: INFO: Wrong image for pod: daemon-set-jv7sw. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Jun 18 12:47:13.929: INFO: Wrong image for pod: daemon-set-lbnhb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Jun 18 12:47:13.929: INFO: Wrong image for pod: daemon-set-r88gk. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Jun 18 12:47:13.929: INFO: Pod daemon-set-r88gk is not available +Jun 18 12:47:14.929: INFO: Wrong image for pod: daemon-set-jv7sw. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Jun 18 12:47:14.929: INFO: Wrong image for pod: daemon-set-lbnhb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Jun 18 12:47:14.929: INFO: Wrong image for pod: daemon-set-r88gk. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Jun 18 12:47:14.929: INFO: Pod daemon-set-r88gk is not available +Jun 18 12:47:15.929: INFO: Wrong image for pod: daemon-set-jv7sw. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Jun 18 12:47:15.929: INFO: Wrong image for pod: daemon-set-lbnhb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Jun 18 12:47:15.929: INFO: Wrong image for pod: daemon-set-r88gk. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Jun 18 12:47:15.929: INFO: Pod daemon-set-r88gk is not available +Jun 18 12:47:16.928: INFO: Wrong image for pod: daemon-set-jv7sw. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Jun 18 12:47:16.928: INFO: Wrong image for pod: daemon-set-lbnhb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Jun 18 12:47:16.928: INFO: Wrong image for pod: daemon-set-r88gk. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Jun 18 12:47:16.928: INFO: Pod daemon-set-r88gk is not available +Jun 18 12:47:17.931: INFO: Wrong image for pod: daemon-set-jv7sw. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Jun 18 12:47:17.931: INFO: Wrong image for pod: daemon-set-lbnhb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Jun 18 12:47:17.931: INFO: Pod daemon-set-w9mt4 is not available +Jun 18 12:47:18.929: INFO: Wrong image for pod: daemon-set-jv7sw. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Jun 18 12:47:18.929: INFO: Wrong image for pod: daemon-set-lbnhb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Jun 18 12:47:18.929: INFO: Pod daemon-set-w9mt4 is not available +Jun 18 12:47:19.937: INFO: Wrong image for pod: daemon-set-jv7sw. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Jun 18 12:47:19.937: INFO: Wrong image for pod: daemon-set-lbnhb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Jun 18 12:47:19.937: INFO: Pod daemon-set-w9mt4 is not available +Jun 18 12:47:20.928: INFO: Wrong image for pod: daemon-set-jv7sw. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Jun 18 12:47:20.928: INFO: Wrong image for pod: daemon-set-lbnhb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Jun 18 12:47:20.928: INFO: Pod daemon-set-w9mt4 is not available +Jun 18 12:47:21.928: INFO: Wrong image for pod: daemon-set-jv7sw. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Jun 18 12:47:21.928: INFO: Wrong image for pod: daemon-set-lbnhb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Jun 18 12:47:21.928: INFO: Pod daemon-set-w9mt4 is not available +Jun 18 12:47:22.928: INFO: Wrong image for pod: daemon-set-jv7sw. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Jun 18 12:47:22.928: INFO: Wrong image for pod: daemon-set-lbnhb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Jun 18 12:47:23.928: INFO: Wrong image for pod: daemon-set-jv7sw. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Jun 18 12:47:23.928: INFO: Wrong image for pod: daemon-set-lbnhb. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Jun 18 12:47:23.928: INFO: Pod daemon-set-lbnhb is not available +Jun 18 12:47:24.929: INFO: Wrong image for pod: daemon-set-jv7sw. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Jun 18 12:47:24.929: INFO: Pod daemon-set-tkdz2 is not available +Jun 18 12:47:25.930: INFO: Wrong image for pod: daemon-set-jv7sw. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Jun 18 12:47:25.930: INFO: Pod daemon-set-tkdz2 is not available +Jun 18 12:47:26.929: INFO: Wrong image for pod: daemon-set-jv7sw. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Jun 18 12:47:27.928: INFO: Wrong image for pod: daemon-set-jv7sw. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Jun 18 12:47:27.928: INFO: Pod daemon-set-jv7sw is not available +Jun 18 12:47:28.931: INFO: Wrong image for pod: daemon-set-jv7sw. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Jun 18 12:47:28.931: INFO: Pod daemon-set-jv7sw is not available +Jun 18 12:47:29.928: INFO: Wrong image for pod: daemon-set-jv7sw. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Jun 18 12:47:29.928: INFO: Pod daemon-set-jv7sw is not available +Jun 18 12:47:30.929: INFO: Wrong image for pod: daemon-set-jv7sw. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Jun 18 12:47:30.929: INFO: Pod daemon-set-jv7sw is not available +Jun 18 12:47:31.928: INFO: Pod daemon-set-v9c9l is not available +STEP: Check that daemon pods are still running on every node of the cluster. +Jun 18 12:47:31.971: INFO: Number of nodes with available pods: 2 +Jun 18 12:47:31.972: INFO: Node 10.72.74.184 is running more than one daemon pod +Jun 18 12:47:33.003: INFO: Number of nodes with available pods: 2 +Jun 18 12:47:33.003: INFO: Node 10.72.74.184 is running more than one daemon pod +Jun 18 12:47:34.000: INFO: Number of nodes with available pods: 3 +Jun 18 12:47:34.000: INFO: Number of running nodes: 3, number of available pods: 3 [AfterEach] [sig-apps] Daemon set [Serial] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:68 + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:68 STEP: Deleting DaemonSet "daemon-set" -STEP: deleting DaemonSet.extensions daemon-set in namespace daemonsets-1795, will wait for the garbage collector to delete the pods -Jun 4 19:32:40.697: INFO: Deleting DaemonSet.extensions daemon-set took: 28.795929ms -Jun 4 19:32:40.797: INFO: Terminating DaemonSet.extensions daemon-set pods took: 100.284616ms -Jun 4 19:32:52.509: INFO: Number of nodes with available pods: 0 -Jun 4 19:32:52.509: INFO: Number of running nodes: 0, number of available pods: 0 -Jun 4 19:32:52.521: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"selfLink":"/apis/apps/v1/namespaces/daemonsets-1795/daemonsets","resourceVersion":"24999"},"items":null} +STEP: deleting DaemonSet.extensions daemon-set in namespace daemonsets-255, will wait for the garbage collector to delete the pods +Jun 18 12:47:34.180: INFO: Deleting DaemonSet.extensions daemon-set took: 46.47552ms +Jun 18 12:47:34.381: INFO: Terminating DaemonSet.extensions daemon-set pods took: 200.339148ms +Jun 18 12:47:47.193: INFO: Number of nodes with available pods: 0 +Jun 18 12:47:47.193: INFO: Number of running nodes: 0, number of available pods: 0 +Jun 18 12:47:47.206: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"selfLink":"/apis/apps/v1/namespaces/daemonsets-255/daemonsets","resourceVersion":"101467"},"items":null} -Jun 4 19:32:52.530: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"selfLink":"/api/v1/namespaces/daemonsets-1795/pods","resourceVersion":"24999"},"items":null} +Jun 18 12:47:47.220: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"selfLink":"/api/v1/namespaces/daemonsets-255/pods","resourceVersion":"101467"},"items":null} [AfterEach] [sig-apps] Daemon set [Serial] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 -Jun 4 19:32:52.681: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "daemonsets-1795" for this suite. -Jun 4 19:33:00.744: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -Jun 4 19:33:01.100: INFO: namespace daemonsets-1795 deletion completed in 8.399396574s + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Jun 18 12:47:47.278: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "daemonsets-255" for this suite. +Jun 18 12:47:55.320: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:47:55.706: INFO: namespace daemonsets-255 deletion completed in 8.415639952s -• [SLOW TEST:41.489 seconds] +• [SLOW TEST:54.265 seconds] [sig-apps] Daemon set [Serial] -/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22 - should rollback without unnecessary restarts [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 ------------------------------- -SSSSSSS ------------------------------- -[sig-apps] Deployment - deployment should support rollover [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -[BeforeEach] [sig-apps] Deployment - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 -STEP: Creating a kubernetes client -Jun 4 19:33:01.101: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504 -STEP: Building a namespace api object, basename deployment -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in deployment-3485 -STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [sig-apps] Deployment - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:65 -[It] deployment should support rollover [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -Jun 4 19:33:01.345: INFO: Pod name rollover-pod: Found 0 pods out of 1 -Jun 4 19:33:06.358: INFO: Pod name rollover-pod: Found 1 pods out of 1 -STEP: ensuring each pod is running -Jun 4 19:33:06.358: INFO: Waiting for pods owned by replica set "test-rollover-controller" to become ready -Jun 4 19:33:08.367: INFO: Creating deployment "test-rollover-deployment" -Jun 4 19:33:08.386: INFO: Make sure deployment "test-rollover-deployment" performs scaling operations -Jun 4 19:33:10.403: INFO: Check revision of new replica set for deployment "test-rollover-deployment" -Jun 4 19:33:10.419: INFO: Ensure that both replica sets have 1 created replica -Jun 4 19:33:10.435: INFO: Rollover old replica sets for deployment "test-rollover-deployment" with new image update -Jun 4 19:33:10.458: INFO: Updating deployment test-rollover-deployment -Jun 4 19:33:10.458: INFO: Wait deployment "test-rollover-deployment" to be observed by the deployment controller -Jun 4 19:33:12.475: INFO: Wait for revision update of deployment "test-rollover-deployment" to 2 -Jun 4 19:33:12.492: INFO: Make sure deployment "test-rollover-deployment" is complete -Jun 4 19:33:12.507: INFO: all replica sets need to contain the pod-template-hash label -Jun 4 19:33:12.507: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:2, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63695273588, loc:(*time.Location)(0x8a140e0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63695273588, loc:(*time.Location)(0x8a140e0)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63695273591, loc:(*time.Location)(0x8a140e0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63695273588, loc:(*time.Location)(0x8a140e0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-766b4d6c9d\" is progressing."}}, CollisionCount:(*int32)(nil)} -Jun 4 19:33:14.556: INFO: all replica sets need to contain the pod-template-hash label -Jun 4 19:33:14.556: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:2, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63695273588, loc:(*time.Location)(0x8a140e0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63695273588, loc:(*time.Location)(0x8a140e0)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63695273591, loc:(*time.Location)(0x8a140e0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63695273588, loc:(*time.Location)(0x8a140e0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-766b4d6c9d\" is progressing."}}, CollisionCount:(*int32)(nil)} -Jun 4 19:33:16.526: INFO: all replica sets need to contain the pod-template-hash label -Jun 4 19:33:16.526: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:2, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63695273588, loc:(*time.Location)(0x8a140e0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63695273588, loc:(*time.Location)(0x8a140e0)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63695273591, loc:(*time.Location)(0x8a140e0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63695273588, loc:(*time.Location)(0x8a140e0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-766b4d6c9d\" is progressing."}}, CollisionCount:(*int32)(nil)} -Jun 4 19:33:18.555: INFO: all replica sets need to contain the pod-template-hash label -Jun 4 19:33:18.556: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:2, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63695273588, loc:(*time.Location)(0x8a140e0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63695273588, loc:(*time.Location)(0x8a140e0)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63695273591, loc:(*time.Location)(0x8a140e0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63695273588, loc:(*time.Location)(0x8a140e0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-766b4d6c9d\" is progressing."}}, CollisionCount:(*int32)(nil)} -Jun 4 19:33:20.524: INFO: all replica sets need to contain the pod-template-hash label -Jun 4 19:33:20.524: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:2, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63695273588, loc:(*time.Location)(0x8a140e0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63695273588, loc:(*time.Location)(0x8a140e0)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63695273591, loc:(*time.Location)(0x8a140e0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63695273588, loc:(*time.Location)(0x8a140e0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-766b4d6c9d\" is progressing."}}, CollisionCount:(*int32)(nil)} -Jun 4 19:33:22.527: INFO: -Jun 4 19:33:22.527: INFO: Ensure that both old replica sets have no replicas -[AfterEach] [sig-apps] Deployment - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:59 -Jun 4 19:33:22.551: INFO: Deployment "test-rollover-deployment": -&Deployment{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-rollover-deployment,GenerateName:,Namespace:deployment-3485,SelfLink:/apis/apps/v1/namespaces/deployment-3485/deployments/test-rollover-deployment,UID:94db69a9-86ff-11e9-8318-1e5386706511,ResourceVersion:25177,Generation:2,CreationTimestamp:2019-06-04 19:33:08 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: rollover-pod,},Annotations:map[string]string{deployment.kubernetes.io/revision: 2,},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:DeploymentSpec{Replicas:*1,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: rollover-pod,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: rollover-pod,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[],Containers:[{redis gcr.io/kubernetes-e2e-test-images/redis:1.0 [] [] [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,},},Strategy:DeploymentStrategy{Type:RollingUpdate,RollingUpdate:&RollingUpdateDeployment{MaxUnavailable:0,MaxSurge:1,},},MinReadySeconds:10,RevisionHistoryLimit:*10,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:2,Replicas:1,UpdatedReplicas:1,AvailableReplicas:1,UnavailableReplicas:0,Conditions:[{Available True 2019-06-04 19:33:08 +0000 UTC 2019-06-04 19:33:08 +0000 UTC MinimumReplicasAvailable Deployment has minimum availability.} {Progressing True 2019-06-04 19:33:21 +0000 UTC 2019-06-04 19:33:08 +0000 UTC NewReplicaSetAvailable ReplicaSet "test-rollover-deployment-766b4d6c9d" has successfully progressed.}],ReadyReplicas:1,CollisionCount:nil,},} - -Jun 4 19:33:22.559: INFO: New ReplicaSet "test-rollover-deployment-766b4d6c9d" of Deployment "test-rollover-deployment": -&ReplicaSet{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-rollover-deployment-766b4d6c9d,GenerateName:,Namespace:deployment-3485,SelfLink:/apis/apps/v1/namespaces/deployment-3485/replicasets/test-rollover-deployment-766b4d6c9d,UID:9619f5c8-86ff-11e9-9509-923735b172b1,ResourceVersion:25166,Generation:2,CreationTimestamp:2019-06-04 19:33:10 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: rollover-pod,pod-template-hash: 766b4d6c9d,},Annotations:map[string]string{deployment.kubernetes.io/desired-replicas: 1,deployment.kubernetes.io/max-replicas: 2,deployment.kubernetes.io/revision: 2,},OwnerReferences:[{apps/v1 Deployment test-rollover-deployment 94db69a9-86ff-11e9-8318-1e5386706511 0xc00285c3a7 0xc00285c3a8}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:ReplicaSetSpec{Replicas:*1,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: rollover-pod,pod-template-hash: 766b4d6c9d,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: rollover-pod,pod-template-hash: 766b4d6c9d,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[],Containers:[{redis gcr.io/kubernetes-e2e-test-images/redis:1.0 [] [] [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,},},MinReadySeconds:10,},Status:ReplicaSetStatus{Replicas:1,FullyLabeledReplicas:1,ObservedGeneration:2,ReadyReplicas:1,AvailableReplicas:1,Conditions:[],},} -Jun 4 19:33:22.559: INFO: All old ReplicaSets of Deployment "test-rollover-deployment": -Jun 4 19:33:22.559: INFO: &ReplicaSet{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-rollover-controller,GenerateName:,Namespace:deployment-3485,SelfLink:/apis/apps/v1/namespaces/deployment-3485/replicasets/test-rollover-controller,UID:90a7dddc-86ff-11e9-8318-1e5386706511,ResourceVersion:25175,Generation:2,CreationTimestamp:2019-06-04 19:33:01 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: rollover-pod,pod: nginx,},Annotations:map[string]string{deployment.kubernetes.io/desired-replicas: 1,deployment.kubernetes.io/max-replicas: 2,},OwnerReferences:[{apps/v1 Deployment test-rollover-deployment 94db69a9-86ff-11e9-8318-1e5386706511 0xc00285c1b7 0xc00285c1b8}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:ReplicaSetSpec{Replicas:*0,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: rollover-pod,pod: nginx,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: rollover-pod,pod: nginx,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,},},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:0,FullyLabeledReplicas:0,ObservedGeneration:2,ReadyReplicas:0,AvailableReplicas:0,Conditions:[],},} -Jun 4 19:33:22.559: INFO: &ReplicaSet{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-rollover-deployment-6455657675,GenerateName:,Namespace:deployment-3485,SelfLink:/apis/apps/v1/namespaces/deployment-3485/replicasets/test-rollover-deployment-6455657675,UID:94df22f0-86ff-11e9-9509-923735b172b1,ResourceVersion:25130,Generation:2,CreationTimestamp:2019-06-04 19:33:08 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: rollover-pod,pod-template-hash: 6455657675,},Annotations:map[string]string{deployment.kubernetes.io/desired-replicas: 1,deployment.kubernetes.io/max-replicas: 2,deployment.kubernetes.io/revision: 1,},OwnerReferences:[{apps/v1 Deployment test-rollover-deployment 94db69a9-86ff-11e9-8318-1e5386706511 0xc00285c287 0xc00285c288}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:ReplicaSetSpec{Replicas:*0,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: rollover-pod,pod-template-hash: 6455657675,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: rollover-pod,pod-template-hash: 6455657675,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[],Containers:[{redis-slave gcr.io/google_samples/gb-redisslave:nonexistent [] [] [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,},},MinReadySeconds:10,},Status:ReplicaSetStatus{Replicas:0,FullyLabeledReplicas:0,ObservedGeneration:2,ReadyReplicas:0,AvailableReplicas:0,Conditions:[],},} -Jun 4 19:33:22.582: INFO: Pod "test-rollover-deployment-766b4d6c9d-995rz" is available: -&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-rollover-deployment-766b4d6c9d-995rz,GenerateName:test-rollover-deployment-766b4d6c9d-,Namespace:deployment-3485,SelfLink:/api/v1/namespaces/deployment-3485/pods/test-rollover-deployment-766b4d6c9d-995rz,UID:9620feba-86ff-11e9-9509-923735b172b1,ResourceVersion:25146,Generation:0,CreationTimestamp:2019-06-04 19:33:10 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: rollover-pod,pod-template-hash: 766b4d6c9d,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet test-rollover-deployment-766b4d6c9d 9619f5c8-86ff-11e9-9509-923735b172b1 0xc00285cec7 0xc00285cec8}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-jfdl2 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-jfdl2,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{redis gcr.io/kubernetes-e2e-test-images/redis:1.0 [] [] [] [] [] {map[] map[]} [{default-token-jfdl2 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.212.23.164,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc00285cf50} {node.kubernetes.io/unreachable Exists NoExecute 0xc00285cf70}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Running,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 19:33:10 +0000 UTC } {Ready True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 19:33:11 +0000 UTC } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 19:33:11 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 19:33:10 +0000 UTC }],Message:,Reason:,HostIP:10.212.23.164,PodIP:172.30.254.165,StartTime:2019-06-04 19:33:10 +0000 UTC,ContainerStatuses:[{redis {nil ContainerStateRunning{StartedAt:2019-06-04 19:33:11 +0000 UTC,} nil} {nil nil nil} true 0 gcr.io/kubernetes-e2e-test-images/redis:1.0 gcr.io/kubernetes-e2e-test-images/redis@sha256:af4748d1655c08dc54d4be5182135395db9ce87aba2d4699b26b14ae197c5830 containerd://2e52a6d0da95ebf7ab1fb789c4e973b12bd8234c0d1f9cc24a10808ae6b80b8e}],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} -[AfterEach] [sig-apps] Deployment - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 -Jun 4 19:33:22.582: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "deployment-3485" for this suite. -Jun 4 19:33:30.653: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -Jun 4 19:33:31.031: INFO: namespace deployment-3485 deletion completed in 8.419272779s - -• [SLOW TEST:29.930 seconds] -[sig-apps] Deployment -/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22 - deployment should support rollover [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 ------------------------------- -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------- -[sig-apps] Deployment - RecreateDeployment should delete old pods and create new ones [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -[BeforeEach] [sig-apps] Deployment - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 -STEP: Creating a kubernetes client -Jun 4 19:33:31.033: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504 -STEP: Building a namespace api object, basename deployment -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in deployment-873 -STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [sig-apps] Deployment - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:65 -[It] RecreateDeployment should delete old pods and create new ones [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -Jun 4 19:33:31.572: INFO: Creating deployment "test-recreate-deployment" -Jun 4 19:33:31.581: INFO: Waiting deployment "test-recreate-deployment" to be updated to revision 1 -Jun 4 19:33:31.598: INFO: deployment "test-recreate-deployment" doesn't have the required revision set -Jun 4 19:33:33.616: INFO: Waiting deployment "test-recreate-deployment" to complete -Jun 4 19:33:33.623: INFO: Triggering a new rollout for deployment "test-recreate-deployment" -Jun 4 19:33:33.655: INFO: Updating deployment test-recreate-deployment -Jun 4 19:33:33.655: INFO: Watching deployment "test-recreate-deployment" to verify that new pods will not run with olds pods -[AfterEach] [sig-apps] Deployment - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:59 -Jun 4 19:33:33.737: INFO: Deployment "test-recreate-deployment": -&Deployment{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-recreate-deployment,GenerateName:,Namespace:deployment-873,SelfLink:/apis/apps/v1/namespaces/deployment-873/deployments/test-recreate-deployment,UID:a2b03036-86ff-11e9-8318-1e5386706511,ResourceVersion:25289,Generation:2,CreationTimestamp:2019-06-04 19:33:31 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod-3,},Annotations:map[string]string{deployment.kubernetes.io/revision: 2,},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:DeploymentSpec{Replicas:*1,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod-3,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod-3,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,},},Strategy:DeploymentStrategy{Type:Recreate,RollingUpdate:nil,},MinReadySeconds:0,RevisionHistoryLimit:*10,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:2,Replicas:1,UpdatedReplicas:1,AvailableReplicas:0,UnavailableReplicas:1,Conditions:[{Available False 2019-06-04 19:33:33 +0000 UTC 2019-06-04 19:33:33 +0000 UTC MinimumReplicasUnavailable Deployment does not have minimum availability.} {Progressing True 2019-06-04 19:33:33 +0000 UTC 2019-06-04 19:33:31 +0000 UTC ReplicaSetUpdated ReplicaSet "test-recreate-deployment-c9cbd8684" is progressing.}],ReadyReplicas:0,CollisionCount:nil,},} - -Jun 4 19:33:33.745: INFO: New ReplicaSet "test-recreate-deployment-c9cbd8684" of Deployment "test-recreate-deployment": -&ReplicaSet{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-recreate-deployment-c9cbd8684,GenerateName:,Namespace:deployment-873,SelfLink:/apis/apps/v1/namespaces/deployment-873/replicasets/test-recreate-deployment-c9cbd8684,UID:a3f29dd4-86ff-11e9-9509-923735b172b1,ResourceVersion:25286,Generation:1,CreationTimestamp:2019-06-04 19:33:33 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod-3,pod-template-hash: c9cbd8684,},Annotations:map[string]string{deployment.kubernetes.io/desired-replicas: 1,deployment.kubernetes.io/max-replicas: 1,deployment.kubernetes.io/revision: 2,},OwnerReferences:[{apps/v1 Deployment test-recreate-deployment a2b03036-86ff-11e9-8318-1e5386706511 0xc0005d1a60 0xc0005d1a61}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:ReplicaSetSpec{Replicas:*1,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod-3,pod-template-hash: c9cbd8684,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod-3,pod-template-hash: c9cbd8684,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,},},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:1,FullyLabeledReplicas:1,ObservedGeneration:1,ReadyReplicas:0,AvailableReplicas:0,Conditions:[],},} -Jun 4 19:33:33.745: INFO: All old ReplicaSets of Deployment "test-recreate-deployment": -Jun 4 19:33:33.745: INFO: &ReplicaSet{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-recreate-deployment-7d57d5ff7c,GenerateName:,Namespace:deployment-873,SelfLink:/apis/apps/v1/namespaces/deployment-873/replicasets/test-recreate-deployment-7d57d5ff7c,UID:a2b15ca8-86ff-11e9-9509-923735b172b1,ResourceVersion:25277,Generation:2,CreationTimestamp:2019-06-04 19:33:31 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod-3,pod-template-hash: 7d57d5ff7c,},Annotations:map[string]string{deployment.kubernetes.io/desired-replicas: 1,deployment.kubernetes.io/max-replicas: 1,deployment.kubernetes.io/revision: 1,},OwnerReferences:[{apps/v1 Deployment test-recreate-deployment a2b03036-86ff-11e9-8318-1e5386706511 0xc0005d1997 0xc0005d1998}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:ReplicaSetSpec{Replicas:*0,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod-3,pod-template-hash: 7d57d5ff7c,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod-3,pod-template-hash: 7d57d5ff7c,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[],Containers:[{redis gcr.io/kubernetes-e2e-test-images/redis:1.0 [] [] [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,},},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:0,FullyLabeledReplicas:0,ObservedGeneration:2,ReadyReplicas:0,AvailableReplicas:0,Conditions:[],},} -Jun 4 19:33:33.755: INFO: Pod "test-recreate-deployment-c9cbd8684-pc2kr" is not available: -&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-recreate-deployment-c9cbd8684-pc2kr,GenerateName:test-recreate-deployment-c9cbd8684-,Namespace:deployment-873,SelfLink:/api/v1/namespaces/deployment-873/pods/test-recreate-deployment-c9cbd8684-pc2kr,UID:a3f3ebbb-86ff-11e9-9509-923735b172b1,ResourceVersion:25285,Generation:0,CreationTimestamp:2019-06-04 19:33:33 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod-3,pod-template-hash: c9cbd8684,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet test-recreate-deployment-c9cbd8684 a3f29dd4-86ff-11e9-9509-923735b172b1 0xc0027822e0 0xc0027822e1}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-88jz6 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-88jz6,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-88jz6 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.212.23.161,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc002782350} {node.kubernetes.io/unreachable Exists NoExecute 0xc002782370}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-04 19:33:33 +0000 UTC }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} -[AfterEach] [sig-apps] Deployment - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 -Jun 4 19:33:33.755: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "deployment-873" for this suite. -Jun 4 19:33:39.810: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -Jun 4 19:33:40.192: INFO: namespace deployment-873 deletion completed in 6.420835007s - -• [SLOW TEST:9.158 seconds] -[sig-apps] Deployment -/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22 - RecreateDeployment should delete old pods and create new ones [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22 + should update pod when spec was updated and update strategy is RollingUpdate [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 ------------------------------ SSSSSSSSSSSS ------------------------------ -[k8s.io] InitContainer [NodeConformance] - should invoke init containers on a RestartAlways pod [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -[BeforeEach] [k8s.io] InitContainer [NodeConformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 -STEP: Creating a kubernetes client -Jun 4 19:33:40.192: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504 -STEP: Building a namespace api object, basename init-container -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in init-container-6163 -STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [k8s.io] InitContainer [NodeConformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/init_container.go:43 -[It] should invoke init containers on a RestartAlways pod [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -STEP: creating the pod -Jun 4 19:33:40.409: INFO: PodSpec: initContainers in spec.initContainers -[AfterEach] [k8s.io] InitContainer [NodeConformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 -Jun 4 19:33:43.778: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "init-container-6163" for this suite. -Jun 4 19:34:07.898: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -Jun 4 19:34:08.253: INFO: namespace init-container-6163 deletion completed in 24.397202629s - -• [SLOW TEST:28.061 seconds] -[k8s.io] InitContainer [NodeConformance] -/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687 - should invoke init containers on a RestartAlways pod [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 ------------------------------- -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------- -[sig-storage] ConfigMap - should be consumable from pods in volume [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -[BeforeEach] [sig-storage] ConfigMap - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 -STEP: Creating a kubernetes client -Jun 4 19:34:08.255: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504 -STEP: Building a namespace api object, basename configmap -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in configmap-5781 -STEP: Waiting for a default service account to be provisioned in namespace -[It] should be consumable from pods in volume [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -STEP: Creating configMap with name configmap-test-volume-b8aeac08-86ff-11e9-8d1b-467ee19922ac -STEP: Creating a pod to test consume configMaps -Jun 4 19:34:08.509: INFO: Waiting up to 5m0s for pod "pod-configmaps-b8b0c847-86ff-11e9-8d1b-467ee19922ac" in namespace "configmap-5781" to be "success or failure" -Jun 4 19:34:08.522: INFO: Pod "pod-configmaps-b8b0c847-86ff-11e9-8d1b-467ee19922ac": Phase="Pending", Reason="", readiness=false. Elapsed: 12.335482ms -Jun 4 19:34:10.544: INFO: Pod "pod-configmaps-b8b0c847-86ff-11e9-8d1b-467ee19922ac": Phase="Running", Reason="", readiness=true. Elapsed: 2.034290014s -Jun 4 19:34:12.556: INFO: Pod "pod-configmaps-b8b0c847-86ff-11e9-8d1b-467ee19922ac": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.047156263s -STEP: Saw pod success -Jun 4 19:34:12.557: INFO: Pod "pod-configmaps-b8b0c847-86ff-11e9-8d1b-467ee19922ac" satisfied condition "success or failure" -Jun 4 19:34:12.656: INFO: Trying to get logs from node 10.212.23.189 pod pod-configmaps-b8b0c847-86ff-11e9-8d1b-467ee19922ac container configmap-volume-test: -STEP: delete the pod -Jun 4 19:34:12.717: INFO: Waiting for pod pod-configmaps-b8b0c847-86ff-11e9-8d1b-467ee19922ac to disappear -Jun 4 19:34:12.728: INFO: Pod pod-configmaps-b8b0c847-86ff-11e9-8d1b-467ee19922ac no longer exists -[AfterEach] [sig-storage] ConfigMap - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 -Jun 4 19:34:12.728: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "configmap-5781" for this suite. -Jun 4 19:34:18.784: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -Jun 4 19:34:19.196: INFO: namespace configmap-5781 deletion completed in 6.452858397s - -• [SLOW TEST:10.941 seconds] -[sig-storage] ConfigMap -/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:32 - should be consumable from pods in volume [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 ------------------------------- -SSS ------------------------------- -[sig-api-machinery] Secrets - should be consumable from pods in env vars [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -[BeforeEach] [sig-api-machinery] Secrets - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 -STEP: Creating a kubernetes client -Jun 4 19:34:19.197: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504 -STEP: Building a namespace api object, basename secrets -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in secrets-4429 -STEP: Waiting for a default service account to be provisioned in namespace -[It] should be consumable from pods in env vars [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -STEP: Creating secret with name secret-test-bf342e40-86ff-11e9-8d1b-467ee19922ac -STEP: Creating a pod to test consume secrets -Jun 4 19:34:19.441: INFO: Waiting up to 5m0s for pod "pod-secrets-bf356dc5-86ff-11e9-8d1b-467ee19922ac" in namespace "secrets-4429" to be "success or failure" -Jun 4 19:34:19.453: INFO: Pod "pod-secrets-bf356dc5-86ff-11e9-8d1b-467ee19922ac": Phase="Pending", Reason="", readiness=false. Elapsed: 11.496877ms -Jun 4 19:34:21.467: INFO: Pod "pod-secrets-bf356dc5-86ff-11e9-8d1b-467ee19922ac": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.025636516s -STEP: Saw pod success -Jun 4 19:34:21.467: INFO: Pod "pod-secrets-bf356dc5-86ff-11e9-8d1b-467ee19922ac" satisfied condition "success or failure" -Jun 4 19:34:21.479: INFO: Trying to get logs from node 10.212.23.161 pod pod-secrets-bf356dc5-86ff-11e9-8d1b-467ee19922ac container secret-env-test: -STEP: delete the pod -Jun 4 19:34:21.548: INFO: Waiting for pod pod-secrets-bf356dc5-86ff-11e9-8d1b-467ee19922ac to disappear -Jun 4 19:34:21.559: INFO: Pod pod-secrets-bf356dc5-86ff-11e9-8d1b-467ee19922ac no longer exists -[AfterEach] [sig-api-machinery] Secrets - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 -Jun 4 19:34:21.559: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "secrets-4429" for this suite. -Jun 4 19:34:27.725: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -Jun 4 19:34:28.056: INFO: namespace secrets-4429 deletion completed in 6.376920859s - -• [SLOW TEST:8.860 seconds] -[sig-api-machinery] Secrets -/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/secrets.go:32 - should be consumable from pods in env vars [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 ------------------------------- -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------- -[sig-storage] Downward API volume - should provide container's cpu limit [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -[BeforeEach] [sig-storage] Downward API volume - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +[sig-node] Downward API + should provide container's limits.cpu/memory and requests.cpu/memory as env vars [NodeConformance] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [sig-node] Downward API + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 STEP: Creating a kubernetes client -Jun 4 19:34:28.057: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504 +Jun 18 12:47:55.707: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150 STEP: Building a namespace api object, basename downward-api -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in downward-api-1423 -STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [sig-storage] Downward API volume - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:39 -[It] should provide container's cpu limit [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -STEP: Creating a pod to test downward API volume plugin -Jun 4 19:34:28.298: INFO: Waiting up to 5m0s for pod "downwardapi-volume-c47c911f-86ff-11e9-8d1b-467ee19922ac" in namespace "downward-api-1423" to be "success or failure" -Jun 4 19:34:28.309: INFO: Pod "downwardapi-volume-c47c911f-86ff-11e9-8d1b-467ee19922ac": Phase="Pending", Reason="", readiness=false. Elapsed: 11.404505ms -Jun 4 19:34:30.323: INFO: Pod "downwardapi-volume-c47c911f-86ff-11e9-8d1b-467ee19922ac": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.025062899s -STEP: Saw pod success -Jun 4 19:34:30.323: INFO: Pod "downwardapi-volume-c47c911f-86ff-11e9-8d1b-467ee19922ac" satisfied condition "success or failure" -Jun 4 19:34:30.335: INFO: Trying to get logs from node 10.212.23.164 pod downwardapi-volume-c47c911f-86ff-11e9-8d1b-467ee19922ac container client-container: -STEP: delete the pod -Jun 4 19:34:30.402: INFO: Waiting for pod downwardapi-volume-c47c911f-86ff-11e9-8d1b-467ee19922ac to disappear -Jun 4 19:34:30.413: INFO: Pod downwardapi-volume-c47c911f-86ff-11e9-8d1b-467ee19922ac no longer exists -[AfterEach] [sig-storage] Downward API volume - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 -Jun 4 19:34:30.413: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "downward-api-1423" for this suite. -Jun 4 19:34:36.471: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -Jun 4 19:34:36.896: INFO: namespace downward-api-1423 deletion completed in 6.467690838s - -• [SLOW TEST:8.839 seconds] -[sig-storage] Downward API volume -/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:34 - should provide container's cpu limit [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 ------------------------------- -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------- -[sig-network] Proxy version v1 - should proxy through a service and a pod [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -[BeforeEach] version v1 - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 -STEP: Creating a kubernetes client -Jun 4 19:34:36.899: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504 -STEP: Building a namespace api object, basename proxy -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in proxy-3796 -STEP: Waiting for a default service account to be provisioned in namespace -[It] should proxy through a service and a pod [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -STEP: starting an echo server on multiple ports -STEP: creating replication controller proxy-service-fwvr6 in namespace proxy-3796 -I0604 19:34:37.159531 18 runners.go:184] Created replication controller with name: proxy-service-fwvr6, namespace: proxy-3796, replica count: 1 -I0604 19:34:38.210082 18 runners.go:184] proxy-service-fwvr6 Pods: 1 out of 1 created, 0 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady -I0604 19:34:39.210874 18 runners.go:184] proxy-service-fwvr6 Pods: 1 out of 1 created, 0 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady -I0604 19:34:40.211247 18 runners.go:184] proxy-service-fwvr6 Pods: 1 out of 1 created, 0 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady -I0604 19:34:41.211501 18 runners.go:184] proxy-service-fwvr6 Pods: 1 out of 1 created, 0 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 1 runningButNotReady -I0604 19:34:42.211946 18 runners.go:184] proxy-service-fwvr6 Pods: 1 out of 1 created, 0 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 1 runningButNotReady -I0604 19:34:43.212385 18 runners.go:184] proxy-service-fwvr6 Pods: 1 out of 1 created, 0 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 1 runningButNotReady -I0604 19:34:44.212882 18 runners.go:184] proxy-service-fwvr6 Pods: 1 out of 1 created, 0 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 1 runningButNotReady -I0604 19:34:45.213250 18 runners.go:184] proxy-service-fwvr6 Pods: 1 out of 1 created, 0 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 1 runningButNotReady -I0604 19:34:46.213624 18 runners.go:184] proxy-service-fwvr6 Pods: 1 out of 1 created, 0 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 1 runningButNotReady -I0604 19:34:47.213948 18 runners.go:184] proxy-service-fwvr6 Pods: 1 out of 1 created, 0 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 1 runningButNotReady -I0604 19:34:48.214337 18 runners.go:184] proxy-service-fwvr6 Pods: 1 out of 1 created, 0 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 1 runningButNotReady -I0604 19:34:49.214708 18 runners.go:184] proxy-service-fwvr6 Pods: 1 out of 1 created, 0 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 1 runningButNotReady -I0604 19:34:50.214984 18 runners.go:184] proxy-service-fwvr6 Pods: 1 out of 1 created, 0 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 1 runningButNotReady -I0604 19:34:51.215548 18 runners.go:184] proxy-service-fwvr6 Pods: 1 out of 1 created, 1 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady -Jun 4 19:34:51.229: INFO: setup took 14.1122181s, starting test cases -STEP: running 16 cases, 20 attempts per case, 320 total attempts -Jun 4 19:34:51.277: INFO: (0) /api/v1/namespaces/proxy-3796/pods/http:proxy-service-fwvr6-47f9b:1080/proxy/:
... (200; 47.426589ms) -Jun 4 19:34:51.280: INFO: (0) /api/v1/namespaces/proxy-3796/pods/proxy-service-fwvr6-47f9b/proxy/: test (200; 50.243706ms) -Jun 4 19:34:51.285: INFO: (0) /api/v1/namespaces/proxy-3796/pods/http:proxy-service-fwvr6-47f9b:160/proxy/: foo (200; 55.833936ms) -Jun 4 19:34:51.285: INFO: (0) /api/v1/namespaces/proxy-3796/pods/http:proxy-service-fwvr6-47f9b:162/proxy/: bar (200; 55.694759ms) -Jun 4 19:34:51.286: INFO: (0) /api/v1/namespaces/proxy-3796/pods/proxy-service-fwvr6-47f9b:1080/proxy/: test<... (200; 55.596024ms) -Jun 4 19:34:51.286: INFO: (0) /api/v1/namespaces/proxy-3796/pods/proxy-service-fwvr6-47f9b:162/proxy/: bar (200; 56.420685ms) -Jun 4 19:34:51.286: INFO: (0) /api/v1/namespaces/proxy-3796/pods/proxy-service-fwvr6-47f9b:160/proxy/: foo (200; 56.140751ms) -Jun 4 19:34:51.288: INFO: (0) /api/v1/namespaces/proxy-3796/services/http:proxy-service-fwvr6:portname2/proxy/: bar (200; 57.732613ms) -Jun 4 19:34:51.290: INFO: (0) /api/v1/namespaces/proxy-3796/services/proxy-service-fwvr6:portname2/proxy/: bar (200; 60.064263ms) -Jun 4 19:34:51.291: INFO: (0) /api/v1/namespaces/proxy-3796/services/proxy-service-fwvr6:portname1/proxy/: foo (200; 61.084157ms) -Jun 4 19:34:51.291: INFO: (0) /api/v1/namespaces/proxy-3796/services/http:proxy-service-fwvr6:portname1/proxy/: foo (200; 61.36097ms) -Jun 4 19:34:51.293: INFO: (0) /api/v1/namespaces/proxy-3796/pods/https:proxy-service-fwvr6-47f9b:443/proxy/: ... (200; 27.139058ms) -Jun 4 19:34:51.334: INFO: (1) /api/v1/namespaces/proxy-3796/pods/http:proxy-service-fwvr6-47f9b:160/proxy/: foo (200; 27.443065ms) -Jun 4 19:34:51.335: INFO: (1) /api/v1/namespaces/proxy-3796/pods/proxy-service-fwvr6-47f9b:1080/proxy/: test<... (200; 27.408434ms) -Jun 4 19:34:51.335: INFO: (1) /api/v1/namespaces/proxy-3796/pods/proxy-service-fwvr6-47f9b/proxy/: test (200; 27.725878ms) -Jun 4 19:34:51.339: INFO: (1) /api/v1/namespaces/proxy-3796/services/proxy-service-fwvr6:portname1/proxy/: foo (200; 31.63947ms) -Jun 4 19:34:51.339: INFO: (1) /api/v1/namespaces/proxy-3796/services/proxy-service-fwvr6:portname2/proxy/: bar (200; 31.560288ms) -Jun 4 19:34:51.342: INFO: (1) /api/v1/namespaces/proxy-3796/services/http:proxy-service-fwvr6:portname2/proxy/: bar (200; 34.99459ms) -Jun 4 19:34:51.342: INFO: (1) /api/v1/namespaces/proxy-3796/services/https:proxy-service-fwvr6:tlsportname1/proxy/: tls baz (200; 35.128384ms) -Jun 4 19:34:51.343: INFO: (1) /api/v1/namespaces/proxy-3796/services/http:proxy-service-fwvr6:portname1/proxy/: foo (200; 35.645865ms) -Jun 4 19:34:51.343: INFO: (1) /api/v1/namespaces/proxy-3796/services/https:proxy-service-fwvr6:tlsportname2/proxy/: tls qux (200; 35.810527ms) -Jun 4 19:34:51.358: INFO: (2) /api/v1/namespaces/proxy-3796/pods/https:proxy-service-fwvr6-47f9b:460/proxy/: tls baz (200; 14.862881ms) -Jun 4 19:34:51.362: INFO: (2) /api/v1/namespaces/proxy-3796/pods/http:proxy-service-fwvr6-47f9b:160/proxy/: foo (200; 19.05848ms) -Jun 4 19:34:51.363: INFO: (2) /api/v1/namespaces/proxy-3796/pods/http:proxy-service-fwvr6-47f9b:1080/proxy/: ... (200; 19.320361ms) -Jun 4 19:34:51.363: INFO: (2) /api/v1/namespaces/proxy-3796/pods/proxy-service-fwvr6-47f9b/proxy/: test (200; 19.383337ms) -Jun 4 19:34:51.363: INFO: (2) /api/v1/namespaces/proxy-3796/pods/proxy-service-fwvr6-47f9b:162/proxy/: bar (200; 19.870257ms) -Jun 4 19:34:51.363: INFO: (2) /api/v1/namespaces/proxy-3796/pods/proxy-service-fwvr6-47f9b:1080/proxy/: test<... (200; 19.91009ms) -Jun 4 19:34:51.363: INFO: (2) /api/v1/namespaces/proxy-3796/pods/http:proxy-service-fwvr6-47f9b:162/proxy/: bar (200; 19.919265ms) -Jun 4 19:34:51.363: INFO: (2) /api/v1/namespaces/proxy-3796/pods/https:proxy-service-fwvr6-47f9b:443/proxy/: test<... (200; 21.462363ms) -Jun 4 19:34:51.398: INFO: (3) /api/v1/namespaces/proxy-3796/pods/proxy-service-fwvr6-47f9b:160/proxy/: foo (200; 21.624994ms) -Jun 4 19:34:51.399: INFO: (3) /api/v1/namespaces/proxy-3796/pods/https:proxy-service-fwvr6-47f9b:460/proxy/: tls baz (200; 21.877646ms) -Jun 4 19:34:51.399: INFO: (3) /api/v1/namespaces/proxy-3796/pods/https:proxy-service-fwvr6-47f9b:462/proxy/: tls qux (200; 22.006455ms) -Jun 4 19:34:51.399: INFO: (3) /api/v1/namespaces/proxy-3796/pods/http:proxy-service-fwvr6-47f9b:1080/proxy/: ... (200; 22.280564ms) -Jun 4 19:34:51.399: INFO: (3) /api/v1/namespaces/proxy-3796/pods/https:proxy-service-fwvr6-47f9b:443/proxy/: test (200; 21.96996ms) -Jun 4 19:34:51.399: INFO: (3) /api/v1/namespaces/proxy-3796/pods/http:proxy-service-fwvr6-47f9b:160/proxy/: foo (200; 22.584182ms) -Jun 4 19:34:51.402: INFO: (3) /api/v1/namespaces/proxy-3796/services/proxy-service-fwvr6:portname1/proxy/: foo (200; 25.466957ms) -Jun 4 19:34:51.406: INFO: (3) /api/v1/namespaces/proxy-3796/services/http:proxy-service-fwvr6:portname1/proxy/: foo (200; 28.739498ms) -Jun 4 19:34:51.406: INFO: (3) /api/v1/namespaces/proxy-3796/services/https:proxy-service-fwvr6:tlsportname2/proxy/: tls qux (200; 29.340786ms) -Jun 4 19:34:51.406: INFO: (3) /api/v1/namespaces/proxy-3796/services/http:proxy-service-fwvr6:portname2/proxy/: bar (200; 28.923212ms) -Jun 4 19:34:51.406: INFO: (3) /api/v1/namespaces/proxy-3796/services/https:proxy-service-fwvr6:tlsportname1/proxy/: tls baz (200; 29.422302ms) -Jun 4 19:34:51.407: INFO: (3) /api/v1/namespaces/proxy-3796/services/proxy-service-fwvr6:portname2/proxy/: bar (200; 29.319172ms) -Jun 4 19:34:51.421: INFO: (4) /api/v1/namespaces/proxy-3796/pods/http:proxy-service-fwvr6-47f9b:1080/proxy/: ... (200; 14.612624ms) -Jun 4 19:34:51.427: INFO: (4) /api/v1/namespaces/proxy-3796/pods/proxy-service-fwvr6-47f9b:160/proxy/: foo (200; 20.337645ms) -Jun 4 19:34:51.427: INFO: (4) /api/v1/namespaces/proxy-3796/pods/proxy-service-fwvr6-47f9b:162/proxy/: bar (200; 20.340041ms) -Jun 4 19:34:51.427: INFO: (4) /api/v1/namespaces/proxy-3796/pods/proxy-service-fwvr6-47f9b:1080/proxy/: test<... (200; 20.595969ms) -Jun 4 19:34:51.427: INFO: (4) /api/v1/namespaces/proxy-3796/pods/proxy-service-fwvr6-47f9b/proxy/: test (200; 20.420277ms) -Jun 4 19:34:51.427: INFO: (4) /api/v1/namespaces/proxy-3796/pods/http:proxy-service-fwvr6-47f9b:160/proxy/: foo (200; 20.352008ms) -Jun 4 19:34:51.428: INFO: (4) /api/v1/namespaces/proxy-3796/pods/https:proxy-service-fwvr6-47f9b:462/proxy/: tls qux (200; 21.046796ms) -Jun 4 19:34:51.428: INFO: (4) /api/v1/namespaces/proxy-3796/pods/http:proxy-service-fwvr6-47f9b:162/proxy/: bar (200; 21.749521ms) -Jun 4 19:34:51.428: INFO: (4) /api/v1/namespaces/proxy-3796/pods/https:proxy-service-fwvr6-47f9b:443/proxy/: test (200; 27.348444ms) -Jun 4 19:34:51.465: INFO: (5) /api/v1/namespaces/proxy-3796/pods/https:proxy-service-fwvr6-47f9b:460/proxy/: tls baz (200; 27.401859ms) -Jun 4 19:34:51.465: INFO: (5) /api/v1/namespaces/proxy-3796/pods/http:proxy-service-fwvr6-47f9b:162/proxy/: bar (200; 27.171787ms) -Jun 4 19:34:51.465: INFO: (5) /api/v1/namespaces/proxy-3796/pods/https:proxy-service-fwvr6-47f9b:462/proxy/: tls qux (200; 27.542118ms) -Jun 4 19:34:51.465: INFO: (5) /api/v1/namespaces/proxy-3796/pods/http:proxy-service-fwvr6-47f9b:1080/proxy/: ... (200; 27.414146ms) -Jun 4 19:34:51.465: INFO: (5) /api/v1/namespaces/proxy-3796/pods/https:proxy-service-fwvr6-47f9b:443/proxy/: test<... (200; 27.828708ms) -Jun 4 19:34:51.468: INFO: (5) /api/v1/namespaces/proxy-3796/services/http:proxy-service-fwvr6:portname2/proxy/: bar (200; 30.838003ms) -Jun 4 19:34:51.472: INFO: (5) /api/v1/namespaces/proxy-3796/services/https:proxy-service-fwvr6:tlsportname1/proxy/: tls baz (200; 34.079107ms) -Jun 4 19:34:51.472: INFO: (5) /api/v1/namespaces/proxy-3796/services/http:proxy-service-fwvr6:portname1/proxy/: foo (200; 34.532418ms) -Jun 4 19:34:51.472: INFO: (5) /api/v1/namespaces/proxy-3796/services/https:proxy-service-fwvr6:tlsportname2/proxy/: tls qux (200; 34.403558ms) -Jun 4 19:34:51.472: INFO: (5) /api/v1/namespaces/proxy-3796/services/proxy-service-fwvr6:portname2/proxy/: bar (200; 34.536503ms) -Jun 4 19:34:51.472: INFO: (5) /api/v1/namespaces/proxy-3796/services/proxy-service-fwvr6:portname1/proxy/: foo (200; 34.670522ms) -Jun 4 19:34:51.493: INFO: (6) /api/v1/namespaces/proxy-3796/pods/https:proxy-service-fwvr6-47f9b:443/proxy/: test (200; 20.230477ms) -Jun 4 19:34:51.493: INFO: (6) /api/v1/namespaces/proxy-3796/pods/http:proxy-service-fwvr6-47f9b:1080/proxy/: ... (200; 20.28898ms) -Jun 4 19:34:51.493: INFO: (6) /api/v1/namespaces/proxy-3796/pods/proxy-service-fwvr6-47f9b:162/proxy/: bar (200; 20.386961ms) -Jun 4 19:34:51.493: INFO: (6) /api/v1/namespaces/proxy-3796/pods/proxy-service-fwvr6-47f9b:160/proxy/: foo (200; 20.389258ms) -Jun 4 19:34:51.493: INFO: (6) /api/v1/namespaces/proxy-3796/pods/http:proxy-service-fwvr6-47f9b:160/proxy/: foo (200; 20.996734ms) -Jun 4 19:34:51.494: INFO: (6) /api/v1/namespaces/proxy-3796/pods/proxy-service-fwvr6-47f9b:1080/proxy/: test<... (200; 21.060828ms) -Jun 4 19:34:51.496: INFO: (6) /api/v1/namespaces/proxy-3796/services/http:proxy-service-fwvr6:portname1/proxy/: foo (200; 23.93794ms) -Jun 4 19:34:51.500: INFO: (6) /api/v1/namespaces/proxy-3796/services/http:proxy-service-fwvr6:portname2/proxy/: bar (200; 27.285979ms) -Jun 4 19:34:51.500: INFO: (6) /api/v1/namespaces/proxy-3796/services/https:proxy-service-fwvr6:tlsportname1/proxy/: tls baz (200; 27.834858ms) -Jun 4 19:34:51.500: INFO: (6) /api/v1/namespaces/proxy-3796/services/proxy-service-fwvr6:portname2/proxy/: bar (200; 27.781113ms) -Jun 4 19:34:51.500: INFO: (6) /api/v1/namespaces/proxy-3796/services/proxy-service-fwvr6:portname1/proxy/: foo (200; 27.793079ms) -Jun 4 19:34:51.500: INFO: (6) /api/v1/namespaces/proxy-3796/services/https:proxy-service-fwvr6:tlsportname2/proxy/: tls qux (200; 27.709345ms) -Jun 4 19:34:51.515: INFO: (7) /api/v1/namespaces/proxy-3796/pods/http:proxy-service-fwvr6-47f9b:160/proxy/: foo (200; 14.345468ms) -Jun 4 19:34:51.515: INFO: (7) /api/v1/namespaces/proxy-3796/pods/proxy-service-fwvr6-47f9b:162/proxy/: bar (200; 14.264003ms) -Jun 4 19:34:51.521: INFO: (7) /api/v1/namespaces/proxy-3796/pods/proxy-service-fwvr6-47f9b:1080/proxy/: test<... (200; 20.00327ms) -Jun 4 19:34:51.521: INFO: (7) /api/v1/namespaces/proxy-3796/pods/http:proxy-service-fwvr6-47f9b:1080/proxy/: ... (200; 20.022684ms) -Jun 4 19:34:51.521: INFO: (7) /api/v1/namespaces/proxy-3796/pods/proxy-service-fwvr6-47f9b:160/proxy/: foo (200; 20.105261ms) -Jun 4 19:34:51.521: INFO: (7) /api/v1/namespaces/proxy-3796/pods/http:proxy-service-fwvr6-47f9b:162/proxy/: bar (200; 20.579231ms) -Jun 4 19:34:51.521: INFO: (7) /api/v1/namespaces/proxy-3796/pods/proxy-service-fwvr6-47f9b/proxy/: test (200; 20.410551ms) -Jun 4 19:34:51.521: INFO: (7) /api/v1/namespaces/proxy-3796/pods/https:proxy-service-fwvr6-47f9b:443/proxy/: test (200; 21.592945ms) -Jun 4 19:34:51.556: INFO: (8) /api/v1/namespaces/proxy-3796/pods/https:proxy-service-fwvr6-47f9b:462/proxy/: tls qux (200; 21.607187ms) -Jun 4 19:34:51.556: INFO: (8) /api/v1/namespaces/proxy-3796/pods/http:proxy-service-fwvr6-47f9b:1080/proxy/: ... (200; 21.826796ms) -Jun 4 19:34:51.556: INFO: (8) /api/v1/namespaces/proxy-3796/pods/http:proxy-service-fwvr6-47f9b:162/proxy/: bar (200; 22.388884ms) -Jun 4 19:34:51.556: INFO: (8) /api/v1/namespaces/proxy-3796/pods/proxy-service-fwvr6-47f9b:162/proxy/: bar (200; 22.478362ms) -Jun 4 19:34:51.556: INFO: (8) /api/v1/namespaces/proxy-3796/pods/https:proxy-service-fwvr6-47f9b:443/proxy/: test<... (200; 22.542417ms) -Jun 4 19:34:51.557: INFO: (8) /api/v1/namespaces/proxy-3796/pods/https:proxy-service-fwvr6-47f9b:460/proxy/: tls baz (200; 22.355914ms) -Jun 4 19:34:51.560: INFO: (8) /api/v1/namespaces/proxy-3796/services/http:proxy-service-fwvr6:portname2/proxy/: bar (200; 25.298417ms) -Jun 4 19:34:51.564: INFO: (8) /api/v1/namespaces/proxy-3796/services/http:proxy-service-fwvr6:portname1/proxy/: foo (200; 29.814319ms) -Jun 4 19:34:51.564: INFO: (8) /api/v1/namespaces/proxy-3796/services/https:proxy-service-fwvr6:tlsportname2/proxy/: tls qux (200; 29.716747ms) -Jun 4 19:34:51.564: INFO: (8) /api/v1/namespaces/proxy-3796/services/proxy-service-fwvr6:portname2/proxy/: bar (200; 29.739581ms) -Jun 4 19:34:51.564: INFO: (8) /api/v1/namespaces/proxy-3796/services/proxy-service-fwvr6:portname1/proxy/: foo (200; 29.861925ms) -Jun 4 19:34:51.564: INFO: (8) /api/v1/namespaces/proxy-3796/services/https:proxy-service-fwvr6:tlsportname1/proxy/: tls baz (200; 29.957689ms) -Jun 4 19:34:51.579: INFO: (9) /api/v1/namespaces/proxy-3796/pods/proxy-service-fwvr6-47f9b:160/proxy/: foo (200; 14.21105ms) -Jun 4 19:34:51.583: INFO: (9) /api/v1/namespaces/proxy-3796/pods/proxy-service-fwvr6-47f9b:162/proxy/: bar (200; 18.751217ms) -Jun 4 19:34:51.584: INFO: (9) /api/v1/namespaces/proxy-3796/pods/http:proxy-service-fwvr6-47f9b:162/proxy/: bar (200; 19.209643ms) -Jun 4 19:34:51.584: INFO: (9) /api/v1/namespaces/proxy-3796/pods/https:proxy-service-fwvr6-47f9b:443/proxy/: test<... (200; 19.137432ms) -Jun 4 19:34:51.584: INFO: (9) /api/v1/namespaces/proxy-3796/pods/https:proxy-service-fwvr6-47f9b:460/proxy/: tls baz (200; 19.216255ms) -Jun 4 19:34:51.584: INFO: (9) /api/v1/namespaces/proxy-3796/pods/http:proxy-service-fwvr6-47f9b:160/proxy/: foo (200; 19.224047ms) -Jun 4 19:34:51.584: INFO: (9) /api/v1/namespaces/proxy-3796/pods/https:proxy-service-fwvr6-47f9b:462/proxy/: tls qux (200; 19.713594ms) -Jun 4 19:34:51.585: INFO: (9) /api/v1/namespaces/proxy-3796/pods/proxy-service-fwvr6-47f9b/proxy/: test (200; 20.00235ms) -Jun 4 19:34:51.585: INFO: (9) /api/v1/namespaces/proxy-3796/pods/http:proxy-service-fwvr6-47f9b:1080/proxy/: ... (200; 20.856566ms) -Jun 4 19:34:51.590: INFO: (9) /api/v1/namespaces/proxy-3796/services/proxy-service-fwvr6:portname2/proxy/: bar (200; 25.29026ms) -Jun 4 19:34:51.596: INFO: (9) /api/v1/namespaces/proxy-3796/services/proxy-service-fwvr6:portname1/proxy/: foo (200; 30.97219ms) -Jun 4 19:34:51.596: INFO: (9) /api/v1/namespaces/proxy-3796/services/https:proxy-service-fwvr6:tlsportname2/proxy/: tls qux (200; 30.972623ms) -Jun 4 19:34:51.596: INFO: (9) /api/v1/namespaces/proxy-3796/services/http:proxy-service-fwvr6:portname2/proxy/: bar (200; 31.124695ms) -Jun 4 19:34:51.596: INFO: (9) /api/v1/namespaces/proxy-3796/services/http:proxy-service-fwvr6:portname1/proxy/: foo (200; 31.059897ms) -Jun 4 19:34:51.596: INFO: (9) /api/v1/namespaces/proxy-3796/services/https:proxy-service-fwvr6:tlsportname1/proxy/: tls baz (200; 31.480828ms) -Jun 4 19:34:51.614: INFO: (10) /api/v1/namespaces/proxy-3796/pods/http:proxy-service-fwvr6-47f9b:162/proxy/: bar (200; 17.277789ms) -Jun 4 19:34:51.614: INFO: (10) /api/v1/namespaces/proxy-3796/pods/proxy-service-fwvr6-47f9b:162/proxy/: bar (200; 17.418389ms) -Jun 4 19:34:51.615: INFO: (10) /api/v1/namespaces/proxy-3796/pods/https:proxy-service-fwvr6-47f9b:462/proxy/: tls qux (200; 17.822086ms) -Jun 4 19:34:51.615: INFO: (10) /api/v1/namespaces/proxy-3796/pods/proxy-service-fwvr6-47f9b:1080/proxy/: test<... (200; 18.200438ms) -Jun 4 19:34:51.615: INFO: (10) /api/v1/namespaces/proxy-3796/pods/https:proxy-service-fwvr6-47f9b:460/proxy/: tls baz (200; 17.918077ms) -Jun 4 19:34:51.615: INFO: (10) /api/v1/namespaces/proxy-3796/pods/https:proxy-service-fwvr6-47f9b:443/proxy/: ... (200; 18.691033ms) -Jun 4 19:34:51.615: INFO: (10) /api/v1/namespaces/proxy-3796/pods/proxy-service-fwvr6-47f9b/proxy/: test (200; 18.52784ms) -Jun 4 19:34:51.618: INFO: (10) /api/v1/namespaces/proxy-3796/services/https:proxy-service-fwvr6:tlsportname1/proxy/: tls baz (200; 22.216293ms) -Jun 4 19:34:51.622: INFO: (10) /api/v1/namespaces/proxy-3796/services/http:proxy-service-fwvr6:portname1/proxy/: foo (200; 25.573515ms) -Jun 4 19:34:51.622: INFO: (10) /api/v1/namespaces/proxy-3796/services/proxy-service-fwvr6:portname2/proxy/: bar (200; 26.158375ms) -Jun 4 19:34:51.623: INFO: (10) /api/v1/namespaces/proxy-3796/services/proxy-service-fwvr6:portname1/proxy/: foo (200; 26.311933ms) -Jun 4 19:34:51.623: INFO: (10) /api/v1/namespaces/proxy-3796/services/http:proxy-service-fwvr6:portname2/proxy/: bar (200; 26.179208ms) -Jun 4 19:34:51.623: INFO: (10) /api/v1/namespaces/proxy-3796/services/https:proxy-service-fwvr6:tlsportname2/proxy/: tls qux (200; 26.775379ms) -Jun 4 19:34:51.642: INFO: (11) /api/v1/namespaces/proxy-3796/pods/proxy-service-fwvr6-47f9b:160/proxy/: foo (200; 18.760859ms) -Jun 4 19:34:51.642: INFO: (11) /api/v1/namespaces/proxy-3796/pods/https:proxy-service-fwvr6-47f9b:460/proxy/: tls baz (200; 18.810517ms) -Jun 4 19:34:51.642: INFO: (11) /api/v1/namespaces/proxy-3796/pods/proxy-service-fwvr6-47f9b:162/proxy/: bar (200; 18.735151ms) -Jun 4 19:34:51.643: INFO: (11) /api/v1/namespaces/proxy-3796/pods/https:proxy-service-fwvr6-47f9b:462/proxy/: tls qux (200; 19.144584ms) -Jun 4 19:34:51.643: INFO: (11) /api/v1/namespaces/proxy-3796/pods/http:proxy-service-fwvr6-47f9b:162/proxy/: bar (200; 19.663302ms) -Jun 4 19:34:51.643: INFO: (11) /api/v1/namespaces/proxy-3796/pods/http:proxy-service-fwvr6-47f9b:160/proxy/: foo (200; 19.330321ms) -Jun 4 19:34:51.643: INFO: (11) /api/v1/namespaces/proxy-3796/pods/https:proxy-service-fwvr6-47f9b:443/proxy/: ... (200; 19.480576ms) -Jun 4 19:34:51.643: INFO: (11) /api/v1/namespaces/proxy-3796/pods/proxy-service-fwvr6-47f9b:1080/proxy/: test<... (200; 19.460751ms) -Jun 4 19:34:51.643: INFO: (11) /api/v1/namespaces/proxy-3796/pods/proxy-service-fwvr6-47f9b/proxy/: test (200; 19.834369ms) -Jun 4 19:34:51.646: INFO: (11) /api/v1/namespaces/proxy-3796/services/https:proxy-service-fwvr6:tlsportname2/proxy/: tls qux (200; 22.605367ms) -Jun 4 19:34:51.650: INFO: (11) /api/v1/namespaces/proxy-3796/services/proxy-service-fwvr6:portname1/proxy/: foo (200; 27.258936ms) -Jun 4 19:34:51.650: INFO: (11) /api/v1/namespaces/proxy-3796/services/proxy-service-fwvr6:portname2/proxy/: bar (200; 27.00535ms) -Jun 4 19:34:51.650: INFO: (11) /api/v1/namespaces/proxy-3796/services/https:proxy-service-fwvr6:tlsportname1/proxy/: tls baz (200; 26.953964ms) -Jun 4 19:34:51.650: INFO: (11) /api/v1/namespaces/proxy-3796/services/http:proxy-service-fwvr6:portname2/proxy/: bar (200; 26.578344ms) -Jun 4 19:34:51.650: INFO: (11) /api/v1/namespaces/proxy-3796/services/http:proxy-service-fwvr6:portname1/proxy/: foo (200; 27.085789ms) -Jun 4 19:34:51.669: INFO: (12) /api/v1/namespaces/proxy-3796/pods/proxy-service-fwvr6-47f9b/proxy/: test (200; 18.698159ms) -Jun 4 19:34:51.669: INFO: (12) /api/v1/namespaces/proxy-3796/pods/proxy-service-fwvr6-47f9b:162/proxy/: bar (200; 18.131276ms) -Jun 4 19:34:51.670: INFO: (12) /api/v1/namespaces/proxy-3796/pods/http:proxy-service-fwvr6-47f9b:1080/proxy/: ... (200; 18.265871ms) -Jun 4 19:34:51.669: INFO: (12) /api/v1/namespaces/proxy-3796/pods/proxy-service-fwvr6-47f9b:160/proxy/: foo (200; 18.772711ms) -Jun 4 19:34:51.670: INFO: (12) /api/v1/namespaces/proxy-3796/pods/proxy-service-fwvr6-47f9b:1080/proxy/: test<... (200; 18.758332ms) -Jun 4 19:34:51.670: INFO: (12) /api/v1/namespaces/proxy-3796/pods/http:proxy-service-fwvr6-47f9b:160/proxy/: foo (200; 18.326519ms) -Jun 4 19:34:51.670: INFO: (12) /api/v1/namespaces/proxy-3796/pods/http:proxy-service-fwvr6-47f9b:162/proxy/: bar (200; 18.545532ms) -Jun 4 19:34:51.670: INFO: (12) /api/v1/namespaces/proxy-3796/pods/https:proxy-service-fwvr6-47f9b:443/proxy/: test (200; 18.475624ms) -Jun 4 19:34:51.697: INFO: (13) /api/v1/namespaces/proxy-3796/pods/proxy-service-fwvr6-47f9b:1080/proxy/: test<... (200; 18.468968ms) -Jun 4 19:34:51.697: INFO: (13) /api/v1/namespaces/proxy-3796/pods/https:proxy-service-fwvr6-47f9b:460/proxy/: tls baz (200; 18.892544ms) -Jun 4 19:34:51.697: INFO: (13) /api/v1/namespaces/proxy-3796/pods/https:proxy-service-fwvr6-47f9b:443/proxy/: ... (200; 18.694461ms) -Jun 4 19:34:51.697: INFO: (13) /api/v1/namespaces/proxy-3796/pods/http:proxy-service-fwvr6-47f9b:162/proxy/: bar (200; 18.719956ms) -Jun 4 19:34:51.697: INFO: (13) /api/v1/namespaces/proxy-3796/pods/http:proxy-service-fwvr6-47f9b:160/proxy/: foo (200; 19.319588ms) -Jun 4 19:34:51.700: INFO: (13) /api/v1/namespaces/proxy-3796/services/https:proxy-service-fwvr6:tlsportname1/proxy/: tls baz (200; 21.506266ms) -Jun 4 19:34:51.705: INFO: (13) /api/v1/namespaces/proxy-3796/services/proxy-service-fwvr6:portname1/proxy/: foo (200; 26.938628ms) -Jun 4 19:34:51.705: INFO: (13) /api/v1/namespaces/proxy-3796/services/http:proxy-service-fwvr6:portname2/proxy/: bar (200; 26.809632ms) -Jun 4 19:34:51.705: INFO: (13) /api/v1/namespaces/proxy-3796/services/proxy-service-fwvr6:portname2/proxy/: bar (200; 26.83096ms) -Jun 4 19:34:51.705: INFO: (13) /api/v1/namespaces/proxy-3796/services/http:proxy-service-fwvr6:portname1/proxy/: foo (200; 26.905102ms) -Jun 4 19:34:51.706: INFO: (13) /api/v1/namespaces/proxy-3796/services/https:proxy-service-fwvr6:tlsportname2/proxy/: tls qux (200; 27.272843ms) -Jun 4 19:34:51.720: INFO: (14) /api/v1/namespaces/proxy-3796/pods/http:proxy-service-fwvr6-47f9b:160/proxy/: foo (200; 14.048393ms) -Jun 4 19:34:51.726: INFO: (14) /api/v1/namespaces/proxy-3796/pods/http:proxy-service-fwvr6-47f9b:1080/proxy/: ... (200; 19.840777ms) -Jun 4 19:34:51.726: INFO: (14) /api/v1/namespaces/proxy-3796/pods/https:proxy-service-fwvr6-47f9b:460/proxy/: tls baz (200; 19.925228ms) -Jun 4 19:34:51.726: INFO: (14) /api/v1/namespaces/proxy-3796/pods/proxy-service-fwvr6-47f9b/proxy/: test (200; 20.068151ms) -Jun 4 19:34:51.726: INFO: (14) /api/v1/namespaces/proxy-3796/pods/proxy-service-fwvr6-47f9b:1080/proxy/: test<... (200; 20.009913ms) -Jun 4 19:34:51.726: INFO: (14) /api/v1/namespaces/proxy-3796/pods/proxy-service-fwvr6-47f9b:162/proxy/: bar (200; 19.973538ms) -Jun 4 19:34:51.726: INFO: (14) /api/v1/namespaces/proxy-3796/pods/http:proxy-service-fwvr6-47f9b:162/proxy/: bar (200; 20.26605ms) -Jun 4 19:34:51.726: INFO: (14) /api/v1/namespaces/proxy-3796/pods/proxy-service-fwvr6-47f9b:160/proxy/: foo (200; 20.032238ms) -Jun 4 19:34:51.726: INFO: (14) /api/v1/namespaces/proxy-3796/pods/https:proxy-service-fwvr6-47f9b:462/proxy/: tls qux (200; 20.169929ms) -Jun 4 19:34:51.726: INFO: (14) /api/v1/namespaces/proxy-3796/pods/https:proxy-service-fwvr6-47f9b:443/proxy/: test<... (200; 15.404206ms) -Jun 4 19:34:51.753: INFO: (15) /api/v1/namespaces/proxy-3796/pods/https:proxy-service-fwvr6-47f9b:443/proxy/: ... (200; 19.800574ms) -Jun 4 19:34:51.757: INFO: (15) /api/v1/namespaces/proxy-3796/pods/https:proxy-service-fwvr6-47f9b:460/proxy/: tls baz (200; 19.710023ms) -Jun 4 19:34:51.757: INFO: (15) /api/v1/namespaces/proxy-3796/pods/proxy-service-fwvr6-47f9b/proxy/: test (200; 20.191728ms) -Jun 4 19:34:51.757: INFO: (15) /api/v1/namespaces/proxy-3796/pods/https:proxy-service-fwvr6-47f9b:462/proxy/: tls qux (200; 20.38339ms) -Jun 4 19:34:51.757: INFO: (15) /api/v1/namespaces/proxy-3796/pods/http:proxy-service-fwvr6-47f9b:162/proxy/: bar (200; 20.104535ms) -Jun 4 19:34:51.757: INFO: (15) /api/v1/namespaces/proxy-3796/pods/proxy-service-fwvr6-47f9b:160/proxy/: foo (200; 19.85747ms) -Jun 4 19:34:51.758: INFO: (15) /api/v1/namespaces/proxy-3796/pods/http:proxy-service-fwvr6-47f9b:160/proxy/: foo (200; 20.126177ms) -Jun 4 19:34:51.758: INFO: (15) /api/v1/namespaces/proxy-3796/pods/proxy-service-fwvr6-47f9b:162/proxy/: bar (200; 20.274164ms) -Jun 4 19:34:51.760: INFO: (15) /api/v1/namespaces/proxy-3796/services/proxy-service-fwvr6:portname2/proxy/: bar (200; 23.125878ms) -Jun 4 19:34:51.766: INFO: (15) /api/v1/namespaces/proxy-3796/services/https:proxy-service-fwvr6:tlsportname1/proxy/: tls baz (200; 28.404689ms) -Jun 4 19:34:51.767: INFO: (15) /api/v1/namespaces/proxy-3796/services/proxy-service-fwvr6:portname1/proxy/: foo (200; 29.623904ms) -Jun 4 19:34:51.767: INFO: (15) /api/v1/namespaces/proxy-3796/services/http:proxy-service-fwvr6:portname2/proxy/: bar (200; 29.651243ms) -Jun 4 19:34:51.767: INFO: (15) /api/v1/namespaces/proxy-3796/services/http:proxy-service-fwvr6:portname1/proxy/: foo (200; 29.718334ms) -Jun 4 19:34:51.767: INFO: (15) /api/v1/namespaces/proxy-3796/services/https:proxy-service-fwvr6:tlsportname2/proxy/: tls qux (200; 29.672157ms) -Jun 4 19:34:51.782: INFO: (16) /api/v1/namespaces/proxy-3796/pods/https:proxy-service-fwvr6-47f9b:462/proxy/: tls qux (200; 14.775853ms) -Jun 4 19:34:51.788: INFO: (16) /api/v1/namespaces/proxy-3796/pods/http:proxy-service-fwvr6-47f9b:160/proxy/: foo (200; 20.854563ms) -Jun 4 19:34:51.788: INFO: (16) /api/v1/namespaces/proxy-3796/pods/proxy-service-fwvr6-47f9b/proxy/: test (200; 21.029299ms) -Jun 4 19:34:51.788: INFO: (16) /api/v1/namespaces/proxy-3796/pods/proxy-service-fwvr6-47f9b:162/proxy/: bar (200; 20.8028ms) -Jun 4 19:34:51.788: INFO: (16) /api/v1/namespaces/proxy-3796/pods/http:proxy-service-fwvr6-47f9b:1080/proxy/: ... (200; 20.731705ms) -Jun 4 19:34:51.788: INFO: (16) /api/v1/namespaces/proxy-3796/pods/proxy-service-fwvr6-47f9b:160/proxy/: foo (200; 21.422725ms) -Jun 4 19:34:51.788: INFO: (16) /api/v1/namespaces/proxy-3796/pods/https:proxy-service-fwvr6-47f9b:443/proxy/: test<... (200; 21.17955ms) -Jun 4 19:34:51.789: INFO: (16) /api/v1/namespaces/proxy-3796/pods/http:proxy-service-fwvr6-47f9b:162/proxy/: bar (200; 20.668211ms) -Jun 4 19:34:51.793: INFO: (16) /api/v1/namespaces/proxy-3796/services/proxy-service-fwvr6:portname2/proxy/: bar (200; 25.35358ms) -Jun 4 19:34:51.793: INFO: (16) /api/v1/namespaces/proxy-3796/services/http:proxy-service-fwvr6:portname1/proxy/: foo (200; 25.957091ms) -Jun 4 19:34:51.798: INFO: (16) /api/v1/namespaces/proxy-3796/services/proxy-service-fwvr6:portname1/proxy/: foo (200; 30.239423ms) -Jun 4 19:34:51.798: INFO: (16) /api/v1/namespaces/proxy-3796/services/http:proxy-service-fwvr6:portname2/proxy/: bar (200; 30.666811ms) -Jun 4 19:34:51.798: INFO: (16) /api/v1/namespaces/proxy-3796/services/https:proxy-service-fwvr6:tlsportname1/proxy/: tls baz (200; 30.410959ms) -Jun 4 19:34:51.798: INFO: (16) /api/v1/namespaces/proxy-3796/services/https:proxy-service-fwvr6:tlsportname2/proxy/: tls qux (200; 30.367855ms) -Jun 4 19:34:51.816: INFO: (17) /api/v1/namespaces/proxy-3796/pods/proxy-service-fwvr6-47f9b:162/proxy/: bar (200; 17.281972ms) -Jun 4 19:34:51.816: INFO: (17) /api/v1/namespaces/proxy-3796/pods/https:proxy-service-fwvr6-47f9b:462/proxy/: tls qux (200; 17.636367ms) -Jun 4 19:34:51.816: INFO: (17) /api/v1/namespaces/proxy-3796/pods/http:proxy-service-fwvr6-47f9b:162/proxy/: bar (200; 17.77855ms) -Jun 4 19:34:51.816: INFO: (17) /api/v1/namespaces/proxy-3796/pods/proxy-service-fwvr6-47f9b:160/proxy/: foo (200; 17.879458ms) -Jun 4 19:34:51.816: INFO: (17) /api/v1/namespaces/proxy-3796/pods/https:proxy-service-fwvr6-47f9b:443/proxy/: ... (200; 17.87351ms) -Jun 4 19:34:51.816: INFO: (17) /api/v1/namespaces/proxy-3796/pods/proxy-service-fwvr6-47f9b/proxy/: test (200; 17.79474ms) -Jun 4 19:34:51.816: INFO: (17) /api/v1/namespaces/proxy-3796/pods/proxy-service-fwvr6-47f9b:1080/proxy/: test<... (200; 18.093135ms) -Jun 4 19:34:51.816: INFO: (17) /api/v1/namespaces/proxy-3796/pods/http:proxy-service-fwvr6-47f9b:160/proxy/: foo (200; 18.103703ms) -Jun 4 19:34:51.817: INFO: (17) /api/v1/namespaces/proxy-3796/pods/https:proxy-service-fwvr6-47f9b:460/proxy/: tls baz (200; 18.133197ms) -Jun 4 19:34:51.824: INFO: (17) /api/v1/namespaces/proxy-3796/services/proxy-service-fwvr6:portname1/proxy/: foo (200; 25.555291ms) -Jun 4 19:34:51.828: INFO: (17) /api/v1/namespaces/proxy-3796/services/proxy-service-fwvr6:portname2/proxy/: bar (200; 29.992565ms) -Jun 4 19:34:51.829: INFO: (17) /api/v1/namespaces/proxy-3796/services/https:proxy-service-fwvr6:tlsportname2/proxy/: tls qux (200; 30.294376ms) -Jun 4 19:34:51.829: INFO: (17) /api/v1/namespaces/proxy-3796/services/http:proxy-service-fwvr6:portname2/proxy/: bar (200; 30.469306ms) -Jun 4 19:34:51.829: INFO: (17) /api/v1/namespaces/proxy-3796/services/https:proxy-service-fwvr6:tlsportname1/proxy/: tls baz (200; 30.359146ms) -Jun 4 19:34:51.829: INFO: (17) /api/v1/namespaces/proxy-3796/services/http:proxy-service-fwvr6:portname1/proxy/: foo (200; 30.284293ms) -Jun 4 19:34:51.852: INFO: (18) /api/v1/namespaces/proxy-3796/pods/https:proxy-service-fwvr6-47f9b:443/proxy/: test (200; 23.113042ms) -Jun 4 19:34:51.852: INFO: (18) /api/v1/namespaces/proxy-3796/pods/proxy-service-fwvr6-47f9b:162/proxy/: bar (200; 22.917863ms) -Jun 4 19:34:51.852: INFO: (18) /api/v1/namespaces/proxy-3796/pods/http:proxy-service-fwvr6-47f9b:162/proxy/: bar (200; 22.888877ms) -Jun 4 19:34:51.852: INFO: (18) /api/v1/namespaces/proxy-3796/pods/http:proxy-service-fwvr6-47f9b:1080/proxy/: ... (200; 23.645645ms) -Jun 4 19:34:51.853: INFO: (18) /api/v1/namespaces/proxy-3796/pods/proxy-service-fwvr6-47f9b:1080/proxy/: test<... (200; 23.360762ms) -Jun 4 19:34:51.853: INFO: (18) /api/v1/namespaces/proxy-3796/pods/http:proxy-service-fwvr6-47f9b:160/proxy/: foo (200; 23.332719ms) -Jun 4 19:34:51.853: INFO: (18) /api/v1/namespaces/proxy-3796/pods/proxy-service-fwvr6-47f9b:160/proxy/: foo (200; 23.955068ms) -Jun 4 19:34:51.853: INFO: (18) /api/v1/namespaces/proxy-3796/pods/https:proxy-service-fwvr6-47f9b:460/proxy/: tls baz (200; 24.161336ms) -Jun 4 19:34:51.856: INFO: (18) /api/v1/namespaces/proxy-3796/services/https:proxy-service-fwvr6:tlsportname2/proxy/: tls qux (200; 25.930955ms) -Jun 4 19:34:51.860: INFO: (18) /api/v1/namespaces/proxy-3796/services/https:proxy-service-fwvr6:tlsportname1/proxy/: tls baz (200; 30.386216ms) -Jun 4 19:34:51.860: INFO: (18) /api/v1/namespaces/proxy-3796/services/proxy-service-fwvr6:portname2/proxy/: bar (200; 30.932043ms) -Jun 4 19:34:51.860: INFO: (18) /api/v1/namespaces/proxy-3796/services/http:proxy-service-fwvr6:portname2/proxy/: bar (200; 30.606067ms) -Jun 4 19:34:51.860: INFO: (18) /api/v1/namespaces/proxy-3796/services/proxy-service-fwvr6:portname1/proxy/: foo (200; 30.527618ms) -Jun 4 19:34:51.860: INFO: (18) /api/v1/namespaces/proxy-3796/services/http:proxy-service-fwvr6:portname1/proxy/: foo (200; 31.118631ms) -Jun 4 19:34:51.883: INFO: (19) /api/v1/namespaces/proxy-3796/pods/proxy-service-fwvr6-47f9b:162/proxy/: bar (200; 22.403198ms) -Jun 4 19:34:51.883: INFO: (19) /api/v1/namespaces/proxy-3796/pods/http:proxy-service-fwvr6-47f9b:160/proxy/: foo (200; 22.38378ms) -Jun 4 19:34:51.883: INFO: (19) /api/v1/namespaces/proxy-3796/pods/proxy-service-fwvr6-47f9b:160/proxy/: foo (200; 22.73085ms) -Jun 4 19:34:51.883: INFO: (19) /api/v1/namespaces/proxy-3796/pods/https:proxy-service-fwvr6-47f9b:443/proxy/: test<... (200; 22.781097ms) -Jun 4 19:34:51.883: INFO: (19) /api/v1/namespaces/proxy-3796/pods/http:proxy-service-fwvr6-47f9b:162/proxy/: bar (200; 22.599266ms) -Jun 4 19:34:51.883: INFO: (19) /api/v1/namespaces/proxy-3796/pods/proxy-service-fwvr6-47f9b/proxy/: test (200; 22.650025ms) -Jun 4 19:34:51.883: INFO: (19) /api/v1/namespaces/proxy-3796/pods/http:proxy-service-fwvr6-47f9b:1080/proxy/: ... (200; 22.649125ms) -Jun 4 19:34:51.883: INFO: (19) /api/v1/namespaces/proxy-3796/pods/https:proxy-service-fwvr6-47f9b:462/proxy/: tls qux (200; 22.8307ms) -Jun 4 19:34:51.883: INFO: (19) /api/v1/namespaces/proxy-3796/pods/https:proxy-service-fwvr6-47f9b:460/proxy/: tls baz (200; 22.877341ms) -Jun 4 19:34:51.887: INFO: (19) /api/v1/namespaces/proxy-3796/services/https:proxy-service-fwvr6:tlsportname2/proxy/: tls qux (200; 26.100455ms) -Jun 4 19:34:51.890: INFO: (19) /api/v1/namespaces/proxy-3796/services/http:proxy-service-fwvr6:portname2/proxy/: bar (200; 29.782372ms) -Jun 4 19:34:51.891: INFO: (19) /api/v1/namespaces/proxy-3796/services/proxy-service-fwvr6:portname2/proxy/: bar (200; 29.878309ms) -Jun 4 19:34:51.891: INFO: (19) /api/v1/namespaces/proxy-3796/services/http:proxy-service-fwvr6:portname1/proxy/: foo (200; 30.369699ms) -Jun 4 19:34:51.891: INFO: (19) /api/v1/namespaces/proxy-3796/services/proxy-service-fwvr6:portname1/proxy/: foo (200; 30.556753ms) -Jun 4 19:34:51.891: INFO: (19) /api/v1/namespaces/proxy-3796/services/https:proxy-service-fwvr6:tlsportname1/proxy/: tls baz (200; 30.89121ms) -STEP: deleting ReplicationController proxy-service-fwvr6 in namespace proxy-3796, will wait for the garbage collector to delete the pods -Jun 4 19:34:51.981: INFO: Deleting ReplicationController proxy-service-fwvr6 took: 28.791068ms -Jun 4 19:34:52.082: INFO: Terminating ReplicationController proxy-service-fwvr6 pods took: 100.87046ms -[AfterEach] version v1 - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 -Jun 4 19:35:02.582: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "proxy-3796" for this suite. -Jun 4 19:35:10.652: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -Jun 4 19:35:10.993: INFO: namespace proxy-3796 deletion completed in 8.393410314s - -• [SLOW TEST:34.094 seconds] -[sig-network] Proxy -/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:22 - version v1 - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/proxy.go:56 - should proxy through a service and a pod [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 ------------------------------- -SSSSSSSSSSSS ------------------------------- -[k8s.io] Pods - should be updated [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -[BeforeEach] [k8s.io] Pods - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 -STEP: Creating a kubernetes client -Jun 4 19:35:10.993: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504 -STEP: Building a namespace api object, basename pods -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in pods-2718 -STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [k8s.io] Pods - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:135 -[It] should be updated [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -STEP: creating the pod -STEP: submitting the pod to kubernetes -STEP: verifying the pod is in kubernetes -STEP: updating the pod -Jun 4 19:35:13.855: INFO: Successfully updated pod "pod-update-de143db4-86ff-11e9-8d1b-467ee19922ac" -STEP: verifying the updated pod is in kubernetes -Jun 4 19:35:13.884: INFO: Pod update OK -[AfterEach] [k8s.io] Pods - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 -Jun 4 19:35:13.884: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "pods-2718" for this suite. -Jun 4 19:35:37.951: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -Jun 4 19:35:38.469: INFO: namespace pods-2718 deletion completed in 24.566753606s - -• [SLOW TEST:27.476 seconds] -[k8s.io] Pods -/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687 - should be updated [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 ------------------------------- -SSS ------------------------------- -[sig-storage] Subpath Atomic writer volumes - should support subpaths with downward pod [LinuxOnly] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -[BeforeEach] [sig-storage] Subpath - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 -STEP: Creating a kubernetes client -Jun 4 19:35:38.469: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504 -STEP: Building a namespace api object, basename subpath -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in subpath-3420 +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in downward-api-919 STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] Atomic writer volumes - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:38 -STEP: Setting up data -[It] should support subpaths with downward pod [LinuxOnly] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -STEP: Creating pod pod-subpath-test-downwardapi-gqmv -STEP: Creating a pod to test atomic-volume-subpath -Jun 4 19:35:38.805: INFO: Waiting up to 5m0s for pod "pod-subpath-test-downwardapi-gqmv" in namespace "subpath-3420" to be "success or failure" -Jun 4 19:35:38.817: INFO: Pod "pod-subpath-test-downwardapi-gqmv": Phase="Pending", Reason="", readiness=false. Elapsed: 12.029893ms -Jun 4 19:35:40.831: INFO: Pod "pod-subpath-test-downwardapi-gqmv": Phase="Running", Reason="", readiness=true. Elapsed: 2.025707687s -Jun 4 19:35:42.843: INFO: Pod "pod-subpath-test-downwardapi-gqmv": Phase="Running", Reason="", readiness=true. Elapsed: 4.038512545s -Jun 4 19:35:44.856: INFO: Pod "pod-subpath-test-downwardapi-gqmv": Phase="Running", Reason="", readiness=true. Elapsed: 6.051045808s -Jun 4 19:35:46.869: INFO: Pod "pod-subpath-test-downwardapi-gqmv": Phase="Running", Reason="", readiness=true. Elapsed: 8.063904749s -Jun 4 19:35:49.037: INFO: Pod "pod-subpath-test-downwardapi-gqmv": Phase="Running", Reason="", readiness=true. Elapsed: 10.232474564s -Jun 4 19:35:51.050: INFO: Pod "pod-subpath-test-downwardapi-gqmv": Phase="Running", Reason="", readiness=true. Elapsed: 12.24565155s -Jun 4 19:35:53.064: INFO: Pod "pod-subpath-test-downwardapi-gqmv": Phase="Running", Reason="", readiness=true. Elapsed: 14.25879908s -Jun 4 19:35:55.077: INFO: Pod "pod-subpath-test-downwardapi-gqmv": Phase="Running", Reason="", readiness=true. Elapsed: 16.272051015s -Jun 4 19:35:57.089: INFO: Pod "pod-subpath-test-downwardapi-gqmv": Phase="Running", Reason="", readiness=true. Elapsed: 18.284650221s -Jun 4 19:35:59.102: INFO: Pod "pod-subpath-test-downwardapi-gqmv": Phase="Running", Reason="", readiness=true. Elapsed: 20.297525721s -Jun 4 19:36:01.117: INFO: Pod "pod-subpath-test-downwardapi-gqmv": Phase="Succeeded", Reason="", readiness=false. Elapsed: 22.312158654s +[It] should provide container's limits.cpu/memory and requests.cpu/memory as env vars [NodeConformance] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: Creating a pod to test downward api env vars +Jun 18 12:47:55.938: INFO: Waiting up to 5m0s for pod "downward-api-4b48f846-91c7-11e9-a25d-8608290c688a" in namespace "downward-api-919" to be "success or failure" +Jun 18 12:47:55.953: INFO: Pod "downward-api-4b48f846-91c7-11e9-a25d-8608290c688a": Phase="Pending", Reason="", readiness=false. Elapsed: 15.201302ms +Jun 18 12:47:57.968: INFO: Pod "downward-api-4b48f846-91c7-11e9-a25d-8608290c688a": Phase="Running", Reason="", readiness=true. Elapsed: 2.029515189s +Jun 18 12:47:59.981: INFO: Pod "downward-api-4b48f846-91c7-11e9-a25d-8608290c688a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.042966858s STEP: Saw pod success -Jun 4 19:36:01.117: INFO: Pod "pod-subpath-test-downwardapi-gqmv" satisfied condition "success or failure" -Jun 4 19:36:01.127: INFO: Trying to get logs from node 10.212.23.164 pod pod-subpath-test-downwardapi-gqmv container test-container-subpath-downwardapi-gqmv: -STEP: delete the pod -Jun 4 19:36:01.216: INFO: Waiting for pod pod-subpath-test-downwardapi-gqmv to disappear -Jun 4 19:36:01.232: INFO: Pod pod-subpath-test-downwardapi-gqmv no longer exists -STEP: Deleting pod pod-subpath-test-downwardapi-gqmv -Jun 4 19:36:01.232: INFO: Deleting pod "pod-subpath-test-downwardapi-gqmv" in namespace "subpath-3420" -[AfterEach] [sig-storage] Subpath - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 -Jun 4 19:36:01.250: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "subpath-3420" for this suite. -Jun 4 19:36:07.344: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -Jun 4 19:36:07.687: INFO: namespace subpath-3420 deletion completed in 6.405123297s - -• [SLOW TEST:29.218 seconds] -[sig-storage] Subpath -/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:22 - Atomic writer volumes - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:34 - should support subpaths with downward pod [LinuxOnly] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 ------------------------------- -SSSSSSSSSSSSS ------------------------------- -[sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] - should perform canary updates and phased rolling updates of template modifications [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -[BeforeEach] [sig-apps] StatefulSet - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 -STEP: Creating a kubernetes client -Jun 4 19:36:07.689: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504 -STEP: Building a namespace api object, basename statefulset -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in statefulset-248 -STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [sig-apps] StatefulSet - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:59 -[BeforeEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:74 -STEP: Creating service test in namespace statefulset-248 -[It] should perform canary updates and phased rolling updates of template modifications [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -STEP: Creating a new StatefulSet -Jun 4 19:36:07.940: INFO: Found 0 stateful pods, waiting for 3 -Jun 4 19:36:17.955: INFO: Waiting for pod ss2-0 to enter Running - Ready=true, currently Running - Ready=true -Jun 4 19:36:17.955: INFO: Waiting for pod ss2-1 to enter Running - Ready=true, currently Running - Ready=true -Jun 4 19:36:17.955: INFO: Waiting for pod ss2-2 to enter Running - Ready=true, currently Running - Ready=true -STEP: Updating stateful set template: update image from docker.io/library/nginx:1.14-alpine to docker.io/library/nginx:1.15-alpine -Jun 4 19:36:18.065: INFO: Updating stateful set ss2 -STEP: Creating a new revision -STEP: Not applying an update when the partition is greater than the number of replicas -STEP: Performing a canary update -Jun 4 19:36:28.138: INFO: Updating stateful set ss2 -Jun 4 19:36:28.159: INFO: Waiting for Pod statefulset-248/ss2-2 to have revision ss2-c79899b9 update revision ss2-787997d666 -STEP: Restoring Pods to the correct revision when they are deleted -Jun 4 19:36:38.272: INFO: Found 2 stateful pods, waiting for 3 -Jun 4 19:36:48.291: INFO: Waiting for pod ss2-0 to enter Running - Ready=true, currently Running - Ready=true -Jun 4 19:36:48.291: INFO: Waiting for pod ss2-1 to enter Running - Ready=true, currently Running - Ready=true -Jun 4 19:36:48.291: INFO: Waiting for pod ss2-2 to enter Running - Ready=true, currently Running - Ready=true -STEP: Performing a phased rolling update -Jun 4 19:36:48.365: INFO: Updating stateful set ss2 -Jun 4 19:36:48.393: INFO: Waiting for Pod statefulset-248/ss2-1 to have revision ss2-c79899b9 update revision ss2-787997d666 -Jun 4 19:36:58.450: INFO: Updating stateful set ss2 -Jun 4 19:36:58.470: INFO: Waiting for StatefulSet statefulset-248/ss2 to complete update -Jun 4 19:36:58.470: INFO: Waiting for Pod statefulset-248/ss2-0 to have revision ss2-c79899b9 update revision ss2-787997d666 -Jun 4 19:37:08.556: INFO: Waiting for StatefulSet statefulset-248/ss2 to complete update -[AfterEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:85 -Jun 4 19:37:18.494: INFO: Deleting all statefulset in ns statefulset-248 -Jun 4 19:37:18.503: INFO: Scaling statefulset ss2 to 0 -Jun 4 19:37:38.546: INFO: Waiting for statefulset status.replicas updated to 0 -Jun 4 19:37:38.556: INFO: Deleting statefulset ss2 -[AfterEach] [sig-apps] StatefulSet - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 -Jun 4 19:37:38.596: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "statefulset-248" for this suite. -Jun 4 19:37:46.653: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -Jun 4 19:37:47.080: INFO: namespace statefulset-248 deletion completed in 8.468513422s +Jun 18 12:47:59.981: INFO: Pod "downward-api-4b48f846-91c7-11e9-a25d-8608290c688a" satisfied condition "success or failure" +Jun 18 12:47:59.998: INFO: Trying to get logs from node 10.72.74.138 pod downward-api-4b48f846-91c7-11e9-a25d-8608290c688a container dapi-container: +STEP: delete the pod +Jun 18 12:48:00.088: INFO: Waiting for pod downward-api-4b48f846-91c7-11e9-a25d-8608290c688a to disappear +Jun 18 12:48:00.107: INFO: Pod downward-api-4b48f846-91c7-11e9-a25d-8608290c688a no longer exists +[AfterEach] [sig-node] Downward API + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Jun 18 12:48:00.107: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "downward-api-919" for this suite. +Jun 18 12:48:06.145: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:48:06.531: INFO: namespace downward-api-919 deletion completed in 6.4110965s -• [SLOW TEST:99.391 seconds] -[sig-apps] StatefulSet -/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22 - [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687 - should perform canary updates and phased rolling updates of template modifications [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +• [SLOW TEST:10.825 seconds] +[sig-node] Downward API +/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downward_api.go:38 + should provide container's limits.cpu/memory and requests.cpu/memory as env vars [NodeConformance] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 ------------------------------ -SS +SSSSSSSSSSSSS ------------------------------ [k8s.io] Docker Containers - should use the image defaults if command and args are blank [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 + should be able to override the image's default arguments (docker cmd) [NodeConformance] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 [BeforeEach] [k8s.io] Docker Containers - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 STEP: Creating a kubernetes client -Jun 4 19:37:47.081: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504 +Jun 18 12:48:06.532: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150 STEP: Building a namespace api object, basename containers -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in containers-4545 +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in containers-5149 STEP: Waiting for a default service account to be provisioned in namespace -[It] should use the image defaults if command and args are blank [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -STEP: Creating a pod to test use defaults -Jun 4 19:37:47.319: INFO: Waiting up to 5m0s for pod "client-containers-3b1c6e11-8700-11e9-8d1b-467ee19922ac" in namespace "containers-4545" to be "success or failure" -Jun 4 19:37:47.331: INFO: Pod "client-containers-3b1c6e11-8700-11e9-8d1b-467ee19922ac": Phase="Pending", Reason="", readiness=false. Elapsed: 11.724276ms -Jun 4 19:37:49.368: INFO: Pod "client-containers-3b1c6e11-8700-11e9-8d1b-467ee19922ac": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.048427807s +[It] should be able to override the image's default arguments (docker cmd) [NodeConformance] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: Creating a pod to test override arguments +Jun 18 12:48:06.757: INFO: Waiting up to 5m0s for pod "client-containers-51bc31f7-91c7-11e9-a25d-8608290c688a" in namespace "containers-5149" to be "success or failure" +Jun 18 12:48:06.783: INFO: Pod "client-containers-51bc31f7-91c7-11e9-a25d-8608290c688a": Phase="Pending", Reason="", readiness=false. Elapsed: 25.824101ms +Jun 18 12:48:08.798: INFO: Pod "client-containers-51bc31f7-91c7-11e9-a25d-8608290c688a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.040758473s STEP: Saw pod success -Jun 4 19:37:49.368: INFO: Pod "client-containers-3b1c6e11-8700-11e9-8d1b-467ee19922ac" satisfied condition "success or failure" -Jun 4 19:37:49.379: INFO: Trying to get logs from node 10.212.23.164 pod client-containers-3b1c6e11-8700-11e9-8d1b-467ee19922ac container test-container: +Jun 18 12:48:08.798: INFO: Pod "client-containers-51bc31f7-91c7-11e9-a25d-8608290c688a" satisfied condition "success or failure" +Jun 18 12:48:08.816: INFO: Trying to get logs from node 10.72.74.184 pod client-containers-51bc31f7-91c7-11e9-a25d-8608290c688a container test-container: STEP: delete the pod -Jun 4 19:37:49.446: INFO: Waiting for pod client-containers-3b1c6e11-8700-11e9-8d1b-467ee19922ac to disappear -Jun 4 19:37:49.459: INFO: Pod client-containers-3b1c6e11-8700-11e9-8d1b-467ee19922ac no longer exists +Jun 18 12:48:08.901: INFO: Waiting for pod client-containers-51bc31f7-91c7-11e9-a25d-8608290c688a to disappear +Jun 18 12:48:08.914: INFO: Pod client-containers-51bc31f7-91c7-11e9-a25d-8608290c688a no longer exists [AfterEach] [k8s.io] Docker Containers - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 -Jun 4 19:37:49.459: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "containers-4545" for this suite. -Jun 4 19:37:55.519: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -Jun 4 19:37:55.883: INFO: namespace containers-4545 deletion completed in 6.407243026s + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Jun 18 12:48:08.914: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "containers-5149" for this suite. +Jun 18 12:48:16.957: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:48:17.339: INFO: namespace containers-5149 deletion completed in 8.408062618s -• [SLOW TEST:8.802 seconds] +• [SLOW TEST:10.807 seconds] [k8s.io] Docker Containers -/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687 - should use the image defaults if command and args are blank [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 ------------------------------- -SSSSS ------------------------------- -[sig-storage] Projected configMap - should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -[BeforeEach] [sig-storage] Projected configMap - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 -STEP: Creating a kubernetes client -Jun 4 19:37:55.883: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504 -STEP: Building a namespace api object, basename projected -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-5462 -STEP: Waiting for a default service account to be provisioned in namespace -[It] should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -STEP: Creating configMap with name projected-configmap-test-volume-map-405b5c78-8700-11e9-8d1b-467ee19922ac -STEP: Creating a pod to test consume configMaps -Jun 4 19:37:56.195: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-4066ce88-8700-11e9-8d1b-467ee19922ac" in namespace "projected-5462" to be "success or failure" -Jun 4 19:37:56.211: INFO: Pod "pod-projected-configmaps-4066ce88-8700-11e9-8d1b-467ee19922ac": Phase="Pending", Reason="", readiness=false. Elapsed: 16.229163ms -Jun 4 19:37:58.224: INFO: Pod "pod-projected-configmaps-4066ce88-8700-11e9-8d1b-467ee19922ac": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.029014824s -STEP: Saw pod success -Jun 4 19:37:58.224: INFO: Pod "pod-projected-configmaps-4066ce88-8700-11e9-8d1b-467ee19922ac" satisfied condition "success or failure" -Jun 4 19:37:58.242: INFO: Trying to get logs from node 10.212.23.189 pod pod-projected-configmaps-4066ce88-8700-11e9-8d1b-467ee19922ac container projected-configmap-volume-test: -STEP: delete the pod -Jun 4 19:37:58.306: INFO: Waiting for pod pod-projected-configmaps-4066ce88-8700-11e9-8d1b-467ee19922ac to disappear -Jun 4 19:37:58.317: INFO: Pod pod-projected-configmaps-4066ce88-8700-11e9-8d1b-467ee19922ac no longer exists -[AfterEach] [sig-storage] Projected configMap - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 -Jun 4 19:37:58.317: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "projected-5462" for this suite. -Jun 4 19:38:04.378: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -Jun 4 19:38:04.712: INFO: namespace projected-5462 deletion completed in 6.379395121s - -• [SLOW TEST:8.829 seconds] -[sig-storage] Projected configMap -/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_configmap.go:33 - should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687 + should be able to override the image's default arguments (docker cmd) [NodeConformance] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 ------------------------------ -SSSSSSSSSSSSSSS +SSSSSSSSSS ------------------------------ -[sig-storage] Projected configMap - should be consumable from pods in volume with mappings as non-root [LinuxOnly] [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -[BeforeEach] [sig-storage] Projected configMap - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +[k8s.io] Probing container + with readiness probe should not be ready before initial delay and never restart [NodeConformance] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [k8s.io] Probing container + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 STEP: Creating a kubernetes client -Jun 4 19:38:04.713: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504 -STEP: Building a namespace api object, basename projected -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-6411 +Jun 18 12:48:17.339: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150 +STEP: Building a namespace api object, basename container-probe +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in container-probe-450 STEP: Waiting for a default service account to be provisioned in namespace -[It] should be consumable from pods in volume with mappings as non-root [LinuxOnly] [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -STEP: Creating configMap with name projected-configmap-test-volume-map-459ebda9-8700-11e9-8d1b-467ee19922ac -STEP: Creating a pod to test consume configMaps -Jun 4 19:38:04.965: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-45a13470-8700-11e9-8d1b-467ee19922ac" in namespace "projected-6411" to be "success or failure" -Jun 4 19:38:04.980: INFO: Pod "pod-projected-configmaps-45a13470-8700-11e9-8d1b-467ee19922ac": Phase="Pending", Reason="", readiness=false. Elapsed: 14.917951ms -Jun 4 19:38:06.992: INFO: Pod "pod-projected-configmaps-45a13470-8700-11e9-8d1b-467ee19922ac": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.02696129s -STEP: Saw pod success -Jun 4 19:38:06.992: INFO: Pod "pod-projected-configmaps-45a13470-8700-11e9-8d1b-467ee19922ac" satisfied condition "success or failure" -Jun 4 19:38:07.055: INFO: Trying to get logs from node 10.212.23.161 pod pod-projected-configmaps-45a13470-8700-11e9-8d1b-467ee19922ac container projected-configmap-volume-test: -STEP: delete the pod -Jun 4 19:38:07.124: INFO: Waiting for pod pod-projected-configmaps-45a13470-8700-11e9-8d1b-467ee19922ac to disappear -Jun 4 19:38:07.135: INFO: Pod pod-projected-configmaps-45a13470-8700-11e9-8d1b-467ee19922ac no longer exists -[AfterEach] [sig-storage] Projected configMap - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 -Jun 4 19:38:07.135: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "projected-6411" for this suite. -Jun 4 19:38:13.197: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -Jun 4 19:38:13.562: INFO: namespace projected-6411 deletion completed in 6.411057637s +[BeforeEach] [k8s.io] Probing container + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/container_probe.go:51 +[It] with readiness probe should not be ready before initial delay and never restart [NodeConformance] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +Jun 18 12:48:41.600: INFO: Container started at 2019-06-18 12:48:18 +0000 UTC, pod became ready at 2019-06-18 12:48:39 +0000 UTC +[AfterEach] [k8s.io] Probing container + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Jun 18 12:48:41.600: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "container-probe-450" for this suite. +Jun 18 12:49:05.646: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:49:06.035: INFO: namespace container-probe-450 deletion completed in 24.415906874s -• [SLOW TEST:8.850 seconds] -[sig-storage] Projected configMap -/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_configmap.go:33 - should be consumable from pods in volume with mappings as non-root [LinuxOnly] [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +• [SLOW TEST:48.696 seconds] +[k8s.io] Probing container +/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687 + with readiness probe should not be ready before initial delay and never restart [NodeConformance] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 ------------------------------ -SSSSSS +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [k8s.io] Probing container should *not* be restarted with a /healthz http liveness probe [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 [BeforeEach] [k8s.io] Probing container - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 STEP: Creating a kubernetes client -Jun 4 19:38:13.563: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504 +Jun 18 12:49:06.036: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150 STEP: Building a namespace api object, basename container-probe -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in container-probe-588 +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in container-probe-6793 STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Probing container - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/container_probe.go:51 + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/container_probe.go:51 [It] should *not* be restarted with a /healthz http liveness probe [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -STEP: Creating pod liveness-http in namespace container-probe-588 -Jun 4 19:38:15.825: INFO: Started pod liveness-http in namespace container-probe-588 + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: Creating pod liveness-http in namespace container-probe-6793 +Jun 18 12:49:10.497: INFO: Started pod liveness-http in namespace container-probe-6793 STEP: checking the pod's current state and verifying that restartCount is present -Jun 4 19:38:15.840: INFO: Initial restart count of pod liveness-http is 0 +Jun 18 12:49:10.511: INFO: Initial restart count of pod liveness-http is 0 STEP: deleting the pod [AfterEach] [k8s.io] Probing container - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 -Jun 4 19:42:17.856: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "container-probe-588" for this suite. -Jun 4 19:42:23.914: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -Jun 4 19:42:24.406: INFO: namespace container-probe-588 deletion completed in 6.535173042s + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Jun 18 12:53:11.712: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "container-probe-6793" for this suite. +Jun 18 12:53:17.758: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:53:18.189: INFO: namespace container-probe-6793 deletion completed in 6.462697058s -• [SLOW TEST:250.843 seconds] +• [SLOW TEST:252.153 seconds] [k8s.io] Probing container -/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687 +/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687 should *not* be restarted with a /healthz http liveness probe [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 ------------------------------ -SSSSS +SSSSSSSSSS ------------------------------ -[sig-storage] Projected downwardAPI - should provide container's cpu limit [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -[BeforeEach] [sig-storage] Projected downwardAPI - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +[sig-storage] Projected secret + should be consumable from pods in volume with mappings [NodeConformance] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [sig-storage] Projected secret + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 STEP: Creating a kubernetes client -Jun 4 19:42:24.406: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504 +Jun 18 12:53:18.191: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150 STEP: Building a namespace api object, basename projected -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-8992 +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-1412 STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [sig-storage] Projected downwardAPI - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:39 -[It] should provide container's cpu limit [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -STEP: Creating a pod to test downward API volume plugin -Jun 4 19:42:24.682: INFO: Waiting up to 5m0s for pod "downwardapi-volume-e06e8835-8700-11e9-8d1b-467ee19922ac" in namespace "projected-8992" to be "success or failure" -Jun 4 19:42:24.694: INFO: Pod "downwardapi-volume-e06e8835-8700-11e9-8d1b-467ee19922ac": Phase="Pending", Reason="", readiness=false. Elapsed: 12.334718ms -Jun 4 19:42:26.705: INFO: Pod "downwardapi-volume-e06e8835-8700-11e9-8d1b-467ee19922ac": Phase="Pending", Reason="", readiness=false. Elapsed: 2.022976309s -Jun 4 19:42:28.719: INFO: Pod "downwardapi-volume-e06e8835-8700-11e9-8d1b-467ee19922ac": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.037041226s +[It] should be consumable from pods in volume with mappings [NodeConformance] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: Creating projection with secret that has name projected-secret-test-map-0b81100b-91c8-11e9-a25d-8608290c688a +STEP: Creating a pod to test consume secrets +Jun 18 12:53:18.441: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-0b82f2bb-91c8-11e9-a25d-8608290c688a" in namespace "projected-1412" to be "success or failure" +Jun 18 12:53:18.461: INFO: Pod "pod-projected-secrets-0b82f2bb-91c8-11e9-a25d-8608290c688a": Phase="Pending", Reason="", readiness=false. Elapsed: 19.58402ms +Jun 18 12:53:20.475: INFO: Pod "pod-projected-secrets-0b82f2bb-91c8-11e9-a25d-8608290c688a": Phase="Running", Reason="", readiness=true. Elapsed: 2.033622007s +Jun 18 12:53:22.493: INFO: Pod "pod-projected-secrets-0b82f2bb-91c8-11e9-a25d-8608290c688a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.051719242s STEP: Saw pod success -Jun 4 19:42:28.719: INFO: Pod "downwardapi-volume-e06e8835-8700-11e9-8d1b-467ee19922ac" satisfied condition "success or failure" -Jun 4 19:42:28.731: INFO: Trying to get logs from node 10.212.23.164 pod downwardapi-volume-e06e8835-8700-11e9-8d1b-467ee19922ac container client-container: +Jun 18 12:53:22.493: INFO: Pod "pod-projected-secrets-0b82f2bb-91c8-11e9-a25d-8608290c688a" satisfied condition "success or failure" +Jun 18 12:53:22.506: INFO: Trying to get logs from node 10.72.74.138 pod pod-projected-secrets-0b82f2bb-91c8-11e9-a25d-8608290c688a container projected-secret-volume-test: STEP: delete the pod -Jun 4 19:42:28.795: INFO: Waiting for pod downwardapi-volume-e06e8835-8700-11e9-8d1b-467ee19922ac to disappear -Jun 4 19:42:28.805: INFO: Pod downwardapi-volume-e06e8835-8700-11e9-8d1b-467ee19922ac no longer exists -[AfterEach] [sig-storage] Projected downwardAPI - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 -Jun 4 19:42:28.805: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "projected-8992" for this suite. -Jun 4 19:42:34.867: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -Jun 4 19:42:35.209: INFO: namespace projected-8992 deletion completed in 6.388101602s +Jun 18 12:53:22.591: INFO: Waiting for pod pod-projected-secrets-0b82f2bb-91c8-11e9-a25d-8608290c688a to disappear +Jun 18 12:53:22.603: INFO: Pod pod-projected-secrets-0b82f2bb-91c8-11e9-a25d-8608290c688a no longer exists +[AfterEach] [sig-storage] Projected secret + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Jun 18 12:53:22.603: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "projected-1412" for this suite. +Jun 18 12:53:28.643: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:53:29.027: INFO: namespace projected-1412 deletion completed in 6.4115535s -• [SLOW TEST:10.803 seconds] -[sig-storage] Projected downwardAPI -/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:33 - should provide container's cpu limit [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +• [SLOW TEST:10.836 seconds] +[sig-storage] Projected secret +/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_secret.go:33 + should be consumable from pods in volume with mappings [NodeConformance] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +------------------------------ +SSSSSSSSSSSSSSSSSSSS +------------------------------ +[k8s.io] [sig-node] PreStop + should call prestop when killing a pod [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [k8s.io] [sig-node] PreStop + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +STEP: Creating a kubernetes client +Jun 18 12:53:29.028: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150 +STEP: Building a namespace api object, basename prestop +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in prestop-142 +STEP: Waiting for a default service account to be provisioned in namespace +[BeforeEach] [k8s.io] [sig-node] PreStop + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/node/pre_stop.go:167 +[It] should call prestop when killing a pod [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: Creating server pod server in namespace prestop-142 +STEP: Waiting for pods to come up. +STEP: Creating tester pod tester in namespace prestop-142 +STEP: Deleting pre-stop pod +Jun 18 12:53:42.392: INFO: Saw: { + "Hostname": "server", + "Sent": null, + "Received": { + "prestop": 1 + }, + "Errors": null, + "Log": [ + "default/nettest has 0 endpoints ([]), which is less than 8 as expected. Waiting for all endpoints to come up.", + "default/nettest has 0 endpoints ([]), which is less than 8 as expected. Waiting for all endpoints to come up." + ], + "StillContactingPeers": true +} +STEP: Deleting the server pod +[AfterEach] [k8s.io] [sig-node] PreStop + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Jun 18 12:53:42.422: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "prestop-142" for this suite. +Jun 18 12:54:29.042: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:54:29.585: INFO: namespace prestop-142 deletion completed in 46.57550628s + +• [SLOW TEST:60.557 seconds] +[k8s.io] [sig-node] PreStop +/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687 + should call prestop when killing a pod [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 ------------------------------ S ------------------------------ -[sig-storage] EmptyDir volumes - should support (root,0644,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -[BeforeEach] [sig-storage] EmptyDir volumes - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +[sig-network] Service endpoints latency + should not be very high [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [sig-network] Service endpoints latency + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 STEP: Creating a kubernetes client -Jun 4 19:42:35.209: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504 -STEP: Building a namespace api object, basename emptydir -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in emptydir-4083 +Jun 18 12:54:29.585: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150 +STEP: Building a namespace api object, basename svc-latency +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in svc-latency-755 STEP: Waiting for a default service account to be provisioned in namespace -[It] should support (root,0644,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -STEP: Creating a pod to test emptydir 0644 on tmpfs -Jun 4 19:42:35.451: INFO: Waiting up to 5m0s for pod "pod-e6d9fe5e-8700-11e9-8d1b-467ee19922ac" in namespace "emptydir-4083" to be "success or failure" -Jun 4 19:42:35.465: INFO: Pod "pod-e6d9fe5e-8700-11e9-8d1b-467ee19922ac": Phase="Pending", Reason="", readiness=false. Elapsed: 14.3462ms -Jun 4 19:42:37.478: INFO: Pod "pod-e6d9fe5e-8700-11e9-8d1b-467ee19922ac": Phase="Running", Reason="", readiness=true. Elapsed: 2.026545151s -Jun 4 19:42:39.489: INFO: Pod "pod-e6d9fe5e-8700-11e9-8d1b-467ee19922ac": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.038132427s +[It] should not be very high [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: creating replication controller svc-latency-rc in namespace svc-latency-755 +I0618 12:54:29.797063 20 runners.go:184] Created replication controller with name: svc-latency-rc, namespace: svc-latency-755, replica count: 1 +I0618 12:54:30.847617 20 runners.go:184] svc-latency-rc Pods: 1 out of 1 created, 0 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady +I0618 12:54:31.847931 20 runners.go:184] svc-latency-rc Pods: 1 out of 1 created, 0 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady +I0618 12:54:32.848243 20 runners.go:184] svc-latency-rc Pods: 1 out of 1 created, 1 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady +Jun 18 12:54:32.975: INFO: Created: latency-svc-qfsz2 +Jun 18 12:54:32.981: INFO: Got endpoints: latency-svc-qfsz2 [32.622766ms] +Jun 18 12:54:33.007: INFO: Created: latency-svc-kwh74 +Jun 18 12:54:33.015: INFO: Got endpoints: latency-svc-kwh74 [33.759427ms] +Jun 18 12:54:33.023: INFO: Created: latency-svc-xnz9g +Jun 18 12:54:33.029: INFO: Got endpoints: latency-svc-xnz9g [47.738656ms] +Jun 18 12:54:33.043: INFO: Created: latency-svc-nhvz9 +Jun 18 12:54:33.050: INFO: Got endpoints: latency-svc-nhvz9 [34.91323ms] +Jun 18 12:54:33.060: INFO: Created: latency-svc-6hb54 +Jun 18 12:54:33.066: INFO: Got endpoints: latency-svc-6hb54 [84.608769ms] +Jun 18 12:54:33.077: INFO: Created: latency-svc-9rjnh +Jun 18 12:54:33.082: INFO: Got endpoints: latency-svc-9rjnh [100.593166ms] +Jun 18 12:54:33.092: INFO: Created: latency-svc-t688z +Jun 18 12:54:33.099: INFO: Got endpoints: latency-svc-t688z [116.531114ms] +Jun 18 12:54:33.111: INFO: Created: latency-svc-nvw9g +Jun 18 12:54:33.117: INFO: Got endpoints: latency-svc-nvw9g [135.197731ms] +Jun 18 12:54:33.132: INFO: Created: latency-svc-tq45x +Jun 18 12:54:33.147: INFO: Got endpoints: latency-svc-tq45x [164.924029ms] +Jun 18 12:54:33.155: INFO: Created: latency-svc-xtc2x +Jun 18 12:54:33.163: INFO: Got endpoints: latency-svc-xtc2x [180.886928ms] +Jun 18 12:54:33.173: INFO: Created: latency-svc-ftks7 +Jun 18 12:54:33.179: INFO: Got endpoints: latency-svc-ftks7 [196.407738ms] +Jun 18 12:54:33.188: INFO: Created: latency-svc-8b889 +Jun 18 12:54:33.194: INFO: Got endpoints: latency-svc-8b889 [211.605826ms] +Jun 18 12:54:33.203: INFO: Created: latency-svc-hlnh5 +Jun 18 12:54:33.210: INFO: Got endpoints: latency-svc-hlnh5 [227.5212ms] +Jun 18 12:54:33.220: INFO: Created: latency-svc-hjkkt +Jun 18 12:54:33.227: INFO: Got endpoints: latency-svc-hjkkt [244.69176ms] +Jun 18 12:54:33.236: INFO: Created: latency-svc-k685w +Jun 18 12:54:33.241: INFO: Got endpoints: latency-svc-k685w [258.317057ms] +Jun 18 12:54:33.250: INFO: Created: latency-svc-j4h8c +Jun 18 12:54:33.257: INFO: Got endpoints: latency-svc-j4h8c [274.929116ms] +Jun 18 12:54:33.266: INFO: Created: latency-svc-rvwd6 +Jun 18 12:54:33.272: INFO: Got endpoints: latency-svc-rvwd6 [290.183214ms] +Jun 18 12:54:33.281: INFO: Created: latency-svc-b79cr +Jun 18 12:54:33.288: INFO: Got endpoints: latency-svc-b79cr [258.091069ms] +Jun 18 12:54:33.298: INFO: Created: latency-svc-xr9b9 +Jun 18 12:54:33.305: INFO: Got endpoints: latency-svc-xr9b9 [254.618775ms] +Jun 18 12:54:33.315: INFO: Created: latency-svc-jxqvm +Jun 18 12:54:33.321: INFO: Got endpoints: latency-svc-jxqvm [255.34747ms] +Jun 18 12:54:33.331: INFO: Created: latency-svc-zrtnd +Jun 18 12:54:33.338: INFO: Got endpoints: latency-svc-zrtnd [255.002528ms] +Jun 18 12:54:33.346: INFO: Created: latency-svc-s65bh +Jun 18 12:54:33.353: INFO: Got endpoints: latency-svc-s65bh [254.797363ms] +Jun 18 12:54:33.361: INFO: Created: latency-svc-27kgs +Jun 18 12:54:33.370: INFO: Got endpoints: latency-svc-27kgs [252.506399ms] +Jun 18 12:54:33.376: INFO: Created: latency-svc-flnmj +Jun 18 12:54:33.382: INFO: Got endpoints: latency-svc-flnmj [235.078324ms] +Jun 18 12:54:33.394: INFO: Created: latency-svc-h96lk +Jun 18 12:54:33.400: INFO: Got endpoints: latency-svc-h96lk [236.722417ms] +Jun 18 12:54:33.409: INFO: Created: latency-svc-c4zhb +Jun 18 12:54:33.415: INFO: Got endpoints: latency-svc-c4zhb [236.434982ms] +Jun 18 12:54:33.423: INFO: Created: latency-svc-qh2p6 +Jun 18 12:54:33.430: INFO: Got endpoints: latency-svc-qh2p6 [235.66045ms] +Jun 18 12:54:33.441: INFO: Created: latency-svc-wkhgs +Jun 18 12:54:33.448: INFO: Got endpoints: latency-svc-wkhgs [238.448296ms] +Jun 18 12:54:33.457: INFO: Created: latency-svc-mfmhl +Jun 18 12:54:33.463: INFO: Got endpoints: latency-svc-mfmhl [236.352608ms] +Jun 18 12:54:33.472: INFO: Created: latency-svc-vdqbv +Jun 18 12:54:33.478: INFO: Got endpoints: latency-svc-vdqbv [237.290478ms] +Jun 18 12:54:33.487: INFO: Created: latency-svc-8qdfh +Jun 18 12:54:33.493: INFO: Got endpoints: latency-svc-8qdfh [236.132076ms] +Jun 18 12:54:33.510: INFO: Created: latency-svc-phq54 +Jun 18 12:54:33.518: INFO: Got endpoints: latency-svc-phq54 [245.36725ms] +Jun 18 12:54:33.527: INFO: Created: latency-svc-8bkt8 +Jun 18 12:54:33.535: INFO: Got endpoints: latency-svc-8bkt8 [247.263389ms] +Jun 18 12:54:33.543: INFO: Created: latency-svc-w944w +Jun 18 12:54:33.551: INFO: Got endpoints: latency-svc-w944w [245.520621ms] +Jun 18 12:54:33.565: INFO: Created: latency-svc-2jszf +Jun 18 12:54:33.578: INFO: Got endpoints: latency-svc-2jszf [256.908373ms] +Jun 18 12:54:33.582: INFO: Created: latency-svc-6lpst +Jun 18 12:54:33.592: INFO: Got endpoints: latency-svc-6lpst [254.187224ms] +Jun 18 12:54:33.605: INFO: Created: latency-svc-c2pf4 +Jun 18 12:54:33.612: INFO: Got endpoints: latency-svc-c2pf4 [258.97333ms] +Jun 18 12:54:33.622: INFO: Created: latency-svc-h296k +Jun 18 12:54:33.630: INFO: Got endpoints: latency-svc-h296k [259.973249ms] +Jun 18 12:54:33.637: INFO: Created: latency-svc-qkdgn +Jun 18 12:54:33.643: INFO: Got endpoints: latency-svc-qkdgn [260.106ms] +Jun 18 12:54:33.653: INFO: Created: latency-svc-mdj58 +Jun 18 12:54:33.658: INFO: Got endpoints: latency-svc-mdj58 [258.534419ms] +Jun 18 12:54:33.669: INFO: Created: latency-svc-594bg +Jun 18 12:54:33.675: INFO: Got endpoints: latency-svc-594bg [260.135596ms] +Jun 18 12:54:33.685: INFO: Created: latency-svc-lrb5v +Jun 18 12:54:33.691: INFO: Got endpoints: latency-svc-lrb5v [260.547923ms] +Jun 18 12:54:33.704: INFO: Created: latency-svc-mskc5 +Jun 18 12:54:33.712: INFO: Got endpoints: latency-svc-mskc5 [263.304037ms] +Jun 18 12:54:33.721: INFO: Created: latency-svc-ptj2m +Jun 18 12:54:33.727: INFO: Got endpoints: latency-svc-ptj2m [263.535737ms] +Jun 18 12:54:33.736: INFO: Created: latency-svc-sn4nw +Jun 18 12:54:33.741: INFO: Got endpoints: latency-svc-sn4nw [263.599741ms] +Jun 18 12:54:33.751: INFO: Created: latency-svc-wc8mb +Jun 18 12:54:33.757: INFO: Got endpoints: latency-svc-wc8mb [263.697736ms] +Jun 18 12:54:33.768: INFO: Created: latency-svc-6l4sh +Jun 18 12:54:33.773: INFO: Got endpoints: latency-svc-6l4sh [255.064316ms] +Jun 18 12:54:33.783: INFO: Created: latency-svc-hhkqg +Jun 18 12:54:33.790: INFO: Got endpoints: latency-svc-hhkqg [254.995286ms] +Jun 18 12:54:33.800: INFO: Created: latency-svc-kbpgz +Jun 18 12:54:33.806: INFO: Got endpoints: latency-svc-kbpgz [255.535927ms] +Jun 18 12:54:33.815: INFO: Created: latency-svc-zs6k4 +Jun 18 12:54:33.823: INFO: Got endpoints: latency-svc-zs6k4 [244.173347ms] +Jun 18 12:54:33.831: INFO: Created: latency-svc-gpvz5 +Jun 18 12:54:33.837: INFO: Got endpoints: latency-svc-gpvz5 [245.050519ms] +Jun 18 12:54:33.847: INFO: Created: latency-svc-ngq7t +Jun 18 12:54:33.853: INFO: Got endpoints: latency-svc-ngq7t [240.55891ms] +Jun 18 12:54:33.861: INFO: Created: latency-svc-vqsx2 +Jun 18 12:54:33.869: INFO: Got endpoints: latency-svc-vqsx2 [238.337226ms] +Jun 18 12:54:33.879: INFO: Created: latency-svc-g77hv +Jun 18 12:54:33.885: INFO: Got endpoints: latency-svc-g77hv [242.392299ms] +Jun 18 12:54:33.895: INFO: Created: latency-svc-qsvsx +Jun 18 12:54:33.904: INFO: Got endpoints: latency-svc-qsvsx [245.263929ms] +Jun 18 12:54:33.912: INFO: Created: latency-svc-gwxvg +Jun 18 12:54:33.919: INFO: Got endpoints: latency-svc-gwxvg [243.32288ms] +Jun 18 12:54:33.927: INFO: Created: latency-svc-hvm9t +Jun 18 12:54:33.936: INFO: Got endpoints: latency-svc-hvm9t [245.219179ms] +Jun 18 12:54:33.947: INFO: Created: latency-svc-6bc9r +Jun 18 12:54:33.954: INFO: Got endpoints: latency-svc-6bc9r [242.247215ms] +Jun 18 12:54:33.963: INFO: Created: latency-svc-8bd9w +Jun 18 12:54:33.973: INFO: Got endpoints: latency-svc-8bd9w [246.498381ms] +Jun 18 12:54:33.980: INFO: Created: latency-svc-8bpc9 +Jun 18 12:54:33.986: INFO: Got endpoints: latency-svc-8bpc9 [244.660798ms] +Jun 18 12:54:33.997: INFO: Created: latency-svc-h9tcv +Jun 18 12:54:34.003: INFO: Got endpoints: latency-svc-h9tcv [246.155099ms] +Jun 18 12:54:34.026: INFO: Created: latency-svc-q572g +Jun 18 12:54:34.034: INFO: Got endpoints: latency-svc-q572g [260.504227ms] +Jun 18 12:54:34.047: INFO: Created: latency-svc-xxwc9 +Jun 18 12:54:34.059: INFO: Got endpoints: latency-svc-xxwc9 [268.420306ms] +Jun 18 12:54:34.064: INFO: Created: latency-svc-hp77v +Jun 18 12:54:34.072: INFO: Got endpoints: latency-svc-hp77v [265.398712ms] +Jun 18 12:54:34.081: INFO: Created: latency-svc-88hw5 +Jun 18 12:54:34.090: INFO: Got endpoints: latency-svc-88hw5 [267.089962ms] +Jun 18 12:54:34.108: INFO: Created: latency-svc-kx67j +Jun 18 12:54:34.117: INFO: Got endpoints: latency-svc-kx67j [279.482332ms] +Jun 18 12:54:34.125: INFO: Created: latency-svc-zcsk6 +Jun 18 12:54:34.132: INFO: Got endpoints: latency-svc-zcsk6 [278.549026ms] +Jun 18 12:54:34.144: INFO: Created: latency-svc-9f659 +Jun 18 12:54:34.152: INFO: Got endpoints: latency-svc-9f659 [283.165076ms] +Jun 18 12:54:34.160: INFO: Created: latency-svc-7zjtf +Jun 18 12:54:34.167: INFO: Got endpoints: latency-svc-7zjtf [281.948323ms] +Jun 18 12:54:34.181: INFO: Created: latency-svc-lzjz8 +Jun 18 12:54:34.191: INFO: Got endpoints: latency-svc-lzjz8 [286.811469ms] +Jun 18 12:54:34.197: INFO: Created: latency-svc-txxzd +Jun 18 12:54:34.203: INFO: Got endpoints: latency-svc-txxzd [284.24075ms] +Jun 18 12:54:34.213: INFO: Created: latency-svc-qchpq +Jun 18 12:54:34.219: INFO: Got endpoints: latency-svc-qchpq [264.935802ms] +Jun 18 12:54:34.241: INFO: Created: latency-svc-q4p7x +Jun 18 12:54:34.242: INFO: Got endpoints: latency-svc-q4p7x [306.035369ms] +Jun 18 12:54:34.256: INFO: Created: latency-svc-ntpzz +Jun 18 12:54:34.262: INFO: Got endpoints: latency-svc-ntpzz [288.018945ms] +Jun 18 12:54:34.270: INFO: Created: latency-svc-t5tbp +Jun 18 12:54:34.278: INFO: Got endpoints: latency-svc-t5tbp [291.534729ms] +Jun 18 12:54:34.286: INFO: Created: latency-svc-x7f7h +Jun 18 12:54:34.293: INFO: Got endpoints: latency-svc-x7f7h [289.157723ms] +Jun 18 12:54:34.308: INFO: Created: latency-svc-gkhqn +Jun 18 12:54:34.324: INFO: Got endpoints: latency-svc-gkhqn [290.01039ms] +Jun 18 12:54:34.334: INFO: Created: latency-svc-hvgff +Jun 18 12:54:34.343: INFO: Got endpoints: latency-svc-hvgff [284.408473ms] +Jun 18 12:54:34.359: INFO: Created: latency-svc-4wv9g +Jun 18 12:54:34.368: INFO: Got endpoints: latency-svc-4wv9g [296.300018ms] +Jun 18 12:54:34.379: INFO: Created: latency-svc-kdm9d +Jun 18 12:54:34.399: INFO: Got endpoints: latency-svc-kdm9d [309.107167ms] +Jun 18 12:54:34.415: INFO: Created: latency-svc-9gqp7 +Jun 18 12:54:34.423: INFO: Got endpoints: latency-svc-9gqp7 [306.52534ms] +Jun 18 12:54:34.426: INFO: Created: latency-svc-5mw9v +Jun 18 12:54:34.435: INFO: Got endpoints: latency-svc-5mw9v [303.60293ms] +Jun 18 12:54:34.453: INFO: Created: latency-svc-5gj48 +Jun 18 12:54:34.461: INFO: Got endpoints: latency-svc-5gj48 [308.486285ms] +Jun 18 12:54:34.468: INFO: Created: latency-svc-j6dcq +Jun 18 12:54:34.476: INFO: Got endpoints: latency-svc-j6dcq [308.502241ms] +Jun 18 12:54:34.491: INFO: Created: latency-svc-zwdw7 +Jun 18 12:54:34.503: INFO: Got endpoints: latency-svc-zwdw7 [312.126307ms] +Jun 18 12:54:34.511: INFO: Created: latency-svc-np4l9 +Jun 18 12:54:34.514: INFO: Got endpoints: latency-svc-np4l9 [311.223302ms] +Jun 18 12:54:34.532: INFO: Created: latency-svc-k9bj2 +Jun 18 12:54:34.542: INFO: Got endpoints: latency-svc-k9bj2 [322.715945ms] +Jun 18 12:54:34.547: INFO: Created: latency-svc-dbl9s +Jun 18 12:54:34.552: INFO: Got endpoints: latency-svc-dbl9s [310.064925ms] +Jun 18 12:54:34.562: INFO: Created: latency-svc-lszzf +Jun 18 12:54:34.569: INFO: Got endpoints: latency-svc-lszzf [307.154816ms] +Jun 18 12:54:34.579: INFO: Created: latency-svc-ml6z7 +Jun 18 12:54:34.585: INFO: Got endpoints: latency-svc-ml6z7 [306.893768ms] +Jun 18 12:54:34.595: INFO: Created: latency-svc-687s9 +Jun 18 12:54:34.602: INFO: Got endpoints: latency-svc-687s9 [309.413696ms] +Jun 18 12:54:34.614: INFO: Created: latency-svc-4cgtn +Jun 18 12:54:34.620: INFO: Got endpoints: latency-svc-4cgtn [296.20073ms] +Jun 18 12:54:34.630: INFO: Created: latency-svc-mx5jg +Jun 18 12:54:34.636: INFO: Got endpoints: latency-svc-mx5jg [293.389808ms] +Jun 18 12:54:34.647: INFO: Created: latency-svc-hzk2b +Jun 18 12:54:34.654: INFO: Got endpoints: latency-svc-hzk2b [285.964161ms] +Jun 18 12:54:34.664: INFO: Created: latency-svc-vrv6p +Jun 18 12:54:34.671: INFO: Got endpoints: latency-svc-vrv6p [271.735136ms] +Jun 18 12:54:34.683: INFO: Created: latency-svc-x9b7b +Jun 18 12:54:34.694: INFO: Got endpoints: latency-svc-x9b7b [270.702941ms] +Jun 18 12:54:34.700: INFO: Created: latency-svc-nqcks +Jun 18 12:54:34.707: INFO: Got endpoints: latency-svc-nqcks [271.184615ms] +Jun 18 12:54:34.716: INFO: Created: latency-svc-dpltf +Jun 18 12:54:34.725: INFO: Got endpoints: latency-svc-dpltf [263.888003ms] +Jun 18 12:54:34.731: INFO: Created: latency-svc-gdtbz +Jun 18 12:54:34.738: INFO: Got endpoints: latency-svc-gdtbz [262.274975ms] +Jun 18 12:54:34.748: INFO: Created: latency-svc-5dsnr +Jun 18 12:54:34.755: INFO: Got endpoints: latency-svc-5dsnr [251.456109ms] +Jun 18 12:54:34.765: INFO: Created: latency-svc-d65qd +Jun 18 12:54:34.771: INFO: Got endpoints: latency-svc-d65qd [256.468018ms] +Jun 18 12:54:34.780: INFO: Created: latency-svc-wtzzm +Jun 18 12:54:34.787: INFO: Got endpoints: latency-svc-wtzzm [245.213487ms] +Jun 18 12:54:34.797: INFO: Created: latency-svc-h9cxg +Jun 18 12:54:34.804: INFO: Got endpoints: latency-svc-h9cxg [250.846746ms] +Jun 18 12:54:34.814: INFO: Created: latency-svc-96sn8 +Jun 18 12:54:34.821: INFO: Got endpoints: latency-svc-96sn8 [252.271609ms] +Jun 18 12:54:34.829: INFO: Created: latency-svc-tsrmn +Jun 18 12:54:34.843: INFO: Got endpoints: latency-svc-tsrmn [257.624183ms] +Jun 18 12:54:34.848: INFO: Created: latency-svc-w86g9 +Jun 18 12:54:34.855: INFO: Got endpoints: latency-svc-w86g9 [252.998922ms] +Jun 18 12:54:34.865: INFO: Created: latency-svc-p42fk +Jun 18 12:54:34.872: INFO: Got endpoints: latency-svc-p42fk [252.026046ms] +Jun 18 12:54:34.881: INFO: Created: latency-svc-rx9qs +Jun 18 12:54:34.887: INFO: Got endpoints: latency-svc-rx9qs [250.494442ms] +Jun 18 12:54:34.897: INFO: Created: latency-svc-mln2f +Jun 18 12:54:34.904: INFO: Got endpoints: latency-svc-mln2f [249.571766ms] +Jun 18 12:54:34.919: INFO: Created: latency-svc-7gr6x +Jun 18 12:54:34.926: INFO: Got endpoints: latency-svc-7gr6x [255.340016ms] +Jun 18 12:54:34.935: INFO: Created: latency-svc-jqr9c +Jun 18 12:54:34.940: INFO: Got endpoints: latency-svc-jqr9c [246.157166ms] +Jun 18 12:54:34.950: INFO: Created: latency-svc-kvs79 +Jun 18 12:54:34.956: INFO: Got endpoints: latency-svc-kvs79 [249.141417ms] +Jun 18 12:54:34.966: INFO: Created: latency-svc-b48zz +Jun 18 12:54:34.972: INFO: Got endpoints: latency-svc-b48zz [247.879235ms] +Jun 18 12:54:34.981: INFO: Created: latency-svc-wv8wg +Jun 18 12:54:34.987: INFO: Got endpoints: latency-svc-wv8wg [248.972199ms] +Jun 18 12:54:35.000: INFO: Created: latency-svc-lgtpw +Jun 18 12:54:35.006: INFO: Got endpoints: latency-svc-lgtpw [251.244193ms] +Jun 18 12:54:35.018: INFO: Created: latency-svc-hh949 +Jun 18 12:54:35.025: INFO: Got endpoints: latency-svc-hh949 [253.6728ms] +Jun 18 12:54:35.032: INFO: Created: latency-svc-gbmn5 +Jun 18 12:54:35.039: INFO: Got endpoints: latency-svc-gbmn5 [252.327495ms] +Jun 18 12:54:35.048: INFO: Created: latency-svc-7smz5 +Jun 18 12:54:35.054: INFO: Got endpoints: latency-svc-7smz5 [250.358112ms] +Jun 18 12:54:35.067: INFO: Created: latency-svc-bgwxz +Jun 18 12:54:35.074: INFO: Got endpoints: latency-svc-bgwxz [252.371297ms] +Jun 18 12:54:35.080: INFO: Created: latency-svc-4vfnh +Jun 18 12:54:35.086: INFO: Got endpoints: latency-svc-4vfnh [243.568528ms] +Jun 18 12:54:35.095: INFO: Created: latency-svc-vw75g +Jun 18 12:54:35.101: INFO: Got endpoints: latency-svc-vw75g [245.405625ms] +Jun 18 12:54:35.111: INFO: Created: latency-svc-wl6tm +Jun 18 12:54:35.117: INFO: Got endpoints: latency-svc-wl6tm [245.436123ms] +Jun 18 12:54:35.128: INFO: Created: latency-svc-fsk72 +Jun 18 12:54:35.136: INFO: Got endpoints: latency-svc-fsk72 [249.240669ms] +Jun 18 12:54:35.143: INFO: Created: latency-svc-28x5c +Jun 18 12:54:35.150: INFO: Got endpoints: latency-svc-28x5c [246.367599ms] +Jun 18 12:54:35.158: INFO: Created: latency-svc-fxcqb +Jun 18 12:54:35.165: INFO: Got endpoints: latency-svc-fxcqb [238.340245ms] +Jun 18 12:54:35.174: INFO: Created: latency-svc-bwtd7 +Jun 18 12:54:35.181: INFO: Got endpoints: latency-svc-bwtd7 [240.401179ms] +Jun 18 12:54:35.190: INFO: Created: latency-svc-nn5j6 +Jun 18 12:54:35.197: INFO: Got endpoints: latency-svc-nn5j6 [241.412891ms] +Jun 18 12:54:35.206: INFO: Created: latency-svc-whg4n +Jun 18 12:54:35.213: INFO: Got endpoints: latency-svc-whg4n [240.719498ms] +Jun 18 12:54:35.222: INFO: Created: latency-svc-gd6vj +Jun 18 12:54:35.228: INFO: Got endpoints: latency-svc-gd6vj [241.032574ms] +Jun 18 12:54:35.237: INFO: Created: latency-svc-zn65l +Jun 18 12:54:35.242: INFO: Got endpoints: latency-svc-zn65l [235.69882ms] +Jun 18 12:54:35.253: INFO: Created: latency-svc-lft5z +Jun 18 12:54:35.259: INFO: Got endpoints: latency-svc-lft5z [234.071625ms] +Jun 18 12:54:35.270: INFO: Created: latency-svc-p4w5k +Jun 18 12:54:35.275: INFO: Got endpoints: latency-svc-p4w5k [235.710418ms] +Jun 18 12:54:35.283: INFO: Created: latency-svc-ccg4g +Jun 18 12:54:35.290: INFO: Got endpoints: latency-svc-ccg4g [236.165368ms] +Jun 18 12:54:35.299: INFO: Created: latency-svc-jzt4h +Jun 18 12:54:35.305: INFO: Got endpoints: latency-svc-jzt4h [230.637688ms] +Jun 18 12:54:35.319: INFO: Created: latency-svc-x5pxs +Jun 18 12:54:35.326: INFO: Got endpoints: latency-svc-x5pxs [239.681883ms] +Jun 18 12:54:35.335: INFO: Created: latency-svc-8tw9z +Jun 18 12:54:35.341: INFO: Got endpoints: latency-svc-8tw9z [240.658523ms] +Jun 18 12:54:35.350: INFO: Created: latency-svc-4zmw7 +Jun 18 12:54:35.356: INFO: Got endpoints: latency-svc-4zmw7 [238.934707ms] +Jun 18 12:54:35.365: INFO: Created: latency-svc-znl72 +Jun 18 12:54:35.372: INFO: Got endpoints: latency-svc-znl72 [235.59447ms] +Jun 18 12:54:35.381: INFO: Created: latency-svc-2ggvd +Jun 18 12:54:35.387: INFO: Got endpoints: latency-svc-2ggvd [236.829125ms] +Jun 18 12:54:35.396: INFO: Created: latency-svc-d9p7z +Jun 18 12:54:35.401: INFO: Got endpoints: latency-svc-d9p7z [236.516724ms] +Jun 18 12:54:35.414: INFO: Created: latency-svc-xnpt6 +Jun 18 12:54:35.421: INFO: Got endpoints: latency-svc-xnpt6 [239.778545ms] +Jun 18 12:54:35.432: INFO: Created: latency-svc-4sw2t +Jun 18 12:54:35.438: INFO: Got endpoints: latency-svc-4sw2t [240.567388ms] +Jun 18 12:54:35.449: INFO: Created: latency-svc-tg4c6 +Jun 18 12:54:35.454: INFO: Got endpoints: latency-svc-tg4c6 [240.731992ms] +Jun 18 12:54:35.465: INFO: Created: latency-svc-84njp +Jun 18 12:54:35.470: INFO: Got endpoints: latency-svc-84njp [242.062116ms] +Jun 18 12:54:35.482: INFO: Created: latency-svc-9s2p9 +Jun 18 12:54:35.489: INFO: Got endpoints: latency-svc-9s2p9 [246.05439ms] +Jun 18 12:54:35.501: INFO: Created: latency-svc-mvk4s +Jun 18 12:54:35.508: INFO: Got endpoints: latency-svc-mvk4s [248.737476ms] +Jun 18 12:54:35.518: INFO: Created: latency-svc-nrfgp +Jun 18 12:54:35.524: INFO: Got endpoints: latency-svc-nrfgp [248.308488ms] +Jun 18 12:54:35.532: INFO: Created: latency-svc-bmkmd +Jun 18 12:54:35.537: INFO: Got endpoints: latency-svc-bmkmd [246.939498ms] +Jun 18 12:54:35.548: INFO: Created: latency-svc-268vb +Jun 18 12:54:35.555: INFO: Got endpoints: latency-svc-268vb [250.644075ms] +Jun 18 12:54:35.563: INFO: Created: latency-svc-nd8tw +Jun 18 12:54:35.570: INFO: Got endpoints: latency-svc-nd8tw [243.539282ms] +Jun 18 12:54:35.578: INFO: Created: latency-svc-6gdzc +Jun 18 12:54:35.585: INFO: Got endpoints: latency-svc-6gdzc [243.560009ms] +Jun 18 12:54:35.594: INFO: Created: latency-svc-bwh85 +Jun 18 12:54:35.598: INFO: Got endpoints: latency-svc-bwh85 [241.983251ms] +Jun 18 12:54:35.610: INFO: Created: latency-svc-tlstq +Jun 18 12:54:35.616: INFO: Got endpoints: latency-svc-tlstq [244.261273ms] +Jun 18 12:54:35.626: INFO: Created: latency-svc-ksh5k +Jun 18 12:54:35.632: INFO: Got endpoints: latency-svc-ksh5k [244.79147ms] +Jun 18 12:54:35.642: INFO: Created: latency-svc-gkgwp +Jun 18 12:54:35.648: INFO: Got endpoints: latency-svc-gkgwp [246.832862ms] +Jun 18 12:54:35.660: INFO: Created: latency-svc-9f9ch +Jun 18 12:54:35.666: INFO: Got endpoints: latency-svc-9f9ch [245.634806ms] +Jun 18 12:54:35.676: INFO: Created: latency-svc-nsk8q +Jun 18 12:54:35.682: INFO: Got endpoints: latency-svc-nsk8q [244.213814ms] +Jun 18 12:54:35.692: INFO: Created: latency-svc-s9t5p +Jun 18 12:54:35.699: INFO: Got endpoints: latency-svc-s9t5p [244.822945ms] +Jun 18 12:54:35.710: INFO: Created: latency-svc-wt44q +Jun 18 12:54:35.716: INFO: Got endpoints: latency-svc-wt44q [245.42215ms] +Jun 18 12:54:35.731: INFO: Created: latency-svc-7j7f8 +Jun 18 12:54:35.737: INFO: Got endpoints: latency-svc-7j7f8 [248.256517ms] +Jun 18 12:54:35.747: INFO: Created: latency-svc-cl5lm +Jun 18 12:54:35.752: INFO: Got endpoints: latency-svc-cl5lm [244.677219ms] +Jun 18 12:54:35.761: INFO: Created: latency-svc-mw962 +Jun 18 12:54:35.767: INFO: Got endpoints: latency-svc-mw962 [243.812034ms] +Jun 18 12:54:35.778: INFO: Created: latency-svc-tlbcl +Jun 18 12:54:35.783: INFO: Got endpoints: latency-svc-tlbcl [246.27379ms] +Jun 18 12:54:35.793: INFO: Created: latency-svc-bknhk +Jun 18 12:54:35.799: INFO: Got endpoints: latency-svc-bknhk [244.118795ms] +Jun 18 12:54:35.811: INFO: Created: latency-svc-nl9wq +Jun 18 12:54:35.818: INFO: Got endpoints: latency-svc-nl9wq [247.83826ms] +Jun 18 12:54:35.827: INFO: Created: latency-svc-cc8vv +Jun 18 12:54:35.832: INFO: Got endpoints: latency-svc-cc8vv [247.407064ms] +Jun 18 12:54:35.842: INFO: Created: latency-svc-htcms +Jun 18 12:54:35.849: INFO: Got endpoints: latency-svc-htcms [250.114831ms] +Jun 18 12:54:35.858: INFO: Created: latency-svc-fgnpx +Jun 18 12:54:35.864: INFO: Got endpoints: latency-svc-fgnpx [247.892694ms] +Jun 18 12:54:35.874: INFO: Created: latency-svc-4v4k7 +Jun 18 12:54:35.879: INFO: Got endpoints: latency-svc-4v4k7 [247.560737ms] +Jun 18 12:54:35.890: INFO: Created: latency-svc-thfhz +Jun 18 12:54:35.898: INFO: Got endpoints: latency-svc-thfhz [250.082481ms] +Jun 18 12:54:35.907: INFO: Created: latency-svc-knxzv +Jun 18 12:54:35.915: INFO: Got endpoints: latency-svc-knxzv [248.531482ms] +Jun 18 12:54:35.923: INFO: Created: latency-svc-4b55v +Jun 18 12:54:35.929: INFO: Got endpoints: latency-svc-4b55v [246.661543ms] +Jun 18 12:54:35.940: INFO: Created: latency-svc-c2mvk +Jun 18 12:54:35.948: INFO: Got endpoints: latency-svc-c2mvk [248.644609ms] +Jun 18 12:54:35.956: INFO: Created: latency-svc-lsr2q +Jun 18 12:54:35.963: INFO: Got endpoints: latency-svc-lsr2q [246.746709ms] +Jun 18 12:54:35.972: INFO: Created: latency-svc-qmkwv +Jun 18 12:54:35.979: INFO: Got endpoints: latency-svc-qmkwv [241.800529ms] +Jun 18 12:54:35.986: INFO: Created: latency-svc-dvppp +Jun 18 12:54:35.992: INFO: Got endpoints: latency-svc-dvppp [240.000413ms] +Jun 18 12:54:36.003: INFO: Created: latency-svc-frj9h +Jun 18 12:54:36.009: INFO: Got endpoints: latency-svc-frj9h [241.563052ms] +Jun 18 12:54:36.021: INFO: Created: latency-svc-gmfwn +Jun 18 12:54:36.028: INFO: Got endpoints: latency-svc-gmfwn [244.102742ms] +Jun 18 12:54:36.038: INFO: Created: latency-svc-8b66g +Jun 18 12:54:36.045: INFO: Got endpoints: latency-svc-8b66g [245.384407ms] +Jun 18 12:54:36.058: INFO: Created: latency-svc-fn4th +Jun 18 12:54:36.068: INFO: Got endpoints: latency-svc-fn4th [250.411234ms] +Jun 18 12:54:36.078: INFO: Created: latency-svc-zlcbc +Jun 18 12:54:36.084: INFO: Got endpoints: latency-svc-zlcbc [252.038695ms] +Jun 18 12:54:36.093: INFO: Created: latency-svc-pd2tl +Jun 18 12:54:36.102: INFO: Got endpoints: latency-svc-pd2tl [253.00192ms] +Jun 18 12:54:36.115: INFO: Created: latency-svc-s676x +Jun 18 12:54:36.123: INFO: Got endpoints: latency-svc-s676x [258.672368ms] +Jun 18 12:54:36.132: INFO: Created: latency-svc-d4wwh +Jun 18 12:54:36.139: INFO: Got endpoints: latency-svc-d4wwh [259.261071ms] +Jun 18 12:54:36.146: INFO: Created: latency-svc-x65m2 +Jun 18 12:54:36.152: INFO: Got endpoints: latency-svc-x65m2 [253.746842ms] +Jun 18 12:54:36.166: INFO: Created: latency-svc-vlmjc +Jun 18 12:54:36.172: INFO: Got endpoints: latency-svc-vlmjc [257.086567ms] +Jun 18 12:54:36.182: INFO: Created: latency-svc-v8tn4 +Jun 18 12:54:36.187: INFO: Got endpoints: latency-svc-v8tn4 [258.555127ms] +Jun 18 12:54:36.201: INFO: Created: latency-svc-lgscg +Jun 18 12:54:36.208: INFO: Got endpoints: latency-svc-lgscg [260.105189ms] +Jun 18 12:54:36.217: INFO: Created: latency-svc-4mlpw +Jun 18 12:54:36.221: INFO: Got endpoints: latency-svc-4mlpw [258.585665ms] +Jun 18 12:54:36.232: INFO: Created: latency-svc-hpqqv +Jun 18 12:54:36.239: INFO: Got endpoints: latency-svc-hpqqv [260.52679ms] +Jun 18 12:54:36.249: INFO: Created: latency-svc-bvldg +Jun 18 12:54:36.256: INFO: Got endpoints: latency-svc-bvldg [263.317443ms] +Jun 18 12:54:36.266: INFO: Created: latency-svc-pns5g +Jun 18 12:54:36.272: INFO: Got endpoints: latency-svc-pns5g [263.39274ms] +Jun 18 12:54:36.283: INFO: Created: latency-svc-b9b6d +Jun 18 12:54:36.290: INFO: Got endpoints: latency-svc-b9b6d [261.965416ms] +Jun 18 12:54:36.318: INFO: Created: latency-svc-d7z5h +Jun 18 12:54:36.323: INFO: Got endpoints: latency-svc-d7z5h [278.37483ms] +Jun 18 12:54:36.327: INFO: Created: latency-svc-fkqls +Jun 18 12:54:36.337: INFO: Got endpoints: latency-svc-fkqls [268.402156ms] +Jun 18 12:54:36.349: INFO: Created: latency-svc-dcbsr +Jun 18 12:54:36.359: INFO: Got endpoints: latency-svc-dcbsr [274.794468ms] +Jun 18 12:54:36.371: INFO: Created: latency-svc-2jqj8 +Jun 18 12:54:36.378: INFO: Got endpoints: latency-svc-2jqj8 [276.099356ms] +Jun 18 12:54:36.388: INFO: Created: latency-svc-qzv8c +Jun 18 12:54:36.394: INFO: Got endpoints: latency-svc-qzv8c [271.595152ms] +Jun 18 12:54:36.404: INFO: Created: latency-svc-xzdz5 +Jun 18 12:54:36.412: INFO: Got endpoints: latency-svc-xzdz5 [272.871843ms] +Jun 18 12:54:36.420: INFO: Created: latency-svc-ztz4h +Jun 18 12:54:36.427: INFO: Got endpoints: latency-svc-ztz4h [274.515408ms] +Jun 18 12:54:36.435: INFO: Created: latency-svc-qddfn +Jun 18 12:54:36.441: INFO: Got endpoints: latency-svc-qddfn [269.308715ms] +Jun 18 12:54:36.441: INFO: Latencies: [33.759427ms 34.91323ms 47.738656ms 84.608769ms 100.593166ms 116.531114ms 135.197731ms 164.924029ms 180.886928ms 196.407738ms 211.605826ms 227.5212ms 230.637688ms 234.071625ms 235.078324ms 235.59447ms 235.66045ms 235.69882ms 235.710418ms 236.132076ms 236.165368ms 236.352608ms 236.434982ms 236.516724ms 236.722417ms 236.829125ms 237.290478ms 238.337226ms 238.340245ms 238.448296ms 238.934707ms 239.681883ms 239.778545ms 240.000413ms 240.401179ms 240.55891ms 240.567388ms 240.658523ms 240.719498ms 240.731992ms 241.032574ms 241.412891ms 241.563052ms 241.800529ms 241.983251ms 242.062116ms 242.247215ms 242.392299ms 243.32288ms 243.539282ms 243.560009ms 243.568528ms 243.812034ms 244.102742ms 244.118795ms 244.173347ms 244.213814ms 244.261273ms 244.660798ms 244.677219ms 244.69176ms 244.79147ms 244.822945ms 245.050519ms 245.213487ms 245.219179ms 245.263929ms 245.36725ms 245.384407ms 245.405625ms 245.42215ms 245.436123ms 245.520621ms 245.634806ms 246.05439ms 246.155099ms 246.157166ms 246.27379ms 246.367599ms 246.498381ms 246.661543ms 246.746709ms 246.832862ms 246.939498ms 247.263389ms 247.407064ms 247.560737ms 247.83826ms 247.879235ms 247.892694ms 248.256517ms 248.308488ms 248.531482ms 248.644609ms 248.737476ms 248.972199ms 249.141417ms 249.240669ms 249.571766ms 250.082481ms 250.114831ms 250.358112ms 250.411234ms 250.494442ms 250.644075ms 250.846746ms 251.244193ms 251.456109ms 252.026046ms 252.038695ms 252.271609ms 252.327495ms 252.371297ms 252.506399ms 252.998922ms 253.00192ms 253.6728ms 253.746842ms 254.187224ms 254.618775ms 254.797363ms 254.995286ms 255.002528ms 255.064316ms 255.340016ms 255.34747ms 255.535927ms 256.468018ms 256.908373ms 257.086567ms 257.624183ms 258.091069ms 258.317057ms 258.534419ms 258.555127ms 258.585665ms 258.672368ms 258.97333ms 259.261071ms 259.973249ms 260.105189ms 260.106ms 260.135596ms 260.504227ms 260.52679ms 260.547923ms 261.965416ms 262.274975ms 263.304037ms 263.317443ms 263.39274ms 263.535737ms 263.599741ms 263.697736ms 263.888003ms 264.935802ms 265.398712ms 267.089962ms 268.402156ms 268.420306ms 269.308715ms 270.702941ms 271.184615ms 271.595152ms 271.735136ms 272.871843ms 274.515408ms 274.794468ms 274.929116ms 276.099356ms 278.37483ms 278.549026ms 279.482332ms 281.948323ms 283.165076ms 284.24075ms 284.408473ms 285.964161ms 286.811469ms 288.018945ms 289.157723ms 290.01039ms 290.183214ms 291.534729ms 293.389808ms 296.20073ms 296.300018ms 303.60293ms 306.035369ms 306.52534ms 306.893768ms 307.154816ms 308.486285ms 308.502241ms 309.107167ms 309.413696ms 310.064925ms 311.223302ms 312.126307ms 322.715945ms] +Jun 18 12:54:36.442: INFO: 50 %ile: 250.114831ms +Jun 18 12:54:36.442: INFO: 90 %ile: 289.157723ms +Jun 18 12:54:36.442: INFO: 99 %ile: 312.126307ms +Jun 18 12:54:36.442: INFO: Total sample count: 200 +[AfterEach] [sig-network] Service endpoints latency + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Jun 18 12:54:36.442: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "svc-latency-755" for this suite. +Jun 18 12:54:54.482: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:54:54.896: INFO: namespace svc-latency-755 deletion completed in 18.440460519s + +• [SLOW TEST:25.311 seconds] +[sig-network] Service endpoints latency +/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:22 + should not be very high [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +------------------------------ +SSSSSSSSSSS +------------------------------ +[sig-storage] Projected downwardAPI + should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [sig-storage] Projected downwardAPI + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +STEP: Creating a kubernetes client +Jun 18 12:54:54.897: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150 +STEP: Building a namespace api object, basename projected +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-9699 +STEP: Waiting for a default service account to be provisioned in namespace +[BeforeEach] [sig-storage] Projected downwardAPI + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:39 +[It] should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: Creating a pod to test downward API volume plugin +Jun 18 12:54:55.137: INFO: Waiting up to 5m0s for pod "downwardapi-volume-45258321-91c8-11e9-a25d-8608290c688a" in namespace "projected-9699" to be "success or failure" +Jun 18 12:54:55.153: INFO: Pod "downwardapi-volume-45258321-91c8-11e9-a25d-8608290c688a": Phase="Pending", Reason="", readiness=false. Elapsed: 15.902803ms +Jun 18 12:54:57.167: INFO: Pod "downwardapi-volume-45258321-91c8-11e9-a25d-8608290c688a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.030510995s STEP: Saw pod success -Jun 4 19:42:39.489: INFO: Pod "pod-e6d9fe5e-8700-11e9-8d1b-467ee19922ac" satisfied condition "success or failure" -Jun 4 19:42:39.500: INFO: Trying to get logs from node 10.212.23.189 pod pod-e6d9fe5e-8700-11e9-8d1b-467ee19922ac container test-container: +Jun 18 12:54:57.167: INFO: Pod "downwardapi-volume-45258321-91c8-11e9-a25d-8608290c688a" satisfied condition "success or failure" +Jun 18 12:54:57.180: INFO: Trying to get logs from node 10.72.74.138 pod downwardapi-volume-45258321-91c8-11e9-a25d-8608290c688a container client-container: STEP: delete the pod -Jun 4 19:42:39.560: INFO: Waiting for pod pod-e6d9fe5e-8700-11e9-8d1b-467ee19922ac to disappear -Jun 4 19:42:39.571: INFO: Pod pod-e6d9fe5e-8700-11e9-8d1b-467ee19922ac no longer exists -[AfterEach] [sig-storage] EmptyDir volumes - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 -Jun 4 19:42:39.571: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "emptydir-4083" for this suite. -Jun 4 19:42:45.701: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -Jun 4 19:42:46.097: INFO: namespace emptydir-4083 deletion completed in 6.441321966s +Jun 18 12:54:57.272: INFO: Waiting for pod downwardapi-volume-45258321-91c8-11e9-a25d-8608290c688a to disappear +Jun 18 12:54:57.288: INFO: Pod downwardapi-volume-45258321-91c8-11e9-a25d-8608290c688a no longer exists +[AfterEach] [sig-storage] Projected downwardAPI + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Jun 18 12:54:57.288: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "projected-9699" for this suite. +Jun 18 12:55:05.329: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:55:06.958: INFO: namespace projected-9699 deletion completed in 9.656869046s -• [SLOW TEST:10.888 seconds] -[sig-storage] EmptyDir volumes -/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:41 - should support (root,0644,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +• [SLOW TEST:12.062 seconds] +[sig-storage] Projected downwardAPI +/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:33 + should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 ------------------------------ -SSSSSSSSSS +SSSSSSS ------------------------------ -[k8s.io] Pods - should support remote command execution over websockets [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -[BeforeEach] [k8s.io] Pods - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +[sig-api-machinery] Namespaces [Serial] + should ensure that all services are removed when a namespace is deleted [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [sig-api-machinery] Namespaces [Serial] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 STEP: Creating a kubernetes client -Jun 4 19:42:46.097: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504 -STEP: Building a namespace api object, basename pods -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in pods-3676 +Jun 18 12:55:06.958: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150 +STEP: Building a namespace api object, basename namespaces +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in namespaces-1662 STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [k8s.io] Pods - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:135 -[It] should support remote command execution over websockets [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -Jun 4 19:42:46.315: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504 -STEP: creating the pod -STEP: submitting the pod to kubernetes -[AfterEach] [k8s.io] Pods - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 -Jun 4 19:42:48.584: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "pods-3676" for this suite. -Jun 4 19:43:28.647: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -Jun 4 19:43:28.990: INFO: namespace pods-3676 deletion completed in 40.389229912s - -• [SLOW TEST:42.893 seconds] -[k8s.io] Pods -/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687 - should support remote command execution over websockets [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[It] should ensure that all services are removed when a namespace is deleted [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: Creating a test namespace +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in nsdeletetest-6765 +STEP: Waiting for a default service account to be provisioned in namespace +STEP: Creating a service in the namespace +STEP: Deleting the namespace +STEP: Waiting for the namespace to be removed. +STEP: Recreating the namespace +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in nsdeletetest-7846 +STEP: Verifying there is no service in the namespace +[AfterEach] [sig-api-machinery] Namespaces [Serial] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Jun 18 12:55:13.612: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "namespaces-1662" for this suite. +Jun 18 12:55:19.654: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:55:20.050: INFO: namespace namespaces-1662 deletion completed in 6.423523689s +STEP: Destroying namespace "nsdeletetest-6765" for this suite. +Jun 18 12:55:20.057: INFO: Namespace nsdeletetest-6765 was already deleted +STEP: Destroying namespace "nsdeletetest-7846" for this suite. +Jun 18 12:55:28.087: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:55:29.260: INFO: namespace nsdeletetest-7846 deletion completed in 9.202974194s + +• [SLOW TEST:22.302 seconds] +[sig-api-machinery] Namespaces [Serial] +/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:22 + should ensure that all services are removed when a namespace is deleted [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 ------------------------------ -[sig-cli] Kubectl client [k8s.io] Kubectl patch - should add annotations for pods in rc [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -[BeforeEach] [sig-cli] Kubectl client - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +SSSSSSSSSSSSSSSSSSSSSSSSSSSSS +------------------------------ +[k8s.io] Probing container + should be restarted with a /healthz http liveness probe [NodeConformance] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [k8s.io] Probing container + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 STEP: Creating a kubernetes client -Jun 4 19:43:28.990: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504 -STEP: Building a namespace api object, basename kubectl -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-8989 +Jun 18 12:55:29.260: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150 +STEP: Building a namespace api object, basename container-probe +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in container-probe-8880 STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [sig-cli] Kubectl client - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:213 -[It] should add annotations for pods in rc [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -STEP: creating Redis RC -Jun 4 19:43:29.260: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 create -f - --namespace=kubectl-8989' -Jun 4 19:43:29.577: INFO: stderr: "" -Jun 4 19:43:29.577: INFO: stdout: "replicationcontroller/redis-master created\n" -STEP: Waiting for Redis master to start. -Jun 4 19:43:30.590: INFO: Selector matched 1 pods for map[app:redis] -Jun 4 19:43:30.590: INFO: Found 0 / 1 -Jun 4 19:43:31.591: INFO: Selector matched 1 pods for map[app:redis] -Jun 4 19:43:31.591: INFO: Found 1 / 1 -Jun 4 19:43:31.591: INFO: WaitFor completed with timeout 5m0s. Pods found = 1 out of 1 -STEP: patching all pods -Jun 4 19:43:31.602: INFO: Selector matched 1 pods for map[app:redis] -Jun 4 19:43:31.603: INFO: ForEach: Found 1 pods from the filter. Now looping through them. -Jun 4 19:43:31.603: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 patch pod redis-master-xj4xg --namespace=kubectl-8989 -p {"metadata":{"annotations":{"x":"y"}}}' -Jun 4 19:43:31.719: INFO: stderr: "" -Jun 4 19:43:31.719: INFO: stdout: "pod/redis-master-xj4xg patched\n" -STEP: checking annotations -Jun 4 19:43:31.731: INFO: Selector matched 1 pods for map[app:redis] -Jun 4 19:43:31.731: INFO: ForEach: Found 1 pods from the filter. Now looping through them. -[AfterEach] [sig-cli] Kubectl client - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 -Jun 4 19:43:31.731: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "kubectl-8989" for this suite. -Jun 4 19:43:55.790: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -Jun 4 19:43:56.316: INFO: namespace kubectl-8989 deletion completed in 24.568724488s +[BeforeEach] [k8s.io] Probing container + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/container_probe.go:51 +[It] should be restarted with a /healthz http liveness probe [NodeConformance] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: Creating pod liveness-http in namespace container-probe-8880 +Jun 18 12:55:31.534: INFO: Started pod liveness-http in namespace container-probe-8880 +STEP: checking the pod's current state and verifying that restartCount is present +Jun 18 12:55:31.547: INFO: Initial restart count of pod liveness-http is 0 +Jun 18 12:55:56.131: INFO: Restart count of pod container-probe-8880/liveness-http is now 1 (24.583613359s elapsed) +STEP: deleting the pod +[AfterEach] [k8s.io] Probing container + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Jun 18 12:55:56.180: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "container-probe-8880" for this suite. +Jun 18 12:56:04.223: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:56:04.640: INFO: namespace container-probe-8880 deletion completed in 8.445206251s -• [SLOW TEST:27.326 seconds] -[sig-cli] Kubectl client -/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 - [k8s.io] Kubectl patch - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687 - should add annotations for pods in rc [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +• [SLOW TEST:35.379 seconds] +[k8s.io] Probing container +/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687 + should be restarted with a /healthz http liveness probe [NodeConformance] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 ------------------------------ -SSSSSSSSSSSSSSSSSSSSSS +SSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-storage] Secrets - should be consumable from pods in volume [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -[BeforeEach] [sig-storage] Secrets - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +[sig-storage] Projected configMap + should be consumable from pods in volume with mappings as non-root [LinuxOnly] [NodeConformance] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [sig-storage] Projected configMap + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 STEP: Creating a kubernetes client -Jun 4 19:43:56.318: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504 -STEP: Building a namespace api object, basename secrets -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in secrets-3251 +Jun 18 12:56:04.640: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150 +STEP: Building a namespace api object, basename projected +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-2584 STEP: Waiting for a default service account to be provisioned in namespace -[It] should be consumable from pods in volume [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -STEP: Creating secret with name secret-test-17357083-8701-11e9-8d1b-467ee19922ac -STEP: Creating a pod to test consume secrets -Jun 4 19:43:56.592: INFO: Waiting up to 5m0s for pod "pod-secrets-1736fc98-8701-11e9-8d1b-467ee19922ac" in namespace "secrets-3251" to be "success or failure" -Jun 4 19:43:56.603: INFO: Pod "pod-secrets-1736fc98-8701-11e9-8d1b-467ee19922ac": Phase="Pending", Reason="", readiness=false. Elapsed: 11.151635ms -Jun 4 19:43:58.618: INFO: Pod "pod-secrets-1736fc98-8701-11e9-8d1b-467ee19922ac": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.026016158s +[It] should be consumable from pods in volume with mappings as non-root [LinuxOnly] [NodeConformance] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: Creating configMap with name projected-configmap-test-volume-map-6eb8102a-91c8-11e9-a25d-8608290c688a +STEP: Creating a pod to test consume configMaps +Jun 18 12:56:04.897: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-6eba7fc9-91c8-11e9-a25d-8608290c688a" in namespace "projected-2584" to be "success or failure" +Jun 18 12:56:04.921: INFO: Pod "pod-projected-configmaps-6eba7fc9-91c8-11e9-a25d-8608290c688a": Phase="Pending", Reason="", readiness=false. Elapsed: 23.966392ms +Jun 18 12:56:06.940: INFO: Pod "pod-projected-configmaps-6eba7fc9-91c8-11e9-a25d-8608290c688a": Phase="Running", Reason="", readiness=true. Elapsed: 2.042354292s +Jun 18 12:56:08.954: INFO: Pod "pod-projected-configmaps-6eba7fc9-91c8-11e9-a25d-8608290c688a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.056766311s STEP: Saw pod success -Jun 4 19:43:58.618: INFO: Pod "pod-secrets-1736fc98-8701-11e9-8d1b-467ee19922ac" satisfied condition "success or failure" -Jun 4 19:43:58.631: INFO: Trying to get logs from node 10.212.23.161 pod pod-secrets-1736fc98-8701-11e9-8d1b-467ee19922ac container secret-volume-test: +Jun 18 12:56:08.954: INFO: Pod "pod-projected-configmaps-6eba7fc9-91c8-11e9-a25d-8608290c688a" satisfied condition "success or failure" +Jun 18 12:56:08.973: INFO: Trying to get logs from node 10.72.74.138 pod pod-projected-configmaps-6eba7fc9-91c8-11e9-a25d-8608290c688a container projected-configmap-volume-test: STEP: delete the pod -Jun 4 19:43:58.696: INFO: Waiting for pod pod-secrets-1736fc98-8701-11e9-8d1b-467ee19922ac to disappear -Jun 4 19:43:58.708: INFO: Pod pod-secrets-1736fc98-8701-11e9-8d1b-467ee19922ac no longer exists -[AfterEach] [sig-storage] Secrets - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 -Jun 4 19:43:58.708: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "secrets-3251" for this suite. -Jun 4 19:44:04.775: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -Jun 4 19:44:05.504: INFO: namespace secrets-3251 deletion completed in 6.781348767s +Jun 18 12:56:09.051: INFO: Waiting for pod pod-projected-configmaps-6eba7fc9-91c8-11e9-a25d-8608290c688a to disappear +Jun 18 12:56:09.067: INFO: Pod pod-projected-configmaps-6eba7fc9-91c8-11e9-a25d-8608290c688a no longer exists +[AfterEach] [sig-storage] Projected configMap + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Jun 18 12:56:09.067: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "projected-2584" for this suite. +Jun 18 12:56:15.110: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:56:16.970: INFO: namespace projected-2584 deletion completed in 7.886644773s -• [SLOW TEST:9.187 seconds] -[sig-storage] Secrets -/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/secrets_volume.go:33 - should be consumable from pods in volume [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +• [SLOW TEST:12.330 seconds] +[sig-storage] Projected configMap +/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_configmap.go:33 + should be consumable from pods in volume with mappings as non-root [LinuxOnly] [NodeConformance] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 ------------------------------ SSSS ------------------------------ -[sig-node] Downward API - should provide default limits.cpu/memory from node allocatable [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -[BeforeEach] [sig-node] Downward API - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +[sig-storage] ConfigMap + updates should be reflected in volume [NodeConformance] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [sig-storage] ConfigMap + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 STEP: Creating a kubernetes client -Jun 4 19:44:05.504: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504 -STEP: Building a namespace api object, basename downward-api -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in downward-api-7381 +Jun 18 12:56:16.971: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150 +STEP: Building a namespace api object, basename configmap +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in configmap-6872 STEP: Waiting for a default service account to be provisioned in namespace -[It] should provide default limits.cpu/memory from node allocatable [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -STEP: Creating a pod to test downward api env vars -Jun 4 19:44:05.755: INFO: Waiting up to 5m0s for pod "downward-api-1cad9bb8-8701-11e9-8d1b-467ee19922ac" in namespace "downward-api-7381" to be "success or failure" -Jun 4 19:44:05.765: INFO: Pod "downward-api-1cad9bb8-8701-11e9-8d1b-467ee19922ac": Phase="Pending", Reason="", readiness=false. Elapsed: 10.158604ms -Jun 4 19:44:07.778: INFO: Pod "downward-api-1cad9bb8-8701-11e9-8d1b-467ee19922ac": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.022856577s -STEP: Saw pod success -Jun 4 19:44:07.778: INFO: Pod "downward-api-1cad9bb8-8701-11e9-8d1b-467ee19922ac" satisfied condition "success or failure" -Jun 4 19:44:07.790: INFO: Trying to get logs from node 10.212.23.189 pod downward-api-1cad9bb8-8701-11e9-8d1b-467ee19922ac container dapi-container: -STEP: delete the pod -Jun 4 19:44:07.890: INFO: Waiting for pod downward-api-1cad9bb8-8701-11e9-8d1b-467ee19922ac to disappear -Jun 4 19:44:07.901: INFO: Pod downward-api-1cad9bb8-8701-11e9-8d1b-467ee19922ac no longer exists -[AfterEach] [sig-node] Downward API - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 -Jun 4 19:44:07.901: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "downward-api-7381" for this suite. -Jun 4 19:44:13.960: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -Jun 4 19:44:14.349: INFO: namespace downward-api-7381 deletion completed in 6.432800778s +[It] updates should be reflected in volume [NodeConformance] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: Creating configMap with name configmap-test-upd-76115999-91c8-11e9-a25d-8608290c688a +STEP: Creating the pod +STEP: Updating configmap configmap-test-upd-76115999-91c8-11e9-a25d-8608290c688a +STEP: waiting to observe update in volume +[AfterEach] [sig-storage] ConfigMap + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Jun 18 12:57:35.951: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "configmap-6872" for this suite. +Jun 18 12:58:00.007: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:58:00.393: INFO: namespace configmap-6872 deletion completed in 24.412495211s -• [SLOW TEST:8.844 seconds] -[sig-node] Downward API -/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downward_api.go:38 - should provide default limits.cpu/memory from node allocatable [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +• [SLOW TEST:103.423 seconds] +[sig-storage] ConfigMap +/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:32 + updates should be reflected in volume [NodeConformance] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 ------------------------------ -SSS +S ------------------------------ -[k8s.io] Kubelet when scheduling a busybox Pod with hostAliases - should write entries to /etc/hosts [LinuxOnly] [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -[BeforeEach] [k8s.io] Kubelet - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +[sig-cli] Kubectl client [k8s.io] Kubectl run job + should create a job from an image when restart is OnFailure [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [sig-cli] Kubectl client + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 STEP: Creating a kubernetes client -Jun 4 19:44:14.349: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504 -STEP: Building a namespace api object, basename kubelet-test -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubelet-test-3021 +Jun 18 12:58:00.394: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150 +STEP: Building a namespace api object, basename kubectl +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-3120 STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [k8s.io] Kubelet - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:37 -[It] should write entries to /etc/hosts [LinuxOnly] [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -[AfterEach] [k8s.io] Kubelet - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 -Jun 4 19:44:16.751: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "kubelet-test-3021" for this suite. -Jun 4 19:44:58.901: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -Jun 4 19:44:59.241: INFO: namespace kubelet-test-3021 deletion completed in 42.385778844s +[BeforeEach] [sig-cli] Kubectl client + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:213 +[BeforeEach] [k8s.io] Kubectl run job + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1510 +[It] should create a job from an image when restart is OnFailure [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: running the image docker.io/library/nginx:1.14-alpine +Jun 18 12:58:00.595: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 run e2e-test-nginx-job --restart=OnFailure --generator=job/v1 --image=docker.io/library/nginx:1.14-alpine --namespace=kubectl-3120' +Jun 18 12:58:00.866: INFO: stderr: "kubectl run --generator=job/v1 is DEPRECATED and will be removed in a future version. Use kubectl run --generator=run-pod/v1 or kubectl create instead.\n" +Jun 18 12:58:00.866: INFO: stdout: "job.batch/e2e-test-nginx-job created\n" +STEP: verifying the job e2e-test-nginx-job was created +[AfterEach] [k8s.io] Kubectl run job + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1515 +Jun 18 12:58:00.879: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 delete jobs e2e-test-nginx-job --namespace=kubectl-3120' +Jun 18 12:58:02.198: INFO: stderr: "" +Jun 18 12:58:02.198: INFO: stdout: "job.batch \"e2e-test-nginx-job\" deleted\n" +[AfterEach] [sig-cli] Kubectl client + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Jun 18 12:58:02.198: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "kubectl-3120" for this suite. +Jun 18 12:58:10.700: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:58:11.191: INFO: namespace kubectl-3120 deletion completed in 8.973852075s -• [SLOW TEST:44.892 seconds] -[k8s.io] Kubelet -/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687 - when scheduling a busybox Pod with hostAliases - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:136 - should write entries to /etc/hosts [LinuxOnly] [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +• [SLOW TEST:10.797 seconds] +[sig-cli] Kubectl client +/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 + [k8s.io] Kubectl run job + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687 + should create a job from an image when restart is OnFailure [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSSSS +SSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-network] Services - should serve a basic endpoint from pods [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -[BeforeEach] [sig-network] Services - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +[sig-storage] Downward API volume + should update labels on modification [NodeConformance] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [sig-storage] Downward API volume + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 STEP: Creating a kubernetes client -Jun 4 19:44:59.242: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504 -STEP: Building a namespace api object, basename services -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in services-2015 +Jun 18 12:58:11.192: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150 +STEP: Building a namespace api object, basename downward-api +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in downward-api-3111 STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [sig-network] Services - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:86 -[It] should serve a basic endpoint from pods [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -STEP: creating service endpoint-test2 in namespace services-2015 -STEP: waiting up to 3m0s for service endpoint-test2 in namespace services-2015 to expose endpoints map[] -Jun 4 19:44:59.513: INFO: successfully validated that service endpoint-test2 in namespace services-2015 exposes endpoints map[] (10.431323ms elapsed) -STEP: Creating pod pod1 in namespace services-2015 -STEP: waiting up to 3m0s for service endpoint-test2 in namespace services-2015 to expose endpoints map[pod1:[80]] -Jun 4 19:45:01.667: INFO: successfully validated that service endpoint-test2 in namespace services-2015 exposes endpoints map[pod1:[80]] (2.135383836s elapsed) -STEP: Creating pod pod2 in namespace services-2015 -STEP: waiting up to 3m0s for service endpoint-test2 in namespace services-2015 to expose endpoints map[pod1:[80] pod2:[80]] -Jun 4 19:45:03.786: INFO: successfully validated that service endpoint-test2 in namespace services-2015 exposes endpoints map[pod1:[80] pod2:[80]] (2.104563464s elapsed) -STEP: Deleting pod pod1 in namespace services-2015 -STEP: waiting up to 3m0s for service endpoint-test2 in namespace services-2015 to expose endpoints map[pod2:[80]] -Jun 4 19:45:03.832: INFO: successfully validated that service endpoint-test2 in namespace services-2015 exposes endpoints map[pod2:[80]] (26.143211ms elapsed) -STEP: Deleting pod pod2 in namespace services-2015 -STEP: waiting up to 3m0s for service endpoint-test2 in namespace services-2015 to expose endpoints map[] -Jun 4 19:45:03.864: INFO: successfully validated that service endpoint-test2 in namespace services-2015 exposes endpoints map[] (10.339696ms elapsed) -[AfterEach] [sig-network] Services - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 -Jun 4 19:45:03.926: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "services-2015" for this suite. -Jun 4 19:45:27.993: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -Jun 4 19:45:28.489: INFO: namespace services-2015 deletion completed in 24.538141033s -[AfterEach] [sig-network] Services - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:91 +[BeforeEach] [sig-storage] Downward API volume + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:39 +[It] should update labels on modification [NodeConformance] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: Creating the pod +Jun 18 12:58:16.047: INFO: Successfully updated pod "labelsupdateba23d6c3-91c8-11e9-a25d-8608290c688a" +[AfterEach] [sig-storage] Downward API volume + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Jun 18 12:58:18.125: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "downward-api-3111" for this suite. +Jun 18 12:58:42.187: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:58:42.585: INFO: namespace downward-api-3111 deletion completed in 24.435163099s -• [SLOW TEST:29.247 seconds] -[sig-network] Services -/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:22 - should serve a basic endpoint from pods [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +• [SLOW TEST:31.393 seconds] +[sig-storage] Downward API volume +/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:34 + should update labels on modification [NodeConformance] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 ------------------------------ -SSSSS +SSSSSSSS ------------------------------ -[sig-network] Networking Granular Checks: Pods - should function for node-pod communication: udp [LinuxOnly] [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -[BeforeEach] [sig-network] Networking - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +[k8s.io] Pods + should support retrieving logs from the container over websockets [NodeConformance] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [k8s.io] Pods + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 STEP: Creating a kubernetes client -Jun 4 19:45:28.490: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504 -STEP: Building a namespace api object, basename pod-network-test -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in pod-network-test-5364 +Jun 18 12:58:42.585: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150 +STEP: Building a namespace api object, basename pods +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in pods-5502 STEP: Waiting for a default service account to be provisioned in namespace -[It] should function for node-pod communication: udp [LinuxOnly] [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -STEP: Performing setup for networking test in namespace pod-network-test-5364 -STEP: creating a selector -STEP: Creating the service pods in kubernetes -Jun 4 19:45:28.759: INFO: Waiting up to 10m0s for all (but 0) nodes to be schedulable -STEP: Creating test pods -Jun 4 19:45:53.154: INFO: ExecWithOptions {Command:[/bin/sh -c echo hostName | nc -w 1 -u 172.30.254.177 8081 | grep -v '^\s*$'] Namespace:pod-network-test-5364 PodName:host-test-container-pod ContainerName:hostexec Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} -Jun 4 19:45:53.154: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504 -Jun 4 19:45:54.355: INFO: Found all expected endpoints: [netserver-0] -Jun 4 19:45:54.368: INFO: ExecWithOptions {Command:[/bin/sh -c echo hostName | nc -w 1 -u 172.30.169.186 8081 | grep -v '^\s*$'] Namespace:pod-network-test-5364 PodName:host-test-container-pod ContainerName:hostexec Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} -Jun 4 19:45:54.368: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504 -Jun 4 19:45:55.558: INFO: Found all expected endpoints: [netserver-1] -Jun 4 19:45:55.571: INFO: ExecWithOptions {Command:[/bin/sh -c echo hostName | nc -w 1 -u 172.30.193.39 8081 | grep -v '^\s*$'] Namespace:pod-network-test-5364 PodName:host-test-container-pod ContainerName:hostexec Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} -Jun 4 19:45:55.571: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504 -Jun 4 19:45:56.740: INFO: Found all expected endpoints: [netserver-2] -[AfterEach] [sig-network] Networking - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 -Jun 4 19:45:56.740: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "pod-network-test-5364" for this suite. -Jun 4 19:46:20.804: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -Jun 4 19:46:21.155: INFO: namespace pod-network-test-5364 deletion completed in 24.396968496s +[BeforeEach] [k8s.io] Pods + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:135 +[It] should support retrieving logs from the container over websockets [NodeConformance] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +Jun 18 12:58:42.778: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150 +STEP: creating the pod +STEP: submitting the pod to kubernetes +[AfterEach] [k8s.io] Pods + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Jun 18 12:58:44.929: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "pods-5502" for this suite. +Jun 18 12:59:24.974: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 12:59:25.359: INFO: namespace pods-5502 deletion completed in 40.414181965s -• [SLOW TEST:52.665 seconds] -[sig-network] Networking -/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:25 - Granular Checks: Pods - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:28 - should function for node-pod communication: udp [LinuxOnly] [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +• [SLOW TEST:42.774 seconds] +[k8s.io] Pods +/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687 + should support retrieving logs from the container over websockets [NodeConformance] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 ------------------------------ -SSSSSSSS +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-cli] Kubectl client [k8s.io] Kubectl run rc - should create an rc from an image [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[sig-cli] Kubectl client [k8s.io] Update Demo + should scale a replication controller [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 [BeforeEach] [sig-cli] Kubectl client - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 STEP: Creating a kubernetes client -Jun 4 19:46:21.155: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504 +Jun 18 12:59:25.360: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150 STEP: Building a namespace api object, basename kubectl -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-5700 +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-5536 STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:213 -[BeforeEach] [k8s.io] Kubectl run rc - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1354 -[It] should create an rc from an image [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -STEP: running the image docker.io/library/nginx:1.14-alpine -Jun 4 19:46:21.366: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 run e2e-test-nginx-rc --image=docker.io/library/nginx:1.14-alpine --generator=run/v1 --namespace=kubectl-5700' -Jun 4 19:46:21.469: INFO: stderr: "kubectl run --generator=run/v1 is DEPRECATED and will be removed in a future version. Use kubectl run --generator=run-pod/v1 or kubectl create instead.\n" -Jun 4 19:46:21.469: INFO: stdout: "replicationcontroller/e2e-test-nginx-rc created\n" -STEP: verifying the rc e2e-test-nginx-rc was created -STEP: verifying the pod controlled by rc e2e-test-nginx-rc was created -STEP: confirm that you can get logs from an rc -Jun 4 19:46:21.494: INFO: Waiting up to 5m0s for 1 pods to be running and ready: [e2e-test-nginx-rc-7d2nb] -Jun 4 19:46:21.494: INFO: Waiting up to 5m0s for pod "e2e-test-nginx-rc-7d2nb" in namespace "kubectl-5700" to be "running and ready" -Jun 4 19:46:21.508: INFO: Pod "e2e-test-nginx-rc-7d2nb": Phase="Pending", Reason="", readiness=false. Elapsed: 13.907912ms -Jun 4 19:46:23.522: INFO: Pod "e2e-test-nginx-rc-7d2nb": Phase="Running", Reason="", readiness=true. Elapsed: 2.028683161s -Jun 4 19:46:23.522: INFO: Pod "e2e-test-nginx-rc-7d2nb" satisfied condition "running and ready" -Jun 4 19:46:23.522: INFO: Wanted all 1 pods to be running and ready. Result: true. Pods: [e2e-test-nginx-rc-7d2nb] -Jun 4 19:46:23.522: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 logs rc/e2e-test-nginx-rc --namespace=kubectl-5700' -Jun 4 19:46:23.640: INFO: stderr: "" -Jun 4 19:46:23.640: INFO: stdout: "" -[AfterEach] [k8s.io] Kubectl run rc - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1359 -Jun 4 19:46:23.640: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 delete rc e2e-test-nginx-rc --namespace=kubectl-5700' -Jun 4 19:46:23.749: INFO: stderr: "" -Jun 4 19:46:23.749: INFO: stdout: "replicationcontroller \"e2e-test-nginx-rc\" deleted\n" + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:213 +[BeforeEach] [k8s.io] Update Demo + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:265 +[It] should scale a replication controller [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: creating a replication controller +Jun 18 12:59:25.553: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 create -f - --namespace=kubectl-5536' +Jun 18 12:59:26.032: INFO: stderr: "" +Jun 18 12:59:26.032: INFO: stdout: "replicationcontroller/update-demo-nautilus created\n" +STEP: waiting for all containers in name=update-demo pods to come up. +Jun 18 12:59:26.032: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-5536' +Jun 18 12:59:26.171: INFO: stderr: "" +Jun 18 12:59:26.171: INFO: stdout: "update-demo-nautilus-crtpx update-demo-nautilus-vlrbb " +Jun 18 12:59:26.171: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 get pods update-demo-nautilus-crtpx -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-5536' +Jun 18 12:59:26.299: INFO: stderr: "" +Jun 18 12:59:26.299: INFO: stdout: "" +Jun 18 12:59:26.299: INFO: update-demo-nautilus-crtpx is created but not running +Jun 18 12:59:31.299: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-5536' +Jun 18 12:59:31.443: INFO: stderr: "" +Jun 18 12:59:31.443: INFO: stdout: "update-demo-nautilus-crtpx update-demo-nautilus-vlrbb " +Jun 18 12:59:31.443: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 get pods update-demo-nautilus-crtpx -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-5536' +Jun 18 12:59:31.573: INFO: stderr: "" +Jun 18 12:59:31.573: INFO: stdout: "true" +Jun 18 12:59:31.573: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 get pods update-demo-nautilus-crtpx -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-5536' +Jun 18 12:59:31.711: INFO: stderr: "" +Jun 18 12:59:31.711: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0" +Jun 18 12:59:31.711: INFO: validating pod update-demo-nautilus-crtpx +Jun 18 12:59:31.738: INFO: got data: { + "image": "nautilus.jpg" +} + +Jun 18 12:59:31.738: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . +Jun 18 12:59:31.738: INFO: update-demo-nautilus-crtpx is verified up and running +Jun 18 12:59:31.738: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 get pods update-demo-nautilus-vlrbb -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-5536' +Jun 18 12:59:31.867: INFO: stderr: "" +Jun 18 12:59:31.867: INFO: stdout: "true" +Jun 18 12:59:31.867: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 get pods update-demo-nautilus-vlrbb -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-5536' +Jun 18 12:59:32.001: INFO: stderr: "" +Jun 18 12:59:32.001: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0" +Jun 18 12:59:32.001: INFO: validating pod update-demo-nautilus-vlrbb +Jun 18 12:59:32.027: INFO: got data: { + "image": "nautilus.jpg" +} + +Jun 18 12:59:32.027: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . +Jun 18 12:59:32.027: INFO: update-demo-nautilus-vlrbb is verified up and running +STEP: scaling down the replication controller +Jun 18 12:59:32.029: INFO: scanned /root for discovery docs: +Jun 18 12:59:32.029: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 scale rc update-demo-nautilus --replicas=1 --timeout=5m --namespace=kubectl-5536' +Jun 18 12:59:33.329: INFO: stderr: "" +Jun 18 12:59:33.329: INFO: stdout: "replicationcontroller/update-demo-nautilus scaled\n" +STEP: waiting for all containers in name=update-demo pods to come up. +Jun 18 12:59:33.329: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-5536' +Jun 18 12:59:33.452: INFO: stderr: "" +Jun 18 12:59:33.452: INFO: stdout: "update-demo-nautilus-crtpx update-demo-nautilus-vlrbb " +STEP: Replicas for name=update-demo: expected=1 actual=2 +Jun 18 12:59:38.452: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-5536' +Jun 18 12:59:45.823: INFO: stderr: "" +Jun 18 12:59:45.823: INFO: stdout: "update-demo-nautilus-vlrbb " +Jun 18 12:59:45.823: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 get pods update-demo-nautilus-vlrbb -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-5536' +Jun 18 12:59:45.933: INFO: stderr: "" +Jun 18 12:59:45.933: INFO: stdout: "true" +Jun 18 12:59:45.933: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 get pods update-demo-nautilus-vlrbb -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-5536' +Jun 18 12:59:46.571: INFO: stderr: "" +Jun 18 12:59:46.571: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0" +Jun 18 12:59:46.571: INFO: validating pod update-demo-nautilus-vlrbb +Jun 18 12:59:46.592: INFO: got data: { + "image": "nautilus.jpg" +} + +Jun 18 12:59:46.592: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . +Jun 18 12:59:46.592: INFO: update-demo-nautilus-vlrbb is verified up and running +STEP: scaling up the replication controller +Jun 18 12:59:46.594: INFO: scanned /root for discovery docs: +Jun 18 12:59:46.594: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 scale rc update-demo-nautilus --replicas=2 --timeout=5m --namespace=kubectl-5536' +Jun 18 12:59:48.206: INFO: stderr: "" +Jun 18 12:59:48.206: INFO: stdout: "replicationcontroller/update-demo-nautilus scaled\n" +STEP: waiting for all containers in name=update-demo pods to come up. +Jun 18 12:59:48.206: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-5536' +Jun 18 12:59:48.340: INFO: stderr: "" +Jun 18 12:59:48.340: INFO: stdout: "update-demo-nautilus-vlrbb update-demo-nautilus-x5hx9 " +Jun 18 12:59:48.340: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 get pods update-demo-nautilus-vlrbb -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-5536' +Jun 18 12:59:48.463: INFO: stderr: "" +Jun 18 12:59:48.463: INFO: stdout: "true" +Jun 18 12:59:48.463: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 get pods update-demo-nautilus-vlrbb -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-5536' +Jun 18 12:59:48.581: INFO: stderr: "" +Jun 18 12:59:48.581: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0" +Jun 18 12:59:48.581: INFO: validating pod update-demo-nautilus-vlrbb +Jun 18 12:59:48.602: INFO: got data: { + "image": "nautilus.jpg" +} + +Jun 18 12:59:48.602: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . +Jun 18 12:59:48.602: INFO: update-demo-nautilus-vlrbb is verified up and running +Jun 18 12:59:48.602: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 get pods update-demo-nautilus-x5hx9 -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-5536' +Jun 18 12:59:48.721: INFO: stderr: "" +Jun 18 12:59:48.721: INFO: stdout: "true" +Jun 18 12:59:48.721: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 get pods update-demo-nautilus-x5hx9 -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-5536' +Jun 18 12:59:48.845: INFO: stderr: "" +Jun 18 12:59:48.845: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0" +Jun 18 12:59:48.845: INFO: validating pod update-demo-nautilus-x5hx9 +Jun 18 12:59:48.873: INFO: got data: { + "image": "nautilus.jpg" +} + +Jun 18 12:59:48.873: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . +Jun 18 12:59:48.873: INFO: update-demo-nautilus-x5hx9 is verified up and running +STEP: using delete to clean up resources +Jun 18 12:59:48.874: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 delete --grace-period=0 --force -f - --namespace=kubectl-5536' +Jun 18 12:59:49.032: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" +Jun 18 12:59:49.032: INFO: stdout: "replicationcontroller \"update-demo-nautilus\" force deleted\n" +Jun 18 12:59:49.032: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 get rc,svc -l name=update-demo --no-headers --namespace=kubectl-5536' +Jun 18 12:59:49.166: INFO: stderr: "No resources found.\n" +Jun 18 12:59:49.166: INFO: stdout: "" +Jun 18 12:59:49.166: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 get pods -l name=update-demo --namespace=kubectl-5536 -o go-template={{ range .items }}{{ if not .metadata.deletionTimestamp }}{{ .metadata.name }}{{ "\n" }}{{ end }}{{ end }}' +Jun 18 12:59:49.317: INFO: stderr: "" +Jun 18 12:59:49.317: INFO: stdout: "" [AfterEach] [sig-cli] Kubectl client - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 -Jun 4 19:46:23.749: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "kubectl-5700" for this suite. -Jun 4 19:46:47.807: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -Jun 4 19:46:48.156: INFO: namespace kubectl-5700 deletion completed in 24.391967535s + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Jun 18 12:59:49.317: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "kubectl-5536" for this suite. +Jun 18 13:00:13.356: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 13:00:13.754: INFO: namespace kubectl-5536 deletion completed in 24.423602091s -• [SLOW TEST:27.002 seconds] +• [SLOW TEST:48.394 seconds] [sig-cli] Kubectl client -/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 - [k8s.io] Kubectl run rc - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687 - should create an rc from an image [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 ------------------------------- -SSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------- -[sig-api-machinery] Namespaces [Serial] - should ensure that all pods are removed when a namespace is deleted [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -[BeforeEach] [sig-api-machinery] Namespaces [Serial] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 -STEP: Creating a kubernetes client -Jun 4 19:46:48.157: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504 -STEP: Building a namespace api object, basename namespaces -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in namespaces-4516 -STEP: Waiting for a default service account to be provisioned in namespace -[It] should ensure that all pods are removed when a namespace is deleted [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -STEP: Creating a test namespace -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in nsdeletetest-2772 -STEP: Waiting for a default service account to be provisioned in namespace -STEP: Creating a pod in the namespace -STEP: Waiting for the pod to have running status -STEP: Deleting the namespace -STEP: Waiting for the namespace to be removed. -STEP: Recreating the namespace -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in nsdeletetest-5574 -STEP: Verifying there are no pods in the namespace -[AfterEach] [sig-api-machinery] Namespaces [Serial] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 -Jun 4 19:47:14.093: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "namespaces-4516" for this suite. -Jun 4 19:47:20.217: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -Jun 4 19:47:20.611: INFO: namespace namespaces-4516 deletion completed in 6.438719769s -STEP: Destroying namespace "nsdeletetest-2772" for this suite. -Jun 4 19:47:20.621: INFO: Namespace nsdeletetest-2772 was already deleted -STEP: Destroying namespace "nsdeletetest-5574" for this suite. -Jun 4 19:47:26.678: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -Jun 4 19:47:27.120: INFO: namespace nsdeletetest-5574 deletion completed in 6.498994872s - -• [SLOW TEST:38.963 seconds] -[sig-api-machinery] Namespaces [Serial] -/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:22 - should ensure that all pods are removed when a namespace is deleted [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 + [k8s.io] Update Demo + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687 + should scale a replication controller [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 ------------------------------ -SSSSSSS +SSSS ------------------------------ [sig-cli] Kubectl client [k8s.io] Kubectl run default should create an rc or deployment from an image [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 [BeforeEach] [sig-cli] Kubectl client - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 STEP: Creating a kubernetes client -Jun 4 19:47:27.121: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504 +Jun 18 13:00:13.756: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150 STEP: Building a namespace api object, basename kubectl -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-1853 +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-7643 STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:213 + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:213 [BeforeEach] [k8s.io] Kubectl run default - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1318 + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1318 [It] should create an rc or deployment from an image [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 STEP: running the image docker.io/library/nginx:1.14-alpine -Jun 4 19:47:27.341: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 run e2e-test-nginx-deployment --image=docker.io/library/nginx:1.14-alpine --namespace=kubectl-1853' -Jun 4 19:47:27.459: INFO: stderr: "kubectl run --generator=deployment/apps.v1 is DEPRECATED and will be removed in a future version. Use kubectl run --generator=run-pod/v1 or kubectl create instead.\n" -Jun 4 19:47:27.459: INFO: stdout: "deployment.apps/e2e-test-nginx-deployment created\n" +Jun 18 13:00:13.950: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 run e2e-test-nginx-deployment --image=docker.io/library/nginx:1.14-alpine --namespace=kubectl-7643' +Jun 18 13:00:14.118: INFO: stderr: "kubectl run --generator=deployment/apps.v1 is DEPRECATED and will be removed in a future version. Use kubectl run --generator=run-pod/v1 or kubectl create instead.\n" +Jun 18 13:00:14.118: INFO: stdout: "deployment.apps/e2e-test-nginx-deployment created\n" STEP: verifying the pod controlled by e2e-test-nginx-deployment gets created [AfterEach] [k8s.io] Kubectl run default - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1324 -Jun 4 19:47:29.487: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 delete deployment e2e-test-nginx-deployment --namespace=kubectl-1853' -Jun 4 19:47:29.600: INFO: stderr: "" -Jun 4 19:47:29.600: INFO: stdout: "deployment.extensions \"e2e-test-nginx-deployment\" deleted\n" + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1324 +Jun 18 13:00:16.146: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 delete deployment e2e-test-nginx-deployment --namespace=kubectl-7643' +Jun 18 13:00:16.297: INFO: stderr: "" +Jun 18 13:00:16.297: INFO: stdout: "deployment.extensions \"e2e-test-nginx-deployment\" deleted\n" [AfterEach] [sig-cli] Kubectl client - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 -Jun 4 19:47:29.600: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "kubectl-1853" for this suite. -Jun 4 19:47:53.660: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -Jun 4 19:47:54.015: INFO: namespace kubectl-1853 deletion completed in 24.398074017s + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Jun 18 13:00:16.297: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "kubectl-7643" for this suite. +Jun 18 13:00:40.344: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 13:00:41.336: INFO: namespace kubectl-7643 deletion completed in 25.018235781s -• [SLOW TEST:26.894 seconds] +• [SLOW TEST:27.580 seconds] [sig-cli] Kubectl client -/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 +/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 [k8s.io] Kubectl run default - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687 + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687 should create an rc or deployment from an image [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 ------------------------------ -SSSSSSSSSSSS +S ------------------------------ [sig-storage] EmptyDir volumes - should support (non-root,0666,default) [LinuxOnly] [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 + should support (non-root,0644,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 [BeforeEach] [sig-storage] EmptyDir volumes - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 STEP: Creating a kubernetes client -Jun 4 19:47:54.015: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504 +Jun 18 13:00:41.336: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150 STEP: Building a namespace api object, basename emptydir -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in emptydir-5364 +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in emptydir-6508 STEP: Waiting for a default service account to be provisioned in namespace -[It] should support (non-root,0666,default) [LinuxOnly] [NodeConformance] [Conformance] - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 -STEP: Creating a pod to test emptydir 0666 on node default medium -Jun 4 19:47:54.277: INFO: Waiting up to 5m0s for pod "pod-a4defd55-8701-11e9-8d1b-467ee19922ac" in namespace "emptydir-5364" to be "success or failure" -Jun 4 19:47:54.290: INFO: Pod "pod-a4defd55-8701-11e9-8d1b-467ee19922ac": Phase="Pending", Reason="", readiness=false. Elapsed: 12.965851ms -Jun 4 19:47:56.302: INFO: Pod "pod-a4defd55-8701-11e9-8d1b-467ee19922ac": Phase="Pending", Reason="", readiness=false. Elapsed: 2.024816723s -Jun 4 19:47:58.316: INFO: Pod "pod-a4defd55-8701-11e9-8d1b-467ee19922ac": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.038985893s +[It] should support (non-root,0644,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: Creating a pod to test emptydir 0644 on tmpfs +Jun 18 13:00:41.554: INFO: Waiting up to 5m0s for pod "pod-13a0b8f0-91c9-11e9-a25d-8608290c688a" in namespace "emptydir-6508" to be "success or failure" +Jun 18 13:00:41.570: INFO: Pod "pod-13a0b8f0-91c9-11e9-a25d-8608290c688a": Phase="Pending", Reason="", readiness=false. Elapsed: 15.536589ms +Jun 18 13:00:43.584: INFO: Pod "pod-13a0b8f0-91c9-11e9-a25d-8608290c688a": Phase="Pending", Reason="", readiness=false. Elapsed: 2.029493594s +Jun 18 13:00:45.598: INFO: Pod "pod-13a0b8f0-91c9-11e9-a25d-8608290c688a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.044135755s STEP: Saw pod success -Jun 4 19:47:58.316: INFO: Pod "pod-a4defd55-8701-11e9-8d1b-467ee19922ac" satisfied condition "success or failure" -Jun 4 19:47:58.328: INFO: Trying to get logs from node 10.212.23.164 pod pod-a4defd55-8701-11e9-8d1b-467ee19922ac container test-container: +Jun 18 13:00:45.598: INFO: Pod "pod-13a0b8f0-91c9-11e9-a25d-8608290c688a" satisfied condition "success or failure" +Jun 18 13:00:45.611: INFO: Trying to get logs from node 10.72.74.184 pod pod-13a0b8f0-91c9-11e9-a25d-8608290c688a container test-container: STEP: delete the pod -Jun 4 19:47:58.421: INFO: Waiting for pod pod-a4defd55-8701-11e9-8d1b-467ee19922ac to disappear -Jun 4 19:47:58.432: INFO: Pod pod-a4defd55-8701-11e9-8d1b-467ee19922ac no longer exists +Jun 18 13:00:45.696: INFO: Waiting for pod pod-13a0b8f0-91c9-11e9-a25d-8608290c688a to disappear +Jun 18 13:00:45.708: INFO: Pod pod-13a0b8f0-91c9-11e9-a25d-8608290c688a no longer exists [AfterEach] [sig-storage] EmptyDir volumes - /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 -Jun 4 19:47:58.432: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "emptydir-5364" for this suite. -Jun 4 19:48:04.491: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered -Jun 4 19:48:04.844: INFO: namespace emptydir-5364 deletion completed in 6.396998368s + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Jun 18 13:00:45.708: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "emptydir-6508" for this suite. +Jun 18 13:00:51.747: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Jun 18 13:00:52.153: INFO: namespace emptydir-6508 deletion completed in 6.43251356s + +• [SLOW TEST:10.818 seconds] +[sig-storage] EmptyDir volumes +/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:41 + should support (non-root,0644,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +------------------------------ +SSSSSSSS +------------------------------ +[sig-network] Proxy version v1 + should proxy logs on node with explicit kubelet port using proxy subresource [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] version v1 + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +STEP: Creating a kubernetes client +Jun 18 13:00:52.154: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150 +STEP: Building a namespace api object, basename proxy +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in proxy-5062 +STEP: Waiting for a default service account to be provisioned in namespace +[It] should proxy logs on node with explicit kubelet port using proxy subresource [Conformance] + /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +Jun 18 13:00:52.390: INFO: (0) /api/v1/nodes/10.72.74.138:10250/proxy/logs/:
+alb/
+alternatives.log
+apt/... (200; 24.871798ms)
+Jun 18 13:00:52.407: INFO: (1) /api/v1/nodes/10.72.74.138:10250/proxy/logs/: 
+alb/
+alternatives.log
+apt/... (200; 17.281225ms)
+Jun 18 13:00:52.425: INFO: (2) /api/v1/nodes/10.72.74.138:10250/proxy/logs/: 
+alb/
+alternatives.log
+apt/... (200; 17.962816ms)
+Jun 18 13:00:52.442: INFO: (3) /api/v1/nodes/10.72.74.138:10250/proxy/logs/: 
+alb/
+alternatives.log
+apt/... (200; 17.470045ms)
+Jun 18 13:00:52.461: INFO: (4) /api/v1/nodes/10.72.74.138:10250/proxy/logs/: 
+alb/
+alternatives.log
+apt/... (200; 18.812765ms)
+Jun 18 13:00:52.480: INFO: (5) /api/v1/nodes/10.72.74.138:10250/proxy/logs/: 
+alb/
+alternatives.log
+apt/... (200; 18.640074ms)
+Jun 18 13:00:52.499: INFO: (6) /api/v1/nodes/10.72.74.138:10250/proxy/logs/: 
+alb/
+alternatives.log
+apt/... (200; 18.807068ms)
+Jun 18 13:00:52.516: INFO: (7) /api/v1/nodes/10.72.74.138:10250/proxy/logs/: 
+alb/
+alternatives.log
+apt/... (200; 17.296595ms)
+Jun 18 13:00:52.537: INFO: (8) /api/v1/nodes/10.72.74.138:10250/proxy/logs/: 
+alb/
+alternatives.log
+apt/... (200; 20.529995ms)
+Jun 18 13:00:52.559: INFO: (9) /api/v1/nodes/10.72.74.138:10250/proxy/logs/: 
+alb/
+alternatives.log
+apt/... (200; 21.994434ms)
+Jun 18 13:00:52.582: INFO: (10) /api/v1/nodes/10.72.74.138:10250/proxy/logs/: 
+alb/
+alternatives.log
+apt/... (200; 22.994367ms)
+Jun 18 13:00:52.599: INFO: (11) /api/v1/nodes/10.72.74.138:10250/proxy/logs/: 
+alb/
+alternatives.log
+apt/... (200; 17.414632ms)
+Jun 18 13:00:52.617: INFO: (12) /api/v1/nodes/10.72.74.138:10250/proxy/logs/: 
+alb/
+alternatives.log
+apt/... (200; 18.034034ms)
+Jun 18 13:00:52.635: INFO: (13) /api/v1/nodes/10.72.74.138:10250/proxy/logs/: 
+alb/
+alternatives.log
+apt/... (200; 17.764701ms)
+Jun 18 13:00:52.653: INFO: (14) /api/v1/nodes/10.72.74.138:10250/proxy/logs/: 
+alb/
+alternatives.log
+apt/... (200; 18.024616ms)
+Jun 18 13:00:52.671: INFO: (15) /api/v1/nodes/10.72.74.138:10250/proxy/logs/: 
+alb/
+alternatives.log
+apt/... (200; 17.841827ms)
+Jun 18 13:00:52.689: INFO: (16) /api/v1/nodes/10.72.74.138:10250/proxy/logs/: 
+alb/
+alternatives.log
+apt/... (200; 18.027306ms)
+Jun 18 13:00:52.707: INFO: (17) /api/v1/nodes/10.72.74.138:10250/proxy/logs/: 
+alb/
+alternatives.log
+apt/... (200; 17.675494ms)
+Jun 18 13:00:52.725: INFO: (18) /api/v1/nodes/10.72.74.138:10250/proxy/logs/: 
+alb/
+alternatives.log
+apt/... (200; 18.427263ms)
+Jun 18 13:00:52.743: INFO: (19) /api/v1/nodes/10.72.74.138:10250/proxy/logs/: 
+alb/
+alternatives.log
+apt/... (200; 17.635209ms)
+[AfterEach] version v1
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 13:00:52.743: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "proxy-5062" for this suite.
+Jun 18 13:00:58.783: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 13:00:59.490: INFO: namespace proxy-5062 deletion completed in 6.733635497s
+
+• [SLOW TEST:7.337 seconds]
+[sig-network] Proxy
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:22
+  version v1
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/proxy.go:56
+    should proxy logs on node with explicit kubelet port using proxy subresource  [Conformance]
+    /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+------------------------------
+[sig-api-machinery] Watchers 
+  should observe add, update, and delete watch notifications on configmaps [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-api-machinery] Watchers
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Jun 18 13:00:59.492: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename watch
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in watch-8781
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should observe add, update, and delete watch notifications on configmaps [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: creating a watch on configmaps with label A
+STEP: creating a watch on configmaps with label B
+STEP: creating a watch on configmaps with label A or B
+STEP: creating a configmap with label A and ensuring the correct watchers observe the notification
+Jun 18 13:00:59.713: INFO: Got : ADDED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-configmap-a,GenerateName:,Namespace:watch-8781,SelfLink:/api/v1/namespaces/watch-8781/configmaps/e2e-watch-test-configmap-a,UID:1e7612a4-91c9-11e9-a08a-ee7a14707756,ResourceVersion:104859,Generation:0,CreationTimestamp:2019-06-18 13:00:59 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: multiple-watchers-A,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{},BinaryData:map[string][]byte{},}
+Jun 18 13:00:59.713: INFO: Got : ADDED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-configmap-a,GenerateName:,Namespace:watch-8781,SelfLink:/api/v1/namespaces/watch-8781/configmaps/e2e-watch-test-configmap-a,UID:1e7612a4-91c9-11e9-a08a-ee7a14707756,ResourceVersion:104859,Generation:0,CreationTimestamp:2019-06-18 13:00:59 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: multiple-watchers-A,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{},BinaryData:map[string][]byte{},}
+STEP: modifying configmap A and ensuring the correct watchers observe the notification
+Jun 18 13:01:09.741: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-configmap-a,GenerateName:,Namespace:watch-8781,SelfLink:/api/v1/namespaces/watch-8781/configmaps/e2e-watch-test-configmap-a,UID:1e7612a4-91c9-11e9-a08a-ee7a14707756,ResourceVersion:104876,Generation:0,CreationTimestamp:2019-06-18 13:00:59 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: multiple-watchers-A,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{mutation: 1,},BinaryData:map[string][]byte{},}
+Jun 18 13:01:09.741: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-configmap-a,GenerateName:,Namespace:watch-8781,SelfLink:/api/v1/namespaces/watch-8781/configmaps/e2e-watch-test-configmap-a,UID:1e7612a4-91c9-11e9-a08a-ee7a14707756,ResourceVersion:104876,Generation:0,CreationTimestamp:2019-06-18 13:00:59 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: multiple-watchers-A,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{mutation: 1,},BinaryData:map[string][]byte{},}
+STEP: modifying configmap A again and ensuring the correct watchers observe the notification
+Jun 18 13:01:19.769: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-configmap-a,GenerateName:,Namespace:watch-8781,SelfLink:/api/v1/namespaces/watch-8781/configmaps/e2e-watch-test-configmap-a,UID:1e7612a4-91c9-11e9-a08a-ee7a14707756,ResourceVersion:104891,Generation:0,CreationTimestamp:2019-06-18 13:00:59 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: multiple-watchers-A,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},}
+Jun 18 13:01:19.769: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-configmap-a,GenerateName:,Namespace:watch-8781,SelfLink:/api/v1/namespaces/watch-8781/configmaps/e2e-watch-test-configmap-a,UID:1e7612a4-91c9-11e9-a08a-ee7a14707756,ResourceVersion:104891,Generation:0,CreationTimestamp:2019-06-18 13:00:59 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: multiple-watchers-A,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},}
+STEP: deleting configmap A and ensuring the correct watchers observe the notification
+Jun 18 13:01:29.805: INFO: Got : DELETED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-configmap-a,GenerateName:,Namespace:watch-8781,SelfLink:/api/v1/namespaces/watch-8781/configmaps/e2e-watch-test-configmap-a,UID:1e7612a4-91c9-11e9-a08a-ee7a14707756,ResourceVersion:104908,Generation:0,CreationTimestamp:2019-06-18 13:00:59 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: multiple-watchers-A,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},}
+Jun 18 13:01:29.805: INFO: Got : DELETED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-configmap-a,GenerateName:,Namespace:watch-8781,SelfLink:/api/v1/namespaces/watch-8781/configmaps/e2e-watch-test-configmap-a,UID:1e7612a4-91c9-11e9-a08a-ee7a14707756,ResourceVersion:104908,Generation:0,CreationTimestamp:2019-06-18 13:00:59 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: multiple-watchers-A,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},}
+STEP: creating a configmap with label B and ensuring the correct watchers observe the notification
+Jun 18 13:01:39.924: INFO: Got : ADDED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-configmap-b,GenerateName:,Namespace:watch-8781,SelfLink:/api/v1/namespaces/watch-8781/configmaps/e2e-watch-test-configmap-b,UID:365e1b05-91c9-11e9-a08a-ee7a14707756,ResourceVersion:104927,Generation:0,CreationTimestamp:2019-06-18 13:01:39 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: multiple-watchers-B,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{},BinaryData:map[string][]byte{},}
+Jun 18 13:01:39.924: INFO: Got : ADDED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-configmap-b,GenerateName:,Namespace:watch-8781,SelfLink:/api/v1/namespaces/watch-8781/configmaps/e2e-watch-test-configmap-b,UID:365e1b05-91c9-11e9-a08a-ee7a14707756,ResourceVersion:104927,Generation:0,CreationTimestamp:2019-06-18 13:01:39 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: multiple-watchers-B,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{},BinaryData:map[string][]byte{},}
+STEP: deleting configmap B and ensuring the correct watchers observe the notification
+Jun 18 13:01:50.405: INFO: Got : DELETED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-configmap-b,GenerateName:,Namespace:watch-8781,SelfLink:/api/v1/namespaces/watch-8781/configmaps/e2e-watch-test-configmap-b,UID:365e1b05-91c9-11e9-a08a-ee7a14707756,ResourceVersion:104946,Generation:0,CreationTimestamp:2019-06-18 13:01:39 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: multiple-watchers-B,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{},BinaryData:map[string][]byte{},}
+Jun 18 13:01:50.405: INFO: Got : DELETED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-configmap-b,GenerateName:,Namespace:watch-8781,SelfLink:/api/v1/namespaces/watch-8781/configmaps/e2e-watch-test-configmap-b,UID:365e1b05-91c9-11e9-a08a-ee7a14707756,ResourceVersion:104946,Generation:0,CreationTimestamp:2019-06-18 13:01:39 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: multiple-watchers-B,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{},BinaryData:map[string][]byte{},}
+[AfterEach] [sig-api-machinery] Watchers
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 13:02:00.405: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "watch-8781" for this suite.
+Jun 18 13:02:06.453: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 13:02:06.858: INFO: namespace watch-8781 deletion completed in 6.430315122s
+
+• [SLOW TEST:67.366 seconds]
+[sig-api-machinery] Watchers
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:22
+  should observe add, update, and delete watch notifications on configmaps [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSS
+------------------------------
+[k8s.io] Probing container 
+  with readiness probe that fails should never be ready and never restart [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [k8s.io] Probing container
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Jun 18 13:02:06.858: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename container-probe
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in container-probe-1976
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [k8s.io] Probing container
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/container_probe.go:51
+[It] with readiness probe that fails should never be ready and never restart [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[AfterEach] [k8s.io] Probing container
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 13:03:07.111: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "container-probe-1976" for this suite.
+Jun 18 13:03:33.167: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 13:03:33.567: INFO: namespace container-probe-1976 deletion completed in 26.441750008s
+
+• [SLOW TEST:86.709 seconds]
+[k8s.io] Probing container
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
+  with readiness probe that fails should never be ready and never restart [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSSSSS
+------------------------------
+[sig-network] Networking Granular Checks: Pods 
+  should function for intra-pod communication: http [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-network] Networking
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Jun 18 13:03:33.568: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename pod-network-test
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in pod-network-test-6856
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should function for intra-pod communication: http [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Performing setup for networking test in namespace pod-network-test-6856
+STEP: creating a selector
+STEP: Creating the service pods in kubernetes
+Jun 18 13:03:33.765: INFO: Waiting up to 10m0s for all (but 0) nodes to be schedulable
+STEP: Creating test pods
+Jun 18 13:03:56.127: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s 'http://172.30.142.63:8080/dial?request=hostName&protocol=http&host=172.30.206.182&port=8080&tries=1'] Namespace:pod-network-test-6856 PodName:host-test-container-pod ContainerName:hostexec Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
+Jun 18 13:03:56.127: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+Jun 18 13:03:56.429: INFO: Waiting for endpoints: map[]
+Jun 18 13:03:56.443: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s 'http://172.30.142.63:8080/dial?request=hostName&protocol=http&host=172.30.142.62&port=8080&tries=1'] Namespace:pod-network-test-6856 PodName:host-test-container-pod ContainerName:hostexec Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
+Jun 18 13:03:56.443: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+Jun 18 13:03:56.655: INFO: Waiting for endpoints: map[]
+Jun 18 13:03:56.668: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s 'http://172.30.142.63:8080/dial?request=hostName&protocol=http&host=172.30.80.159&port=8080&tries=1'] Namespace:pod-network-test-6856 PodName:host-test-container-pod ContainerName:hostexec Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
+Jun 18 13:03:56.668: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+Jun 18 13:03:56.897: INFO: Waiting for endpoints: map[]
+[AfterEach] [sig-network] Networking
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 13:03:56.897: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "pod-network-test-6856" for this suite.
+Jun 18 13:04:12.947: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 13:04:13.359: INFO: namespace pod-network-test-6856 deletion completed in 16.448235614s
+
+• [SLOW TEST:39.791 seconds]
+[sig-network] Networking
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:25
+  Granular Checks: Pods
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:28
+    should function for intra-pod communication: http [LinuxOnly] [NodeConformance] [Conformance]
+    /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSSSSSS
+------------------------------
+[sig-apps] Deployment 
+  deployment should support proportional scaling [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-apps] Deployment
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Jun 18 13:04:13.360: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename deployment
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in deployment-3117
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [sig-apps] Deployment
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:65
+[It] deployment should support proportional scaling [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+Jun 18 13:04:13.556: INFO: Creating deployment "nginx-deployment"
+Jun 18 13:04:13.565: INFO: Waiting for observed generation 1
+Jun 18 13:04:15.579: INFO: Waiting for all required pods to come up
+Jun 18 13:04:15.595: INFO: Pod name nginx: Found 10 pods out of 10
+STEP: ensuring each pod is running
+Jun 18 13:04:17.642: INFO: Waiting for deployment "nginx-deployment" to complete
+Jun 18 13:04:17.661: INFO: Updating deployment "nginx-deployment" with a non-existent image
+Jun 18 13:04:17.679: INFO: Updating deployment nginx-deployment
+Jun 18 13:04:17.679: INFO: Waiting for observed generation 2
+Jun 18 13:04:19.705: INFO: Waiting for the first rollout's replicaset to have .status.availableReplicas = 8
+Jun 18 13:04:19.716: INFO: Waiting for the first rollout's replicaset to have .spec.replicas = 8
+Jun 18 13:04:19.729: INFO: Waiting for the first rollout's replicaset of deployment "nginx-deployment" to have desired number of replicas
+Jun 18 13:04:19.764: INFO: Verifying that the second rollout's replicaset has .status.availableReplicas = 0
+Jun 18 13:04:19.765: INFO: Waiting for the second rollout's replicaset to have .spec.replicas = 5
+Jun 18 13:04:19.775: INFO: Waiting for the second rollout's replicaset of deployment "nginx-deployment" to have desired number of replicas
+Jun 18 13:04:19.797: INFO: Verifying that deployment "nginx-deployment" has minimum required number of available replicas
+Jun 18 13:04:19.797: INFO: Scaling up the deployment "nginx-deployment" from 10 to 30
+Jun 18 13:04:19.814: INFO: Updating deployment nginx-deployment
+Jun 18 13:04:19.814: INFO: Waiting for the replicasets of deployment "nginx-deployment" to have desired number of replicas
+Jun 18 13:04:19.838: INFO: Verifying that first rollout's replicaset has .spec.replicas = 20
+Jun 18 13:04:19.856: INFO: Verifying that second rollout's replicaset has .spec.replicas = 13
+[AfterEach] [sig-apps] Deployment
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:59
+Jun 18 13:04:19.879: INFO: Deployment "nginx-deployment":
+&Deployment{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment,GenerateName:,Namespace:deployment-3117,SelfLink:/apis/apps/v1/namespaces/deployment-3117/deployments/nginx-deployment,UID:92026614-91c9-11e9-a08a-ee7a14707756,ResourceVersion:105602,Generation:3,CreationTimestamp:2019-06-18 13:04:13 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,},Annotations:map[string]string{deployment.kubernetes.io/revision: 2,},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:DeploymentSpec{Replicas:*30,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: nginx,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[],Containers:[{nginx nginx:404 [] []  [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,},},Strategy:DeploymentStrategy{Type:RollingUpdate,RollingUpdate:&RollingUpdateDeployment{MaxUnavailable:2,MaxSurge:3,},},MinReadySeconds:0,RevisionHistoryLimit:*10,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:3,Replicas:13,UpdatedReplicas:5,AvailableReplicas:8,UnavailableReplicas:25,Conditions:[{Progressing True 2019-06-18 13:04:17 +0000 UTC 2019-06-18 13:04:13 +0000 UTC ReplicaSetUpdated ReplicaSet "nginx-deployment-5f9595f595" is progressing.} {Available False 2019-06-18 13:04:19 +0000 UTC 2019-06-18 13:04:19 +0000 UTC MinimumReplicasUnavailable Deployment does not have minimum availability.}],ReadyReplicas:8,CollisionCount:nil,},}
+
+Jun 18 13:04:19.894: INFO: New ReplicaSet "nginx-deployment-5f9595f595" of Deployment "nginx-deployment":
+&ReplicaSet{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-5f9595f595,GenerateName:,Namespace:deployment-3117,SelfLink:/apis/apps/v1/namespaces/deployment-3117/replicasets/nginx-deployment-5f9595f595,UID:94775170-91c9-11e9-a08a-ee7a14707756,ResourceVersion:105595,Generation:3,CreationTimestamp:2019-06-18 13:04:17 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 5f9595f595,},Annotations:map[string]string{deployment.kubernetes.io/desired-replicas: 30,deployment.kubernetes.io/max-replicas: 33,deployment.kubernetes.io/revision: 2,},OwnerReferences:[{apps/v1 Deployment nginx-deployment 92026614-91c9-11e9-a08a-ee7a14707756 0xc002eeb6c7 0xc002eeb6c8}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:ReplicaSetSpec{Replicas:*13,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: nginx,pod-template-hash: 5f9595f595,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 5f9595f595,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[],Containers:[{nginx nginx:404 [] []  [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,},},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:5,FullyLabeledReplicas:5,ObservedGeneration:2,ReadyReplicas:0,AvailableReplicas:0,Conditions:[],},}
+Jun 18 13:04:19.894: INFO: All old ReplicaSets of Deployment "nginx-deployment":
+Jun 18 13:04:19.894: INFO: &ReplicaSet{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-6f478d8d8,GenerateName:,Namespace:deployment-3117,SelfLink:/apis/apps/v1/namespaces/deployment-3117/replicasets/nginx-deployment-6f478d8d8,UID:92038df1-91c9-11e9-a08a-ee7a14707756,ResourceVersion:105592,Generation:3,CreationTimestamp:2019-06-18 13:04:13 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 6f478d8d8,},Annotations:map[string]string{deployment.kubernetes.io/desired-replicas: 30,deployment.kubernetes.io/max-replicas: 33,deployment.kubernetes.io/revision: 1,},OwnerReferences:[{apps/v1 Deployment nginx-deployment 92026614-91c9-11e9-a08a-ee7a14707756 0xc002eeb7b7 0xc002eeb7b8}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:ReplicaSetSpec{Replicas:*20,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: nginx,pod-template-hash: 6f478d8d8,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 6f478d8d8,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] []  [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,},},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:8,FullyLabeledReplicas:8,ObservedGeneration:2,ReadyReplicas:8,AvailableReplicas:8,Conditions:[],},}
+Jun 18 13:04:19.923: INFO: Pod "nginx-deployment-5f9595f595-6zc6n" is not available:
+&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-5f9595f595-6zc6n,GenerateName:nginx-deployment-5f9595f595-,Namespace:deployment-3117,SelfLink:/api/v1/namespaces/deployment-3117/pods/nginx-deployment-5f9595f595-6zc6n,UID:9482226e-91c9-11e9-a08a-ee7a14707756,ResourceVersion:105562,Generation:0,CreationTimestamp:2019-06-18 13:04:17 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 5f9595f595,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-5f9595f595 94775170-91c9-11e9-a08a-ee7a14707756 0xc002a022d7 0xc002a022d8}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-8crrs {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-8crrs,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] []  [] [] [] {map[] map[]} [{default-token-8crrs true /var/run/secrets/kubernetes.io/serviceaccount   }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.72.74.184,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists  NoExecute 0xc002a023f0} {node.kubernetes.io/unreachable Exists  NoExecute 0xc002a02410}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 13:04:17 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 13:04:17 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 13:04:17 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 13:04:17 +0000 UTC  }],Message:,Reason:,HostIP:10.72.74.184,PodIP:,StartTime:2019-06-18 13:04:17 +0000 UTC,ContainerStatuses:[{nginx {ContainerStateWaiting{Reason:ContainerCreating,Message:,} nil nil} {nil nil nil} false 0 nginx:404  }],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},}
+Jun 18 13:04:19.923: INFO: Pod "nginx-deployment-5f9595f595-b8d5k" is not available:
+&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-5f9595f595-b8d5k,GenerateName:nginx-deployment-5f9595f595-,Namespace:deployment-3117,SelfLink:/api/v1/namespaces/deployment-3117/pods/nginx-deployment-5f9595f595-b8d5k,UID:95c71b7b-91c9-11e9-a08a-ee7a14707756,ResourceVersion:105646,Generation:0,CreationTimestamp:2019-06-18 13:04:19 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 5f9595f595,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-5f9595f595 94775170-91c9-11e9-a08a-ee7a14707756 0xc002a02550 0xc002a02551}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-8crrs {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-8crrs,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] []  [] [] [] {map[] map[]} [{default-token-8crrs true /var/run/secrets/kubernetes.io/serviceaccount   }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.72.74.189,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists  NoExecute 0xc002a025d0} {node.kubernetes.io/unreachable Exists  NoExecute 0xc002a02650}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 13:04:19 +0000 UTC  }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},}
+Jun 18 13:04:19.923: INFO: Pod "nginx-deployment-5f9595f595-dmv4p" is not available:
+&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-5f9595f595-dmv4p,GenerateName:nginx-deployment-5f9595f595-,Namespace:deployment-3117,SelfLink:/api/v1/namespaces/deployment-3117/pods/nginx-deployment-5f9595f595-dmv4p,UID:95c95862-91c9-11e9-a08a-ee7a14707756,ResourceVersion:105640,Generation:0,CreationTimestamp:2019-06-18 13:04:19 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 5f9595f595,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-5f9595f595 94775170-91c9-11e9-a08a-ee7a14707756 0xc002a02740 0xc002a02741}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-8crrs {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-8crrs,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] []  [] [] [] {map[] map[]} [{default-token-8crrs true /var/run/secrets/kubernetes.io/serviceaccount   }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists  NoExecute 0xc002a02880} {node.kubernetes.io/unreachable Exists  NoExecute 0xc002a028a0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},}
+Jun 18 13:04:19.924: INFO: Pod "nginx-deployment-5f9595f595-j6nv9" is not available:
+&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-5f9595f595-j6nv9,GenerateName:nginx-deployment-5f9595f595-,Namespace:deployment-3117,SelfLink:/api/v1/namespaces/deployment-3117/pods/nginx-deployment-5f9595f595-j6nv9,UID:947b1a75-91c9-11e9-a08a-ee7a14707756,ResourceVersion:105536,Generation:0,CreationTimestamp:2019-06-18 13:04:17 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 5f9595f595,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-5f9595f595 94775170-91c9-11e9-a08a-ee7a14707756 0xc002a02940 0xc002a02941}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-8crrs {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-8crrs,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] []  [] [] [] {map[] map[]} [{default-token-8crrs true /var/run/secrets/kubernetes.io/serviceaccount   }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.72.74.138,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists  NoExecute 0xc002a029c0} {node.kubernetes.io/unreachable Exists  NoExecute 0xc002a029e0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 13:04:17 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 13:04:17 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 13:04:17 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 13:04:17 +0000 UTC  }],Message:,Reason:,HostIP:10.72.74.138,PodIP:,StartTime:2019-06-18 13:04:17 +0000 UTC,ContainerStatuses:[{nginx {ContainerStateWaiting{Reason:ContainerCreating,Message:,} nil nil} {nil nil nil} false 0 nginx:404  }],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},}
+Jun 18 13:04:19.924: INFO: Pod "nginx-deployment-5f9595f595-lnz8w" is not available:
+&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-5f9595f595-lnz8w,GenerateName:nginx-deployment-5f9595f595-,Namespace:deployment-3117,SelfLink:/api/v1/namespaces/deployment-3117/pods/nginx-deployment-5f9595f595-lnz8w,UID:95c34a44-91c9-11e9-a08a-ee7a14707756,ResourceVersion:105618,Generation:0,CreationTimestamp:2019-06-18 13:04:19 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 5f9595f595,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-5f9595f595 94775170-91c9-11e9-a08a-ee7a14707756 0xc002a02c20 0xc002a02c21}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-8crrs {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-8crrs,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] []  [] [] [] {map[] map[]} [{default-token-8crrs true /var/run/secrets/kubernetes.io/serviceaccount   }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.72.74.184,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists  NoExecute 0xc002a02ca0} {node.kubernetes.io/unreachable Exists  NoExecute 0xc002a02cc0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 13:04:19 +0000 UTC  }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},}
+Jun 18 13:04:19.924: INFO: Pod "nginx-deployment-5f9595f595-m2v7m" is not available:
+&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-5f9595f595-m2v7m,GenerateName:nginx-deployment-5f9595f595-,Namespace:deployment-3117,SelfLink:/api/v1/namespaces/deployment-3117/pods/nginx-deployment-5f9595f595-m2v7m,UID:95c664a8-91c9-11e9-a08a-ee7a14707756,ResourceVersion:105639,Generation:0,CreationTimestamp:2019-06-18 13:04:19 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 5f9595f595,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-5f9595f595 94775170-91c9-11e9-a08a-ee7a14707756 0xc002a02e40 0xc002a02e41}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-8crrs {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-8crrs,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] []  [] [] [] {map[] map[]} [{default-token-8crrs true /var/run/secrets/kubernetes.io/serviceaccount   }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.72.74.138,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists  NoExecute 0xc002a02f30} {node.kubernetes.io/unreachable Exists  NoExecute 0xc002a02f50}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 13:04:19 +0000 UTC  }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},}
+Jun 18 13:04:19.924: INFO: Pod "nginx-deployment-5f9595f595-n6gb2" is not available:
+&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-5f9595f595-n6gb2,GenerateName:nginx-deployment-5f9595f595-,Namespace:deployment-3117,SelfLink:/api/v1/namespaces/deployment-3117/pods/nginx-deployment-5f9595f595-n6gb2,UID:95c638b0-91c9-11e9-a08a-ee7a14707756,ResourceVersion:105634,Generation:0,CreationTimestamp:2019-06-18 13:04:19 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 5f9595f595,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-5f9595f595 94775170-91c9-11e9-a08a-ee7a14707756 0xc002a03070 0xc002a03071}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-8crrs {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-8crrs,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] []  [] [] [] {map[] map[]} [{default-token-8crrs true /var/run/secrets/kubernetes.io/serviceaccount   }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.72.74.138,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists  NoExecute 0xc002a031a0} {node.kubernetes.io/unreachable Exists  NoExecute 0xc002a031c0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 13:04:19 +0000 UTC  }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},}
+Jun 18 13:04:19.924: INFO: Pod "nginx-deployment-5f9595f595-n7lm4" is not available:
+&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-5f9595f595-n7lm4,GenerateName:nginx-deployment-5f9595f595-,Namespace:deployment-3117,SelfLink:/api/v1/namespaces/deployment-3117/pods/nginx-deployment-5f9595f595-n7lm4,UID:95c67fb9-91c9-11e9-a08a-ee7a14707756,ResourceVersion:105645,Generation:0,CreationTimestamp:2019-06-18 13:04:19 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 5f9595f595,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-5f9595f595 94775170-91c9-11e9-a08a-ee7a14707756 0xc002a03280 0xc002a03281}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-8crrs {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-8crrs,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] []  [] [] [] {map[] map[]} [{default-token-8crrs true /var/run/secrets/kubernetes.io/serviceaccount   }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.72.74.184,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists  NoExecute 0xc002a03310} {node.kubernetes.io/unreachable Exists  NoExecute 0xc002a03340}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 13:04:19 +0000 UTC  }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},}
+Jun 18 13:04:19.924: INFO: Pod "nginx-deployment-5f9595f595-nbgdp" is not available:
+&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-5f9595f595-nbgdp,GenerateName:nginx-deployment-5f9595f595-,Namespace:deployment-3117,SelfLink:/api/v1/namespaces/deployment-3117/pods/nginx-deployment-5f9595f595-nbgdp,UID:95c34772-91c9-11e9-a08a-ee7a14707756,ResourceVersion:105616,Generation:0,CreationTimestamp:2019-06-18 13:04:19 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 5f9595f595,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-5f9595f595 94775170-91c9-11e9-a08a-ee7a14707756 0xc002a033e0 0xc002a033e1}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-8crrs {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-8crrs,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] []  [] [] [] {map[] map[]} [{default-token-8crrs true /var/run/secrets/kubernetes.io/serviceaccount   }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.72.74.138,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists  NoExecute 0xc002a03460} {node.kubernetes.io/unreachable Exists  NoExecute 0xc002a03480}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 13:04:19 +0000 UTC  }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},}
+Jun 18 13:04:19.924: INFO: Pod "nginx-deployment-5f9595f595-scvpm" is not available:
+&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-5f9595f595-scvpm,GenerateName:nginx-deployment-5f9595f595-,Namespace:deployment-3117,SelfLink:/api/v1/namespaces/deployment-3117/pods/nginx-deployment-5f9595f595-scvpm,UID:947b48e2-91c9-11e9-a08a-ee7a14707756,ResourceVersion:105548,Generation:0,CreationTimestamp:2019-06-18 13:04:17 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 5f9595f595,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-5f9595f595 94775170-91c9-11e9-a08a-ee7a14707756 0xc002a035f0 0xc002a035f1}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-8crrs {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-8crrs,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] []  [] [] [] {map[] map[]} [{default-token-8crrs true /var/run/secrets/kubernetes.io/serviceaccount   }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.72.74.189,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists  NoExecute 0xc002a03670} {node.kubernetes.io/unreachable Exists  NoExecute 0xc002a03690}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 13:04:17 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 13:04:17 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 13:04:17 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 13:04:17 +0000 UTC  }],Message:,Reason:,HostIP:10.72.74.189,PodIP:,StartTime:2019-06-18 13:04:17 +0000 UTC,ContainerStatuses:[{nginx {ContainerStateWaiting{Reason:ContainerCreating,Message:,} nil nil} {nil nil nil} false 0 nginx:404  }],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},}
+Jun 18 13:04:19.925: INFO: Pod "nginx-deployment-5f9595f595-sxfdz" is not available:
+&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-5f9595f595-sxfdz,GenerateName:nginx-deployment-5f9595f595-,Namespace:deployment-3117,SelfLink:/api/v1/namespaces/deployment-3117/pods/nginx-deployment-5f9595f595-sxfdz,UID:95c0ed30-91c9-11e9-a08a-ee7a14707756,ResourceVersion:105607,Generation:0,CreationTimestamp:2019-06-18 13:04:19 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 5f9595f595,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-5f9595f595 94775170-91c9-11e9-a08a-ee7a14707756 0xc002a03770 0xc002a03771}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-8crrs {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-8crrs,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] []  [] [] [] {map[] map[]} [{default-token-8crrs true /var/run/secrets/kubernetes.io/serviceaccount   }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.72.74.189,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists  NoExecute 0xc002a037f0} {node.kubernetes.io/unreachable Exists  NoExecute 0xc002a03810}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 13:04:19 +0000 UTC  }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},}
+Jun 18 13:04:19.925: INFO: Pod "nginx-deployment-5f9595f595-tl9gx" is not available:
+&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-5f9595f595-tl9gx,GenerateName:nginx-deployment-5f9595f595-,Namespace:deployment-3117,SelfLink:/api/v1/namespaces/deployment-3117/pods/nginx-deployment-5f9595f595-tl9gx,UID:948464dc-91c9-11e9-a08a-ee7a14707756,ResourceVersion:105561,Generation:0,CreationTimestamp:2019-06-18 13:04:17 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 5f9595f595,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-5f9595f595 94775170-91c9-11e9-a08a-ee7a14707756 0xc002a03890 0xc002a03891}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-8crrs {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-8crrs,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] []  [] [] [] {map[] map[]} [{default-token-8crrs true /var/run/secrets/kubernetes.io/serviceaccount   }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.72.74.138,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists  NoExecute 0xc002a03910} {node.kubernetes.io/unreachable Exists  NoExecute 0xc002a03930}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 13:04:17 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 13:04:17 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 13:04:17 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 13:04:17 +0000 UTC  }],Message:,Reason:,HostIP:10.72.74.138,PodIP:,StartTime:2019-06-18 13:04:17 +0000 UTC,ContainerStatuses:[{nginx {ContainerStateWaiting{Reason:ContainerCreating,Message:,} nil nil} {nil nil nil} false 0 nginx:404  }],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},}
+Jun 18 13:04:19.925: INFO: Pod "nginx-deployment-5f9595f595-x8wm5" is not available:
+&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-5f9595f595-x8wm5,GenerateName:nginx-deployment-5f9595f595-,Namespace:deployment-3117,SelfLink:/api/v1/namespaces/deployment-3117/pods/nginx-deployment-5f9595f595-x8wm5,UID:9478f15c-91c9-11e9-a08a-ee7a14707756,ResourceVersion:105537,Generation:0,CreationTimestamp:2019-06-18 13:04:17 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 5f9595f595,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-5f9595f595 94775170-91c9-11e9-a08a-ee7a14707756 0xc002a03a00 0xc002a03a01}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-8crrs {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-8crrs,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] []  [] [] [] {map[] map[]} [{default-token-8crrs true /var/run/secrets/kubernetes.io/serviceaccount   }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.72.74.184,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists  NoExecute 0xc002a03a80} {node.kubernetes.io/unreachable Exists  NoExecute 0xc002a03aa0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 13:04:17 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 13:04:17 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 13:04:17 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 13:04:17 +0000 UTC  }],Message:,Reason:,HostIP:10.72.74.184,PodIP:,StartTime:2019-06-18 13:04:17 +0000 UTC,ContainerStatuses:[{nginx {ContainerStateWaiting{Reason:ContainerCreating,Message:,} nil nil} {nil nil nil} false 0 nginx:404  }],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},}
+Jun 18 13:04:19.925: INFO: Pod "nginx-deployment-6f478d8d8-5rfz5" is not available:
+&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-6f478d8d8-5rfz5,GenerateName:nginx-deployment-6f478d8d8-,Namespace:deployment-3117,SelfLink:/api/v1/namespaces/deployment-3117/pods/nginx-deployment-6f478d8d8-5rfz5,UID:95c6e7f9-91c9-11e9-a08a-ee7a14707756,ResourceVersion:105644,Generation:0,CreationTimestamp:2019-06-18 13:04:19 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 6f478d8d8,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-6f478d8d8 92038df1-91c9-11e9-a08a-ee7a14707756 0xc002a03b80 0xc002a03b81}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-8crrs {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-8crrs,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] []  [] [] [] {map[] map[]} [{default-token-8crrs true /var/run/secrets/kubernetes.io/serviceaccount   }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.72.74.184,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists  NoExecute 0xc002a03bf0} {node.kubernetes.io/unreachable Exists  NoExecute 0xc002a03c10}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 13:04:19 +0000 UTC  }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},}
+Jun 18 13:04:19.925: INFO: Pod "nginx-deployment-6f478d8d8-65nqq" is available:
+&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-6f478d8d8-65nqq,GenerateName:nginx-deployment-6f478d8d8-,Namespace:deployment-3117,SelfLink:/api/v1/namespaces/deployment-3117/pods/nginx-deployment-6f478d8d8-65nqq,UID:920cb83c-91c9-11e9-a08a-ee7a14707756,ResourceVersion:105495,Generation:0,CreationTimestamp:2019-06-18 13:04:13 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 6f478d8d8,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-6f478d8d8 92038df1-91c9-11e9-a08a-ee7a14707756 0xc002a03c90 0xc002a03c91}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-8crrs {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-8crrs,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] []  [] [] [] {map[] map[]} [{default-token-8crrs true /var/run/secrets/kubernetes.io/serviceaccount   }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.72.74.184,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists  NoExecute 0xc002a03d00} {node.kubernetes.io/unreachable Exists  NoExecute 0xc002a03d20}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Running,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 13:04:13 +0000 UTC  } {Ready True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 13:04:15 +0000 UTC  } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 13:04:15 +0000 UTC  } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 13:04:13 +0000 UTC  }],Message:,Reason:,HostIP:10.72.74.184,PodIP:172.30.142.7,StartTime:2019-06-18 13:04:13 +0000 UTC,ContainerStatuses:[{nginx {nil ContainerStateRunning{StartedAt:2019-06-18 13:04:15 +0000 UTC,} nil} {nil nil nil} true 0 docker.io/library/nginx:1.14-alpine docker.io/library/nginx@sha256:485b610fefec7ff6c463ced9623314a04ed67e3945b9c08d7e53a47f6d108dc7 containerd://10c32d71bce35de7c952701db62ef95429dee84390c30eadca4a12ca514069d5}],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},}
+Jun 18 13:04:19.925: INFO: Pod "nginx-deployment-6f478d8d8-6dfg2" is not available:
+&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-6f478d8d8-6dfg2,GenerateName:nginx-deployment-6f478d8d8-,Namespace:deployment-3117,SelfLink:/api/v1/namespaces/deployment-3117/pods/nginx-deployment-6f478d8d8-6dfg2,UID:95c6615c-91c9-11e9-a08a-ee7a14707756,ResourceVersion:105642,Generation:0,CreationTimestamp:2019-06-18 13:04:19 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 6f478d8d8,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-6f478d8d8 92038df1-91c9-11e9-a08a-ee7a14707756 0xc002a03df7 0xc002a03df8}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-8crrs {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-8crrs,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] []  [] [] [] {map[] map[]} [{default-token-8crrs true /var/run/secrets/kubernetes.io/serviceaccount   }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.72.74.184,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists  NoExecute 0xc002a03e70} {node.kubernetes.io/unreachable Exists  NoExecute 0xc002a03e90}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 13:04:19 +0000 UTC  }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},}
+Jun 18 13:04:19.926: INFO: Pod "nginx-deployment-6f478d8d8-7dtvz" is available:
+&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-6f478d8d8-7dtvz,GenerateName:nginx-deployment-6f478d8d8-,Namespace:deployment-3117,SelfLink:/api/v1/namespaces/deployment-3117/pods/nginx-deployment-6f478d8d8-7dtvz,UID:9209799f-91c9-11e9-a08a-ee7a14707756,ResourceVersion:105491,Generation:0,CreationTimestamp:2019-06-18 13:04:13 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 6f478d8d8,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-6f478d8d8 92038df1-91c9-11e9-a08a-ee7a14707756 0xc002a03f10 0xc002a03f11}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-8crrs {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-8crrs,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] []  [] [] [] {map[] map[]} [{default-token-8crrs true /var/run/secrets/kubernetes.io/serviceaccount   }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.72.74.184,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists  NoExecute 0xc002a03f80} {node.kubernetes.io/unreachable Exists  NoExecute 0xc002a03fa0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Running,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 13:04:13 +0000 UTC  } {Ready True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 13:04:15 +0000 UTC  } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 13:04:15 +0000 UTC  } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 13:04:13 +0000 UTC  }],Message:,Reason:,HostIP:10.72.74.184,PodIP:172.30.142.3,StartTime:2019-06-18 13:04:13 +0000 UTC,ContainerStatuses:[{nginx {nil ContainerStateRunning{StartedAt:2019-06-18 13:04:15 +0000 UTC,} nil} {nil nil nil} true 0 docker.io/library/nginx:1.14-alpine docker.io/library/nginx@sha256:485b610fefec7ff6c463ced9623314a04ed67e3945b9c08d7e53a47f6d108dc7 containerd://10623d473f1a70f3ae8c3b0b92a462eb66234cd936431973325563c6cc54ee22}],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},}
+Jun 18 13:04:19.926: INFO: Pod "nginx-deployment-6f478d8d8-7wnns" is available:
+&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-6f478d8d8-7wnns,GenerateName:nginx-deployment-6f478d8d8-,Namespace:deployment-3117,SelfLink:/api/v1/namespaces/deployment-3117/pods/nginx-deployment-6f478d8d8-7wnns,UID:92072060-91c9-11e9-a08a-ee7a14707756,ResourceVersion:105473,Generation:0,CreationTimestamp:2019-06-18 13:04:13 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 6f478d8d8,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-6f478d8d8 92038df1-91c9-11e9-a08a-ee7a14707756 0xc0024be4a7 0xc0024be4a8}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-8crrs {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-8crrs,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] []  [] [] [] {map[] map[]} [{default-token-8crrs true /var/run/secrets/kubernetes.io/serviceaccount   }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.72.74.138,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists  NoExecute 0xc0024be520} {node.kubernetes.io/unreachable Exists  NoExecute 0xc0024be540}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Running,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 13:04:13 +0000 UTC  } {Ready True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 13:04:15 +0000 UTC  } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 13:04:15 +0000 UTC  } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 13:04:13 +0000 UTC  }],Message:,Reason:,HostIP:10.72.74.138,PodIP:172.30.206.183,StartTime:2019-06-18 13:04:13 +0000 UTC,ContainerStatuses:[{nginx {nil ContainerStateRunning{StartedAt:2019-06-18 13:04:14 +0000 UTC,} nil} {nil nil nil} true 0 docker.io/library/nginx:1.14-alpine docker.io/library/nginx@sha256:485b610fefec7ff6c463ced9623314a04ed67e3945b9c08d7e53a47f6d108dc7 containerd://51f0cba4b2efb19e493d3f612d7a8edd7e5d648791b533500a46b7ea37a72a2e}],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},}
+Jun 18 13:04:19.926: INFO: Pod "nginx-deployment-6f478d8d8-84g5t" is available:
+&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-6f478d8d8-84g5t,GenerateName:nginx-deployment-6f478d8d8-,Namespace:deployment-3117,SelfLink:/api/v1/namespaces/deployment-3117/pods/nginx-deployment-6f478d8d8-84g5t,UID:920c9fb0-91c9-11e9-a08a-ee7a14707756,ResourceVersion:105499,Generation:0,CreationTimestamp:2019-06-18 13:04:13 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 6f478d8d8,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-6f478d8d8 92038df1-91c9-11e9-a08a-ee7a14707756 0xc0024be617 0xc0024be618}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-8crrs {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-8crrs,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] []  [] [] [] {map[] map[]} [{default-token-8crrs true /var/run/secrets/kubernetes.io/serviceaccount   }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.72.74.138,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists  NoExecute 0xc0024be690} {node.kubernetes.io/unreachable Exists  NoExecute 0xc0024be6b0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Running,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 13:04:13 +0000 UTC  } {Ready True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 13:04:16 +0000 UTC  } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 13:04:16 +0000 UTC  } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 13:04:13 +0000 UTC  }],Message:,Reason:,HostIP:10.72.74.138,PodIP:172.30.206.184,StartTime:2019-06-18 13:04:13 +0000 UTC,ContainerStatuses:[{nginx {nil ContainerStateRunning{StartedAt:2019-06-18 13:04:15 +0000 UTC,} nil} {nil nil nil} true 0 docker.io/library/nginx:1.14-alpine docker.io/library/nginx@sha256:485b610fefec7ff6c463ced9623314a04ed67e3945b9c08d7e53a47f6d108dc7 containerd://f79756078b13750c0b67940503315d997b4cbb156ce1a85cdd6416cacfea58de}],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},}
+Jun 18 13:04:19.926: INFO: Pod "nginx-deployment-6f478d8d8-8bdld" is not available:
+&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-6f478d8d8-8bdld,GenerateName:nginx-deployment-6f478d8d8-,Namespace:deployment-3117,SelfLink:/api/v1/namespaces/deployment-3117/pods/nginx-deployment-6f478d8d8-8bdld,UID:95c0d0d4-91c9-11e9-a08a-ee7a14707756,ResourceVersion:105631,Generation:0,CreationTimestamp:2019-06-18 13:04:19 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 6f478d8d8,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-6f478d8d8 92038df1-91c9-11e9-a08a-ee7a14707756 0xc0024be787 0xc0024be788}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-8crrs {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-8crrs,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] []  [] [] [] {map[] map[]} [{default-token-8crrs true /var/run/secrets/kubernetes.io/serviceaccount   }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.72.74.138,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists  NoExecute 0xc0024be800} {node.kubernetes.io/unreachable Exists  NoExecute 0xc0024be820}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 13:04:19 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 13:04:19 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 13:04:19 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 13:04:19 +0000 UTC  }],Message:,Reason:,HostIP:10.72.74.138,PodIP:,StartTime:2019-06-18 13:04:19 +0000 UTC,ContainerStatuses:[{nginx {ContainerStateWaiting{Reason:ContainerCreating,Message:,} nil nil} {nil nil nil} false 0 docker.io/library/nginx:1.14-alpine  }],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},}
+Jun 18 13:04:19.927: INFO: Pod "nginx-deployment-6f478d8d8-brx48" is available:
+&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-6f478d8d8-brx48,GenerateName:nginx-deployment-6f478d8d8-,Namespace:deployment-3117,SelfLink:/api/v1/namespaces/deployment-3117/pods/nginx-deployment-6f478d8d8-brx48,UID:920cc1be-91c9-11e9-a08a-ee7a14707756,ResourceVersion:105477,Generation:0,CreationTimestamp:2019-06-18 13:04:13 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 6f478d8d8,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-6f478d8d8 92038df1-91c9-11e9-a08a-ee7a14707756 0xc0024be8e7 0xc0024be8e8}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-8crrs {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-8crrs,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] []  [] [] [] {map[] map[]} [{default-token-8crrs true /var/run/secrets/kubernetes.io/serviceaccount   }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.72.74.138,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists  NoExecute 0xc0024be970} {node.kubernetes.io/unreachable Exists  NoExecute 0xc0024be990}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Running,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 13:04:13 +0000 UTC  } {Ready True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 13:04:15 +0000 UTC  } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 13:04:15 +0000 UTC  } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 13:04:13 +0000 UTC  }],Message:,Reason:,HostIP:10.72.74.138,PodIP:172.30.206.185,StartTime:2019-06-18 13:04:13 +0000 UTC,ContainerStatuses:[{nginx {nil ContainerStateRunning{StartedAt:2019-06-18 13:04:15 +0000 UTC,} nil} {nil nil nil} true 0 docker.io/library/nginx:1.14-alpine docker.io/library/nginx@sha256:485b610fefec7ff6c463ced9623314a04ed67e3945b9c08d7e53a47f6d108dc7 containerd://93982a3d03c2649b558d3dab6f1a81cdb1b0c41330c287843e9d6022a4075ccc}],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},}
+Jun 18 13:04:19.928: INFO: Pod "nginx-deployment-6f478d8d8-cqvk4" is available:
+&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-6f478d8d8-cqvk4,GenerateName:nginx-deployment-6f478d8d8-,Namespace:deployment-3117,SelfLink:/api/v1/namespaces/deployment-3117/pods/nginx-deployment-6f478d8d8-cqvk4,UID:920cd6cc-91c9-11e9-a08a-ee7a14707756,ResourceVersion:105487,Generation:0,CreationTimestamp:2019-06-18 13:04:13 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 6f478d8d8,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-6f478d8d8 92038df1-91c9-11e9-a08a-ee7a14707756 0xc0024bea67 0xc0024bea68}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-8crrs {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-8crrs,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] []  [] [] [] {map[] map[]} [{default-token-8crrs true /var/run/secrets/kubernetes.io/serviceaccount   }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.72.74.184,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists  NoExecute 0xc0024beae0} {node.kubernetes.io/unreachable Exists  NoExecute 0xc0024beb00}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Running,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 13:04:13 +0000 UTC  } {Ready True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 13:04:15 +0000 UTC  } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 13:04:15 +0000 UTC  } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 13:04:13 +0000 UTC  }],Message:,Reason:,HostIP:10.72.74.184,PodIP:172.30.142.6,StartTime:2019-06-18 13:04:13 +0000 UTC,ContainerStatuses:[{nginx {nil ContainerStateRunning{StartedAt:2019-06-18 13:04:15 +0000 UTC,} nil} {nil nil nil} true 0 docker.io/library/nginx:1.14-alpine docker.io/library/nginx@sha256:485b610fefec7ff6c463ced9623314a04ed67e3945b9c08d7e53a47f6d108dc7 containerd://67a8cd3f89affb19bd3858cb72186ede544e45a0647505b2394997530745e413}],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},}
+Jun 18 13:04:19.928: INFO: Pod "nginx-deployment-6f478d8d8-df7xq" is available:
+&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-6f478d8d8-df7xq,GenerateName:nginx-deployment-6f478d8d8-,Namespace:deployment-3117,SelfLink:/api/v1/namespaces/deployment-3117/pods/nginx-deployment-6f478d8d8-df7xq,UID:9209b1fc-91c9-11e9-a08a-ee7a14707756,ResourceVersion:105472,Generation:0,CreationTimestamp:2019-06-18 13:04:13 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 6f478d8d8,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-6f478d8d8 92038df1-91c9-11e9-a08a-ee7a14707756 0xc0024bebf7 0xc0024bebf8}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-8crrs {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-8crrs,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] []  [] [] [] {map[] map[]} [{default-token-8crrs true /var/run/secrets/kubernetes.io/serviceaccount   }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.72.74.189,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists  NoExecute 0xc0024bec80} {node.kubernetes.io/unreachable Exists  NoExecute 0xc0024beca0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Running,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 13:04:13 +0000 UTC  } {Ready True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 13:04:15 +0000 UTC  } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 13:04:15 +0000 UTC  } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 13:04:13 +0000 UTC  }],Message:,Reason:,HostIP:10.72.74.189,PodIP:172.30.80.160,StartTime:2019-06-18 13:04:13 +0000 UTC,ContainerStatuses:[{nginx {nil ContainerStateRunning{StartedAt:2019-06-18 13:04:14 +0000 UTC,} nil} {nil nil nil} true 0 docker.io/library/nginx:1.14-alpine docker.io/library/nginx@sha256:485b610fefec7ff6c463ced9623314a04ed67e3945b9c08d7e53a47f6d108dc7 containerd://89535a27265361b69454396f8088bb80c2bfb5137eb22bf3fc6f22ead9b4197d}],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},}
+Jun 18 13:04:19.928: INFO: Pod "nginx-deployment-6f478d8d8-h6zp4" is not available:
+&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-6f478d8d8-h6zp4,GenerateName:nginx-deployment-6f478d8d8-,Namespace:deployment-3117,SelfLink:/api/v1/namespaces/deployment-3117/pods/nginx-deployment-6f478d8d8-h6zp4,UID:95c30f76-91c9-11e9-a08a-ee7a14707756,ResourceVersion:105647,Generation:0,CreationTimestamp:2019-06-18 13:04:19 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 6f478d8d8,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-6f478d8d8 92038df1-91c9-11e9-a08a-ee7a14707756 0xc0024bed77 0xc0024bed78}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-8crrs {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-8crrs,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] []  [] [] [] {map[] map[]} [{default-token-8crrs true /var/run/secrets/kubernetes.io/serviceaccount   }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.72.74.138,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists  NoExecute 0xc0024bedf0} {node.kubernetes.io/unreachable Exists  NoExecute 0xc0024bee10}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 13:04:19 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 13:04:19 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 13:04:19 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 13:04:19 +0000 UTC  }],Message:,Reason:,HostIP:10.72.74.138,PodIP:,StartTime:2019-06-18 13:04:19 +0000 UTC,ContainerStatuses:[{nginx {ContainerStateWaiting{Reason:ContainerCreating,Message:,} nil nil} {nil nil nil} false 0 docker.io/library/nginx:1.14-alpine  }],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},}
+Jun 18 13:04:19.928: INFO: Pod "nginx-deployment-6f478d8d8-hthvz" is not available:
+&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-6f478d8d8-hthvz,GenerateName:nginx-deployment-6f478d8d8-,Namespace:deployment-3117,SelfLink:/api/v1/namespaces/deployment-3117/pods/nginx-deployment-6f478d8d8-hthvz,UID:95c65ce2-91c9-11e9-a08a-ee7a14707756,ResourceVersion:105635,Generation:0,CreationTimestamp:2019-06-18 13:04:19 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 6f478d8d8,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-6f478d8d8 92038df1-91c9-11e9-a08a-ee7a14707756 0xc0024beed7 0xc0024beed8}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-8crrs {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-8crrs,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] []  [] [] [] {map[] map[]} [{default-token-8crrs true /var/run/secrets/kubernetes.io/serviceaccount   }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.72.74.189,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists  NoExecute 0xc0024bef50} {node.kubernetes.io/unreachable Exists  NoExecute 0xc0024bef70}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 13:04:19 +0000 UTC  }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},}
+Jun 18 13:04:19.928: INFO: Pod "nginx-deployment-6f478d8d8-jkg7b" is not available:
+&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-6f478d8d8-jkg7b,GenerateName:nginx-deployment-6f478d8d8-,Namespace:deployment-3117,SelfLink:/api/v1/namespaces/deployment-3117/pods/nginx-deployment-6f478d8d8-jkg7b,UID:95c6becd-91c9-11e9-a08a-ee7a14707756,ResourceVersion:105643,Generation:0,CreationTimestamp:2019-06-18 13:04:19 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 6f478d8d8,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-6f478d8d8 92038df1-91c9-11e9-a08a-ee7a14707756 0xc0024beff0 0xc0024beff1}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-8crrs {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-8crrs,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] []  [] [] [] {map[] map[]} [{default-token-8crrs true /var/run/secrets/kubernetes.io/serviceaccount   }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.72.74.138,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists  NoExecute 0xc0024bf060} {node.kubernetes.io/unreachable Exists  NoExecute 0xc0024bf080}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 13:04:19 +0000 UTC  }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},}
+Jun 18 13:04:19.928: INFO: Pod "nginx-deployment-6f478d8d8-kfq8p" is not available:
+&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-6f478d8d8-kfq8p,GenerateName:nginx-deployment-6f478d8d8-,Namespace:deployment-3117,SelfLink:/api/v1/namespaces/deployment-3117/pods/nginx-deployment-6f478d8d8-kfq8p,UID:95be9cc7-91c9-11e9-a08a-ee7a14707756,ResourceVersion:105628,Generation:0,CreationTimestamp:2019-06-18 13:04:19 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 6f478d8d8,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-6f478d8d8 92038df1-91c9-11e9-a08a-ee7a14707756 0xc0024bf100 0xc0024bf101}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-8crrs {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-8crrs,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] []  [] [] [] {map[] map[]} [{default-token-8crrs true /var/run/secrets/kubernetes.io/serviceaccount   }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.72.74.189,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists  NoExecute 0xc0024bf170} {node.kubernetes.io/unreachable Exists  NoExecute 0xc0024bf190}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 13:04:19 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 13:04:19 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 13:04:19 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 13:04:19 +0000 UTC  }],Message:,Reason:,HostIP:10.72.74.189,PodIP:,StartTime:2019-06-18 13:04:19 +0000 UTC,ContainerStatuses:[{nginx {ContainerStateWaiting{Reason:ContainerCreating,Message:,} nil nil} {nil nil nil} false 0 docker.io/library/nginx:1.14-alpine  }],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},}
+Jun 18 13:04:19.929: INFO: Pod "nginx-deployment-6f478d8d8-lbj6v" is not available:
+&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-6f478d8d8-lbj6v,GenerateName:nginx-deployment-6f478d8d8-,Namespace:deployment-3117,SelfLink:/api/v1/namespaces/deployment-3117/pods/nginx-deployment-6f478d8d8-lbj6v,UID:95c6624f-91c9-11e9-a08a-ee7a14707756,ResourceVersion:105641,Generation:0,CreationTimestamp:2019-06-18 13:04:19 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 6f478d8d8,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-6f478d8d8 92038df1-91c9-11e9-a08a-ee7a14707756 0xc0024bf257 0xc0024bf258}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-8crrs {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-8crrs,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] []  [] [] [] {map[] map[]} [{default-token-8crrs true /var/run/secrets/kubernetes.io/serviceaccount   }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.72.74.189,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists  NoExecute 0xc0024bf2d0} {node.kubernetes.io/unreachable Exists  NoExecute 0xc0024bf2f0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 13:04:19 +0000 UTC  }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},}
+Jun 18 13:04:19.929: INFO: Pod "nginx-deployment-6f478d8d8-lgjrf" is not available:
+&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-6f478d8d8-lgjrf,GenerateName:nginx-deployment-6f478d8d8-,Namespace:deployment-3117,SelfLink:/api/v1/namespaces/deployment-3117/pods/nginx-deployment-6f478d8d8-lgjrf,UID:95c33e26-91c9-11e9-a08a-ee7a14707756,ResourceVersion:105619,Generation:0,CreationTimestamp:2019-06-18 13:04:19 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 6f478d8d8,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-6f478d8d8 92038df1-91c9-11e9-a08a-ee7a14707756 0xc0024bf370 0xc0024bf371}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-8crrs {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-8crrs,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] []  [] [] [] {map[] map[]} [{default-token-8crrs true /var/run/secrets/kubernetes.io/serviceaccount   }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.72.74.184,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists  NoExecute 0xc0024bf3e0} {node.kubernetes.io/unreachable Exists  NoExecute 0xc0024bf400}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 13:04:19 +0000 UTC  }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},}
+Jun 18 13:04:19.929: INFO: Pod "nginx-deployment-6f478d8d8-ltfzm" is available:
+&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-6f478d8d8-ltfzm,GenerateName:nginx-deployment-6f478d8d8-,Namespace:deployment-3117,SelfLink:/api/v1/namespaces/deployment-3117/pods/nginx-deployment-6f478d8d8-ltfzm,UID:920f2d50-91c9-11e9-a08a-ee7a14707756,ResourceVersion:105466,Generation:0,CreationTimestamp:2019-06-18 13:04:13 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 6f478d8d8,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-6f478d8d8 92038df1-91c9-11e9-a08a-ee7a14707756 0xc0024bf480 0xc0024bf481}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-8crrs {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-8crrs,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] []  [] [] [] {map[] map[]} [{default-token-8crrs true /var/run/secrets/kubernetes.io/serviceaccount   }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.72.74.189,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists  NoExecute 0xc0024bf4f0} {node.kubernetes.io/unreachable Exists  NoExecute 0xc0024bf510}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Running,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 13:04:13 +0000 UTC  } {Ready True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 13:04:15 +0000 UTC  } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 13:04:15 +0000 UTC  } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 13:04:13 +0000 UTC  }],Message:,Reason:,HostIP:10.72.74.189,PodIP:172.30.80.161,StartTime:2019-06-18 13:04:13 +0000 UTC,ContainerStatuses:[{nginx {nil ContainerStateRunning{StartedAt:2019-06-18 13:04:15 +0000 UTC,} nil} {nil nil nil} true 0 docker.io/library/nginx:1.14-alpine docker.io/library/nginx@sha256:485b610fefec7ff6c463ced9623314a04ed67e3945b9c08d7e53a47f6d108dc7 containerd://5e77b0fc57b6323f94b2dc475ab7e8fd4f1bfd4b4cb70d15e0881551b1343c15}],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},}
+Jun 18 13:04:19.929: INFO: Pod "nginx-deployment-6f478d8d8-n9nxq" is not available:
+&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-6f478d8d8-n9nxq,GenerateName:nginx-deployment-6f478d8d8-,Namespace:deployment-3117,SelfLink:/api/v1/namespaces/deployment-3117/pods/nginx-deployment-6f478d8d8-n9nxq,UID:95c3459f-91c9-11e9-a08a-ee7a14707756,ResourceVersion:105624,Generation:0,CreationTimestamp:2019-06-18 13:04:19 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 6f478d8d8,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-6f478d8d8 92038df1-91c9-11e9-a08a-ee7a14707756 0xc0024bf5e7 0xc0024bf5e8}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-8crrs {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-8crrs,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] []  [] [] [] {map[] map[]} [{default-token-8crrs true /var/run/secrets/kubernetes.io/serviceaccount   }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.72.74.184,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists  NoExecute 0xc0024bf660} {node.kubernetes.io/unreachable Exists  NoExecute 0xc0024bf680}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 13:04:19 +0000 UTC  }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},}
+Jun 18 13:04:19.929: INFO: Pod "nginx-deployment-6f478d8d8-nw89w" is not available:
+&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-6f478d8d8-nw89w,GenerateName:nginx-deployment-6f478d8d8-,Namespace:deployment-3117,SelfLink:/api/v1/namespaces/deployment-3117/pods/nginx-deployment-6f478d8d8-nw89w,UID:95c3918d-91c9-11e9-a08a-ee7a14707756,ResourceVersion:105622,Generation:0,CreationTimestamp:2019-06-18 13:04:19 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 6f478d8d8,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-6f478d8d8 92038df1-91c9-11e9-a08a-ee7a14707756 0xc0024bf870 0xc0024bf871}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-8crrs {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-8crrs,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] []  [] [] [] {map[] map[]} [{default-token-8crrs true /var/run/secrets/kubernetes.io/serviceaccount   }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.72.74.189,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists  NoExecute 0xc0024bf8e0} {node.kubernetes.io/unreachable Exists  NoExecute 0xc0024bf900}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 13:04:19 +0000 UTC  }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},}
+Jun 18 13:04:19.929: INFO: Pod "nginx-deployment-6f478d8d8-pq72q" is not available:
+&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-6f478d8d8-pq72q,GenerateName:nginx-deployment-6f478d8d8-,Namespace:deployment-3117,SelfLink:/api/v1/namespaces/deployment-3117/pods/nginx-deployment-6f478d8d8-pq72q,UID:95c0c36b-91c9-11e9-a08a-ee7a14707756,ResourceVersion:105648,Generation:0,CreationTimestamp:2019-06-18 13:04:19 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 6f478d8d8,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-6f478d8d8 92038df1-91c9-11e9-a08a-ee7a14707756 0xc0024bf980 0xc0024bf981}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-8crrs {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-8crrs,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] []  [] [] [] {map[] map[]} [{default-token-8crrs true /var/run/secrets/kubernetes.io/serviceaccount   }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.72.74.184,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists  NoExecute 0xc0024bf9f0} {node.kubernetes.io/unreachable Exists  NoExecute 0xc0024bfa10}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 13:04:19 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 13:04:19 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-06-18 13:04:19 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 13:04:19 +0000 UTC  }],Message:,Reason:,HostIP:10.72.74.184,PodIP:,StartTime:2019-06-18 13:04:19 +0000 UTC,ContainerStatuses:[{nginx {ContainerStateWaiting{Reason:ContainerCreating,Message:,} nil nil} {nil nil nil} false 0 docker.io/library/nginx:1.14-alpine  }],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},}
+[AfterEach] [sig-apps] Deployment
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 13:04:19.929: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "deployment-3117" for this suite.
+Jun 18 13:04:29.969: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 13:04:30.377: INFO: namespace deployment-3117 deletion completed in 10.433149191s
 
-• [SLOW TEST:10.829 seconds]
-[sig-storage] EmptyDir volumes
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:41
-  should support (non-root,0666,default) [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+• [SLOW TEST:17.017 seconds]
+[sig-apps] Deployment
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22
+  deployment should support proportional scaling [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-SSSSSSS
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
 [sig-storage] Projected secret 
-  should be consumable from pods in volume [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+  optional updates should be reflected in volume [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 [BeforeEach] [sig-storage] Projected secret
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 19:48:04.845: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
+Jun 18 13:04:30.378: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
 STEP: Building a namespace api object, basename projected
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-4086
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-7952
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should be consumable from pods in volume [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: Creating projection with secret that has name projected-secret-test-ab54969f-8701-11e9-8d1b-467ee19922ac
-STEP: Creating a pod to test consume secrets
-Jun  4 19:48:05.098: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-ab5633de-8701-11e9-8d1b-467ee19922ac" in namespace "projected-4086" to be "success or failure"
-Jun  4 19:48:05.110: INFO: Pod "pod-projected-secrets-ab5633de-8701-11e9-8d1b-467ee19922ac": Phase="Pending", Reason="", readiness=false. Elapsed: 12.238473ms
-Jun  4 19:48:07.124: INFO: Pod "pod-projected-secrets-ab5633de-8701-11e9-8d1b-467ee19922ac": Phase="Running", Reason="", readiness=true. Elapsed: 2.025874732s
-Jun  4 19:48:09.137: INFO: Pod "pod-projected-secrets-ab5633de-8701-11e9-8d1b-467ee19922ac": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.03929088s
-STEP: Saw pod success
-Jun  4 19:48:09.137: INFO: Pod "pod-projected-secrets-ab5633de-8701-11e9-8d1b-467ee19922ac" satisfied condition "success or failure"
-Jun  4 19:48:09.150: INFO: Trying to get logs from node 10.212.23.189 pod pod-projected-secrets-ab5633de-8701-11e9-8d1b-467ee19922ac container projected-secret-volume-test: 
-STEP: delete the pod
-Jun  4 19:48:09.255: INFO: Waiting for pod pod-projected-secrets-ab5633de-8701-11e9-8d1b-467ee19922ac to disappear
-Jun  4 19:48:09.268: INFO: Pod pod-projected-secrets-ab5633de-8701-11e9-8d1b-467ee19922ac no longer exists
+[It] optional updates should be reflected in volume [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating secret with name s-test-opt-del-9c295a1b-91c9-11e9-a25d-8608290c688a
+STEP: Creating secret with name s-test-opt-upd-9c295a7d-91c9-11e9-a25d-8608290c688a
+STEP: Creating the pod
+STEP: Deleting secret s-test-opt-del-9c295a1b-91c9-11e9-a25d-8608290c688a
+STEP: Updating secret s-test-opt-upd-9c295a7d-91c9-11e9-a25d-8608290c688a
+STEP: Creating secret with name s-test-opt-create-9c295ac7-91c9-11e9-a25d-8608290c688a
+STEP: waiting to observe update in volume
 [AfterEach] [sig-storage] Projected secret
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 19:48:09.268: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "projected-4086" for this suite.
-Jun  4 19:48:15.326: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 19:48:15.706: INFO: namespace projected-4086 deletion completed in 6.422298068s
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 13:04:34.965: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "projected-7952" for this suite.
+Jun 18 13:05:01.008: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 13:05:01.609: INFO: namespace projected-7952 deletion completed in 26.627934318s
 
-• [SLOW TEST:10.861 seconds]
+• [SLOW TEST:31.232 seconds]
 [sig-storage] Projected secret
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_secret.go:33
-  should be consumable from pods in volume [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-------------------------------
-SSSSSSSSSSSSSSSSSS
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_secret.go:33
+  optional updates should be reflected in volume [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-[sig-cli] Kubectl client [k8s.io] Update Demo 
-  should scale a replication controller  [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+[sig-network] Services 
+  should provide secure master service  [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-network] Services
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 19:48:15.706: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-STEP: Building a namespace api object, basename kubectl
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-6533
+Jun 18 13:05:01.611: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename services
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in services-4430
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:213
-[BeforeEach] [k8s.io] Update Demo
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:265
-[It] should scale a replication controller  [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: creating a replication controller
-Jun  4 19:48:15.923: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 create -f - --namespace=kubectl-6533'
-Jun  4 19:48:16.256: INFO: stderr: ""
-Jun  4 19:48:16.256: INFO: stdout: "replicationcontroller/update-demo-nautilus created\n"
-STEP: waiting for all containers in name=update-demo pods to come up.
-Jun  4 19:48:16.256: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-6533'
-Jun  4 19:48:16.346: INFO: stderr: ""
-Jun  4 19:48:16.346: INFO: stdout: "update-demo-nautilus-nbtgq update-demo-nautilus-wr24j "
-Jun  4 19:48:16.346: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 get pods update-demo-nautilus-nbtgq -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-6533'
-Jun  4 19:48:16.457: INFO: stderr: ""
-Jun  4 19:48:16.457: INFO: stdout: ""
-Jun  4 19:48:16.457: INFO: update-demo-nautilus-nbtgq is created but not running
-Jun  4 19:48:21.457: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-6533'
-Jun  4 19:48:21.596: INFO: stderr: ""
-Jun  4 19:48:21.596: INFO: stdout: "update-demo-nautilus-nbtgq update-demo-nautilus-wr24j "
-Jun  4 19:48:21.596: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 get pods update-demo-nautilus-nbtgq -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-6533'
-Jun  4 19:48:21.680: INFO: stderr: ""
-Jun  4 19:48:21.680: INFO: stdout: "true"
-Jun  4 19:48:21.680: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 get pods update-demo-nautilus-nbtgq -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-6533'
-Jun  4 19:48:21.765: INFO: stderr: ""
-Jun  4 19:48:21.765: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0"
-Jun  4 19:48:21.765: INFO: validating pod update-demo-nautilus-nbtgq
-Jun  4 19:48:21.788: INFO: got data: {
-  "image": "nautilus.jpg"
-}
-
-Jun  4 19:48:21.788: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg .
-Jun  4 19:48:21.788: INFO: update-demo-nautilus-nbtgq is verified up and running
-Jun  4 19:48:21.788: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 get pods update-demo-nautilus-wr24j -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-6533'
-Jun  4 19:48:21.869: INFO: stderr: ""
-Jun  4 19:48:21.869: INFO: stdout: "true"
-Jun  4 19:48:21.869: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 get pods update-demo-nautilus-wr24j -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-6533'
-Jun  4 19:48:21.952: INFO: stderr: ""
-Jun  4 19:48:21.952: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0"
-Jun  4 19:48:21.952: INFO: validating pod update-demo-nautilus-wr24j
-Jun  4 19:48:21.971: INFO: got data: {
-  "image": "nautilus.jpg"
-}
-
-Jun  4 19:48:21.971: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg .
-Jun  4 19:48:21.971: INFO: update-demo-nautilus-wr24j is verified up and running
-STEP: scaling down the replication controller
-Jun  4 19:48:21.972: INFO: scanned /root for discovery docs: 
-Jun  4 19:48:21.972: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 scale rc update-demo-nautilus --replicas=1 --timeout=5m --namespace=kubectl-6533'
-Jun  4 19:48:23.115: INFO: stderr: ""
-Jun  4 19:48:23.115: INFO: stdout: "replicationcontroller/update-demo-nautilus scaled\n"
-STEP: waiting for all containers in name=update-demo pods to come up.
-Jun  4 19:48:23.115: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-6533'
-Jun  4 19:48:23.214: INFO: stderr: ""
-Jun  4 19:48:23.214: INFO: stdout: "update-demo-nautilus-nbtgq update-demo-nautilus-wr24j "
-STEP: Replicas for name=update-demo: expected=1 actual=2
-Jun  4 19:48:28.215: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-6533'
-Jun  4 19:48:28.312: INFO: stderr: ""
-Jun  4 19:48:28.312: INFO: stdout: "update-demo-nautilus-wr24j "
-Jun  4 19:48:28.312: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 get pods update-demo-nautilus-wr24j -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-6533'
-Jun  4 19:48:28.433: INFO: stderr: ""
-Jun  4 19:48:28.433: INFO: stdout: "true"
-Jun  4 19:48:28.434: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 get pods update-demo-nautilus-wr24j -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-6533'
-Jun  4 19:48:28.528: INFO: stderr: ""
-Jun  4 19:48:28.528: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0"
-Jun  4 19:48:28.528: INFO: validating pod update-demo-nautilus-wr24j
-Jun  4 19:48:28.544: INFO: got data: {
-  "image": "nautilus.jpg"
-}
-
-Jun  4 19:48:28.545: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg .
-Jun  4 19:48:28.545: INFO: update-demo-nautilus-wr24j is verified up and running
-STEP: scaling up the replication controller
-Jun  4 19:48:28.546: INFO: scanned /root for discovery docs: 
-Jun  4 19:48:28.546: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 scale rc update-demo-nautilus --replicas=2 --timeout=5m --namespace=kubectl-6533'
-Jun  4 19:48:29.690: INFO: stderr: ""
-Jun  4 19:48:29.690: INFO: stdout: "replicationcontroller/update-demo-nautilus scaled\n"
-STEP: waiting for all containers in name=update-demo pods to come up.
-Jun  4 19:48:29.690: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-6533'
-Jun  4 19:48:29.786: INFO: stderr: ""
-Jun  4 19:48:29.786: INFO: stdout: "update-demo-nautilus-hvqdf update-demo-nautilus-wr24j "
-Jun  4 19:48:29.787: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 get pods update-demo-nautilus-hvqdf -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-6533'
-Jun  4 19:48:29.887: INFO: stderr: ""
-Jun  4 19:48:29.887: INFO: stdout: ""
-Jun  4 19:48:29.887: INFO: update-demo-nautilus-hvqdf is created but not running
-Jun  4 19:48:34.887: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-6533'
-Jun  4 19:48:34.993: INFO: stderr: ""
-Jun  4 19:48:34.993: INFO: stdout: "update-demo-nautilus-hvqdf update-demo-nautilus-wr24j "
-Jun  4 19:48:34.993: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 get pods update-demo-nautilus-hvqdf -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-6533'
-Jun  4 19:48:35.078: INFO: stderr: ""
-Jun  4 19:48:35.078: INFO: stdout: "true"
-Jun  4 19:48:35.078: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 get pods update-demo-nautilus-hvqdf -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-6533'
-Jun  4 19:48:35.173: INFO: stderr: ""
-Jun  4 19:48:35.173: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0"
-Jun  4 19:48:35.173: INFO: validating pod update-demo-nautilus-hvqdf
-Jun  4 19:48:35.194: INFO: got data: {
-  "image": "nautilus.jpg"
-}
-
-Jun  4 19:48:35.194: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg .
-Jun  4 19:48:35.194: INFO: update-demo-nautilus-hvqdf is verified up and running
-Jun  4 19:48:35.194: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 get pods update-demo-nautilus-wr24j -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-6533'
-Jun  4 19:48:35.290: INFO: stderr: ""
-Jun  4 19:48:35.290: INFO: stdout: "true"
-Jun  4 19:48:35.290: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 get pods update-demo-nautilus-wr24j -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-6533'
-Jun  4 19:48:35.375: INFO: stderr: ""
-Jun  4 19:48:35.375: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0"
-Jun  4 19:48:35.375: INFO: validating pod update-demo-nautilus-wr24j
-Jun  4 19:48:35.389: INFO: got data: {
-  "image": "nautilus.jpg"
-}
-
-Jun  4 19:48:35.389: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg .
-Jun  4 19:48:35.389: INFO: update-demo-nautilus-wr24j is verified up and running
-STEP: using delete to clean up resources
-Jun  4 19:48:35.389: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 delete --grace-period=0 --force -f - --namespace=kubectl-6533'
-Jun  4 19:48:35.485: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n"
-Jun  4 19:48:35.485: INFO: stdout: "replicationcontroller \"update-demo-nautilus\" force deleted\n"
-Jun  4 19:48:35.485: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 get rc,svc -l name=update-demo --no-headers --namespace=kubectl-6533'
-Jun  4 19:48:35.584: INFO: stderr: "No resources found.\n"
-Jun  4 19:48:35.584: INFO: stdout: ""
-Jun  4 19:48:35.584: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 get pods -l name=update-demo --namespace=kubectl-6533 -o go-template={{ range .items }}{{ if not .metadata.deletionTimestamp }}{{ .metadata.name }}{{ "\n" }}{{ end }}{{ end }}'
-Jun  4 19:48:35.680: INFO: stderr: ""
-Jun  4 19:48:35.680: INFO: stdout: ""
-[AfterEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 19:48:35.680: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "kubectl-6533" for this suite.
-Jun  4 19:48:59.740: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 19:49:00.198: INFO: namespace kubectl-6533 deletion completed in 24.501814005s
+[BeforeEach] [sig-network] Services
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:86
+[It] should provide secure master service  [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[AfterEach] [sig-network] Services
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 13:05:01.820: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "services-4430" for this suite.
+Jun 18 13:05:09.859: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 13:05:10.241: INFO: namespace services-4430 deletion completed in 8.408179185s
+[AfterEach] [sig-network] Services
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:91
 
-• [SLOW TEST:44.491 seconds]
-[sig-cli] Kubectl client
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23
-  [k8s.io] Update Demo
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
-    should scale a replication controller  [Conformance]
-    /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+• [SLOW TEST:8.630 seconds]
+[sig-network] Services
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:22
+  should provide secure master service  [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-SSSSSSSSSSSSSSSSSSSSSSS
+SSSSSSSS
 ------------------------------
-[sig-apps] ReplicationController 
-  should serve a basic image on each replica with a public image  [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [sig-apps] ReplicationController
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+[sig-storage] Projected downwardAPI 
+  should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-storage] Projected downwardAPI
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 19:49:00.198: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-STEP: Building a namespace api object, basename replication-controller
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in replication-controller-184
+Jun 18 13:05:10.241: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename projected
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-1256
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should serve a basic image on each replica with a public image  [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: Creating replication controller my-hostname-basic-cc59600d-8701-11e9-8d1b-467ee19922ac
-Jun  4 19:49:00.490: INFO: Pod name my-hostname-basic-cc59600d-8701-11e9-8d1b-467ee19922ac: Found 0 pods out of 1
-Jun  4 19:49:05.502: INFO: Pod name my-hostname-basic-cc59600d-8701-11e9-8d1b-467ee19922ac: Found 1 pods out of 1
-Jun  4 19:49:05.502: INFO: Ensuring all pods for ReplicationController "my-hostname-basic-cc59600d-8701-11e9-8d1b-467ee19922ac" are running
-Jun  4 19:49:05.513: INFO: Pod "my-hostname-basic-cc59600d-8701-11e9-8d1b-467ee19922ac-jfktm" is running (conditions: [{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2019-06-04 19:49:00 +0000 UTC Reason: Message:} {Type:Ready Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2019-06-04 19:49:01 +0000 UTC Reason: Message:} {Type:ContainersReady Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2019-06-04 19:49:01 +0000 UTC Reason: Message:} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2019-06-04 19:49:00 +0000 UTC Reason: Message:}])
-Jun  4 19:49:05.513: INFO: Trying to dial the pod
-Jun  4 19:49:10.563: INFO: Controller my-hostname-basic-cc59600d-8701-11e9-8d1b-467ee19922ac: Got expected result from replica 1 [my-hostname-basic-cc59600d-8701-11e9-8d1b-467ee19922ac-jfktm]: "my-hostname-basic-cc59600d-8701-11e9-8d1b-467ee19922ac-jfktm", 1 of 1 required successes so far
-[AfterEach] [sig-apps] ReplicationController
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 19:49:10.563: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "replication-controller-184" for this suite.
-Jun  4 19:49:16.621: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 19:49:16.979: INFO: namespace replication-controller-184 deletion completed in 6.400589719s
+[BeforeEach] [sig-storage] Projected downwardAPI
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:39
+[It] should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating a pod to test downward API volume plugin
+Jun 18 13:05:10.478: INFO: Waiting up to 5m0s for pod "downwardapi-volume-b3eb7ebe-91c9-11e9-a25d-8608290c688a" in namespace "projected-1256" to be "success or failure"
+Jun 18 13:05:10.491: INFO: Pod "downwardapi-volume-b3eb7ebe-91c9-11e9-a25d-8608290c688a": Phase="Pending", Reason="", readiness=false. Elapsed: 12.601853ms
+Jun 18 13:05:12.504: INFO: Pod "downwardapi-volume-b3eb7ebe-91c9-11e9-a25d-8608290c688a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.026043458s
+STEP: Saw pod success
+Jun 18 13:05:12.504: INFO: Pod "downwardapi-volume-b3eb7ebe-91c9-11e9-a25d-8608290c688a" satisfied condition "success or failure"
+Jun 18 13:05:12.517: INFO: Trying to get logs from node 10.72.74.184 pod downwardapi-volume-b3eb7ebe-91c9-11e9-a25d-8608290c688a container client-container: 
+STEP: delete the pod
+Jun 18 13:05:12.607: INFO: Waiting for pod downwardapi-volume-b3eb7ebe-91c9-11e9-a25d-8608290c688a to disappear
+Jun 18 13:05:12.627: INFO: Pod downwardapi-volume-b3eb7ebe-91c9-11e9-a25d-8608290c688a no longer exists
+[AfterEach] [sig-storage] Projected downwardAPI
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 13:05:12.627: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "projected-1256" for this suite.
+Jun 18 13:05:18.674: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 13:05:20.750: INFO: namespace projected-1256 deletion completed in 8.106987964s
 
-• [SLOW TEST:16.781 seconds]
-[sig-apps] ReplicationController
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22
-  should serve a basic image on each replica with a public image  [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+• [SLOW TEST:10.509 seconds]
+[sig-storage] Projected downwardAPI
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:33
+  should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-SSSSSS
+SSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-storage] Downward API volume 
-  should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [sig-storage] Downward API volume
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+[k8s.io] InitContainer [NodeConformance] 
+  should invoke init containers on a RestartNever pod [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [k8s.io] InitContainer [NodeConformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 19:49:16.980: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-STEP: Building a namespace api object, basename downward-api
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in downward-api-9135
+Jun 18 13:05:20.751: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename init-container
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in init-container-1955
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-storage] Downward API volume
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:39
-[It] should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: Creating a pod to test downward API volume plugin
-Jun  4 19:49:17.217: INFO: Waiting up to 5m0s for pod "downwardapi-volume-d6527237-8701-11e9-8d1b-467ee19922ac" in namespace "downward-api-9135" to be "success or failure"
-Jun  4 19:49:17.232: INFO: Pod "downwardapi-volume-d6527237-8701-11e9-8d1b-467ee19922ac": Phase="Pending", Reason="", readiness=false. Elapsed: 15.458431ms
-Jun  4 19:49:19.248: INFO: Pod "downwardapi-volume-d6527237-8701-11e9-8d1b-467ee19922ac": Phase="Running", Reason="", readiness=true. Elapsed: 2.031621818s
-Jun  4 19:49:21.262: INFO: Pod "downwardapi-volume-d6527237-8701-11e9-8d1b-467ee19922ac": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.045257791s
-STEP: Saw pod success
-Jun  4 19:49:21.262: INFO: Pod "downwardapi-volume-d6527237-8701-11e9-8d1b-467ee19922ac" satisfied condition "success or failure"
-Jun  4 19:49:21.274: INFO: Trying to get logs from node 10.212.23.164 pod downwardapi-volume-d6527237-8701-11e9-8d1b-467ee19922ac container client-container: 
-STEP: delete the pod
-Jun  4 19:49:21.343: INFO: Waiting for pod downwardapi-volume-d6527237-8701-11e9-8d1b-467ee19922ac to disappear
-Jun  4 19:49:21.353: INFO: Pod downwardapi-volume-d6527237-8701-11e9-8d1b-467ee19922ac no longer exists
-[AfterEach] [sig-storage] Downward API volume
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 19:49:21.353: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "downward-api-9135" for this suite.
-Jun  4 19:49:27.411: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 19:49:27.784: INFO: namespace downward-api-9135 deletion completed in 6.415491016s
+[BeforeEach] [k8s.io] InitContainer [NodeConformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/init_container.go:43
+[It] should invoke init containers on a RestartNever pod [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: creating the pod
+Jun 18 13:05:20.946: INFO: PodSpec: initContainers in spec.initContainers
+[AfterEach] [k8s.io] InitContainer [NodeConformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 13:05:24.583: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "init-container-1955" for this suite.
+Jun 18 13:05:30.679: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 13:05:31.558: INFO: namespace init-container-1955 deletion completed in 6.906323589s
 
-• [SLOW TEST:10.805 seconds]
-[sig-storage] Downward API volume
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:34
-  should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+• [SLOW TEST:10.807 seconds]
+[k8s.io] InitContainer [NodeConformance]
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
+  should invoke init containers on a RestartNever pod [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-SSSS
+SSSSSSSSSSSSSSSSSSS
 ------------------------------
 [k8s.io] Kubelet when scheduling a busybox command that always fails in a pod 
   should be possible to delete [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 [BeforeEach] [k8s.io] Kubelet
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 19:49:27.785: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
+Jun 18 13:05:31.558: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
 STEP: Building a namespace api object, basename kubelet-test
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubelet-test-5304
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubelet-test-6443
 STEP: Waiting for a default service account to be provisioned in namespace
 [BeforeEach] [k8s.io] Kubelet
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:37
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:37
 [BeforeEach] when scheduling a busybox command that always fails in a pod
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:81
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:81
 [It] should be possible to delete [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 [AfterEach] [k8s.io] Kubelet
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 19:49:28.038: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "kubelet-test-5304" for this suite.
-Jun  4 19:49:34.093: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 19:49:34.483: INFO: namespace kubelet-test-5304 deletion completed in 6.429983844s
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 13:05:31.861: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "kubelet-test-6443" for this suite.
+Jun 18 13:05:39.900: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 13:05:41.034: INFO: namespace kubelet-test-6443 deletion completed in 9.159682494s
 
-• [SLOW TEST:6.698 seconds]
+• [SLOW TEST:9.476 seconds]
 [k8s.io] Kubelet
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
   when scheduling a busybox command that always fails in a pod
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:78
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:78
     should be possible to delete [NodeConformance] [Conformance]
-    /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+    /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-SSSSSS
+SSSSSSSSS
 ------------------------------
 [sig-storage] EmptyDir volumes 
-  should support (root,0666,default) [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+  should support (root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 [BeforeEach] [sig-storage] EmptyDir volumes
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 19:49:34.483: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
+Jun 18 13:05:41.034: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
 STEP: Building a namespace api object, basename emptydir
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in emptydir-1226
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in emptydir-5544
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should support (root,0666,default) [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: Creating a pod to test emptydir 0666 on node default medium
-Jun  4 19:49:34.720: INFO: Waiting up to 5m0s for pod "pod-e0c15f0c-8701-11e9-8d1b-467ee19922ac" in namespace "emptydir-1226" to be "success or failure"
-Jun  4 19:49:34.731: INFO: Pod "pod-e0c15f0c-8701-11e9-8d1b-467ee19922ac": Phase="Pending", Reason="", readiness=false. Elapsed: 10.757477ms
-Jun  4 19:49:36.743: INFO: Pod "pod-e0c15f0c-8701-11e9-8d1b-467ee19922ac": Phase="Running", Reason="", readiness=true. Elapsed: 2.023476051s
-Jun  4 19:49:38.757: INFO: Pod "pod-e0c15f0c-8701-11e9-8d1b-467ee19922ac": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.037307675s
+[It] should support (root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating a pod to test emptydir 0666 on tmpfs
+Jun 18 13:05:41.264: INFO: Waiting up to 5m0s for pod "pod-c6447e53-91c9-11e9-a25d-8608290c688a" in namespace "emptydir-5544" to be "success or failure"
+Jun 18 13:05:41.277: INFO: Pod "pod-c6447e53-91c9-11e9-a25d-8608290c688a": Phase="Pending", Reason="", readiness=false. Elapsed: 13.314797ms
+Jun 18 13:05:43.291: INFO: Pod "pod-c6447e53-91c9-11e9-a25d-8608290c688a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.026642153s
 STEP: Saw pod success
-Jun  4 19:49:38.757: INFO: Pod "pod-e0c15f0c-8701-11e9-8d1b-467ee19922ac" satisfied condition "success or failure"
-Jun  4 19:49:38.769: INFO: Trying to get logs from node 10.212.23.161 pod pod-e0c15f0c-8701-11e9-8d1b-467ee19922ac container test-container: 
+Jun 18 13:05:43.291: INFO: Pod "pod-c6447e53-91c9-11e9-a25d-8608290c688a" satisfied condition "success or failure"
+Jun 18 13:05:43.324: INFO: Trying to get logs from node 10.72.74.184 pod pod-c6447e53-91c9-11e9-a25d-8608290c688a container test-container: 
 STEP: delete the pod
-Jun  4 19:49:38.847: INFO: Waiting for pod pod-e0c15f0c-8701-11e9-8d1b-467ee19922ac to disappear
-Jun  4 19:49:38.856: INFO: Pod pod-e0c15f0c-8701-11e9-8d1b-467ee19922ac no longer exists
+Jun 18 13:05:43.402: INFO: Waiting for pod pod-c6447e53-91c9-11e9-a25d-8608290c688a to disappear
+Jun 18 13:05:43.414: INFO: Pod pod-c6447e53-91c9-11e9-a25d-8608290c688a no longer exists
 [AfterEach] [sig-storage] EmptyDir volumes
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 19:49:38.856: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "emptydir-1226" for this suite.
-Jun  4 19:49:44.918: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 19:49:45.270: INFO: namespace emptydir-1226 deletion completed in 6.399496169s
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 13:05:43.414: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "emptydir-5544" for this suite.
+Jun 18 13:05:49.454: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 13:05:50.343: INFO: namespace emptydir-5544 deletion completed in 6.914111428s
 
-• [SLOW TEST:10.787 seconds]
+• [SLOW TEST:9.309 seconds]
 [sig-storage] EmptyDir volumes
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:41
-  should support (root,0666,default) [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:41
+  should support (root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-SSSSSSSSSSSSSSSSSSSSSSSS
+SSSSSSSSSS
 ------------------------------
-[sig-storage] Projected secret 
-  optional updates should be reflected in volume [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [sig-storage] Projected secret
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+[sig-storage] Projected downwardAPI 
+  should provide container's cpu request [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-storage] Projected downwardAPI
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 19:49:45.272: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
+Jun 18 13:05:50.344: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
 STEP: Building a namespace api object, basename projected
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-8377
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-1704
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] optional updates should be reflected in volume [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: Creating secret with name s-test-opt-del-e731da84-8701-11e9-8d1b-467ee19922ac
-STEP: Creating secret with name s-test-opt-upd-e731dac7-8701-11e9-8d1b-467ee19922ac
-STEP: Creating the pod
-STEP: Deleting secret s-test-opt-del-e731da84-8701-11e9-8d1b-467ee19922ac
-STEP: Updating secret s-test-opt-upd-e731dac7-8701-11e9-8d1b-467ee19922ac
-STEP: Creating secret with name s-test-opt-create-e731dae2-8701-11e9-8d1b-467ee19922ac
-STEP: waiting to observe update in volume
-[AfterEach] [sig-storage] Projected secret
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 19:50:54.855: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "projected-8377" for this suite.
-Jun  4 19:51:18.913: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 19:51:19.292: INFO: namespace projected-8377 deletion completed in 24.421563271s
+[BeforeEach] [sig-storage] Projected downwardAPI
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:39
+[It] should provide container's cpu request [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating a pod to test downward API volume plugin
+Jun 18 13:05:50.578: INFO: Waiting up to 5m0s for pod "downwardapi-volume-cbd23c7c-91c9-11e9-a25d-8608290c688a" in namespace "projected-1704" to be "success or failure"
+Jun 18 13:05:50.591: INFO: Pod "downwardapi-volume-cbd23c7c-91c9-11e9-a25d-8608290c688a": Phase="Pending", Reason="", readiness=false. Elapsed: 13.485532ms
+Jun 18 13:05:52.607: INFO: Pod "downwardapi-volume-cbd23c7c-91c9-11e9-a25d-8608290c688a": Phase="Running", Reason="", readiness=true. Elapsed: 2.029356244s
+Jun 18 13:05:54.926: INFO: Pod "downwardapi-volume-cbd23c7c-91c9-11e9-a25d-8608290c688a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.348219005s
+STEP: Saw pod success
+Jun 18 13:05:54.926: INFO: Pod "downwardapi-volume-cbd23c7c-91c9-11e9-a25d-8608290c688a" satisfied condition "success or failure"
+Jun 18 13:05:54.940: INFO: Trying to get logs from node 10.72.74.184 pod downwardapi-volume-cbd23c7c-91c9-11e9-a25d-8608290c688a container client-container: 
+STEP: delete the pod
+Jun 18 13:05:55.013: INFO: Waiting for pod downwardapi-volume-cbd23c7c-91c9-11e9-a25d-8608290c688a to disappear
+Jun 18 13:05:55.030: INFO: Pod downwardapi-volume-cbd23c7c-91c9-11e9-a25d-8608290c688a no longer exists
+[AfterEach] [sig-storage] Projected downwardAPI
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 13:05:55.030: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "projected-1704" for this suite.
+Jun 18 13:06:03.639: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 13:06:04.106: INFO: namespace projected-1704 deletion completed in 9.062615478s
 
-• [SLOW TEST:94.021 seconds]
-[sig-storage] Projected secret
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_secret.go:33
-  optional updates should be reflected in volume [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+• [SLOW TEST:13.763 seconds]
+[sig-storage] Projected downwardAPI
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:33
+  should provide container's cpu request [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSSSSSSSSSSSSSSSSS
+------------------------------
+[k8s.io] [sig-node] Pods Extended [k8s.io] Pods Set QOS Class 
+  should be submitted and removed  [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [k8s.io] [sig-node] Pods Extended
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Jun 18 13:06:04.107: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename pods
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in pods-4778
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [k8s.io] Pods Set QOS Class
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/node/pods.go:177
+[It] should be submitted and removed  [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: creating the pod
+STEP: submitting the pod to kubernetes
+STEP: verifying QOS class is set on the pod
+[AfterEach] [k8s.io] [sig-node] Pods Extended
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 13:06:04.355: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "pods-4778" for this suite.
+Jun 18 13:06:28.395: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 13:06:28.880: INFO: namespace pods-4778 deletion completed in 24.510909443s
+
+• [SLOW TEST:24.772 seconds]
+[k8s.io] [sig-node] Pods Extended
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
+  [k8s.io] Pods Set QOS Class
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
+    should be submitted and removed  [Conformance]
+    /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSSSSSS
+------------------------------
+[sig-storage] Projected downwardAPI 
+  should set mode on item file [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-storage] Projected downwardAPI
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Jun 18 13:06:28.880: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename projected
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-9350
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [sig-storage] Projected downwardAPI
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:39
+[It] should set mode on item file [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating a pod to test downward API volume plugin
+Jun 18 13:06:29.103: INFO: Waiting up to 5m0s for pod "downwardapi-volume-e2c89e92-91c9-11e9-a25d-8608290c688a" in namespace "projected-9350" to be "success or failure"
+Jun 18 13:06:29.116: INFO: Pod "downwardapi-volume-e2c89e92-91c9-11e9-a25d-8608290c688a": Phase="Pending", Reason="", readiness=false. Elapsed: 13.052404ms
+Jun 18 13:06:31.129: INFO: Pod "downwardapi-volume-e2c89e92-91c9-11e9-a25d-8608290c688a": Phase="Pending", Reason="", readiness=false. Elapsed: 2.026548737s
+Jun 18 13:06:33.143: INFO: Pod "downwardapi-volume-e2c89e92-91c9-11e9-a25d-8608290c688a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.040460399s
+STEP: Saw pod success
+Jun 18 13:06:33.143: INFO: Pod "downwardapi-volume-e2c89e92-91c9-11e9-a25d-8608290c688a" satisfied condition "success or failure"
+Jun 18 13:06:33.160: INFO: Trying to get logs from node 10.72.74.184 pod downwardapi-volume-e2c89e92-91c9-11e9-a25d-8608290c688a container client-container: 
+STEP: delete the pod
+Jun 18 13:06:33.238: INFO: Waiting for pod downwardapi-volume-e2c89e92-91c9-11e9-a25d-8608290c688a to disappear
+Jun 18 13:06:33.252: INFO: Pod downwardapi-volume-e2c89e92-91c9-11e9-a25d-8608290c688a no longer exists
+[AfterEach] [sig-storage] Projected downwardAPI
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 13:06:33.252: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "projected-9350" for this suite.
+Jun 18 13:06:39.297: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 13:06:39.695: INFO: namespace projected-9350 deletion completed in 6.427752551s
+
+• [SLOW TEST:10.815 seconds]
+[sig-storage] Projected downwardAPI
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:33
+  should set mode on item file [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
 SSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-node] Downward API 
-  should provide pod UID as env vars [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [sig-node] Downward API
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+[sig-storage] EmptyDir volumes 
+  should support (non-root,0777,default) [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-storage] EmptyDir volumes
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 19:51:19.293: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-STEP: Building a namespace api object, basename downward-api
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in downward-api-3534
+Jun 18 13:06:39.696: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename emptydir
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in emptydir-4907
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should provide pod UID as env vars [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: Creating a pod to test downward api env vars
-Jun  4 19:51:19.536: INFO: Waiting up to 5m0s for pod "downward-api-1f3b188d-8702-11e9-8d1b-467ee19922ac" in namespace "downward-api-3534" to be "success or failure"
-Jun  4 19:51:19.547: INFO: Pod "downward-api-1f3b188d-8702-11e9-8d1b-467ee19922ac": Phase="Pending", Reason="", readiness=false. Elapsed: 11.072792ms
-Jun  4 19:51:21.560: INFO: Pod "downward-api-1f3b188d-8702-11e9-8d1b-467ee19922ac": Phase="Pending", Reason="", readiness=false. Elapsed: 2.023711678s
-Jun  4 19:51:23.574: INFO: Pod "downward-api-1f3b188d-8702-11e9-8d1b-467ee19922ac": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.037316014s
+[It] should support (non-root,0777,default) [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating a pod to test emptydir 0777 on node default medium
+Jun 18 13:06:39.917: INFO: Waiting up to 5m0s for pod "pod-e93aa044-91c9-11e9-a25d-8608290c688a" in namespace "emptydir-4907" to be "success or failure"
+Jun 18 13:06:39.935: INFO: Pod "pod-e93aa044-91c9-11e9-a25d-8608290c688a": Phase="Pending", Reason="", readiness=false. Elapsed: 17.654471ms
+Jun 18 13:06:41.950: INFO: Pod "pod-e93aa044-91c9-11e9-a25d-8608290c688a": Phase="Pending", Reason="", readiness=false. Elapsed: 2.032531006s
+Jun 18 13:06:43.964: INFO: Pod "pod-e93aa044-91c9-11e9-a25d-8608290c688a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.046200121s
 STEP: Saw pod success
-Jun  4 19:51:23.574: INFO: Pod "downward-api-1f3b188d-8702-11e9-8d1b-467ee19922ac" satisfied condition "success or failure"
-Jun  4 19:51:23.585: INFO: Trying to get logs from node 10.212.23.164 pod downward-api-1f3b188d-8702-11e9-8d1b-467ee19922ac container dapi-container: 
+Jun 18 13:06:43.964: INFO: Pod "pod-e93aa044-91c9-11e9-a25d-8608290c688a" satisfied condition "success or failure"
+Jun 18 13:06:43.976: INFO: Trying to get logs from node 10.72.74.138 pod pod-e93aa044-91c9-11e9-a25d-8608290c688a container test-container: 
 STEP: delete the pod
-Jun  4 19:51:23.656: INFO: Waiting for pod downward-api-1f3b188d-8702-11e9-8d1b-467ee19922ac to disappear
-Jun  4 19:51:23.668: INFO: Pod downward-api-1f3b188d-8702-11e9-8d1b-467ee19922ac no longer exists
-[AfterEach] [sig-node] Downward API
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 19:51:23.668: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "downward-api-3534" for this suite.
-Jun  4 19:51:31.724: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 19:51:32.088: INFO: namespace downward-api-3534 deletion completed in 8.40545936s
+Jun 18 13:06:44.068: INFO: Waiting for pod pod-e93aa044-91c9-11e9-a25d-8608290c688a to disappear
+Jun 18 13:06:44.086: INFO: Pod pod-e93aa044-91c9-11e9-a25d-8608290c688a no longer exists
+[AfterEach] [sig-storage] EmptyDir volumes
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 13:06:44.086: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "emptydir-4907" for this suite.
+Jun 18 13:06:52.127: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 13:06:52.529: INFO: namespace emptydir-4907 deletion completed in 8.429295107s
 
-• [SLOW TEST:12.796 seconds]
-[sig-node] Downward API
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downward_api.go:38
-  should provide pod UID as env vars [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+• [SLOW TEST:12.834 seconds]
+[sig-storage] EmptyDir volumes
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:41
+  should support (non-root,0777,default) [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-SSSSSSSSSSSSSSSSSS
+SSSSSSSSSS
 ------------------------------
-[sig-storage] Subpath Atomic writer volumes 
-  should support subpaths with configmap pod with mountPath of existing file [LinuxOnly] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [sig-storage] Subpath
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+[sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] 
+  should perform rolling updates and roll backs of template modifications [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-apps] StatefulSet
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Jun 18 13:06:52.530: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename statefulset
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in statefulset-7958
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [sig-apps] StatefulSet
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:59
+[BeforeEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:74
+STEP: Creating service test in namespace statefulset-7958
+[It] should perform rolling updates and roll backs of template modifications [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating a new StatefulSet
+Jun 18 13:06:52.771: INFO: Found 0 stateful pods, waiting for 3
+Jun 18 13:07:02.785: INFO: Waiting for pod ss2-0 to enter Running - Ready=true, currently Running - Ready=true
+Jun 18 13:07:02.786: INFO: Waiting for pod ss2-1 to enter Running - Ready=true, currently Running - Ready=true
+Jun 18 13:07:02.786: INFO: Waiting for pod ss2-2 to enter Running - Ready=true, currently Running - Ready=true
+Jun 18 13:07:02.823: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-7958 ss2-1 -- /bin/sh -x -c mv -v /usr/share/nginx/html/index.html /tmp/ || true'
+Jun 18 13:07:03.183: INFO: stderr: "+ mv -v /usr/share/nginx/html/index.html /tmp/\n"
+Jun 18 13:07:03.183: INFO: stdout: "'/usr/share/nginx/html/index.html' -> '/tmp/index.html'\n"
+Jun 18 13:07:03.183: INFO: stdout of mv -v /usr/share/nginx/html/index.html /tmp/ || true on ss2-1: '/usr/share/nginx/html/index.html' -> '/tmp/index.html'
+
+STEP: Updating StatefulSet template: update image from docker.io/library/nginx:1.14-alpine to docker.io/library/nginx:1.15-alpine
+Jun 18 13:07:13.269: INFO: Updating stateful set ss2
+STEP: Creating a new revision
+STEP: Updating Pods in reverse ordinal order
+Jun 18 13:07:23.345: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-7958 ss2-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
+Jun 18 13:07:23.708: INFO: stderr: "+ mv -v /tmp/index.html /usr/share/nginx/html/\n"
+Jun 18 13:07:23.708: INFO: stdout: "'/tmp/index.html' -> '/usr/share/nginx/html/index.html'\n"
+Jun 18 13:07:23.708: INFO: stdout of mv -v /tmp/index.html /usr/share/nginx/html/ || true on ss2-1: '/tmp/index.html' -> '/usr/share/nginx/html/index.html'
+
+STEP: Rolling back to a previous revision
+Jun 18 13:07:43.883: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-7958 ss2-1 -- /bin/sh -x -c mv -v /usr/share/nginx/html/index.html /tmp/ || true'
+Jun 18 13:07:44.303: INFO: stderr: "+ mv -v /usr/share/nginx/html/index.html /tmp/\n"
+Jun 18 13:07:44.303: INFO: stdout: "'/usr/share/nginx/html/index.html' -> '/tmp/index.html'\n"
+Jun 18 13:07:44.303: INFO: stdout of mv -v /usr/share/nginx/html/index.html /tmp/ || true on ss2-1: '/usr/share/nginx/html/index.html' -> '/tmp/index.html'
+
+Jun 18 13:07:54.396: INFO: Updating stateful set ss2
+STEP: Rolling back update in reverse ordinal order
+Jun 18 13:08:04.489: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 exec --namespace=statefulset-7958 ss2-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
+Jun 18 13:08:04.855: INFO: stderr: "+ mv -v /tmp/index.html /usr/share/nginx/html/\n"
+Jun 18 13:08:04.855: INFO: stdout: "'/tmp/index.html' -> '/usr/share/nginx/html/index.html'\n"
+Jun 18 13:08:04.855: INFO: stdout of mv -v /tmp/index.html /usr/share/nginx/html/ || true on ss2-1: '/tmp/index.html' -> '/usr/share/nginx/html/index.html'
+
+Jun 18 13:08:04.907: INFO: Waiting for StatefulSet statefulset-7958/ss2 to complete update
+Jun 18 13:08:04.907: INFO: Waiting for Pod statefulset-7958/ss2-0 to have revision ss2-787997d666 update revision ss2-c79899b9
+Jun 18 13:08:04.907: INFO: Waiting for Pod statefulset-7958/ss2-1 to have revision ss2-787997d666 update revision ss2-c79899b9
+Jun 18 13:08:04.907: INFO: Waiting for Pod statefulset-7958/ss2-2 to have revision ss2-787997d666 update revision ss2-c79899b9
+Jun 18 13:08:14.969: INFO: Waiting for StatefulSet statefulset-7958/ss2 to complete update
+Jun 18 13:08:14.969: INFO: Waiting for Pod statefulset-7958/ss2-0 to have revision ss2-787997d666 update revision ss2-c79899b9
+[AfterEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:85
+Jun 18 13:08:24.941: INFO: Deleting all statefulset in ns statefulset-7958
+Jun 18 13:08:24.952: INFO: Scaling statefulset ss2 to 0
+Jun 18 13:08:45.005: INFO: Waiting for statefulset status.replicas updated to 0
+Jun 18 13:08:45.018: INFO: Deleting statefulset ss2
+[AfterEach] [sig-apps] StatefulSet
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 13:08:45.078: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "statefulset-7958" for this suite.
+Jun 18 13:08:53.122: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 13:08:53.536: INFO: namespace statefulset-7958 deletion completed in 8.441193648s
+
+• [SLOW TEST:121.006 seconds]
+[sig-apps] StatefulSet
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22
+  [k8s.io] Basic StatefulSet functionality [StatefulSetBasic]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
+    should perform rolling updates and roll backs of template modifications [Conformance]
+    /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSSSSSSSSSSSS
+------------------------------
+[sig-cli] Kubectl client [k8s.io] Kubectl label 
+  should update the label on a resource  [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-cli] Kubectl client
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Jun 18 13:08:53.536: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename kubectl
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-8681
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [sig-cli] Kubectl client
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:213
+[BeforeEach] [k8s.io] Kubectl label
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1108
+STEP: creating the pod
+Jun 18 13:08:53.736: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 create -f - --namespace=kubectl-8681'
+Jun 18 13:08:54.152: INFO: stderr: ""
+Jun 18 13:08:54.152: INFO: stdout: "pod/pause created\n"
+Jun 18 13:08:54.152: INFO: Waiting up to 5m0s for 1 pods to be running and ready: [pause]
+Jun 18 13:08:54.152: INFO: Waiting up to 5m0s for pod "pause" in namespace "kubectl-8681" to be "running and ready"
+Jun 18 13:08:54.166: INFO: Pod "pause": Phase="Pending", Reason="", readiness=false. Elapsed: 14.46712ms
+Jun 18 13:08:56.180: INFO: Pod "pause": Phase="Running", Reason="", readiness=true. Elapsed: 2.02802797s
+Jun 18 13:08:56.180: INFO: Pod "pause" satisfied condition "running and ready"
+Jun 18 13:08:56.180: INFO: Wanted all 1 pods to be running and ready. Result: true. Pods: [pause]
+[It] should update the label on a resource  [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: adding the label testing-label with value testing-label-value to a pod
+Jun 18 13:08:56.180: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 label pods pause testing-label=testing-label-value --namespace=kubectl-8681'
+Jun 18 13:08:56.314: INFO: stderr: ""
+Jun 18 13:08:56.314: INFO: stdout: "pod/pause labeled\n"
+STEP: verifying the pod has the label testing-label with the value testing-label-value
+Jun 18 13:08:56.314: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 get pod pause -L testing-label --namespace=kubectl-8681'
+Jun 18 13:08:56.431: INFO: stderr: ""
+Jun 18 13:08:56.431: INFO: stdout: "NAME    READY   STATUS    RESTARTS   AGE   TESTING-LABEL\npause   1/1     Running   0          2s    testing-label-value\n"
+STEP: removing the label testing-label of a pod
+Jun 18 13:08:56.431: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 label pods pause testing-label- --namespace=kubectl-8681'
+Jun 18 13:08:56.582: INFO: stderr: ""
+Jun 18 13:08:56.582: INFO: stdout: "pod/pause labeled\n"
+STEP: verifying the pod doesn't have the label testing-label
+Jun 18 13:08:56.582: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 get pod pause -L testing-label --namespace=kubectl-8681'
+Jun 18 13:08:56.729: INFO: stderr: ""
+Jun 18 13:08:56.729: INFO: stdout: "NAME    READY   STATUS    RESTARTS   AGE   TESTING-LABEL\npause   1/1     Running   0          2s    \n"
+[AfterEach] [k8s.io] Kubectl label
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1115
+STEP: using delete to clean up resources
+Jun 18 13:08:56.730: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 delete --grace-period=0 --force -f - --namespace=kubectl-8681'
+Jun 18 13:08:56.897: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n"
+Jun 18 13:08:56.897: INFO: stdout: "pod \"pause\" force deleted\n"
+Jun 18 13:08:56.897: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 get rc,svc -l name=pause --no-headers --namespace=kubectl-8681'
+Jun 18 13:08:57.034: INFO: stderr: "No resources found.\n"
+Jun 18 13:08:57.034: INFO: stdout: ""
+Jun 18 13:08:57.034: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 get pods -l name=pause --namespace=kubectl-8681 -o go-template={{ range .items }}{{ if not .metadata.deletionTimestamp }}{{ .metadata.name }}{{ "\n" }}{{ end }}{{ end }}'
+Jun 18 13:08:57.169: INFO: stderr: ""
+Jun 18 13:08:57.169: INFO: stdout: ""
+[AfterEach] [sig-cli] Kubectl client
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 13:08:57.169: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "kubectl-8681" for this suite.
+Jun 18 13:09:03.222: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 13:09:03.634: INFO: namespace kubectl-8681 deletion completed in 6.449671254s
+
+• [SLOW TEST:10.098 seconds]
+[sig-cli] Kubectl client
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23
+  [k8s.io] Kubectl label
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
+    should update the label on a resource  [Conformance]
+    /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSSSSSSSSSSSSSSSSSSSS
+------------------------------
+[sig-api-machinery] Watchers 
+  should be able to start watching from a specific resource version [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-api-machinery] Watchers
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 19:51:32.089: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-STEP: Building a namespace api object, basename subpath
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in subpath-7030
+Jun 18 13:09:03.635: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename watch
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in watch-3939
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] Atomic writer volumes
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:38
-STEP: Setting up data
-[It] should support subpaths with configmap pod with mountPath of existing file [LinuxOnly] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: Creating pod pod-subpath-test-configmap-7jnp
-STEP: Creating a pod to test atomic-volume-subpath
-Jun  4 19:51:32.356: INFO: Waiting up to 5m0s for pod "pod-subpath-test-configmap-7jnp" in namespace "subpath-7030" to be "success or failure"
-Jun  4 19:51:32.372: INFO: Pod "pod-subpath-test-configmap-7jnp": Phase="Pending", Reason="", readiness=false. Elapsed: 15.864478ms
-Jun  4 19:51:34.386: INFO: Pod "pod-subpath-test-configmap-7jnp": Phase="Running", Reason="", readiness=true. Elapsed: 2.029583413s
-Jun  4 19:51:36.399: INFO: Pod "pod-subpath-test-configmap-7jnp": Phase="Running", Reason="", readiness=true. Elapsed: 4.042507623s
-Jun  4 19:51:38.412: INFO: Pod "pod-subpath-test-configmap-7jnp": Phase="Running", Reason="", readiness=true. Elapsed: 6.055350758s
-Jun  4 19:51:40.426: INFO: Pod "pod-subpath-test-configmap-7jnp": Phase="Running", Reason="", readiness=true. Elapsed: 8.069099749s
-Jun  4 19:51:42.438: INFO: Pod "pod-subpath-test-configmap-7jnp": Phase="Running", Reason="", readiness=true. Elapsed: 10.081066973s
-Jun  4 19:51:44.451: INFO: Pod "pod-subpath-test-configmap-7jnp": Phase="Running", Reason="", readiness=true. Elapsed: 12.094521199s
-Jun  4 19:51:46.464: INFO: Pod "pod-subpath-test-configmap-7jnp": Phase="Running", Reason="", readiness=true. Elapsed: 14.107160634s
-Jun  4 19:51:48.478: INFO: Pod "pod-subpath-test-configmap-7jnp": Phase="Running", Reason="", readiness=true. Elapsed: 16.121867264s
-Jun  4 19:51:50.491: INFO: Pod "pod-subpath-test-configmap-7jnp": Phase="Running", Reason="", readiness=true. Elapsed: 18.134209239s
-Jun  4 19:51:52.504: INFO: Pod "pod-subpath-test-configmap-7jnp": Phase="Running", Reason="", readiness=true. Elapsed: 20.147156737s
-Jun  4 19:51:54.516: INFO: Pod "pod-subpath-test-configmap-7jnp": Phase="Succeeded", Reason="", readiness=false. Elapsed: 22.160036854s
-STEP: Saw pod success
-Jun  4 19:51:54.517: INFO: Pod "pod-subpath-test-configmap-7jnp" satisfied condition "success or failure"
-Jun  4 19:51:54.529: INFO: Trying to get logs from node 10.212.23.161 pod pod-subpath-test-configmap-7jnp container test-container-subpath-configmap-7jnp: 
-STEP: delete the pod
-Jun  4 19:51:54.594: INFO: Waiting for pod pod-subpath-test-configmap-7jnp to disappear
-Jun  4 19:51:54.605: INFO: Pod pod-subpath-test-configmap-7jnp no longer exists
-STEP: Deleting pod pod-subpath-test-configmap-7jnp
-Jun  4 19:51:54.605: INFO: Deleting pod "pod-subpath-test-configmap-7jnp" in namespace "subpath-7030"
-[AfterEach] [sig-storage] Subpath
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 19:51:54.666: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "subpath-7030" for this suite.
-Jun  4 19:52:00.725: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 19:52:01.069: INFO: namespace subpath-7030 deletion completed in 6.387457834s
+[It] should be able to start watching from a specific resource version [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: creating a new configmap
+STEP: modifying the configmap once
+STEP: modifying the configmap a second time
+STEP: deleting the configmap
+STEP: creating a watch on configmaps from the resource version returned by the first update
+STEP: Expecting to observe notifications for all changes to the configmap after the first update
+Jun 18 13:09:03.986: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-resource-version,GenerateName:,Namespace:watch-3939,SelfLink:/api/v1/namespaces/watch-3939/configmaps/e2e-watch-test-resource-version,UID:3f0b1fa9-91ca-11e9-a08a-ee7a14707756,ResourceVersion:107329,Generation:0,CreationTimestamp:2019-06-18 13:09:03 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: from-resource-version,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},}
+Jun 18 13:09:03.986: INFO: Got : DELETED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-resource-version,GenerateName:,Namespace:watch-3939,SelfLink:/api/v1/namespaces/watch-3939/configmaps/e2e-watch-test-resource-version,UID:3f0b1fa9-91ca-11e9-a08a-ee7a14707756,ResourceVersion:107330,Generation:0,CreationTimestamp:2019-06-18 13:09:03 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: from-resource-version,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},}
+[AfterEach] [sig-api-machinery] Watchers
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 13:09:03.986: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "watch-3939" for this suite.
+Jun 18 13:09:10.026: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 13:09:11.727: INFO: namespace watch-3939 deletion completed in 7.728249928s
 
-• [SLOW TEST:28.981 seconds]
-[sig-storage] Subpath
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:22
-  Atomic writer volumes
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:34
-    should support subpaths with configmap pod with mountPath of existing file [LinuxOnly] [Conformance]
-    /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+• [SLOW TEST:8.092 seconds]
+[sig-api-machinery] Watchers
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:22
+  should be able to start watching from a specific resource version [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-SSSSSSSSSSSSS
+SSSSSSSSSSSSSSSS
 ------------------------------
-[sig-apps] ReplicaSet 
-  should adopt matching pods on creation and release no longer matching pods [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [sig-apps] ReplicaSet
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+[sig-api-machinery] CustomResourceDefinition resources Simple CustomResourceDefinition 
+  creating/deleting custom resource definition objects works  [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-api-machinery] CustomResourceDefinition resources
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 19:52:01.070: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-STEP: Building a namespace api object, basename replicaset
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in replicaset-4762
+Jun 18 13:09:11.727: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename custom-resource-definition
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in custom-resource-definition-9039
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should adopt matching pods on creation and release no longer matching pods [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: Given a Pod with a 'name' label pod-adoption-release is created
-STEP: When a replicaset with a matching selector is created
-STEP: Then the orphan pod is adopted
-STEP: When the matched label of one of its pods change
-Jun  4 19:52:04.426: INFO: Pod name pod-adoption-release: Found 1 pods out of 1
-STEP: Then the pod is released
-[AfterEach] [sig-apps] ReplicaSet
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 19:52:04.472: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "replicaset-4762" for this suite.
-Jun  4 19:52:28.528: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 19:52:28.954: INFO: namespace replicaset-4762 deletion completed in 24.467507189s
+[It] creating/deleting custom resource definition objects works  [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+Jun 18 13:09:11.926: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+[AfterEach] [sig-api-machinery] CustomResourceDefinition resources
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 13:09:13.673: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "custom-resource-definition-9039" for this suite.
+Jun 18 13:09:20.364: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 13:09:20.780: INFO: namespace custom-resource-definition-9039 deletion completed in 7.092067347s
 
-• [SLOW TEST:27.884 seconds]
-[sig-apps] ReplicaSet
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22
-  should adopt matching pods on creation and release no longer matching pods [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+• [SLOW TEST:9.053 seconds]
+[sig-api-machinery] CustomResourceDefinition resources
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:22
+  Simple CustomResourceDefinition
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/custom_resource_definition.go:35
+    creating/deleting custom resource definition objects works  [Conformance]
+    /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-SSSSSSSSSSSSS
+SSSS
 ------------------------------
-[k8s.io] Probing container 
-  should *not* be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [k8s.io] Probing container
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+[k8s.io] Kubelet when scheduling a busybox command in a pod 
+  should print the output to logs [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [k8s.io] Kubelet
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 19:52:28.954: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-STEP: Building a namespace api object, basename container-probe
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in container-probe-6082
+Jun 18 13:09:20.780: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename kubelet-test
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubelet-test-9786
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [k8s.io] Probing container
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/container_probe.go:51
-[It] should *not* be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: Creating pod liveness-exec in namespace container-probe-6082
-Jun  4 19:52:33.226: INFO: Started pod liveness-exec in namespace container-probe-6082
-STEP: checking the pod's current state and verifying that restartCount is present
-Jun  4 19:52:33.239: INFO: Initial restart count of pod liveness-exec is 0
-STEP: deleting the pod
-[AfterEach] [k8s.io] Probing container
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 19:56:35.079: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "container-probe-6082" for this suite.
-Jun  4 19:56:41.178: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 19:56:41.575: INFO: namespace container-probe-6082 deletion completed in 6.480706129s
+[BeforeEach] [k8s.io] Kubelet
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:37
+[It] should print the output to logs [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[AfterEach] [k8s.io] Kubelet
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 13:09:23.074: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "kubelet-test-9786" for this suite.
+Jun 18 13:10:15.431: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 13:10:15.832: INFO: namespace kubelet-test-9786 deletion completed in 52.72458513s
 
-• [SLOW TEST:252.622 seconds]
-[k8s.io] Probing container
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
-  should *not* be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+• [SLOW TEST:55.052 seconds]
+[k8s.io] Kubelet
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
+  when scheduling a busybox command in a pod
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:40
+    should print the output to logs [NodeConformance] [Conformance]
+    /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-node] Downward API 
-  should provide container's limits.cpu/memory and requests.cpu/memory as env vars [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [sig-node] Downward API
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+[sig-apps] Deployment 
+  deployment should delete old replica sets [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-apps] Deployment
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 19:56:41.579: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-STEP: Building a namespace api object, basename downward-api
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in downward-api-6157
+Jun 18 13:10:15.834: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename deployment
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in deployment-6599
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should provide container's limits.cpu/memory and requests.cpu/memory as env vars [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: Creating a pod to test downward api env vars
-Jun  4 19:56:41.882: INFO: Waiting up to 5m0s for pod "downward-api-df5ceb25-8702-11e9-8d1b-467ee19922ac" in namespace "downward-api-6157" to be "success or failure"
-Jun  4 19:56:41.894: INFO: Pod "downward-api-df5ceb25-8702-11e9-8d1b-467ee19922ac": Phase="Pending", Reason="", readiness=false. Elapsed: 11.826138ms
-Jun  4 19:56:43.908: INFO: Pod "downward-api-df5ceb25-8702-11e9-8d1b-467ee19922ac": Phase="Running", Reason="", readiness=true. Elapsed: 2.025421435s
-Jun  4 19:56:45.922: INFO: Pod "downward-api-df5ceb25-8702-11e9-8d1b-467ee19922ac": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.039357737s
-STEP: Saw pod success
-Jun  4 19:56:45.922: INFO: Pod "downward-api-df5ceb25-8702-11e9-8d1b-467ee19922ac" satisfied condition "success or failure"
-Jun  4 19:56:45.933: INFO: Trying to get logs from node 10.212.23.164 pod downward-api-df5ceb25-8702-11e9-8d1b-467ee19922ac container dapi-container: 
-STEP: delete the pod
-Jun  4 19:56:46.002: INFO: Waiting for pod downward-api-df5ceb25-8702-11e9-8d1b-467ee19922ac to disappear
-Jun  4 19:56:46.013: INFO: Pod downward-api-df5ceb25-8702-11e9-8d1b-467ee19922ac no longer exists
-[AfterEach] [sig-node] Downward API
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 19:56:46.013: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "downward-api-6157" for this suite.
-Jun  4 19:56:52.071: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 19:56:52.494: INFO: namespace downward-api-6157 deletion completed in 6.466018594s
+[BeforeEach] [sig-apps] Deployment
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:65
+[It] deployment should delete old replica sets [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+Jun 18 13:10:16.071: INFO: Pod name cleanup-pod: Found 0 pods out of 1
+Jun 18 13:10:21.088: INFO: Pod name cleanup-pod: Found 1 pods out of 1
+STEP: ensuring each pod is running
+Jun 18 13:10:21.088: INFO: Creating deployment test-cleanup-deployment
+STEP: Waiting for deployment test-cleanup-deployment history to be cleaned up
+[AfterEach] [sig-apps] Deployment
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:59
+Jun 18 13:10:21.140: INFO: Deployment "test-cleanup-deployment":
+&Deployment{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-cleanup-deployment,GenerateName:,Namespace:deployment-6599,SelfLink:/apis/apps/v1/namespaces/deployment-6599/deployments/test-cleanup-deployment,UID:6d15d073-91ca-11e9-a08a-ee7a14707756,ResourceVersion:107552,Generation:1,CreationTimestamp:2019-06-18 13:10:21 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: cleanup-pod,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:DeploymentSpec{Replicas:*1,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: cleanup-pod,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: cleanup-pod,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[],Containers:[{redis gcr.io/kubernetes-e2e-test-images/redis:1.0 [] []  [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,},},Strategy:DeploymentStrategy{Type:RollingUpdate,RollingUpdate:&RollingUpdateDeployment{MaxUnavailable:25%!,(MISSING)MaxSurge:25%!,(MISSING)},},MinReadySeconds:0,RevisionHistoryLimit:*0,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:0,Replicas:0,UpdatedReplicas:0,AvailableReplicas:0,UnavailableReplicas:0,Conditions:[],ReadyReplicas:0,CollisionCount:nil,},}
+
+Jun 18 13:10:21.154: INFO: New ReplicaSet "test-cleanup-deployment-55cbfbc8f5" of Deployment "test-cleanup-deployment":
+&ReplicaSet{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-cleanup-deployment-55cbfbc8f5,GenerateName:,Namespace:deployment-6599,SelfLink:/apis/apps/v1/namespaces/deployment-6599/replicasets/test-cleanup-deployment-55cbfbc8f5,UID:6d1aa935-91ca-11e9-a08a-ee7a14707756,ResourceVersion:107554,Generation:1,CreationTimestamp:2019-06-18 13:10:21 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: cleanup-pod,pod-template-hash: 55cbfbc8f5,},Annotations:map[string]string{deployment.kubernetes.io/desired-replicas: 1,deployment.kubernetes.io/max-replicas: 2,deployment.kubernetes.io/revision: 1,},OwnerReferences:[{apps/v1 Deployment test-cleanup-deployment 6d15d073-91ca-11e9-a08a-ee7a14707756 0xc002836d67 0xc002836d68}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:ReplicaSetSpec{Replicas:*1,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: cleanup-pod,pod-template-hash: 55cbfbc8f5,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: cleanup-pod,pod-template-hash: 55cbfbc8f5,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[],Containers:[{redis gcr.io/kubernetes-e2e-test-images/redis:1.0 [] []  [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,},},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:0,FullyLabeledReplicas:0,ObservedGeneration:0,ReadyReplicas:0,AvailableReplicas:0,Conditions:[],},}
+Jun 18 13:10:21.155: INFO: All old ReplicaSets of Deployment "test-cleanup-deployment":
+Jun 18 13:10:21.155: INFO: &ReplicaSet{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-cleanup-controller,GenerateName:,Namespace:deployment-6599,SelfLink:/apis/apps/v1/namespaces/deployment-6599/replicasets/test-cleanup-controller,UID:6a0f9170-91ca-11e9-a08a-ee7a14707756,ResourceVersion:107553,Generation:1,CreationTimestamp:2019-06-18 13:10:16 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: cleanup-pod,pod: nginx,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 Deployment test-cleanup-deployment 6d15d073-91ca-11e9-a08a-ee7a14707756 0xc002836b77 0xc002836b78}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:ReplicaSetSpec{Replicas:*1,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: cleanup-pod,pod: nginx,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: cleanup-pod,pod: nginx,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] []  [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,},},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:1,FullyLabeledReplicas:1,ObservedGeneration:1,ReadyReplicas:1,AvailableReplicas:1,Conditions:[],},}
+Jun 18 13:10:21.167: INFO: Pod "test-cleanup-controller-8f7px" is available:
+&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-cleanup-controller-8f7px,GenerateName:test-cleanup-controller-,Namespace:deployment-6599,SelfLink:/api/v1/namespaces/deployment-6599/pods/test-cleanup-controller-8f7px,UID:6a144fd2-91ca-11e9-a08a-ee7a14707756,ResourceVersion:107547,Generation:0,CreationTimestamp:2019-06-18 13:10:16 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: cleanup-pod,pod: nginx,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet test-cleanup-controller 6a0f9170-91ca-11e9-a08a-ee7a14707756 0xc002837977 0xc002837978}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-xtnlv {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-xtnlv,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] []  [] [] [] {map[] map[]} [{default-token-xtnlv true /var/run/secrets/kubernetes.io/serviceaccount   }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.72.74.184,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists  NoExecute 0xc002837b00} {node.kubernetes.io/unreachable Exists  NoExecute 0xc002837b20}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Running,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 13:10:16 +0000 UTC  } {Ready True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 13:10:17 +0000 UTC  } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 13:10:17 +0000 UTC  } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-06-18 13:10:16 +0000 UTC  }],Message:,Reason:,HostIP:10.72.74.184,PodIP:172.30.142.29,StartTime:2019-06-18 13:10:16 +0000 UTC,ContainerStatuses:[{nginx {nil ContainerStateRunning{StartedAt:2019-06-18 13:10:17 +0000 UTC,} nil} {nil nil nil} true 0 docker.io/library/nginx:1.14-alpine docker.io/library/nginx@sha256:485b610fefec7ff6c463ced9623314a04ed67e3945b9c08d7e53a47f6d108dc7 containerd://4a5d05415918e41b2a55ffd654bb380b6804304f51d74cce4224a9782e7b5b36}],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},}
+Jun 18 13:10:21.168: INFO: Pod "test-cleanup-deployment-55cbfbc8f5-thjbb" is not available:
+&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-cleanup-deployment-55cbfbc8f5-thjbb,GenerateName:test-cleanup-deployment-55cbfbc8f5-,Namespace:deployment-6599,SelfLink:/api/v1/namespaces/deployment-6599/pods/test-cleanup-deployment-55cbfbc8f5-thjbb,UID:6d1cbe1d-91ca-11e9-a08a-ee7a14707756,ResourceVersion:107557,Generation:0,CreationTimestamp:2019-06-18 13:10:21 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: cleanup-pod,pod-template-hash: 55cbfbc8f5,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet test-cleanup-deployment-55cbfbc8f5 6d1aa935-91ca-11e9-a08a-ee7a14707756 0xc002837c57 0xc002837c58}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-xtnlv {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-xtnlv,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{redis gcr.io/kubernetes-e2e-test-images/redis:1.0 [] []  [] [] [] {map[] map[]} [{default-token-xtnlv true /var/run/secrets/kubernetes.io/serviceaccount   }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists  NoExecute 0xc002837d00} {node.kubernetes.io/unreachable Exists  NoExecute 0xc002837dc0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},}
+[AfterEach] [sig-apps] Deployment
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 13:10:21.168: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "deployment-6599" for this suite.
+Jun 18 13:10:29.210: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 13:10:29.599: INFO: namespace deployment-6599 deletion completed in 8.416854061s
 
-• [SLOW TEST:10.915 seconds]
-[sig-node] Downward API
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downward_api.go:38
-  should provide container's limits.cpu/memory and requests.cpu/memory as env vars [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-------------------------------
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+• [SLOW TEST:13.765 seconds]
+[sig-apps] Deployment
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22
+  deployment should delete old replica sets [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-[sig-cli] Kubectl client [k8s.io] Kubectl expose 
-  should create services for rc  [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+[sig-apps] Daemon set [Serial] 
+  should retry creating failed daemon pods [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-apps] Daemon set [Serial]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 19:56:52.495: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-STEP: Building a namespace api object, basename kubectl
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-2736
+Jun 18 13:10:29.599: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename daemonsets
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in daemonsets-9988
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:213
-[It] should create services for rc  [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: creating Redis RC
-Jun  4 19:56:52.709: INFO: namespace kubectl-2736
-Jun  4 19:56:52.709: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 create -f - --namespace=kubectl-2736'
-Jun  4 19:56:53.004: INFO: stderr: ""
-Jun  4 19:56:53.004: INFO: stdout: "replicationcontroller/redis-master created\n"
-STEP: Waiting for Redis master to start.
-Jun  4 19:56:54.016: INFO: Selector matched 1 pods for map[app:redis]
-Jun  4 19:56:54.016: INFO: Found 0 / 1
-Jun  4 19:56:55.018: INFO: Selector matched 1 pods for map[app:redis]
-Jun  4 19:56:55.018: INFO: Found 1 / 1
-Jun  4 19:56:55.018: INFO: WaitFor completed with timeout 5m0s.  Pods found = 1 out of 1
-Jun  4 19:56:55.028: INFO: Selector matched 1 pods for map[app:redis]
-Jun  4 19:56:55.028: INFO: ForEach: Found 1 pods from the filter.  Now looping through them.
-Jun  4 19:56:55.028: INFO: wait on redis-master startup in kubectl-2736 
-Jun  4 19:56:55.028: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 logs redis-master-ls5tc redis-master --namespace=kubectl-2736'
-Jun  4 19:56:55.202: INFO: stderr: ""
-Jun  4 19:56:55.202: INFO: stdout: "                _._                                                  \n           _.-``__ ''-._                                             \n      _.-``    `.  `_.  ''-._           Redis 3.2.12 (35a5711f/0) 64 bit\n  .-`` .-```.  ```\\/    _.,_ ''-._                                   \n (    '      ,       .-`  | `,    )     Running in standalone mode\n |`-._`-...-` __...-.``-._|'` _.-'|     Port: 6379\n |    `-._   `._    /     _.-'    |     PID: 1\n  `-._    `-._  `-./  _.-'    _.-'                                   \n |`-._`-._    `-.__.-'    _.-'_.-'|                                  \n |    `-._`-._        _.-'_.-'    |           http://redis.io        \n  `-._    `-._`-.__.-'_.-'    _.-'                                   \n |`-._`-._    `-.__.-'    _.-'_.-'|                                  \n |    `-._`-._        _.-'_.-'    |                                  \n  `-._    `-._`-.__.-'_.-'    _.-'                                   \n      `-._    `-.__.-'    _.-'                                       \n          `-._        _.-'                                           \n              `-.__.-'                                               \n\n1:M 04 Jun 19:56:54.185 # WARNING: The TCP backlog setting of 511 cannot be enforced because /proc/sys/net/core/somaxconn is set to the lower value of 128.\n1:M 04 Jun 19:56:54.185 # Server started, Redis version 3.2.12\n1:M 04 Jun 19:56:54.185 # WARNING you have Transparent Huge Pages (THP) support enabled in your kernel. This will create latency and memory usage issues with Redis. To fix this issue run the command 'echo never > /sys/kernel/mm/transparent_hugepage/enabled' as root, and add it to your /etc/rc.local in order to retain the setting after a reboot. Redis must be restarted after THP is disabled.\n1:M 04 Jun 19:56:54.185 * The server is now ready to accept connections on port 6379\n"
-STEP: exposing RC
-Jun  4 19:56:55.202: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 expose rc redis-master --name=rm2 --port=1234 --target-port=6379 --namespace=kubectl-2736'
-Jun  4 19:56:55.328: INFO: stderr: ""
-Jun  4 19:56:55.328: INFO: stdout: "service/rm2 exposed\n"
-Jun  4 19:56:55.340: INFO: Service rm2 in namespace kubectl-2736 found.
-STEP: exposing service
-Jun  4 19:56:57.367: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 expose service rm2 --name=rm3 --port=2345 --target-port=6379 --namespace=kubectl-2736'
-Jun  4 19:56:57.503: INFO: stderr: ""
-Jun  4 19:56:57.503: INFO: stdout: "service/rm3 exposed\n"
-Jun  4 19:56:57.513: INFO: Service rm3 in namespace kubectl-2736 found.
-[AfterEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 19:56:59.555: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "kubectl-2736" for this suite.
-Jun  4 19:57:23.627: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 19:57:23.971: INFO: namespace kubectl-2736 deletion completed in 24.399160912s
+[BeforeEach] [sig-apps] Daemon set [Serial]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:102
+[It] should retry creating failed daemon pods [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating a simple DaemonSet "daemon-set"
+STEP: Check that daemon pods launch on every node of the cluster.
+Jun 18 13:10:29.938: INFO: Number of nodes with available pods: 0
+Jun 18 13:10:29.938: INFO: Node 10.72.74.138 is running more than one daemon pod
+Jun 18 13:10:31.026: INFO: Number of nodes with available pods: 0
+Jun 18 13:10:31.027: INFO: Node 10.72.74.138 is running more than one daemon pod
+Jun 18 13:10:32.241: INFO: Number of nodes with available pods: 3
+Jun 18 13:10:32.242: INFO: Number of running nodes: 3, number of available pods: 3
+STEP: Set a daemon pod's phase to 'Failed', check that the daemon pod is revived.
+Jun 18 13:10:32.315: INFO: Number of nodes with available pods: 2
+Jun 18 13:10:32.315: INFO: Node 10.72.74.138 is running more than one daemon pod
+Jun 18 13:10:33.344: INFO: Number of nodes with available pods: 2
+Jun 18 13:10:33.344: INFO: Node 10.72.74.138 is running more than one daemon pod
+Jun 18 13:10:34.348: INFO: Number of nodes with available pods: 3
+Jun 18 13:10:34.348: INFO: Number of running nodes: 3, number of available pods: 3
+STEP: Wait for the failed daemon pod to be completely deleted.
+[AfterEach] [sig-apps] Daemon set [Serial]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:68
+STEP: Deleting DaemonSet "daemon-set"
+STEP: deleting DaemonSet.extensions daemon-set in namespace daemonsets-9988, will wait for the garbage collector to delete the pods
+Jun 18 13:10:34.471: INFO: Deleting DaemonSet.extensions daemon-set took: 32.311252ms
+Jun 18 13:10:34.571: INFO: Terminating DaemonSet.extensions daemon-set pods took: 100.291874ms
+Jun 18 13:10:47.185: INFO: Number of nodes with available pods: 0
+Jun 18 13:10:47.185: INFO: Number of running nodes: 0, number of available pods: 0
+Jun 18 13:10:47.198: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"selfLink":"/apis/apps/v1/namespaces/daemonsets-9988/daemonsets","resourceVersion":"107753"},"items":null}
 
-• [SLOW TEST:31.476 seconds]
-[sig-cli] Kubectl client
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23
-  [k8s.io] Kubectl expose
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
-    should create services for rc  [Conformance]
-    /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+Jun 18 13:10:47.210: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"selfLink":"/api/v1/namespaces/daemonsets-9988/pods","resourceVersion":"107753"},"items":null}
+
+[AfterEach] [sig-apps] Daemon set [Serial]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 13:10:47.260: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "daemonsets-9988" for this suite.
+Jun 18 13:10:55.305: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 13:10:55.736: INFO: namespace daemonsets-9988 deletion completed in 8.461912652s
+
+• [SLOW TEST:26.136 seconds]
+[sig-apps] Daemon set [Serial]
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22
+  should retry creating failed daemon pods [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-SSS
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-storage] Secrets 
-  should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [sig-storage] Secrets
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+[k8s.io] Container Lifecycle Hook when create a pod with lifecycle hook 
+  should execute prestop http hook properly [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [k8s.io] Container Lifecycle Hook
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 19:57:23.972: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-STEP: Building a namespace api object, basename secrets
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in secrets-541
+Jun 18 13:10:55.737: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename container-lifecycle-hook
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in container-lifecycle-hook-1015
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: Creating secret with name secret-test-f897c36c-8702-11e9-8d1b-467ee19922ac
-STEP: Creating a pod to test consume secrets
-Jun  4 19:57:24.286: INFO: Waiting up to 5m0s for pod "pod-secrets-f8a35c7e-8702-11e9-8d1b-467ee19922ac" in namespace "secrets-541" to be "success or failure"
-Jun  4 19:57:24.297: INFO: Pod "pod-secrets-f8a35c7e-8702-11e9-8d1b-467ee19922ac": Phase="Pending", Reason="", readiness=false. Elapsed: 11.202823ms
-Jun  4 19:57:26.310: INFO: Pod "pod-secrets-f8a35c7e-8702-11e9-8d1b-467ee19922ac": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.023830551s
-STEP: Saw pod success
-Jun  4 19:57:26.310: INFO: Pod "pod-secrets-f8a35c7e-8702-11e9-8d1b-467ee19922ac" satisfied condition "success or failure"
-Jun  4 19:57:26.321: INFO: Trying to get logs from node 10.212.23.161 pod pod-secrets-f8a35c7e-8702-11e9-8d1b-467ee19922ac container secret-volume-test: 
-STEP: delete the pod
-Jun  4 19:57:26.389: INFO: Waiting for pod pod-secrets-f8a35c7e-8702-11e9-8d1b-467ee19922ac to disappear
-Jun  4 19:57:26.400: INFO: Pod pod-secrets-f8a35c7e-8702-11e9-8d1b-467ee19922ac no longer exists
-[AfterEach] [sig-storage] Secrets
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 19:57:26.400: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "secrets-541" for this suite.
-Jun  4 19:57:32.467: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 19:57:32.805: INFO: namespace secrets-541 deletion completed in 6.388976023s
+[BeforeEach] when create a pod with lifecycle hook
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:61
+STEP: create the container to handle the HTTPGet hook request.
+[It] should execute prestop http hook properly [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: create the pod with lifecycle hook
+STEP: delete the pod with lifecycle hook
+Jun 18 13:11:02.086: INFO: Waiting for pod pod-with-prestop-http-hook to disappear
+Jun 18 13:11:02.099: INFO: Pod pod-with-prestop-http-hook still exists
+Jun 18 13:11:04.099: INFO: Waiting for pod pod-with-prestop-http-hook to disappear
+Jun 18 13:11:04.145: INFO: Pod pod-with-prestop-http-hook still exists
+Jun 18 13:11:06.099: INFO: Waiting for pod pod-with-prestop-http-hook to disappear
+Jun 18 13:11:06.114: INFO: Pod pod-with-prestop-http-hook still exists
+Jun 18 13:11:08.099: INFO: Waiting for pod pod-with-prestop-http-hook to disappear
+Jun 18 13:11:08.113: INFO: Pod pod-with-prestop-http-hook still exists
+Jun 18 13:11:10.099: INFO: Waiting for pod pod-with-prestop-http-hook to disappear
+Jun 18 13:11:10.113: INFO: Pod pod-with-prestop-http-hook still exists
+Jun 18 13:11:12.099: INFO: Waiting for pod pod-with-prestop-http-hook to disappear
+Jun 18 13:11:12.113: INFO: Pod pod-with-prestop-http-hook no longer exists
+STEP: check prestop hook
+[AfterEach] [k8s.io] Container Lifecycle Hook
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 13:11:12.147: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "container-lifecycle-hook-1015" for this suite.
+Jun 18 13:11:36.198: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 13:11:36.588: INFO: namespace container-lifecycle-hook-1015 deletion completed in 24.426495276s
 
-• [SLOW TEST:8.833 seconds]
-[sig-storage] Secrets
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/secrets_volume.go:33
-  should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+• [SLOW TEST:40.852 seconds]
+[k8s.io] Container Lifecycle Hook
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
+  when create a pod with lifecycle hook
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:40
+    should execute prestop http hook properly [NodeConformance] [Conformance]
+    /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-SSSSSSS
+SSSSSSSSSSSSSSS
 ------------------------------
-[sig-storage] EmptyDir volumes 
-  should support (non-root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [sig-storage] EmptyDir volumes
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+[sig-storage] Projected downwardAPI 
+  should provide container's cpu limit [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-storage] Projected downwardAPI
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 19:57:32.805: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-STEP: Building a namespace api object, basename emptydir
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in emptydir-7693
+Jun 18 13:11:36.588: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename projected
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-5065
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should support (non-root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: Creating a pod to test emptydir 0666 on tmpfs
-Jun  4 19:57:33.041: INFO: Waiting up to 5m0s for pod "pod-fddb33aa-8702-11e9-8d1b-467ee19922ac" in namespace "emptydir-7693" to be "success or failure"
-Jun  4 19:57:33.060: INFO: Pod "pod-fddb33aa-8702-11e9-8d1b-467ee19922ac": Phase="Pending", Reason="", readiness=false. Elapsed: 19.191758ms
-Jun  4 19:57:35.074: INFO: Pod "pod-fddb33aa-8702-11e9-8d1b-467ee19922ac": Phase="Running", Reason="", readiness=true. Elapsed: 2.032701446s
-Jun  4 19:57:37.087: INFO: Pod "pod-fddb33aa-8702-11e9-8d1b-467ee19922ac": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.046004676s
+[BeforeEach] [sig-storage] Projected downwardAPI
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:39
+[It] should provide container's cpu limit [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating a pod to test downward API volume plugin
+Jun 18 13:11:36.806: INFO: Waiting up to 5m0s for pod "downwardapi-volume-9a30a564-91ca-11e9-a25d-8608290c688a" in namespace "projected-5065" to be "success or failure"
+Jun 18 13:11:36.830: INFO: Pod "downwardapi-volume-9a30a564-91ca-11e9-a25d-8608290c688a": Phase="Pending", Reason="", readiness=false. Elapsed: 23.799681ms
+Jun 18 13:11:38.843: INFO: Pod "downwardapi-volume-9a30a564-91ca-11e9-a25d-8608290c688a": Phase="Pending", Reason="", readiness=false. Elapsed: 2.03760054s
+Jun 18 13:11:40.857: INFO: Pod "downwardapi-volume-9a30a564-91ca-11e9-a25d-8608290c688a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.050784496s
 STEP: Saw pod success
-Jun  4 19:57:37.087: INFO: Pod "pod-fddb33aa-8702-11e9-8d1b-467ee19922ac" satisfied condition "success or failure"
-Jun  4 19:57:37.099: INFO: Trying to get logs from node 10.212.23.164 pod pod-fddb33aa-8702-11e9-8d1b-467ee19922ac container test-container: 
+Jun 18 13:11:40.857: INFO: Pod "downwardapi-volume-9a30a564-91ca-11e9-a25d-8608290c688a" satisfied condition "success or failure"
+Jun 18 13:11:40.870: INFO: Trying to get logs from node 10.72.74.138 pod downwardapi-volume-9a30a564-91ca-11e9-a25d-8608290c688a container client-container: 
 STEP: delete the pod
-Jun  4 19:57:37.174: INFO: Waiting for pod pod-fddb33aa-8702-11e9-8d1b-467ee19922ac to disappear
-Jun  4 19:57:37.189: INFO: Pod pod-fddb33aa-8702-11e9-8d1b-467ee19922ac no longer exists
-[AfterEach] [sig-storage] EmptyDir volumes
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 19:57:37.189: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "emptydir-7693" for this suite.
-Jun  4 19:57:43.253: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 19:57:43.691: INFO: namespace emptydir-7693 deletion completed in 6.483303803s
+Jun 18 13:11:40.944: INFO: Waiting for pod downwardapi-volume-9a30a564-91ca-11e9-a25d-8608290c688a to disappear
+Jun 18 13:11:40.959: INFO: Pod downwardapi-volume-9a30a564-91ca-11e9-a25d-8608290c688a no longer exists
+[AfterEach] [sig-storage] Projected downwardAPI
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 13:11:40.959: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "projected-5065" for this suite.
+Jun 18 13:11:47.002: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 13:11:47.397: INFO: namespace projected-5065 deletion completed in 6.424068846s
 
-• [SLOW TEST:10.886 seconds]
-[sig-storage] EmptyDir volumes
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:41
-  should support (non-root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+• [SLOW TEST:10.808 seconds]
+[sig-storage] Projected downwardAPI
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:33
+  should provide container's cpu limit [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+SSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-api-machinery] Garbage collector 
-  should not be blocked by dependency circle [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [sig-api-machinery] Garbage collector
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+[sig-network] Networking Granular Checks: Pods 
+  should function for intra-pod communication: udp [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-network] Networking
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 19:57:43.691: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-STEP: Building a namespace api object, basename gc
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in gc-6888
+Jun 18 13:11:47.398: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename pod-network-test
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in pod-network-test-360
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should not be blocked by dependency circle [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-Jun  4 19:57:43.999: INFO: pod1.ObjectMeta.OwnerReferences=[]v1.OwnerReference{v1.OwnerReference{APIVersion:"v1", Kind:"Pod", Name:"pod3", UID:"04628532-8703-11e9-8318-1e5386706511", Controller:(*bool)(0xc002dc58d6), BlockOwnerDeletion:(*bool)(0xc002dc58d7)}}
-Jun  4 19:57:44.015: INFO: pod2.ObjectMeta.OwnerReferences=[]v1.OwnerReference{v1.OwnerReference{APIVersion:"v1", Kind:"Pod", Name:"pod1", UID:"045b0936-8703-11e9-8318-1e5386706511", Controller:(*bool)(0xc002d665f6), BlockOwnerDeletion:(*bool)(0xc002d665f7)}}
-Jun  4 19:57:44.029: INFO: pod3.ObjectMeta.OwnerReferences=[]v1.OwnerReference{v1.OwnerReference{APIVersion:"v1", Kind:"Pod", Name:"pod2", UID:"046069c1-8703-11e9-8318-1e5386706511", Controller:(*bool)(0xc002c36246), BlockOwnerDeletion:(*bool)(0xc002c36247)}}
-[AfterEach] [sig-api-machinery] Garbage collector
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 19:57:49.063: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "gc-6888" for this suite.
-Jun  4 19:57:55.176: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 19:57:55.623: INFO: namespace gc-6888 deletion completed in 6.544232848s
+[It] should function for intra-pod communication: udp [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Performing setup for networking test in namespace pod-network-test-360
+STEP: creating a selector
+STEP: Creating the service pods in kubernetes
+Jun 18 13:11:47.596: INFO: Waiting up to 10m0s for all (but 0) nodes to be schedulable
+STEP: Creating test pods
+Jun 18 13:12:12.630: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s 'http://172.30.142.32:8080/dial?request=hostName&protocol=udp&host=172.30.80.173&port=8081&tries=1'] Namespace:pod-network-test-360 PodName:host-test-container-pod ContainerName:hostexec Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
+Jun 18 13:12:12.630: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+Jun 18 13:12:12.910: INFO: Waiting for endpoints: map[]
+Jun 18 13:12:12.928: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s 'http://172.30.142.32:8080/dial?request=hostName&protocol=udp&host=172.30.142.31&port=8081&tries=1'] Namespace:pod-network-test-360 PodName:host-test-container-pod ContainerName:hostexec Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
+Jun 18 13:12:12.928: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+Jun 18 13:12:13.148: INFO: Waiting for endpoints: map[]
+Jun 18 13:12:13.162: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s 'http://172.30.142.32:8080/dial?request=hostName&protocol=udp&host=172.30.206.146&port=8081&tries=1'] Namespace:pod-network-test-360 PodName:host-test-container-pod ContainerName:hostexec Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
+Jun 18 13:12:13.162: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+Jun 18 13:12:13.417: INFO: Waiting for endpoints: map[]
+[AfterEach] [sig-network] Networking
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 13:12:13.418: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "pod-network-test-360" for this suite.
+Jun 18 13:12:39.467: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 13:12:39.878: INFO: namespace pod-network-test-360 deletion completed in 26.445354338s
 
-• [SLOW TEST:11.932 seconds]
-[sig-api-machinery] Garbage collector
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:22
-  should not be blocked by dependency circle [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+• [SLOW TEST:52.480 seconds]
+[sig-network] Networking
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:25
+  Granular Checks: Pods
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:28
+    should function for intra-pod communication: udp [LinuxOnly] [NodeConformance] [Conformance]
+    /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-SSSSSS
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-storage] Projected secret 
-  should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [sig-storage] Projected secret
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+[sig-storage] Projected configMap 
+  should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-storage] Projected configMap
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 19:57:55.624: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
+Jun 18 13:12:39.880: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
 STEP: Building a namespace api object, basename projected
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-7222
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-1642
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: Creating projection with secret that has name projected-secret-test-0b753721-8703-11e9-8d1b-467ee19922ac
-STEP: Creating a pod to test consume secrets
-Jun  4 19:57:55.870: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-0b76adfd-8703-11e9-8d1b-467ee19922ac" in namespace "projected-7222" to be "success or failure"
-Jun  4 19:57:55.882: INFO: Pod "pod-projected-secrets-0b76adfd-8703-11e9-8d1b-467ee19922ac": Phase="Pending", Reason="", readiness=false. Elapsed: 11.933163ms
-Jun  4 19:57:57.895: INFO: Pod "pod-projected-secrets-0b76adfd-8703-11e9-8d1b-467ee19922ac": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.024898419s
+[It] should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating configMap with name projected-configmap-test-volume-bfea2ca8-91ca-11e9-a25d-8608290c688a
+STEP: Creating a pod to test consume configMaps
+Jun 18 13:12:40.115: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-bfecb5cd-91ca-11e9-a25d-8608290c688a" in namespace "projected-1642" to be "success or failure"
+Jun 18 13:12:40.129: INFO: Pod "pod-projected-configmaps-bfecb5cd-91ca-11e9-a25d-8608290c688a": Phase="Pending", Reason="", readiness=false. Elapsed: 13.282791ms
+Jun 18 13:12:42.144: INFO: Pod "pod-projected-configmaps-bfecb5cd-91ca-11e9-a25d-8608290c688a": Phase="Running", Reason="", readiness=true. Elapsed: 2.028420139s
+Jun 18 13:12:44.158: INFO: Pod "pod-projected-configmaps-bfecb5cd-91ca-11e9-a25d-8608290c688a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.04238634s
 STEP: Saw pod success
-Jun  4 19:57:57.895: INFO: Pod "pod-projected-secrets-0b76adfd-8703-11e9-8d1b-467ee19922ac" satisfied condition "success or failure"
-Jun  4 19:57:57.907: INFO: Trying to get logs from node 10.212.23.189 pod pod-projected-secrets-0b76adfd-8703-11e9-8d1b-467ee19922ac container projected-secret-volume-test: 
+Jun 18 13:12:44.158: INFO: Pod "pod-projected-configmaps-bfecb5cd-91ca-11e9-a25d-8608290c688a" satisfied condition "success or failure"
+Jun 18 13:12:44.172: INFO: Trying to get logs from node 10.72.74.184 pod pod-projected-configmaps-bfecb5cd-91ca-11e9-a25d-8608290c688a container projected-configmap-volume-test: 
 STEP: delete the pod
-Jun  4 19:57:57.972: INFO: Waiting for pod pod-projected-secrets-0b76adfd-8703-11e9-8d1b-467ee19922ac to disappear
-Jun  4 19:57:57.983: INFO: Pod pod-projected-secrets-0b76adfd-8703-11e9-8d1b-467ee19922ac no longer exists
-[AfterEach] [sig-storage] Projected secret
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 19:57:57.983: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "projected-7222" for this suite.
-Jun  4 19:58:04.041: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 19:58:04.370: INFO: namespace projected-7222 deletion completed in 6.371367995s
+Jun 18 13:12:44.253: INFO: Waiting for pod pod-projected-configmaps-bfecb5cd-91ca-11e9-a25d-8608290c688a to disappear
+Jun 18 13:12:44.265: INFO: Pod pod-projected-configmaps-bfecb5cd-91ca-11e9-a25d-8608290c688a no longer exists
+[AfterEach] [sig-storage] Projected configMap
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 13:12:44.265: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "projected-1642" for this suite.
+Jun 18 13:12:50.308: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 13:12:50.723: INFO: namespace projected-1642 deletion completed in 6.442481745s
 
-• [SLOW TEST:8.746 seconds]
-[sig-storage] Projected secret
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_secret.go:33
-  should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+• [SLOW TEST:10.843 seconds]
+[sig-storage] Projected configMap
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_configmap.go:33
+  should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-SSSSSSSSSSSS
+SSSSSSSSSSS
 ------------------------------
-[sig-cli] Kubectl client [k8s.io] Kubectl logs 
-  should be able to retrieve and filter logs  [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[sig-storage] EmptyDir wrapper volumes 
+  should not cause race condition when used for configmaps [Serial] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-storage] EmptyDir wrapper volumes
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Jun 18 13:12:50.723: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename emptydir-wrapper
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in emptydir-wrapper-9941
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should not cause race condition when used for configmaps [Serial] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating 50 configmaps
+STEP: Creating RC which spawns configmap-volume pods
+Jun 18 13:12:51.702: INFO: Pod name wrapped-volume-race-c6d250fa-91ca-11e9-a25d-8608290c688a: Found 0 pods out of 5
+Jun 18 13:12:56.728: INFO: Pod name wrapped-volume-race-c6d250fa-91ca-11e9-a25d-8608290c688a: Found 5 pods out of 5
+STEP: Ensuring each pod is running
+STEP: deleting ReplicationController wrapped-volume-race-c6d250fa-91ca-11e9-a25d-8608290c688a in namespace emptydir-wrapper-9941, will wait for the garbage collector to delete the pods
+Jun 18 13:13:06.921: INFO: Deleting ReplicationController wrapped-volume-race-c6d250fa-91ca-11e9-a25d-8608290c688a took: 39.074005ms
+Jun 18 13:13:07.221: INFO: Terminating ReplicationController wrapped-volume-race-c6d250fa-91ca-11e9-a25d-8608290c688a pods took: 300.281852ms
+STEP: Creating RC which spawns configmap-volume pods
+Jun 18 13:13:44.454: INFO: Pod name wrapped-volume-race-e635c899-91ca-11e9-a25d-8608290c688a: Found 0 pods out of 5
+Jun 18 13:13:49.474: INFO: Pod name wrapped-volume-race-e635c899-91ca-11e9-a25d-8608290c688a: Found 5 pods out of 5
+STEP: Ensuring each pod is running
+STEP: deleting ReplicationController wrapped-volume-race-e635c899-91ca-11e9-a25d-8608290c688a in namespace emptydir-wrapper-9941, will wait for the garbage collector to delete the pods
+Jun 18 13:13:59.685: INFO: Deleting ReplicationController wrapped-volume-race-e635c899-91ca-11e9-a25d-8608290c688a took: 34.544597ms
+Jun 18 13:13:59.885: INFO: Terminating ReplicationController wrapped-volume-race-e635c899-91ca-11e9-a25d-8608290c688a pods took: 200.373489ms
+STEP: Creating RC which spawns configmap-volume pods
+Jun 18 13:14:44.046: INFO: Pod name wrapped-volume-race-09c5cb48-91cb-11e9-a25d-8608290c688a: Found 0 pods out of 5
+Jun 18 13:14:49.068: INFO: Pod name wrapped-volume-race-09c5cb48-91cb-11e9-a25d-8608290c688a: Found 5 pods out of 5
+STEP: Ensuring each pod is running
+STEP: deleting ReplicationController wrapped-volume-race-09c5cb48-91cb-11e9-a25d-8608290c688a in namespace emptydir-wrapper-9941, will wait for the garbage collector to delete the pods
+Jun 18 13:15:01.287: INFO: Deleting ReplicationController wrapped-volume-race-09c5cb48-91cb-11e9-a25d-8608290c688a took: 35.063787ms
+Jun 18 13:15:01.487: INFO: Terminating ReplicationController wrapped-volume-race-09c5cb48-91cb-11e9-a25d-8608290c688a pods took: 200.385644ms
+STEP: Cleaning up the configMaps
+[AfterEach] [sig-storage] EmptyDir wrapper volumes
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 13:15:46.553: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "emptydir-wrapper-9941" for this suite.
+Jun 18 13:15:58.592: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 13:15:59.002: INFO: namespace emptydir-wrapper-9941 deletion completed in 12.435376526s
+
+• [SLOW TEST:188.279 seconds]
+[sig-storage] EmptyDir wrapper volumes
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:22
+  should not cause race condition when used for configmaps [Serial] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+S
+------------------------------
+[sig-cli] Kubectl client [k8s.io] Kubectl describe 
+  should check if kubectl describe prints relevant information for rc and pods  [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 [BeforeEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 19:58:04.370: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
+Jun 18 13:15:59.003: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
 STEP: Building a namespace api object, basename kubectl
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-2776
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-887
 STEP: Waiting for a default service account to be provisioned in namespace
 [BeforeEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:213
-[BeforeEach] [k8s.io] Kubectl logs
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1190
-STEP: creating an rc
-Jun  4 19:58:04.589: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 create -f - --namespace=kubectl-2776'
-Jun  4 19:58:04.835: INFO: stderr: ""
-Jun  4 19:58:04.835: INFO: stdout: "replicationcontroller/redis-master created\n"
-[It] should be able to retrieve and filter logs  [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:213
+[It] should check if kubectl describe prints relevant information for rc and pods  [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+Jun 18 13:15:59.203: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 version --client'
+Jun 18 13:15:59.272: INFO: stderr: ""
+Jun 18 13:15:59.272: INFO: stdout: "Client Version: version.Info{Major:\"1\", Minor:\"14\", GitVersion:\"v1.14.3\", GitCommit:\"5e53fd6bc17c0dec8434817e69b04a25d8ae0ff0\", GitTreeState:\"clean\", BuildDate:\"2019-06-06T01:44:30Z\", GoVersion:\"go1.12.5\", Compiler:\"gc\", Platform:\"linux/amd64\"}\n"
+Jun 18 13:15:59.276: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 create -f - --namespace=kubectl-887'
+Jun 18 13:15:59.635: INFO: stderr: ""
+Jun 18 13:15:59.635: INFO: stdout: "replicationcontroller/redis-master created\n"
+Jun 18 13:15:59.635: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 create -f - --namespace=kubectl-887'
+Jun 18 13:15:59.998: INFO: stderr: ""
+Jun 18 13:15:59.998: INFO: stdout: "service/redis-master created\n"
 STEP: Waiting for Redis master to start.
-Jun  4 19:58:05.858: INFO: Selector matched 1 pods for map[app:redis]
-Jun  4 19:58:05.858: INFO: Found 0 / 1
-Jun  4 19:58:06.847: INFO: Selector matched 1 pods for map[app:redis]
-Jun  4 19:58:06.847: INFO: Found 1 / 1
-Jun  4 19:58:06.847: INFO: WaitFor completed with timeout 5m0s.  Pods found = 1 out of 1
-Jun  4 19:58:06.859: INFO: Selector matched 1 pods for map[app:redis]
-Jun  4 19:58:06.859: INFO: ForEach: Found 1 pods from the filter.  Now looping through them.
-STEP: checking for a matching strings
-Jun  4 19:58:06.859: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 logs redis-master-ls567 redis-master --namespace=kubectl-2776'
-Jun  4 19:58:06.997: INFO: stderr: ""
-Jun  4 19:58:06.997: INFO: stdout: "                _._                                                  \n           _.-``__ ''-._                                             \n      _.-``    `.  `_.  ''-._           Redis 3.2.12 (35a5711f/0) 64 bit\n  .-`` .-```.  ```\\/    _.,_ ''-._                                   \n (    '      ,       .-`  | `,    )     Running in standalone mode\n |`-._`-...-` __...-.``-._|'` _.-'|     Port: 6379\n |    `-._   `._    /     _.-'    |     PID: 1\n  `-._    `-._  `-./  _.-'    _.-'                                   \n |`-._`-._    `-.__.-'    _.-'_.-'|                                  \n |    `-._`-._        _.-'_.-'    |           http://redis.io        \n  `-._    `-._`-.__.-'_.-'    _.-'                                   \n |`-._`-._    `-.__.-'    _.-'_.-'|                                  \n |    `-._`-._        _.-'_.-'    |                                  \n  `-._    `-._`-.__.-'_.-'    _.-'                                   \n      `-._    `-.__.-'    _.-'                                       \n          `-._        _.-'                                           \n              `-.__.-'                                               \n\n1:M 04 Jun 19:58:06.001 # WARNING: The TCP backlog setting of 511 cannot be enforced because /proc/sys/net/core/somaxconn is set to the lower value of 128.\n1:M 04 Jun 19:58:06.001 # Server started, Redis version 3.2.12\n1:M 04 Jun 19:58:06.001 # WARNING you have Transparent Huge Pages (THP) support enabled in your kernel. This will create latency and memory usage issues with Redis. To fix this issue run the command 'echo never > /sys/kernel/mm/transparent_hugepage/enabled' as root, and add it to your /etc/rc.local in order to retain the setting after a reboot. Redis must be restarted after THP is disabled.\n1:M 04 Jun 19:58:06.001 * The server is now ready to accept connections on port 6379\n"
-STEP: limiting log lines
-Jun  4 19:58:06.997: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 log redis-master-ls567 redis-master --namespace=kubectl-2776 --tail=1'
-Jun  4 19:58:07.104: INFO: stderr: ""
-Jun  4 19:58:07.104: INFO: stdout: "1:M 04 Jun 19:58:06.001 * The server is now ready to accept connections on port 6379\n"
-STEP: limiting log bytes
-Jun  4 19:58:07.104: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 log redis-master-ls567 redis-master --namespace=kubectl-2776 --limit-bytes=1'
-Jun  4 19:58:07.225: INFO: stderr: ""
-Jun  4 19:58:07.225: INFO: stdout: " "
-STEP: exposing timestamps
-Jun  4 19:58:07.225: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 log redis-master-ls567 redis-master --namespace=kubectl-2776 --tail=1 --timestamps'
-Jun  4 19:58:07.357: INFO: stderr: ""
-Jun  4 19:58:07.357: INFO: stdout: "2019-06-04T19:58:06.002171996Z 1:M 04 Jun 19:58:06.001 * The server is now ready to accept connections on port 6379\n"
-STEP: restricting to a time range
-Jun  4 19:58:09.857: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 log redis-master-ls567 redis-master --namespace=kubectl-2776 --since=1s'
-Jun  4 19:58:10.170: INFO: stderr: ""
-Jun  4 19:58:10.170: INFO: stdout: ""
-Jun  4 19:58:10.170: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 log redis-master-ls567 redis-master --namespace=kubectl-2776 --since=24h'
-Jun  4 19:58:10.320: INFO: stderr: ""
-Jun  4 19:58:10.320: INFO: stdout: "                _._                                                  \n           _.-``__ ''-._                                             \n      _.-``    `.  `_.  ''-._           Redis 3.2.12 (35a5711f/0) 64 bit\n  .-`` .-```.  ```\\/    _.,_ ''-._                                   \n (    '      ,       .-`  | `,    )     Running in standalone mode\n |`-._`-...-` __...-.``-._|'` _.-'|     Port: 6379\n |    `-._   `._    /     _.-'    |     PID: 1\n  `-._    `-._  `-./  _.-'    _.-'                                   \n |`-._`-._    `-.__.-'    _.-'_.-'|                                  \n |    `-._`-._        _.-'_.-'    |           http://redis.io        \n  `-._    `-._`-.__.-'_.-'    _.-'                                   \n |`-._`-._    `-.__.-'    _.-'_.-'|                                  \n |    `-._`-._        _.-'_.-'    |                                  \n  `-._    `-._`-.__.-'_.-'    _.-'                                   \n      `-._    `-.__.-'    _.-'                                       \n          `-._        _.-'                                           \n              `-.__.-'                                               \n\n1:M 04 Jun 19:58:06.001 # WARNING: The TCP backlog setting of 511 cannot be enforced because /proc/sys/net/core/somaxconn is set to the lower value of 128.\n1:M 04 Jun 19:58:06.001 # Server started, Redis version 3.2.12\n1:M 04 Jun 19:58:06.001 # WARNING you have Transparent Huge Pages (THP) support enabled in your kernel. This will create latency and memory usage issues with Redis. To fix this issue run the command 'echo never > /sys/kernel/mm/transparent_hugepage/enabled' as root, and add it to your /etc/rc.local in order to retain the setting after a reboot. Redis must be restarted after THP is disabled.\n1:M 04 Jun 19:58:06.001 * The server is now ready to accept connections on port 6379\n"
-[AfterEach] [k8s.io] Kubectl logs
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1196
-STEP: using delete to clean up resources
-Jun  4 19:58:10.320: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 delete --grace-period=0 --force -f - --namespace=kubectl-2776'
-Jun  4 19:58:10.434: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n"
-Jun  4 19:58:10.434: INFO: stdout: "replicationcontroller \"redis-master\" force deleted\n"
-Jun  4 19:58:10.434: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 get rc,svc -l name=nginx --no-headers --namespace=kubectl-2776'
-Jun  4 19:58:10.603: INFO: stderr: "No resources found.\n"
-Jun  4 19:58:10.603: INFO: stdout: ""
-Jun  4 19:58:10.603: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-229234504 get pods -l name=nginx --namespace=kubectl-2776 -o go-template={{ range .items }}{{ if not .metadata.deletionTimestamp }}{{ .metadata.name }}{{ "\n" }}{{ end }}{{ end }}'
-Jun  4 19:58:10.680: INFO: stderr: ""
-Jun  4 19:58:10.680: INFO: stdout: ""
+Jun 18 13:16:01.012: INFO: Selector matched 1 pods for map[app:redis]
+Jun 18 13:16:01.012: INFO: Found 0 / 1
+Jun 18 13:16:02.012: INFO: Selector matched 1 pods for map[app:redis]
+Jun 18 13:16:02.012: INFO: Found 1 / 1
+Jun 18 13:16:02.012: INFO: WaitFor completed with timeout 5m0s.  Pods found = 1 out of 1
+Jun 18 13:16:02.025: INFO: Selector matched 1 pods for map[app:redis]
+Jun 18 13:16:02.025: INFO: ForEach: Found 1 pods from the filter.  Now looping through them.
+Jun 18 13:16:02.025: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 describe pod redis-master-8xm6r --namespace=kubectl-887'
+Jun 18 13:16:02.206: INFO: stderr: ""
+Jun 18 13:16:02.206: INFO: stdout: "Name:               redis-master-8xm6r\nNamespace:          kubectl-887\nPriority:           0\nPriorityClassName:  \nNode:               10.72.74.138/10.72.74.138\nStart Time:         Tue, 18 Jun 2019 13:15:59 +0000\nLabels:             app=redis\n                    role=master\nAnnotations:        kubernetes.io/psp: e2e-test-privileged-psp\nStatus:             Running\nIP:                 172.30.206.161\nControlled By:      ReplicationController/redis-master\nContainers:\n  redis-master:\n    Container ID:   containerd://0c9ff5c3bdbc23ae2da93775740d79db32fd217382393ca3b71c8a43a6978ad8\n    Image:          gcr.io/kubernetes-e2e-test-images/redis:1.0\n    Image ID:       gcr.io/kubernetes-e2e-test-images/redis@sha256:af4748d1655c08dc54d4be5182135395db9ce87aba2d4699b26b14ae197c5830\n    Port:           6379/TCP\n    Host Port:      0/TCP\n    State:          Running\n      Started:      Tue, 18 Jun 2019 13:16:01 +0000\n    Ready:          True\n    Restart Count:  0\n    Environment:    \n    Mounts:\n      /var/run/secrets/kubernetes.io/serviceaccount from default-token-9tffb (ro)\nConditions:\n  Type              Status\n  Initialized       True \n  Ready             True \n  ContainersReady   True \n  PodScheduled      True \nVolumes:\n  default-token-9tffb:\n    Type:        Secret (a volume populated by a Secret)\n    SecretName:  default-token-9tffb\n    Optional:    false\nQoS Class:       BestEffort\nNode-Selectors:  \nTolerations:     node.kubernetes.io/not-ready:NoExecute for 300s\n                 node.kubernetes.io/unreachable:NoExecute for 300s\nEvents:\n  Type    Reason     Age   From                   Message\n  ----    ------     ----  ----                   -------\n  Normal  Scheduled  3s    default-scheduler      Successfully assigned kubectl-887/redis-master-8xm6r to 10.72.74.138\n  Normal  Pulled     2s    kubelet, 10.72.74.138  Container image \"gcr.io/kubernetes-e2e-test-images/redis:1.0\" already present on machine\n  Normal  Created    2s    kubelet, 10.72.74.138  Created container redis-master\n  Normal  Started    1s    kubelet, 10.72.74.138  Started container redis-master\n"
+Jun 18 13:16:02.206: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 describe rc redis-master --namespace=kubectl-887'
+Jun 18 13:16:02.398: INFO: stderr: ""
+Jun 18 13:16:02.398: INFO: stdout: "Name:         redis-master\nNamespace:    kubectl-887\nSelector:     app=redis,role=master\nLabels:       app=redis\n              role=master\nAnnotations:  \nReplicas:     1 current / 1 desired\nPods Status:  1 Running / 0 Waiting / 0 Succeeded / 0 Failed\nPod Template:\n  Labels:  app=redis\n           role=master\n  Containers:\n   redis-master:\n    Image:        gcr.io/kubernetes-e2e-test-images/redis:1.0\n    Port:         6379/TCP\n    Host Port:    0/TCP\n    Environment:  \n    Mounts:       \n  Volumes:        \nEvents:\n  Type    Reason            Age   From                    Message\n  ----    ------            ----  ----                    -------\n  Normal  SuccessfulCreate  3s    replication-controller  Created pod: redis-master-8xm6r\n"
+Jun 18 13:16:02.398: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 describe service redis-master --namespace=kubectl-887'
+Jun 18 13:16:02.555: INFO: stderr: ""
+Jun 18 13:16:02.555: INFO: stdout: "Name:              redis-master\nNamespace:         kubectl-887\nLabels:            app=redis\n                   role=master\nAnnotations:       \nSelector:          app=redis,role=master\nType:              ClusterIP\nIP:                172.21.124.85\nPort:                6379/TCP\nTargetPort:        redis-server/TCP\nEndpoints:         172.30.206.161:6379\nSession Affinity:  None\nEvents:            \n"
+Jun 18 13:16:02.570: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 describe node 10.72.74.138'
+Jun 18 13:16:02.775: INFO: stderr: ""
+Jun 18 13:16:02.775: INFO: stdout: "Name:               10.72.74.138\nRoles:              \nLabels:             arch=amd64\n                    beta.kubernetes.io/arch=amd64\n                    beta.kubernetes.io/instance-type=b3c.4x16.encrypted\n                    beta.kubernetes.io/os=linux\n                    failure-domain.beta.kubernetes.io/region=eu-gb\n                    failure-domain.beta.kubernetes.io/zone=lon06\n                    ibm-cloud.kubernetes.io/encrypted-docker-data=true\n                    ibm-cloud.kubernetes.io/external-ip=158.176.111.57\n                    ibm-cloud.kubernetes.io/ha-worker=true\n                    ibm-cloud.kubernetes.io/iaas-provider=softlayer\n                    ibm-cloud.kubernetes.io/internal-ip=10.72.74.138\n                    ibm-cloud.kubernetes.io/machine-type=b3c.4x16.encrypted\n                    ibm-cloud.kubernetes.io/os=UBUNTU_18_64\n                    ibm-cloud.kubernetes.io/region=eu-gb\n                    ibm-cloud.kubernetes.io/sgx-enabled=false\n                    ibm-cloud.kubernetes.io/worker-id=kube-lon06-crd9ae0c6aaa554167a4f1d2c76b9ad9f3-w1\n                    ibm-cloud.kubernetes.io/worker-pool-id=d9ae0c6aaa554167a4f1d2c76b9ad9f3-c9f86c0\n                    ibm-cloud.kubernetes.io/worker-version=1.14.3_1523\n                    ibm-cloud.kubernetes.io/zone=lon06\n                    kubernetes.io/arch=amd64\n                    kubernetes.io/hostname=10.72.74.138\n                    kubernetes.io/os=linux\n                    privateVLAN=2643595\n                    publicVLAN=2643593\nAnnotations:        node.alpha.kubernetes.io/ttl: 0\n                    volumes.kubernetes.io/controller-managed-attach-detach: true\nCreationTimestamp:  Mon, 17 Jun 2019 21:37:25 +0000\nTaints:             \nUnschedulable:      false\nConditions:\n  Type             Status  LastHeartbeatTime                 LastTransitionTime                Reason                       Message\n  ----             ------  -----------------                 ------------------                ------                       -------\n  MemoryPressure   False   Tue, 18 Jun 2019 13:15:32 +0000   Mon, 17 Jun 2019 21:37:25 +0000   KubeletHasSufficientMemory   kubelet has sufficient memory available\n  DiskPressure     False   Tue, 18 Jun 2019 13:15:32 +0000   Mon, 17 Jun 2019 21:37:25 +0000   KubeletHasNoDiskPressure     kubelet has no disk pressure\n  PIDPressure      False   Tue, 18 Jun 2019 13:15:32 +0000   Mon, 17 Jun 2019 21:37:25 +0000   KubeletHasSufficientPID      kubelet has sufficient PID available\n  Ready            True    Tue, 18 Jun 2019 13:15:32 +0000   Mon, 17 Jun 2019 21:37:35 +0000   KubeletReady                 kubelet is posting ready status. AppArmor enabled\nAddresses:\n  InternalIP:  10.72.74.138\n  ExternalIP:  158.176.111.57\n  Hostname:    10.72.74.138\nCapacity:\n cpu:                4\n ephemeral-storage:  102685624Ki\n hugepages-1Gi:      0\n hugepages-2Mi:      0\n memory:             16419920Ki\n pods:               110\nAllocatable:\n cpu:                3910m\n ephemeral-storage:  99892574949\n hugepages-1Gi:      0\n hugepages-2Mi:      0\n memory:             13627472Ki\n pods:               110\nSystem Info:\n Machine ID:                 cdba7d4da1f04339b876140126a7b7c2\n System UUID:                EE26817E-4741-27D4-473E-A59DAFCFACA9\n Boot ID:                    9ef61385-a585-42b5-aaa1-8c9c44b3d4fb\n Kernel Version:             4.15.0-51-generic\n OS Image:                   Ubuntu 18.04.2 LTS\n Operating System:           linux\n Architecture:               amd64\n Container Runtime Version:  containerd://1.2.6\n Kubelet Version:            v1.14.3+IKS\n Kube-Proxy Version:         v1.14.3+IKS\nProviderID:                  ibm://d18c889395112a40d2f4e3065f237a7d///d9ae0c6aaa554167a4f1d2c76b9ad9f3/kube-lon06-crd9ae0c6aaa554167a4f1d2c76b9ad9f3-w1\nNon-terminated Pods:         (8 in total)\n  Namespace                  Name                                                             CPU Requests  CPU Limits  Memory Requests  Memory Limits  AGE\n  ---------                  ----                                                             ------------  ----------  ---------------  -------------  ---\n  heptio-sonobuoy            sonobuoy-systemd-logs-daemon-set-376e58a3dd534c11-ccq8b          0 (0%)        0 (0%)      0 (0%)           0 (0%)         102m\n  ibm-system                 ibm-cloud-provider-ip-158-176-120-140-d8f5f45f5-4tksx            5m (0%)       0 (0%)      10Mi (0%)        0 (0%)         15h\n  kube-system                calico-node-prxhz                                                250m (6%)     0 (0%)      80Mi (0%)        0 (0%)         15h\n  kube-system                ibm-keepalived-watcher-4l788                                     5m (0%)       0 (0%)      10Mi (0%)        0 (0%)         15h\n  kube-system                ibm-kube-fluentd-gmp54                                           25m (0%)      300m (7%)   150Mi (1%)       800M (5%)      15h\n  kube-system                ibm-master-proxy-static-10.72.74.138                             25m (0%)      300m (7%)   32M (0%)         512M (3%)      15h\n  kube-system                public-crd9ae0c6aaa554167a4f1d2c76b9ad9f3-alb1-fd9fb76d-q646z    0 (0%)        0 (0%)      0 (0%)           0 (0%)         15h\n  kubectl-887                redis-master-8xm6r                                               0 (0%)        0 (0%)      0 (0%)           0 (0%)         3s\nAllocated resources:\n  (Total limits may be over 100 percent, i.e., overcommitted.)\n  Resource           Requests       Limits\n  --------           --------       ------\n  cpu                310m (7%)      600m (15%)\n  memory             287250Ki (2%)  1312M (9%)\n  ephemeral-storage  0 (0%)         0 (0%)\nEvents:              \n"
+Jun 18 13:16:02.775: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-426929150 describe namespace kubectl-887'
+Jun 18 13:16:02.955: INFO: stderr: ""
+Jun 18 13:16:02.955: INFO: stdout: "Name:         kubectl-887\nLabels:       e2e-framework=kubectl\n              e2e-run=0bbd93b1-91bd-11e9-a25d-8608290c688a\nAnnotations:  \nStatus:       Active\n\nNo resource quota.\n\nNo resource limits.\n"
 [AfterEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 19:58:10.680: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "kubectl-2776" for this suite.
-Jun  4 19:58:16.739: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 19:58:17.099: INFO: namespace kubectl-2776 deletion completed in 6.403020461s
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 13:16:02.955: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "kubectl-887" for this suite.
+Jun 18 13:16:26.998: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 13:16:27.399: INFO: namespace kubectl-887 deletion completed in 24.429240941s
 
-• [SLOW TEST:12.728 seconds]
+• [SLOW TEST:28.397 seconds]
 [sig-cli] Kubectl client
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23
-  [k8s.io] Kubectl logs
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
-    should be able to retrieve and filter logs  [Conformance]
-    /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23
+  [k8s.io] Kubectl describe
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
+    should check if kubectl describe prints relevant information for rc and pods  [Conformance]
+    /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-SS
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-api-machinery] Garbage collector 
-  should delete pods created by rc when not orphaning [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [sig-api-machinery] Garbage collector
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+[sig-storage] Projected downwardAPI 
+  should update annotations on modification [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-storage] Projected downwardAPI
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 19:58:17.099: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-STEP: Building a namespace api object, basename gc
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in gc-3866
+Jun 18 13:16:27.402: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename projected
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-2268
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should delete pods created by rc when not orphaning [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: create the rc
-STEP: delete the rc
-STEP: wait for all pods to be garbage collected
-STEP: Gathering metrics
-W0604 19:58:27.432608      18 metrics_grabber.go:79] Master node is not registered. Grabbing metrics from Scheduler, ControllerManager and ClusterAutoscaler is disabled.
-Jun  4 19:58:27.432: INFO: For apiserver_request_total:
-For apiserver_request_latencies_summary:
-For apiserver_init_events_total:
-For garbage_collector_attempt_to_delete_queue_latency:
-For garbage_collector_attempt_to_delete_work_duration:
-For garbage_collector_attempt_to_orphan_queue_latency:
-For garbage_collector_attempt_to_orphan_work_duration:
-For garbage_collector_dirty_processing_latency_microseconds:
-For garbage_collector_event_processing_latency_microseconds:
-For garbage_collector_graph_changes_queue_latency:
-For garbage_collector_graph_changes_work_duration:
-For garbage_collector_orphan_processing_latency_microseconds:
-For namespace_queue_latency:
-For namespace_queue_latency_sum:
-For namespace_queue_latency_count:
-For namespace_retries:
-For namespace_work_duration:
-For namespace_work_duration_sum:
-For namespace_work_duration_count:
-For function_duration_seconds:
-For errors_total:
-For evicted_pods_total:
-
-[AfterEach] [sig-api-machinery] Garbage collector
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 19:58:27.432: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "gc-3866" for this suite.
-Jun  4 19:58:33.486: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 19:58:33.974: INFO: namespace gc-3866 deletion completed in 6.530039034s
+[BeforeEach] [sig-storage] Projected downwardAPI
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:39
+[It] should update annotations on modification [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating the pod
+Jun 18 13:16:30.264: INFO: Successfully updated pod "annotationupdate47899944-91cb-11e9-a25d-8608290c688a"
+[AfterEach] [sig-storage] Projected downwardAPI
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 13:16:34.350: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "projected-2268" for this suite.
+Jun 18 13:16:58.392: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 13:17:00.346: INFO: namespace projected-2268 deletion completed in 25.980700834s
 
-• [SLOW TEST:16.875 seconds]
-[sig-api-machinery] Garbage collector
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:22
-  should delete pods created by rc when not orphaning [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+• [SLOW TEST:32.943 seconds]
+[sig-storage] Projected downwardAPI
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:33
+  should update annotations on modification [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-SSSSSSSSSSSSSSSSSSSSS
+SSSSSSS
 ------------------------------
-[sig-apps] Daemon set [Serial] 
-  should run and stop simple daemon [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [sig-apps] Daemon set [Serial]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+[sig-scheduling] SchedulerPredicates [Serial] 
+  validates that NodeSelector is respected if matching  [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-scheduling] SchedulerPredicates [Serial]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 19:58:33.974: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-STEP: Building a namespace api object, basename daemonsets
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in daemonsets-6913
+Jun 18 13:17:00.346: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename sched-pred
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in sched-pred-6229
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-apps] Daemon set [Serial]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:102
-[It] should run and stop simple daemon [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: Creating simple DaemonSet "daemon-set"
-STEP: Check that daemon pods launch on every node of the cluster.
-Jun  4 19:58:34.311: INFO: Number of nodes with available pods: 0
-Jun  4 19:58:34.311: INFO: Node 10.212.23.161 is running more than one daemon pod
-Jun  4 19:58:35.337: INFO: Number of nodes with available pods: 0
-Jun  4 19:58:35.338: INFO: Node 10.212.23.161 is running more than one daemon pod
-Jun  4 19:58:36.339: INFO: Number of nodes with available pods: 2
-Jun  4 19:58:36.339: INFO: Node 10.212.23.164 is running more than one daemon pod
-Jun  4 19:58:37.368: INFO: Number of nodes with available pods: 3
-Jun  4 19:58:37.368: INFO: Number of running nodes: 3, number of available pods: 3
-STEP: Stop a daemon pod, check that the daemon pod is revived.
-Jun  4 19:58:37.442: INFO: Number of nodes with available pods: 2
-Jun  4 19:58:37.442: INFO: Node 10.212.23.189 is running more than one daemon pod
-Jun  4 19:58:38.766: INFO: Number of nodes with available pods: 2
-Jun  4 19:58:38.766: INFO: Node 10.212.23.189 is running more than one daemon pod
-Jun  4 19:58:39.555: INFO: Number of nodes with available pods: 2
-Jun  4 19:58:39.555: INFO: Node 10.212.23.189 is running more than one daemon pod
-Jun  4 19:58:40.471: INFO: Number of nodes with available pods: 2
-Jun  4 19:58:40.471: INFO: Node 10.212.23.189 is running more than one daemon pod
-Jun  4 19:58:41.555: INFO: Number of nodes with available pods: 2
-Jun  4 19:58:41.556: INFO: Node 10.212.23.189 is running more than one daemon pod
-Jun  4 19:58:42.471: INFO: Number of nodes with available pods: 2
-Jun  4 19:58:42.471: INFO: Node 10.212.23.189 is running more than one daemon pod
-Jun  4 19:58:43.472: INFO: Number of nodes with available pods: 2
-Jun  4 19:58:43.472: INFO: Node 10.212.23.189 is running more than one daemon pod
-Jun  4 19:58:44.474: INFO: Number of nodes with available pods: 2
-Jun  4 19:58:44.474: INFO: Node 10.212.23.189 is running more than one daemon pod
-Jun  4 19:58:45.472: INFO: Number of nodes with available pods: 3
-Jun  4 19:58:45.472: INFO: Number of running nodes: 3, number of available pods: 3
-[AfterEach] [sig-apps] Daemon set [Serial]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:68
-STEP: Deleting DaemonSet "daemon-set"
-STEP: deleting DaemonSet.extensions daemon-set in namespace daemonsets-6913, will wait for the garbage collector to delete the pods
-Jun  4 19:58:45.658: INFO: Deleting DaemonSet.extensions daemon-set took: 26.603228ms
-Jun  4 19:58:45.759: INFO: Terminating DaemonSet.extensions daemon-set pods took: 100.173598ms
-Jun  4 19:58:58.073: INFO: Number of nodes with available pods: 0
-Jun  4 19:58:58.073: INFO: Number of running nodes: 0, number of available pods: 0
-Jun  4 19:58:58.085: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"selfLink":"/apis/apps/v1/namespaces/daemonsets-6913/daemonsets","resourceVersion":"30253"},"items":null}
-
-Jun  4 19:58:58.095: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"selfLink":"/api/v1/namespaces/daemonsets-6913/pods","resourceVersion":"30253"},"items":null}
-
-[AfterEach] [sig-apps] Daemon set [Serial]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 19:58:58.146: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "daemonsets-6913" for this suite.
-Jun  4 19:59:06.206: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 19:59:06.571: INFO: namespace daemonsets-6913 deletion completed in 8.41452819s
+[BeforeEach] [sig-scheduling] SchedulerPredicates [Serial]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:79
+Jun 18 13:17:00.536: INFO: Waiting up to 1m0s for all (but 0) nodes to be ready
+Jun 18 13:17:00.563: INFO: Waiting for terminating namespaces to be deleted...
+Jun 18 13:17:00.571: INFO: 
+Logging pods the kubelet thinks is on node 10.72.74.138 before test
+Jun 18 13:17:00.825: INFO: ibm-cloud-provider-ip-158-176-120-140-d8f5f45f5-4tksx from ibm-system started at 2019-06-17 21:41:02 +0000 UTC (1 container statuses recorded)
+Jun 18 13:17:00.826: INFO: 	Container ibm-cloud-provider-ip-158-176-120-140 ready: true, restart count 0
+Jun 18 13:17:00.826: INFO: sonobuoy-systemd-logs-daemon-set-376e58a3dd534c11-ccq8b from heptio-sonobuoy started at 2019-06-18 11:33:56 +0000 UTC (2 container statuses recorded)
+Jun 18 13:17:00.826: INFO: 	Container sonobuoy-worker ready: true, restart count 1
+Jun 18 13:17:00.826: INFO: 	Container systemd-logs ready: true, restart count 1
+Jun 18 13:17:00.826: INFO: calico-node-prxhz from kube-system started at 2019-06-17 21:37:25 +0000 UTC (1 container statuses recorded)
+Jun 18 13:17:00.826: INFO: 	Container calico-node ready: true, restart count 0
+Jun 18 13:17:00.826: INFO: ibm-kube-fluentd-gmp54 from kube-system started at 2019-06-17 21:37:34 +0000 UTC (1 container statuses recorded)
+Jun 18 13:17:00.826: INFO: 	Container fluentd ready: true, restart count 0
+Jun 18 13:17:00.826: INFO: public-crd9ae0c6aaa554167a4f1d2c76b9ad9f3-alb1-fd9fb76d-q646z from kube-system started at 2019-06-17 21:41:07 +0000 UTC (4 container statuses recorded)
+Jun 18 13:17:00.826: INFO: 	Container ingress-auth-1 ready: true, restart count 0
+Jun 18 13:17:00.827: INFO: 	Container ingress-auth-2 ready: true, restart count 0
+Jun 18 13:17:00.827: INFO: 	Container ingress-auth-3 ready: true, restart count 0
+Jun 18 13:17:00.827: INFO: 	Container nginx-ingress ready: true, restart count 0
+Jun 18 13:17:00.827: INFO: ibm-master-proxy-static-10.72.74.138 from kube-system started at  (0 container statuses recorded)
+Jun 18 13:17:00.827: INFO: ibm-keepalived-watcher-4l788 from kube-system started at 2019-06-17 21:37:25 +0000 UTC (1 container statuses recorded)
+Jun 18 13:17:00.827: INFO: 	Container keepalived-watcher ready: true, restart count 0
+Jun 18 13:17:00.827: INFO: 
+Logging pods the kubelet thinks is on node 10.72.74.184 before test
+Jun 18 13:17:00.875: INFO: sonobuoy-systemd-logs-daemon-set-376e58a3dd534c11-x9mmr from heptio-sonobuoy started at 2019-06-18 11:33:56 +0000 UTC (2 container statuses recorded)
+Jun 18 13:17:00.875: INFO: 	Container sonobuoy-worker ready: true, restart count 1
+Jun 18 13:17:00.875: INFO: 	Container systemd-logs ready: true, restart count 1
+Jun 18 13:17:00.875: INFO: ibm-master-proxy-static-10.72.74.184 from kube-system started at  (0 container statuses recorded)
+Jun 18 13:17:00.875: INFO: test-k8s-e2e-pvg-master-verification from default started at 2019-06-18 11:33:41 +0000 UTC (1 container statuses recorded)
+Jun 18 13:17:00.875: INFO: 	Container test-k8s-e2e-pvg-master-verification ready: true, restart count 0
+Jun 18 13:17:00.875: INFO: ibm-cloud-provider-ip-158-176-120-140-d8f5f45f5-kszmp from ibm-system started at 2019-06-17 21:41:02 +0000 UTC (1 container statuses recorded)
+Jun 18 13:17:00.875: INFO: 	Container ibm-cloud-provider-ip-158-176-120-140 ready: true, restart count 0
+Jun 18 13:17:00.875: INFO: sonobuoy from heptio-sonobuoy started at 2019-06-18 11:33:48 +0000 UTC (1 container statuses recorded)
+Jun 18 13:17:00.875: INFO: 	Container kube-sonobuoy ready: true, restart count 0
+Jun 18 13:17:00.875: INFO: calico-node-2jll6 from kube-system started at 2019-06-17 21:37:03 +0000 UTC (1 container statuses recorded)
+Jun 18 13:17:00.875: INFO: 	Container calico-node ready: true, restart count 0
+Jun 18 13:17:00.875: INFO: ibm-kube-fluentd-66f2t from kube-system started at 2019-06-17 21:37:34 +0000 UTC (1 container statuses recorded)
+Jun 18 13:17:00.875: INFO: 	Container fluentd ready: true, restart count 0
+Jun 18 13:17:00.875: INFO: coredns-78cff85d65-jvwmh from kube-system started at 2019-06-17 21:37:29 +0000 UTC (1 container statuses recorded)
+Jun 18 13:17:00.875: INFO: 	Container coredns ready: true, restart count 0
+Jun 18 13:17:00.875: INFO: metrics-server-5cdb4c5b-67974 from kube-system started at 2019-06-17 21:37:27 +0000 UTC (2 container statuses recorded)
+Jun 18 13:17:00.875: INFO: 	Container metrics-server ready: true, restart count 0
+Jun 18 13:17:00.875: INFO: 	Container metrics-server-nanny ready: true, restart count 0
+Jun 18 13:17:00.875: INFO: ibm-keepalived-watcher-xkcvg from kube-system started at 2019-06-17 21:37:03 +0000 UTC (1 container statuses recorded)
+Jun 18 13:17:00.875: INFO: 	Container keepalived-watcher ready: true, restart count 0
+Jun 18 13:17:00.875: INFO: 
+Logging pods the kubelet thinks is on node 10.72.74.189 before test
+Jun 18 13:17:00.921: INFO: kubernetes-dashboard-6f5f8c6896-rbvmt from kube-system started at 2019-06-17 21:37:09 +0000 UTC (1 container statuses recorded)
+Jun 18 13:17:00.921: INFO: 	Container kubernetes-dashboard ready: true, restart count 0
+Jun 18 13:17:00.921: INFO: sonobuoy-e2e-job-68f893029cbd431f from heptio-sonobuoy started at 2019-06-18 11:33:56 +0000 UTC (2 container statuses recorded)
+Jun 18 13:17:00.921: INFO: 	Container e2e ready: true, restart count 0
+Jun 18 13:17:00.921: INFO: 	Container sonobuoy-worker ready: true, restart count 0
+Jun 18 13:17:00.921: INFO: ibm-master-proxy-static-10.72.74.189 from kube-system started at  (0 container statuses recorded)
+Jun 18 13:17:00.921: INFO: ibm-keepalived-watcher-klm4c from kube-system started at 2019-06-17 21:36:59 +0000 UTC (1 container statuses recorded)
+Jun 18 13:17:00.921: INFO: 	Container keepalived-watcher ready: true, restart count 0
+Jun 18 13:17:00.921: INFO: vpn-cf6ff59b-gqxpw from kube-system started at 2019-06-17 21:37:09 +0000 UTC (1 container statuses recorded)
+Jun 18 13:17:00.921: INFO: 	Container vpn ready: true, restart count 0
+Jun 18 13:17:00.921: INFO: calico-kube-controllers-85d6c48f7-ggvwd from kube-system started at 2019-06-17 21:37:09 +0000 UTC (1 container statuses recorded)
+Jun 18 13:17:00.921: INFO: 	Container calico-kube-controllers ready: true, restart count 0
+Jun 18 13:17:00.921: INFO: public-crd9ae0c6aaa554167a4f1d2c76b9ad9f3-alb1-fd9fb76d-sf25n from kube-system started at 2019-06-17 21:41:07 +0000 UTC (4 container statuses recorded)
+Jun 18 13:17:00.921: INFO: 	Container ingress-auth-1 ready: true, restart count 0
+Jun 18 13:17:00.921: INFO: 	Container ingress-auth-2 ready: true, restart count 0
+Jun 18 13:17:00.921: INFO: 	Container ingress-auth-3 ready: true, restart count 0
+Jun 18 13:17:00.921: INFO: 	Container nginx-ingress ready: true, restart count 0
+Jun 18 13:17:00.921: INFO: ibm-storage-watcher-964758dd-rsgxw from kube-system started at 2019-06-17 21:37:09 +0000 UTC (1 container statuses recorded)
+Jun 18 13:17:00.921: INFO: 	Container ibm-storage-watcher-container ready: true, restart count 0
+Jun 18 13:17:00.921: INFO: coredns-78cff85d65-2mm72 from kube-system started at 2019-06-17 21:37:09 +0000 UTC (1 container statuses recorded)
+Jun 18 13:17:00.921: INFO: 	Container coredns ready: true, restart count 0
+Jun 18 13:17:00.921: INFO: coredns-autoscaler-6854575d6-p69qn from kube-system started at 2019-06-17 21:37:09 +0000 UTC (1 container statuses recorded)
+Jun 18 13:17:00.921: INFO: 	Container autoscaler ready: true, restart count 0
+Jun 18 13:17:00.921: INFO: sonobuoy-systemd-logs-daemon-set-376e58a3dd534c11-rn8dw from heptio-sonobuoy started at 2019-06-18 11:33:56 +0000 UTC (2 container statuses recorded)
+Jun 18 13:17:00.921: INFO: 	Container sonobuoy-worker ready: true, restart count 1
+Jun 18 13:17:00.921: INFO: 	Container systemd-logs ready: true, restart count 1
+Jun 18 13:17:00.921: INFO: calico-node-htlqc from kube-system started at 2019-06-17 21:36:59 +0000 UTC (1 container statuses recorded)
+Jun 18 13:17:00.921: INFO: 	Container calico-node ready: true, restart count 0
+Jun 18 13:17:00.921: INFO: ibm-file-plugin-98f5986d7-bbzfx from kube-system started at 2019-06-17 21:37:09 +0000 UTC (1 container statuses recorded)
+Jun 18 13:17:00.921: INFO: 	Container ibm-file-plugin-container ready: true, restart count 0
+Jun 18 13:17:00.921: INFO: ibm-kube-fluentd-qg9mz from kube-system started at 2019-06-17 21:37:34 +0000 UTC (1 container statuses recorded)
+Jun 18 13:17:00.921: INFO: 	Container fluentd ready: true, restart count 0
+[It] validates that NodeSelector is respected if matching  [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Trying to launch a pod without a label to get a node which can launch it.
+STEP: Explicitly delete pod here to free the resource it takes.
+STEP: Trying to apply a random label on the found node.
+STEP: verifying the node has the label kubernetes.io/e2e-5ddc025b-91cb-11e9-a25d-8608290c688a 42
+STEP: Trying to relaunch the pod, now with labels.
+STEP: removing the label kubernetes.io/e2e-5ddc025b-91cb-11e9-a25d-8608290c688a off the node 10.72.74.138
+STEP: verifying the node doesn't have the label kubernetes.io/e2e-5ddc025b-91cb-11e9-a25d-8608290c688a
+[AfterEach] [sig-scheduling] SchedulerPredicates [Serial]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 13:17:09.198: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "sched-pred-6229" for this suite.
+Jun 18 13:17:29.238: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 13:17:29.640: INFO: namespace sched-pred-6229 deletion completed in 20.429375176s
+[AfterEach] [sig-scheduling] SchedulerPredicates [Serial]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:70
 
-• [SLOW TEST:32.597 seconds]
-[sig-apps] Daemon set [Serial]
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22
-  should run and stop simple daemon [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+• [SLOW TEST:29.295 seconds]
+[sig-scheduling] SchedulerPredicates [Serial]
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/framework.go:22
+  validates that NodeSelector is respected if matching  [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
 SSSSSSSS
 ------------------------------
 [sig-network] DNS 
   should provide DNS for services  [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 [BeforeEach] [sig-network] DNS
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 19:59:06.572: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
+Jun 18 13:17:29.641: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
 STEP: Building a namespace api object, basename dns
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in dns-4783
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in dns-201
 STEP: Waiting for a default service account to be provisioned in namespace
 [It] should provide DNS for services  [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 STEP: Creating a test headless service
-STEP: Running these commands on wheezy: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search dns-test-service.dns-4783.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_udp@dns-test-service.dns-4783.svc.cluster.local;check="$$(dig +tcp +noall +answer +search dns-test-service.dns-4783.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@dns-test-service.dns-4783.svc.cluster.local;check="$$(dig +notcp +noall +answer +search _http._tcp.dns-test-service.dns-4783.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/wheezy_udp@_http._tcp.dns-test-service.dns-4783.svc.cluster.local;check="$$(dig +tcp +noall +answer +search _http._tcp.dns-test-service.dns-4783.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/wheezy_tcp@_http._tcp.dns-test-service.dns-4783.svc.cluster.local;check="$$(dig +notcp +noall +answer +search _http._tcp.test-service-2.dns-4783.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/wheezy_udp@_http._tcp.test-service-2.dns-4783.svc.cluster.local;check="$$(dig +tcp +noall +answer +search _http._tcp.test-service-2.dns-4783.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/wheezy_tcp@_http._tcp.test-service-2.dns-4783.svc.cluster.local;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-4783.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@PodARecord;check="$$(dig +notcp +noall +answer +search 30.27.21.172.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/172.21.27.30_udp@PTR;check="$$(dig +tcp +noall +answer +search 30.27.21.172.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/172.21.27.30_tcp@PTR;sleep 1; done
+STEP: Running these commands on wheezy: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search dns-test-service.dns-201.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_udp@dns-test-service.dns-201.svc.cluster.local;check="$$(dig +tcp +noall +answer +search dns-test-service.dns-201.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@dns-test-service.dns-201.svc.cluster.local;check="$$(dig +notcp +noall +answer +search _http._tcp.dns-test-service.dns-201.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/wheezy_udp@_http._tcp.dns-test-service.dns-201.svc.cluster.local;check="$$(dig +tcp +noall +answer +search _http._tcp.dns-test-service.dns-201.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/wheezy_tcp@_http._tcp.dns-test-service.dns-201.svc.cluster.local;check="$$(dig +notcp +noall +answer +search _http._tcp.test-service-2.dns-201.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/wheezy_udp@_http._tcp.test-service-2.dns-201.svc.cluster.local;check="$$(dig +tcp +noall +answer +search _http._tcp.test-service-2.dns-201.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/wheezy_tcp@_http._tcp.test-service-2.dns-201.svc.cluster.local;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-201.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@PodARecord;check="$$(dig +notcp +noall +answer +search 54.78.21.172.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/172.21.78.54_udp@PTR;check="$$(dig +tcp +noall +answer +search 54.78.21.172.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/172.21.78.54_tcp@PTR;sleep 1; done
 
-STEP: Running these commands on jessie: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search dns-test-service.dns-4783.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_udp@dns-test-service.dns-4783.svc.cluster.local;check="$$(dig +tcp +noall +answer +search dns-test-service.dns-4783.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_tcp@dns-test-service.dns-4783.svc.cluster.local;check="$$(dig +notcp +noall +answer +search _http._tcp.dns-test-service.dns-4783.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/jessie_udp@_http._tcp.dns-test-service.dns-4783.svc.cluster.local;check="$$(dig +tcp +noall +answer +search _http._tcp.dns-test-service.dns-4783.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/jessie_tcp@_http._tcp.dns-test-service.dns-4783.svc.cluster.local;check="$$(dig +notcp +noall +answer +search _http._tcp.test-service-2.dns-4783.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/jessie_udp@_http._tcp.test-service-2.dns-4783.svc.cluster.local;check="$$(dig +tcp +noall +answer +search _http._tcp.test-service-2.dns-4783.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/jessie_tcp@_http._tcp.test-service-2.dns-4783.svc.cluster.local;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-4783.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_tcp@PodARecord;check="$$(dig +notcp +noall +answer +search 30.27.21.172.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/172.21.27.30_udp@PTR;check="$$(dig +tcp +noall +answer +search 30.27.21.172.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/172.21.27.30_tcp@PTR;sleep 1; done
+STEP: Running these commands on jessie: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search dns-test-service.dns-201.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_udp@dns-test-service.dns-201.svc.cluster.local;check="$$(dig +tcp +noall +answer +search dns-test-service.dns-201.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_tcp@dns-test-service.dns-201.svc.cluster.local;check="$$(dig +notcp +noall +answer +search _http._tcp.dns-test-service.dns-201.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/jessie_udp@_http._tcp.dns-test-service.dns-201.svc.cluster.local;check="$$(dig +tcp +noall +answer +search _http._tcp.dns-test-service.dns-201.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/jessie_tcp@_http._tcp.dns-test-service.dns-201.svc.cluster.local;check="$$(dig +notcp +noall +answer +search _http._tcp.test-service-2.dns-201.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/jessie_udp@_http._tcp.test-service-2.dns-201.svc.cluster.local;check="$$(dig +tcp +noall +answer +search _http._tcp.test-service-2.dns-201.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/jessie_tcp@_http._tcp.test-service-2.dns-201.svc.cluster.local;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-201.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_tcp@PodARecord;check="$$(dig +notcp +noall +answer +search 54.78.21.172.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/172.21.78.54_udp@PTR;check="$$(dig +tcp +noall +answer +search 54.78.21.172.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/172.21.78.54_tcp@PTR;sleep 1; done
 
 STEP: creating a pod to probe DNS
 STEP: submitting the pod to kubernetes
 STEP: retrieving the pod
 STEP: looking for the results for each expected name from probers
-Jun  4 19:59:18.918: INFO: Unable to read wheezy_udp@dns-test-service.dns-4783.svc.cluster.local from pod dns-4783/dns-test-35c643f3-8703-11e9-8d1b-467ee19922ac: the server could not find the requested resource (get pods dns-test-35c643f3-8703-11e9-8d1b-467ee19922ac)
-Jun  4 19:59:18.934: INFO: Unable to read wheezy_tcp@dns-test-service.dns-4783.svc.cluster.local from pod dns-4783/dns-test-35c643f3-8703-11e9-8d1b-467ee19922ac: the server could not find the requested resource (get pods dns-test-35c643f3-8703-11e9-8d1b-467ee19922ac)
-Jun  4 19:59:18.951: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.dns-4783.svc.cluster.local from pod dns-4783/dns-test-35c643f3-8703-11e9-8d1b-467ee19922ac: the server could not find the requested resource (get pods dns-test-35c643f3-8703-11e9-8d1b-467ee19922ac)
-Jun  4 19:59:18.967: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-4783.svc.cluster.local from pod dns-4783/dns-test-35c643f3-8703-11e9-8d1b-467ee19922ac: the server could not find the requested resource (get pods dns-test-35c643f3-8703-11e9-8d1b-467ee19922ac)
-Jun  4 19:59:19.079: INFO: Unable to read jessie_udp@dns-test-service.dns-4783.svc.cluster.local from pod dns-4783/dns-test-35c643f3-8703-11e9-8d1b-467ee19922ac: the server could not find the requested resource (get pods dns-test-35c643f3-8703-11e9-8d1b-467ee19922ac)
-Jun  4 19:59:19.095: INFO: Unable to read jessie_tcp@dns-test-service.dns-4783.svc.cluster.local from pod dns-4783/dns-test-35c643f3-8703-11e9-8d1b-467ee19922ac: the server could not find the requested resource (get pods dns-test-35c643f3-8703-11e9-8d1b-467ee19922ac)
-Jun  4 19:59:19.110: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.dns-4783.svc.cluster.local from pod dns-4783/dns-test-35c643f3-8703-11e9-8d1b-467ee19922ac: the server could not find the requested resource (get pods dns-test-35c643f3-8703-11e9-8d1b-467ee19922ac)
-Jun  4 19:59:19.126: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.dns-4783.svc.cluster.local from pod dns-4783/dns-test-35c643f3-8703-11e9-8d1b-467ee19922ac: the server could not find the requested resource (get pods dns-test-35c643f3-8703-11e9-8d1b-467ee19922ac)
-Jun  4 19:59:19.307: INFO: Lookups using dns-4783/dns-test-35c643f3-8703-11e9-8d1b-467ee19922ac failed for: [wheezy_udp@dns-test-service.dns-4783.svc.cluster.local wheezy_tcp@dns-test-service.dns-4783.svc.cluster.local wheezy_udp@_http._tcp.dns-test-service.dns-4783.svc.cluster.local wheezy_tcp@_http._tcp.dns-test-service.dns-4783.svc.cluster.local jessie_udp@dns-test-service.dns-4783.svc.cluster.local jessie_tcp@dns-test-service.dns-4783.svc.cluster.local jessie_udp@_http._tcp.dns-test-service.dns-4783.svc.cluster.local jessie_tcp@_http._tcp.dns-test-service.dns-4783.svc.cluster.local]
-
-Jun  4 19:59:24.623: INFO: DNS probes using dns-4783/dns-test-35c643f3-8703-11e9-8d1b-467ee19922ac succeeded
+Jun 18 13:17:34.215: INFO: Unable to read wheezy_udp@dns-test-service.dns-201.svc.cluster.local from pod dns-201/dns-test-6ca54d7b-91cb-11e9-a25d-8608290c688a: the server could not find the requested resource (get pods dns-test-6ca54d7b-91cb-11e9-a25d-8608290c688a)
+Jun 18 13:17:34.236: INFO: Unable to read wheezy_tcp@dns-test-service.dns-201.svc.cluster.local from pod dns-201/dns-test-6ca54d7b-91cb-11e9-a25d-8608290c688a: the server could not find the requested resource (get pods dns-test-6ca54d7b-91cb-11e9-a25d-8608290c688a)
+Jun 18 13:17:34.258: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.dns-201.svc.cluster.local from pod dns-201/dns-test-6ca54d7b-91cb-11e9-a25d-8608290c688a: the server could not find the requested resource (get pods dns-test-6ca54d7b-91cb-11e9-a25d-8608290c688a)
+Jun 18 13:17:34.280: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-201.svc.cluster.local from pod dns-201/dns-test-6ca54d7b-91cb-11e9-a25d-8608290c688a: the server could not find the requested resource (get pods dns-test-6ca54d7b-91cb-11e9-a25d-8608290c688a)
+Jun 18 13:17:34.426: INFO: Unable to read jessie_udp@dns-test-service.dns-201.svc.cluster.local from pod dns-201/dns-test-6ca54d7b-91cb-11e9-a25d-8608290c688a: the server could not find the requested resource (get pods dns-test-6ca54d7b-91cb-11e9-a25d-8608290c688a)
+Jun 18 13:17:34.446: INFO: Unable to read jessie_tcp@dns-test-service.dns-201.svc.cluster.local from pod dns-201/dns-test-6ca54d7b-91cb-11e9-a25d-8608290c688a: the server could not find the requested resource (get pods dns-test-6ca54d7b-91cb-11e9-a25d-8608290c688a)
+Jun 18 13:17:34.468: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.dns-201.svc.cluster.local from pod dns-201/dns-test-6ca54d7b-91cb-11e9-a25d-8608290c688a: the server could not find the requested resource (get pods dns-test-6ca54d7b-91cb-11e9-a25d-8608290c688a)
+Jun 18 13:17:34.487: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.dns-201.svc.cluster.local from pod dns-201/dns-test-6ca54d7b-91cb-11e9-a25d-8608290c688a: the server could not find the requested resource (get pods dns-test-6ca54d7b-91cb-11e9-a25d-8608290c688a)
+Jun 18 13:17:34.611: INFO: Lookups using dns-201/dns-test-6ca54d7b-91cb-11e9-a25d-8608290c688a failed for: [wheezy_udp@dns-test-service.dns-201.svc.cluster.local wheezy_tcp@dns-test-service.dns-201.svc.cluster.local wheezy_udp@_http._tcp.dns-test-service.dns-201.svc.cluster.local wheezy_tcp@_http._tcp.dns-test-service.dns-201.svc.cluster.local jessie_udp@dns-test-service.dns-201.svc.cluster.local jessie_tcp@dns-test-service.dns-201.svc.cluster.local jessie_udp@_http._tcp.dns-test-service.dns-201.svc.cluster.local jessie_tcp@_http._tcp.dns-test-service.dns-201.svc.cluster.local]
+
+Jun 18 13:17:40.285: INFO: DNS probes using dns-201/dns-test-6ca54d7b-91cb-11e9-a25d-8608290c688a succeeded
 
 STEP: deleting the pod
 STEP: deleting the test service
 STEP: deleting the test headless service
 [AfterEach] [sig-network] DNS
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 19:59:24.771: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "dns-4783" for this suite.
-Jun  4 19:59:30.835: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 19:59:31.282: INFO: namespace dns-4783 deletion completed in 6.494920565s
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 13:17:40.409: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "dns-201" for this suite.
+Jun 18 13:17:48.450: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 13:17:48.891: INFO: namespace dns-201 deletion completed in 8.466929131s
 
-• [SLOW TEST:24.711 seconds]
+• [SLOW TEST:19.251 seconds]
 [sig-network] DNS
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:22
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:22
   should provide DNS for services  [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-SSSSSSSSSS
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-storage] Downward API volume 
-  should set mode on item file [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [sig-storage] Downward API volume
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+[k8s.io] Kubelet when scheduling a busybox Pod with hostAliases 
+  should write entries to /etc/hosts [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [k8s.io] Kubelet
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 19:59:31.284: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-STEP: Building a namespace api object, basename downward-api
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in downward-api-6695
+Jun 18 13:17:48.895: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename kubelet-test
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubelet-test-2186
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-storage] Downward API volume
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:39
-[It] should set mode on item file [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: Creating a pod to test downward API volume plugin
-Jun  4 19:59:31.585: INFO: Waiting up to 5m0s for pod "downwardapi-volume-44833729-8703-11e9-8d1b-467ee19922ac" in namespace "downward-api-6695" to be "success or failure"
-Jun  4 19:59:31.596: INFO: Pod "downwardapi-volume-44833729-8703-11e9-8d1b-467ee19922ac": Phase="Pending", Reason="", readiness=false. Elapsed: 11.033618ms
-Jun  4 19:59:33.608: INFO: Pod "downwardapi-volume-44833729-8703-11e9-8d1b-467ee19922ac": Phase="Running", Reason="", readiness=true. Elapsed: 2.022561801s
-Jun  4 19:59:35.621: INFO: Pod "downwardapi-volume-44833729-8703-11e9-8d1b-467ee19922ac": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.035661356s
+[BeforeEach] [k8s.io] Kubelet
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:37
+[It] should write entries to /etc/hosts [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[AfterEach] [k8s.io] Kubelet
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 13:17:51.195: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "kubelet-test-2186" for this suite.
+Jun 18 13:18:35.243: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 13:18:35.621: INFO: namespace kubelet-test-2186 deletion completed in 44.411475562s
+
+• [SLOW TEST:46.726 seconds]
+[k8s.io] Kubelet
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
+  when scheduling a busybox Pod with hostAliases
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:136
+    should write entries to /etc/hosts [LinuxOnly] [NodeConformance] [Conformance]
+    /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSSSSSSSSSSSSS
+------------------------------
+[k8s.io] Pods 
+  should contain environment variables for services [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [k8s.io] Pods
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Jun 18 13:18:35.621: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename pods
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in pods-6674
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [k8s.io] Pods
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:135
+[It] should contain environment variables for services [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+Jun 18 13:18:37.929: INFO: Waiting up to 5m0s for pod "client-envvars-95348bdd-91cb-11e9-a25d-8608290c688a" in namespace "pods-6674" to be "success or failure"
+Jun 18 13:18:37.965: INFO: Pod "client-envvars-95348bdd-91cb-11e9-a25d-8608290c688a": Phase="Pending", Reason="", readiness=false. Elapsed: 35.409127ms
+Jun 18 13:18:39.979: INFO: Pod "client-envvars-95348bdd-91cb-11e9-a25d-8608290c688a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.049285346s
 STEP: Saw pod success
-Jun  4 19:59:35.621: INFO: Pod "downwardapi-volume-44833729-8703-11e9-8d1b-467ee19922ac" satisfied condition "success or failure"
-Jun  4 19:59:35.633: INFO: Trying to get logs from node 10.212.23.164 pod downwardapi-volume-44833729-8703-11e9-8d1b-467ee19922ac container client-container: 
+Jun 18 13:18:39.979: INFO: Pod "client-envvars-95348bdd-91cb-11e9-a25d-8608290c688a" satisfied condition "success or failure"
+Jun 18 13:18:39.993: INFO: Trying to get logs from node 10.72.74.138 pod client-envvars-95348bdd-91cb-11e9-a25d-8608290c688a container env3cont: 
 STEP: delete the pod
-Jun  4 19:59:35.755: INFO: Waiting for pod downwardapi-volume-44833729-8703-11e9-8d1b-467ee19922ac to disappear
-Jun  4 19:59:35.767: INFO: Pod downwardapi-volume-44833729-8703-11e9-8d1b-467ee19922ac no longer exists
-[AfterEach] [sig-storage] Downward API volume
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 19:59:35.767: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "downward-api-6695" for this suite.
-Jun  4 19:59:41.827: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 19:59:42.276: INFO: namespace downward-api-6695 deletion completed in 6.49204476s
+Jun 18 13:18:40.084: INFO: Waiting for pod client-envvars-95348bdd-91cb-11e9-a25d-8608290c688a to disappear
+Jun 18 13:18:40.098: INFO: Pod client-envvars-95348bdd-91cb-11e9-a25d-8608290c688a no longer exists
+[AfterEach] [k8s.io] Pods
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 13:18:40.098: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "pods-6674" for this suite.
+Jun 18 13:19:24.137: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 13:19:24.925: INFO: namespace pods-6674 deletion completed in 44.813115902s
 
-• [SLOW TEST:10.992 seconds]
-[sig-storage] Downward API volume
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:34
-  should set mode on item file [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+• [SLOW TEST:49.304 seconds]
+[k8s.io] Pods
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
+  should contain environment variables for services [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+SSSSSSS
 ------------------------------
 [sig-api-machinery] Garbage collector 
-  should delete RS created by deployment when not orphaning [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+  should not be blocked by dependency circle [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 [BeforeEach] [sig-api-machinery] Garbage collector
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 19:59:42.276: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
+Jun 18 13:19:24.928: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
 STEP: Building a namespace api object, basename gc
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in gc-2694
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in gc-8627
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should delete RS created by deployment when not orphaning [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: create the deployment
-STEP: Wait for the Deployment to create new ReplicaSet
-STEP: delete the deployment
-STEP: wait for all rs to be garbage collected
-STEP: expected 0 pods, got 2 pods
-STEP: expected 0 rs, got 1 rs
-STEP: Gathering metrics
-W0604 19:59:43.671422      18 metrics_grabber.go:79] Master node is not registered. Grabbing metrics from Scheduler, ControllerManager and ClusterAutoscaler is disabled.
-Jun  4 19:59:43.671: INFO: For apiserver_request_total:
-For apiserver_request_latencies_summary:
-For apiserver_init_events_total:
-For garbage_collector_attempt_to_delete_queue_latency:
-For garbage_collector_attempt_to_delete_work_duration:
-For garbage_collector_attempt_to_orphan_queue_latency:
-For garbage_collector_attempt_to_orphan_work_duration:
-For garbage_collector_dirty_processing_latency_microseconds:
-For garbage_collector_event_processing_latency_microseconds:
-For garbage_collector_graph_changes_queue_latency:
-For garbage_collector_graph_changes_work_duration:
-For garbage_collector_orphan_processing_latency_microseconds:
-For namespace_queue_latency:
-For namespace_queue_latency_sum:
-For namespace_queue_latency_count:
-For namespace_retries:
-For namespace_work_duration:
-For namespace_work_duration_sum:
-For namespace_work_duration_count:
-For function_duration_seconds:
-For errors_total:
-For evicted_pods_total:
-
+[It] should not be blocked by dependency circle [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+Jun 18 13:19:25.215: INFO: pod1.ObjectMeta.OwnerReferences=[]v1.OwnerReference{v1.OwnerReference{APIVersion:"v1", Kind:"Pod", Name:"pod3", UID:"b161f953-91cb-11e9-a08a-ee7a14707756", Controller:(*bool)(0xc002d0c99a), BlockOwnerDeletion:(*bool)(0xc002d0c99b)}}
+Jun 18 13:19:25.241: INFO: pod2.ObjectMeta.OwnerReferences=[]v1.OwnerReference{v1.OwnerReference{APIVersion:"v1", Kind:"Pod", Name:"pod1", UID:"b15c7b54-91cb-11e9-a08a-ee7a14707756", Controller:(*bool)(0xc0029421a2), BlockOwnerDeletion:(*bool)(0xc0029421a3)}}
+Jun 18 13:19:25.257: INFO: pod3.ObjectMeta.OwnerReferences=[]v1.OwnerReference{v1.OwnerReference{APIVersion:"v1", Kind:"Pod", Name:"pod2", UID:"b15ed98f-91cb-11e9-a08a-ee7a14707756", Controller:(*bool)(0xc00294282a), BlockOwnerDeletion:(*bool)(0xc00294282b)}}
 [AfterEach] [sig-api-machinery] Garbage collector
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 19:59:43.671: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "gc-2694" for this suite.
-Jun  4 19:59:49.727: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 19:59:50.081: INFO: namespace gc-2694 deletion completed in 6.396112679s
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 13:19:30.299: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "gc-8627" for this suite.
+Jun 18 13:19:36.342: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 13:19:36.756: INFO: namespace gc-8627 deletion completed in 6.443031813s
 
-• [SLOW TEST:7.805 seconds]
+• [SLOW TEST:11.829 seconds]
 [sig-api-machinery] Garbage collector
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:22
-  should delete RS created by deployment when not orphaning [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:22
+  should not be blocked by dependency circle [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-SSSSSSSSS
+SSSSSSS
 ------------------------------
-[sig-api-machinery] Watchers 
-  should be able to restart watching from the last resource version observed by the previous watch [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [sig-api-machinery] Watchers
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+[k8s.io] Kubelet when scheduling a busybox command that always fails in a pod 
+  should have an terminated reason [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [k8s.io] Kubelet
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 19:59:50.081: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-STEP: Building a namespace api object, basename watch
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in watch-7749
+Jun 18 13:19:36.757: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename kubelet-test
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubelet-test-108
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should be able to restart watching from the last resource version observed by the previous watch [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: creating a watch on configmaps
-STEP: creating a new configmap
-STEP: modifying the configmap once
-STEP: closing the watch once it receives two notifications
-Jun  4 19:59:50.345: INFO: Got : ADDED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-watch-closed,GenerateName:,Namespace:watch-7749,SelfLink:/api/v1/namespaces/watch-7749/configmaps/e2e-watch-test-watch-closed,UID:4fb055a4-8703-11e9-8318-1e5386706511,ResourceVersion:30541,Generation:0,CreationTimestamp:2019-06-04 19:59:50 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: watch-closed-and-restarted,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{},BinaryData:map[string][]byte{},}
-Jun  4 19:59:50.345: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-watch-closed,GenerateName:,Namespace:watch-7749,SelfLink:/api/v1/namespaces/watch-7749/configmaps/e2e-watch-test-watch-closed,UID:4fb055a4-8703-11e9-8318-1e5386706511,ResourceVersion:30542,Generation:0,CreationTimestamp:2019-06-04 19:59:50 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: watch-closed-and-restarted,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{mutation: 1,},BinaryData:map[string][]byte{},}
-STEP: modifying the configmap a second time, while the watch is closed
-STEP: creating a new watch on configmaps from the last resource version observed by the first watch
-STEP: deleting the configmap
-STEP: Expecting to observe notifications for all changes to the configmap since the first watch closed
-Jun  4 19:59:50.395: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-watch-closed,GenerateName:,Namespace:watch-7749,SelfLink:/api/v1/namespaces/watch-7749/configmaps/e2e-watch-test-watch-closed,UID:4fb055a4-8703-11e9-8318-1e5386706511,ResourceVersion:30543,Generation:0,CreationTimestamp:2019-06-04 19:59:50 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: watch-closed-and-restarted,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},}
-Jun  4 19:59:50.395: INFO: Got : DELETED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-watch-closed,GenerateName:,Namespace:watch-7749,SelfLink:/api/v1/namespaces/watch-7749/configmaps/e2e-watch-test-watch-closed,UID:4fb055a4-8703-11e9-8318-1e5386706511,ResourceVersion:30544,Generation:0,CreationTimestamp:2019-06-04 19:59:50 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: watch-closed-and-restarted,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},}
-[AfterEach] [sig-api-machinery] Watchers
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 19:59:50.395: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "watch-7749" for this suite.
-Jun  4 19:59:56.450: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 19:59:56.788: INFO: namespace watch-7749 deletion completed in 6.379681374s
+[BeforeEach] [k8s.io] Kubelet
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:37
+[BeforeEach] when scheduling a busybox command that always fails in a pod
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:81
+[It] should have an terminated reason [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[AfterEach] [k8s.io] Kubelet
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 13:19:41.016: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "kubelet-test-108" for this suite.
+Jun 18 13:19:47.057: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 13:19:47.440: INFO: namespace kubelet-test-108 deletion completed in 6.40914248s
 
-• [SLOW TEST:6.707 seconds]
-[sig-api-machinery] Watchers
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:22
-  should be able to restart watching from the last resource version observed by the previous watch [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+• [SLOW TEST:10.683 seconds]
+[k8s.io] Kubelet
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
+  when scheduling a busybox command that always fails in a pod
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:78
+    should have an terminated reason [NodeConformance] [Conformance]
+    /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+SSSSSSSSS
 ------------------------------
-[sig-storage] Projected configMap 
-  optional updates should be reflected in volume [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [sig-storage] Projected configMap
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+[sig-storage] Secrets 
+  should be consumable from pods in volume with mappings [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-storage] Secrets
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 19:59:56.789: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-STEP: Building a namespace api object, basename projected
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-3704
+Jun 18 13:19:47.441: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename secrets
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in secrets-5842
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] optional updates should be reflected in volume [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: Creating configMap with name cm-test-opt-del-53b06656-8703-11e9-8d1b-467ee19922ac
-STEP: Creating configMap with name cm-test-opt-upd-53b0668a-8703-11e9-8d1b-467ee19922ac
-STEP: Creating the pod
-STEP: Deleting configmap cm-test-opt-del-53b06656-8703-11e9-8d1b-467ee19922ac
-STEP: Updating configmap cm-test-opt-upd-53b0668a-8703-11e9-8d1b-467ee19922ac
-STEP: Creating configMap with name cm-test-opt-create-53b066a0-8703-11e9-8d1b-467ee19922ac
-STEP: waiting to observe update in volume
-[AfterEach] [sig-storage] Projected configMap
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 20:00:01.425: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "projected-3704" for this suite.
-Jun  4 20:00:25.495: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 20:00:25.912: INFO: namespace projected-3704 deletion completed in 24.472894208s
+[It] should be consumable from pods in volume with mappings [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating secret with name secret-test-map-bec44ca0-91cb-11e9-a25d-8608290c688a
+STEP: Creating a pod to test consume secrets
+Jun 18 13:19:47.683: INFO: Waiting up to 5m0s for pod "pod-secrets-bec6364c-91cb-11e9-a25d-8608290c688a" in namespace "secrets-5842" to be "success or failure"
+Jun 18 13:19:47.696: INFO: Pod "pod-secrets-bec6364c-91cb-11e9-a25d-8608290c688a": Phase="Pending", Reason="", readiness=false. Elapsed: 13.481542ms
+Jun 18 13:19:49.710: INFO: Pod "pod-secrets-bec6364c-91cb-11e9-a25d-8608290c688a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.0270561s
+STEP: Saw pod success
+Jun 18 13:19:49.710: INFO: Pod "pod-secrets-bec6364c-91cb-11e9-a25d-8608290c688a" satisfied condition "success or failure"
+Jun 18 13:19:49.723: INFO: Trying to get logs from node 10.72.74.184 pod pod-secrets-bec6364c-91cb-11e9-a25d-8608290c688a container secret-volume-test: 
+STEP: delete the pod
+Jun 18 13:19:49.802: INFO: Waiting for pod pod-secrets-bec6364c-91cb-11e9-a25d-8608290c688a to disappear
+Jun 18 13:19:49.814: INFO: Pod pod-secrets-bec6364c-91cb-11e9-a25d-8608290c688a no longer exists
+[AfterEach] [sig-storage] Secrets
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 13:19:49.814: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "secrets-5842" for this suite.
+Jun 18 13:19:57.854: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 13:19:58.297: INFO: namespace secrets-5842 deletion completed in 8.468664491s
 
-• [SLOW TEST:29.124 seconds]
-[sig-storage] Projected configMap
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_configmap.go:33
-  optional updates should be reflected in volume [NodeConformance] [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+• [SLOW TEST:10.856 seconds]
+[sig-storage] Secrets
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/secrets_volume.go:33
+  should be consumable from pods in volume with mappings [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-SSSS
+SSSSSSSSSS
 ------------------------------
-[sig-api-machinery] Garbage collector 
-  should not delete dependents that have both valid owner and owner that's waiting for dependents to be deleted [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-[BeforeEach] [sig-api-machinery] Garbage collector
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+[sig-network] Services 
+  should serve a basic endpoint from pods  [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-network] Services
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
 STEP: Creating a kubernetes client
-Jun  4 20:00:25.913: INFO: >>> kubeConfig: /tmp/kubeconfig-229234504
-STEP: Building a namespace api object, basename gc
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in gc-6933
+Jun 18 13:19:58.297: INFO: >>> kubeConfig: /tmp/kubeconfig-426929150
+STEP: Building a namespace api object, basename services
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in services-6038
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should not delete dependents that have both valid owner and owner that's waiting for dependents to be deleted [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
-STEP: create the rc1
-STEP: create the rc2
-STEP: set half of pods created by rc simpletest-rc-to-be-deleted to have rc simpletest-rc-to-stay as owner as well
-STEP: delete the rc simpletest-rc-to-be-deleted
-STEP: wait for the rc to be deleted
-STEP: Gathering metrics
-W0604 20:00:36.483452      18 metrics_grabber.go:79] Master node is not registered. Grabbing metrics from Scheduler, ControllerManager and ClusterAutoscaler is disabled.
-Jun  4 20:00:36.483: INFO: For apiserver_request_total:
-For apiserver_request_latencies_summary:
-For apiserver_init_events_total:
-For garbage_collector_attempt_to_delete_queue_latency:
-For garbage_collector_attempt_to_delete_work_duration:
-For garbage_collector_attempt_to_orphan_queue_latency:
-For garbage_collector_attempt_to_orphan_work_duration:
-For garbage_collector_dirty_processing_latency_microseconds:
-For garbage_collector_event_processing_latency_microseconds:
-For garbage_collector_graph_changes_queue_latency:
-For garbage_collector_graph_changes_work_duration:
-For garbage_collector_orphan_processing_latency_microseconds:
-For namespace_queue_latency:
-For namespace_queue_latency_sum:
-For namespace_queue_latency_count:
-For namespace_retries:
-For namespace_work_duration:
-For namespace_work_duration_sum:
-For namespace_work_duration_count:
-For function_duration_seconds:
-For errors_total:
-For evicted_pods_total:
-
-[AfterEach] [sig-api-machinery] Garbage collector
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
-Jun  4 20:00:36.483: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "gc-6933" for this suite.
-Jun  4 20:00:44.540: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
-Jun  4 20:00:44.927: INFO: namespace gc-6933 deletion completed in 8.430420211s
+[BeforeEach] [sig-network] Services
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:86
+[It] should serve a basic endpoint from pods  [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: creating service endpoint-test2 in namespace services-6038
+STEP: waiting up to 3m0s for service endpoint-test2 in namespace services-6038 to expose endpoints map[]
+Jun 18 13:19:58.554: INFO: successfully validated that service endpoint-test2 in namespace services-6038 exposes endpoints map[] (23.103493ms elapsed)
+STEP: Creating pod pod1 in namespace services-6038
+STEP: waiting up to 3m0s for service endpoint-test2 in namespace services-6038 to expose endpoints map[pod1:[80]]
+Jun 18 13:20:01.662: INFO: successfully validated that service endpoint-test2 in namespace services-6038 exposes endpoints map[pod1:[80]] (3.08281696s elapsed)
+STEP: Creating pod pod2 in namespace services-6038
+STEP: waiting up to 3m0s for service endpoint-test2 in namespace services-6038 to expose endpoints map[pod1:[80] pod2:[80]]
+Jun 18 13:20:03.787: INFO: successfully validated that service endpoint-test2 in namespace services-6038 exposes endpoints map[pod1:[80] pod2:[80]] (2.107150197s elapsed)
+STEP: Deleting pod pod1 in namespace services-6038
+STEP: waiting up to 3m0s for service endpoint-test2 in namespace services-6038 to expose endpoints map[pod2:[80]]
+Jun 18 13:20:03.832: INFO: successfully validated that service endpoint-test2 in namespace services-6038 exposes endpoints map[pod2:[80]] (19.360151ms elapsed)
+STEP: Deleting pod pod2 in namespace services-6038
+STEP: waiting up to 3m0s for service endpoint-test2 in namespace services-6038 to expose endpoints map[]
+Jun 18 13:20:03.864: INFO: successfully validated that service endpoint-test2 in namespace services-6038 exposes endpoints map[] (6.430694ms elapsed)
+[AfterEach] [sig-network] Services
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Jun 18 13:20:03.913: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "services-6038" for this suite.
+Jun 18 13:20:27.975: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Jun 18 13:20:28.390: INFO: namespace services-6038 deletion completed in 24.441893523s
+[AfterEach] [sig-network] Services
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:91
 
-• [SLOW TEST:19.015 seconds]
-[sig-api-machinery] Garbage collector
-/workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:22
-  should not delete dependents that have both valid owner and owner that's waiting for dependents to be deleted [Conformance]
-  /workspace/anago-v1.14.2-beta.0.85+66049e3b21efe1/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+• [SLOW TEST:30.092 seconds]
+[sig-network] Services
+/workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:22
+  should serve a basic endpoint from pods  [Conformance]
+  /workspace/anago-v1.14.3-beta.0.37+5e53fd6bc17c0d/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
 ------------------------------
-SSSSJun  4 20:00:44.927: INFO: Running AfterSuite actions on all nodes
-Jun  4 20:00:44.927: INFO: Running AfterSuite actions on node 1
-Jun  4 20:00:44.927: INFO: Skipping dumping logs from cluster
+SSSSSSSSSSSSSSJun 18 13:20:28.391: INFO: Running AfterSuite actions on all nodes
+Jun 18 13:20:28.391: INFO: Running AfterSuite actions on node 1
+Jun 18 13:20:28.391: INFO: Skipping dumping logs from cluster
 
-Ran 204 of 3585 Specs in 5566.436 seconds
+Ran 204 of 3585 Specs in 6352.755 seconds
 SUCCESS! -- 204 Passed | 0 Failed | 0 Pending | 3381 Skipped PASS
 
-Ginkgo ran 1 suite in 1h32m47.300470535s
+Ginkgo ran 1 suite in 1h45m54.154141461s
 Test Suite Passed
diff --git a/v1.14/iks/junit_01.xml b/v1.14/iks/junit_01.xml
index 08fd658998..c7253f0a0c 100644
--- a/v1.14/iks/junit_01.xml
+++ b/v1.14/iks/junit_01.xml
@@ -1,10350 +1,10350 @@
 
-  
-      
+  
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
-      
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
-      
+      
+      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
+      
           
       
-      
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
   
\ No newline at end of file
diff --git a/v1.14/iks/version.txt b/v1.14/iks/version.txt
index 172803727c..bfa278fdf3 100644
--- a/v1.14/iks/version.txt
+++ b/v1.14/iks/version.txt
@@ -1,2 +1,2 @@
-Client Version: version.Info{Major:"1", Minor:"14", GitVersion:"v1.14.2", GitCommit:"66049e3b21efe110454d67df4fa62b08ea79a19b", GitTreeState:"clean", BuildDate:"2019-05-16T16:23:09Z", GoVersion:"go1.12.5", Compiler:"gc", Platform:"linux/amd64"}
-Server Version: version.Info{Major:"1", Minor:"14", GitVersion:"v1.14.2+IKS", GitCommit:"b3e16abec4c3166d8bb7524d9db770d265f6e470", GitTreeState:"clean", BuildDate:"2019-05-31T08:07:55Z", GoVersion:"go1.12.5", Compiler:"gc", Platform:"linux/amd64"}
+Client Version: version.Info{Major:"1", Minor:"14", GitVersion:"v1.14.3", GitCommit:"5e53fd6bc17c0dec8434817e69b04a25d8ae0ff0", GitTreeState:"clean", BuildDate:"2019-06-06T01:44:30Z", GoVersion:"go1.12.5", Compiler:"gc", Platform:"linux/amd64"}
+Server Version: version.Info{Major:"1", Minor:"14", GitVersion:"v1.14.3+IKS", GitCommit:"1c6446568f0a7d9d62dfe2f6c6a8388b3d4b1a6b", GitTreeState:"clean", BuildDate:"2019-06-13T12:53:41Z", GoVersion:"go1.12.5", Compiler:"gc", Platform:"linux/amd64"}