diff --git a/v1.14/giantswarm-kvm/PRODUCT.yaml b/v1.14/giantswarm-kvm/PRODUCT.yaml new file mode 100644 index 0000000000..f775265b01 --- /dev/null +++ b/v1.14/giantswarm-kvm/PRODUCT.yaml @@ -0,0 +1,9 @@ +vendor: Giant Swarm +name: Managed Kubernetes on KVM +version: 8.0.0 +website_url: https://giantswarm.io +repo_url: https://github.com/giantswarm/kvm-operator/ +documentation_url: https://docs.giantswarm.io +product_logo_url: https://www.dropbox.com/s/kydb137cjsix7du/giantswarm_ant.png +type: distribution +description: 'The Giant Swarm platform enables users to simply and rapidly create and use 24/7 managed Kubernetes clusters on-demand.' diff --git a/v1.14/giantswarm-kvm/README.md b/v1.14/giantswarm-kvm/README.md new file mode 100644 index 0000000000..e18299b880 --- /dev/null +++ b/v1.14/giantswarm-kvm/README.md @@ -0,0 +1,67 @@ +## To Reproduce: + +Note: to reproduce you need a Giant Swarm account. + +### Create cluster + +``` +$ gsctl create cluster --owner=myorg +``` + +This will report back a cluster ID that you need for the next step. + + +### Get Credentials + + +``` +$ gsctl create kubeconfig -c --certificate-organizations=system:masters +``` + +### Run the tests + +Wait a bit for the cluster to come up (depending on the underlying infrastructure this might take a few minutes). + +Download a [binary release](https://github.com/heptio/sonobuoy/releases) of the CLI, or build it yourself by running: + +``` +$ go get -u -v github.com/heptio/sonobuoy +``` + +Deploy a Sonobuoy pod to your cluster with: + +``` +$ sonobuoy run +``` + +View actively running pods: + +``` +$ sonobuoy status +``` + +To inspect the logs: + +``` +$ sonobuoy logs +``` + +Once `sonobuoy status` shows the run as `completed`, copy the output directory from the main Sonobuoy pod to +a local directory: + +``` +$ sonobuoy retrieve . +``` + +This copies a single `.tar.gz` snapshot from the Sonobuoy pod into your local +`.` directory. Extract the contents into `./results` with: + +``` +mkdir ./results; tar xzf *.tar.gz -C ./results +``` + +### Destroy cluster + +``` +$ gsctl delete cluster -c +``` diff --git a/v1.14/giantswarm-kvm/e2e.log b/v1.14/giantswarm-kvm/e2e.log new file mode 100644 index 0000000000..bd2b44f2b0 --- /dev/null +++ b/v1.14/giantswarm-kvm/e2e.log @@ -0,0 +1,10968 @@ +I0429 11:31:17.241479 15 test_context.go:405] Using a temporary kubeconfig file from in-cluster config : /tmp/kubeconfig-185508174 +I0429 11:31:17.241642 15 e2e.go:240] Starting e2e run "4c9f30da-6a72-11e9-b6ef-0e71f8c7e888" on Ginkgo node 1 +Running Suite: Kubernetes e2e suite +=================================== +Random Seed: 1556537475 - Will randomize all specs +Will run 204 of 3584 specs + +Apr 29 11:31:17.447: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174 +Apr 29 11:31:17.450: INFO: Waiting up to 30m0s for all (but 0) nodes to be schedulable +Apr 29 11:31:17.568: INFO: Waiting up to 10m0s for all pods (need at least 0) in namespace 'kube-system' to be running and ready +Apr 29 11:31:17.692: INFO: 33 / 33 pods in namespace 'kube-system' are running and ready (0 seconds elapsed) +Apr 29 11:31:17.692: INFO: expected 10 pod replicas in namespace 'kube-system', 10 are Running and Ready. +Apr 29 11:31:17.693: INFO: Waiting up to 5m0s for all daemonsets in namespace 'kube-system' to start +Apr 29 11:31:17.752: INFO: 4 / 4 pods ready in namespace 'kube-system' in daemonset 'calico-node' (0 seconds elapsed) +Apr 29 11:31:17.752: INFO: 4 / 4 pods ready in namespace 'kube-system' in daemonset 'cert-exporter' (0 seconds elapsed) +Apr 29 11:31:17.753: INFO: 4 / 4 pods ready in namespace 'kube-system' in daemonset 'kube-proxy' (0 seconds elapsed) +Apr 29 11:31:17.753: INFO: 4 / 4 pods ready in namespace 'kube-system' in daemonset 'net-exporter' (0 seconds elapsed) +Apr 29 11:31:17.753: INFO: 4 / 4 pods ready in namespace 'kube-system' in daemonset 'node-exporter' (0 seconds elapsed) +Apr 29 11:31:17.753: INFO: e2e test version: v1.14.1 +Apr 29 11:31:17.769: INFO: kube-apiserver version: v1.14.1 +SS +------------------------------ +[k8s.io] Kubelet when scheduling a busybox command that always fails in a pod + should be possible to delete [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [k8s.io] Kubelet + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +STEP: Creating a kubernetes client +Apr 29 11:31:17.781: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174 +STEP: Building a namespace api object, basename kubelet-test +Apr 29 11:31:17.952: INFO: Found PodSecurityPolicies; assuming PodSecurityPolicy is enabled. +Apr 29 11:31:17.994: INFO: Found ClusterRoles; assuming RBAC is enabled. +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubelet-test-5522 +STEP: Waiting for a default service account to be provisioned in namespace +[BeforeEach] [k8s.io] Kubelet + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:37 +[BeforeEach] when scheduling a busybox command that always fails in a pod + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:81 +[It] should be possible to delete [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[AfterEach] [k8s.io] Kubelet + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Apr 29 11:31:18.203: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "kubelet-test-5522" for this suite. +Apr 29 11:31:26.239: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Apr 29 11:31:26.559: INFO: namespace kubelet-test-5522 deletion completed in 8.343923855s + +• [SLOW TEST:8.778 seconds] +[k8s.io] Kubelet +/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687 + when scheduling a busybox command that always fails in a pod + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:78 + should be possible to delete [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +------------------------------ +SSSS +------------------------------ +[sig-cli] Kubectl client [k8s.io] Proxy server + should support proxy with --port 0 [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [sig-cli] Kubectl client + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +STEP: Creating a kubernetes client +Apr 29 11:31:26.560: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174 +STEP: Building a namespace api object, basename kubectl +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-3699 +STEP: Waiting for a default service account to be provisioned in namespace +[BeforeEach] [sig-cli] Kubectl client + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:213 +[It] should support proxy with --port 0 [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: starting the proxy server +Apr 29 11:31:26.774: INFO: Asynchronously running '/usr/local/bin/kubectl kubectl --kubeconfig=/tmp/kubeconfig-185508174 proxy -p 0 --disable-filter' +STEP: curling proxy /api/ output +[AfterEach] [sig-cli] Kubectl client + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Apr 29 11:31:26.888: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "kubectl-3699" for this suite. +Apr 29 11:31:32.927: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Apr 29 11:31:33.168: INFO: namespace kubectl-3699 deletion completed in 6.262633177s + +• [SLOW TEST:6.608 seconds] +[sig-cli] Kubectl client +/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 + [k8s.io] Proxy server + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687 + should support proxy with --port 0 [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +------------------------------ +SS +------------------------------ +[sig-api-machinery] Namespaces [Serial] + should ensure that all pods are removed when a namespace is deleted [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [sig-api-machinery] Namespaces [Serial] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +STEP: Creating a kubernetes client +Apr 29 11:31:33.175: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174 +STEP: Building a namespace api object, basename namespaces +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in namespaces-8810 +STEP: Waiting for a default service account to be provisioned in namespace +[It] should ensure that all pods are removed when a namespace is deleted [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: Creating a test namespace +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in nsdeletetest-3061 +STEP: Waiting for a default service account to be provisioned in namespace +STEP: Creating a pod in the namespace +STEP: Waiting for the pod to have running status +STEP: Deleting the namespace +STEP: Waiting for the namespace to be removed. +STEP: Recreating the namespace +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in nsdeletetest-3410 +STEP: Verifying there are no pods in the namespace +[AfterEach] [sig-api-machinery] Namespaces [Serial] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Apr 29 11:32:00.891: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "namespaces-8810" for this suite. +Apr 29 11:32:06.920: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Apr 29 11:32:07.197: INFO: namespace namespaces-8810 deletion completed in 6.298715531s +STEP: Destroying namespace "nsdeletetest-3061" for this suite. +Apr 29 11:32:07.206: INFO: Namespace nsdeletetest-3061 was already deleted +STEP: Destroying namespace "nsdeletetest-3410" for this suite. +Apr 29 11:32:13.228: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Apr 29 11:32:13.464: INFO: namespace nsdeletetest-3410 deletion completed in 6.258033298s + +• [SLOW TEST:40.290 seconds] +[sig-api-machinery] Namespaces [Serial] +/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:22 + should ensure that all pods are removed when a namespace is deleted [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +------------------------------ +[sig-storage] HostPath + should give a volume the correct mode [LinuxOnly] [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [sig-storage] HostPath + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +STEP: Creating a kubernetes client +Apr 29 11:32:13.464: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174 +STEP: Building a namespace api object, basename hostpath +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in hostpath-7807 +STEP: Waiting for a default service account to be provisioned in namespace +[BeforeEach] [sig-storage] HostPath + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/host_path.go:37 +[It] should give a volume the correct mode [LinuxOnly] [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: Creating a pod to test hostPath mode +Apr 29 11:32:13.814: INFO: Waiting up to 5m0s for pod "pod-host-path-test" in namespace "hostpath-7807" to be "success or failure" +Apr 29 11:32:13.827: INFO: Pod "pod-host-path-test": Phase="Pending", Reason="", readiness=false. Elapsed: 12.176792ms +Apr 29 11:32:15.838: INFO: Pod "pod-host-path-test": Phase="Pending", Reason="", readiness=false. Elapsed: 2.023640402s +Apr 29 11:32:17.847: INFO: Pod "pod-host-path-test": Phase="Pending", Reason="", readiness=false. Elapsed: 4.032658988s +Apr 29 11:32:19.858: INFO: Pod "pod-host-path-test": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.043655599s +STEP: Saw pod success +Apr 29 11:32:19.858: INFO: Pod "pod-host-path-test" satisfied condition "success or failure" +Apr 29 11:32:19.862: INFO: Trying to get logs from node worker-4av2h-6665d7b5ff-9nc42 pod pod-host-path-test container test-container-1: +STEP: delete the pod +Apr 29 11:32:19.977: INFO: Waiting for pod pod-host-path-test to disappear +Apr 29 11:32:19.987: INFO: Pod pod-host-path-test no longer exists +[AfterEach] [sig-storage] HostPath + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Apr 29 11:32:19.987: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "hostpath-7807" for this suite. +Apr 29 11:32:26.023: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Apr 29 11:32:26.204: INFO: namespace hostpath-7807 deletion completed in 6.208397363s + +• [SLOW TEST:12.740 seconds] +[sig-storage] HostPath +/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/host_path.go:34 + should give a volume the correct mode [LinuxOnly] [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +------------------------------ +[sig-storage] ConfigMap + updates should be reflected in volume [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [sig-storage] ConfigMap + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +STEP: Creating a kubernetes client +Apr 29 11:32:26.205: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174 +STEP: Building a namespace api object, basename configmap +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in configmap-395 +STEP: Waiting for a default service account to be provisioned in namespace +[It] updates should be reflected in volume [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: Creating configMap with name configmap-test-upd-76d69816-6a72-11e9-b6ef-0e71f8c7e888 +STEP: Creating the pod +STEP: Updating configmap configmap-test-upd-76d69816-6a72-11e9-b6ef-0e71f8c7e888 +STEP: waiting to observe update in volume +[AfterEach] [sig-storage] ConfigMap + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Apr 29 11:33:51.337: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "configmap-395" for this suite. +Apr 29 11:34:31.362: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Apr 29 11:34:31.612: INFO: namespace configmap-395 deletion completed in 40.26751645s + +• [SLOW TEST:125.408 seconds] +[sig-storage] ConfigMap +/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:32 + updates should be reflected in volume [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +------------------------------ +SSSSSSSSSSSSSSSSSSSSSSSSSSS +------------------------------ +[sig-network] Proxy version v1 + should proxy through a service and a pod [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] version v1 + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +STEP: Creating a kubernetes client +Apr 29 11:34:31.612: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174 +STEP: Building a namespace api object, basename proxy +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in proxy-383 +STEP: Waiting for a default service account to be provisioned in namespace +[It] should proxy through a service and a pod [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: starting an echo server on multiple ports +STEP: creating replication controller proxy-service-6lw44 in namespace proxy-383 +I0429 11:34:31.862994 15 runners.go:184] Created replication controller with name: proxy-service-6lw44, namespace: proxy-383, replica count: 1 +I0429 11:34:32.914159 15 runners.go:184] proxy-service-6lw44 Pods: 1 out of 1 created, 0 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady +I0429 11:34:33.914576 15 runners.go:184] proxy-service-6lw44 Pods: 1 out of 1 created, 0 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady +I0429 11:34:34.914903 15 runners.go:184] proxy-service-6lw44 Pods: 1 out of 1 created, 0 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady +I0429 11:34:35.915328 15 runners.go:184] proxy-service-6lw44 Pods: 1 out of 1 created, 0 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady +I0429 11:34:36.915869 15 runners.go:184] proxy-service-6lw44 Pods: 1 out of 1 created, 0 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 1 runningButNotReady +I0429 11:34:37.917370 15 runners.go:184] proxy-service-6lw44 Pods: 1 out of 1 created, 1 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady +Apr 29 11:34:37.932: INFO: setup took 6.113519772s, starting test cases +STEP: running 16 cases, 20 attempts per case, 320 total attempts +Apr 29 11:34:37.985: INFO: (0) /api/v1/namespaces/proxy-383/services/proxy-service-6lw44:portname2/proxy/: bar (200; 52.547505ms) +Apr 29 11:34:37.987: INFO: (0) /api/v1/namespaces/proxy-383/pods/https:proxy-service-6lw44-6r54z:460/proxy/: tls baz (200; 54.219361ms) +Apr 29 11:34:37.992: INFO: (0) /api/v1/namespaces/proxy-383/services/proxy-service-6lw44:portname1/proxy/: foo (200; 57.583197ms) +Apr 29 11:34:38.009: INFO: (0) /api/v1/namespaces/proxy-383/pods/proxy-service-6lw44-6r54z:160/proxy/: foo (200; 75.742695ms) +Apr 29 11:34:38.023: INFO: (0) /api/v1/namespaces/proxy-383/pods/proxy-service-6lw44-6r54z/proxy/: test (200; 90.031929ms) +Apr 29 11:34:38.024: INFO: (0) /api/v1/namespaces/proxy-383/pods/http:proxy-service-6lw44-6r54z:160/proxy/: foo (200; 89.969692ms) +Apr 29 11:34:38.029: INFO: (0) /api/v1/namespaces/proxy-383/services/https:proxy-service-6lw44:tlsportname1/proxy/: tls baz (200; 94.219409ms) +Apr 29 11:34:38.051: INFO: (0) /api/v1/namespaces/proxy-383/pods/http:proxy-service-6lw44-6r54z:162/proxy/: bar (200; 116.881233ms) +Apr 29 11:34:38.051: INFO: (0) /api/v1/namespaces/proxy-383/pods/proxy-service-6lw44-6r54z:1080/proxy/: testt... (200; 147.79711ms) +Apr 29 11:34:38.083: INFO: (0) /api/v1/namespaces/proxy-383/pods/https:proxy-service-6lw44-6r54z:443/proxy/: t... (200; 27.061076ms) +Apr 29 11:34:38.125: INFO: (1) /api/v1/namespaces/proxy-383/pods/http:proxy-service-6lw44-6r54z:160/proxy/: foo (200; 28.191788ms) +Apr 29 11:34:38.126: INFO: (1) /api/v1/namespaces/proxy-383/pods/proxy-service-6lw44-6r54z:1080/proxy/: testtest (200; 62.320421ms) +Apr 29 11:34:38.191: INFO: (2) /api/v1/namespaces/proxy-383/pods/http:proxy-service-6lw44-6r54z:160/proxy/: foo (200; 31.022041ms) +Apr 29 11:34:38.191: INFO: (2) /api/v1/namespaces/proxy-383/pods/http:proxy-service-6lw44-6r54z:1080/proxy/: t... (200; 30.56416ms) +Apr 29 11:34:38.192: INFO: (2) /api/v1/namespaces/proxy-383/pods/https:proxy-service-6lw44-6r54z:443/proxy/: test (200; 51.689612ms) +Apr 29 11:34:38.214: INFO: (2) /api/v1/namespaces/proxy-383/pods/proxy-service-6lw44-6r54z:1080/proxy/: testt... (200; 19.242631ms) +Apr 29 11:34:38.248: INFO: (3) /api/v1/namespaces/proxy-383/pods/proxy-service-6lw44-6r54z:162/proxy/: bar (200; 20.871653ms) +Apr 29 11:34:38.248: INFO: (3) /api/v1/namespaces/proxy-383/pods/https:proxy-service-6lw44-6r54z:460/proxy/: tls baz (200; 20.85481ms) +Apr 29 11:34:38.248: INFO: (3) /api/v1/namespaces/proxy-383/pods/http:proxy-service-6lw44-6r54z:162/proxy/: bar (200; 20.963027ms) +Apr 29 11:34:38.248: INFO: (3) /api/v1/namespaces/proxy-383/pods/proxy-service-6lw44-6r54z:160/proxy/: foo (200; 20.809595ms) +Apr 29 11:34:38.248: INFO: (3) /api/v1/namespaces/proxy-383/pods/proxy-service-6lw44-6r54z:1080/proxy/: testtest (200; 21.073627ms) +Apr 29 11:34:38.252: INFO: (3) /api/v1/namespaces/proxy-383/services/proxy-service-6lw44:portname1/proxy/: foo (200; 24.932922ms) +Apr 29 11:34:38.253: INFO: (3) /api/v1/namespaces/proxy-383/services/proxy-service-6lw44:portname2/proxy/: bar (200; 25.855281ms) +Apr 29 11:34:38.255: INFO: (3) /api/v1/namespaces/proxy-383/services/https:proxy-service-6lw44:tlsportname1/proxy/: tls baz (200; 28.330582ms) +Apr 29 11:34:38.255: INFO: (3) /api/v1/namespaces/proxy-383/services/http:proxy-service-6lw44:portname2/proxy/: bar (200; 28.262221ms) +Apr 29 11:34:38.255: INFO: (3) /api/v1/namespaces/proxy-383/services/http:proxy-service-6lw44:portname1/proxy/: foo (200; 28.217278ms) +Apr 29 11:34:38.255: INFO: (3) /api/v1/namespaces/proxy-383/services/https:proxy-service-6lw44:tlsportname2/proxy/: tls qux (200; 28.193517ms) +Apr 29 11:34:38.276: INFO: (4) /api/v1/namespaces/proxy-383/pods/http:proxy-service-6lw44-6r54z:162/proxy/: bar (200; 20.703419ms) +Apr 29 11:34:38.276: INFO: (4) /api/v1/namespaces/proxy-383/pods/https:proxy-service-6lw44-6r54z:443/proxy/: t... (200; 26.994506ms) +Apr 29 11:34:38.291: INFO: (4) /api/v1/namespaces/proxy-383/pods/http:proxy-service-6lw44-6r54z:160/proxy/: foo (200; 31.700917ms) +Apr 29 11:34:38.291: INFO: (4) /api/v1/namespaces/proxy-383/pods/proxy-service-6lw44-6r54z/proxy/: test (200; 34.388886ms) +Apr 29 11:34:38.291: INFO: (4) /api/v1/namespaces/proxy-383/pods/https:proxy-service-6lw44-6r54z:462/proxy/: tls qux (200; 33.860985ms) +Apr 29 11:34:38.291: INFO: (4) /api/v1/namespaces/proxy-383/pods/proxy-service-6lw44-6r54z:1080/proxy/: testtesttest (200; 35.510134ms) +Apr 29 11:34:38.344: INFO: (5) /api/v1/namespaces/proxy-383/pods/http:proxy-service-6lw44-6r54z:1080/proxy/: t... (200; 35.870727ms) +Apr 29 11:34:38.344: INFO: (5) /api/v1/namespaces/proxy-383/pods/http:proxy-service-6lw44-6r54z:162/proxy/: bar (200; 37.726281ms) +Apr 29 11:34:38.372: INFO: (5) /api/v1/namespaces/proxy-383/services/proxy-service-6lw44:portname1/proxy/: foo (200; 64.88788ms) +Apr 29 11:34:38.374: INFO: (5) /api/v1/namespaces/proxy-383/services/https:proxy-service-6lw44:tlsportname2/proxy/: tls qux (200; 66.533822ms) +Apr 29 11:34:38.377: INFO: (5) /api/v1/namespaces/proxy-383/services/http:proxy-service-6lw44:portname1/proxy/: foo (200; 68.68758ms) +Apr 29 11:34:38.385: INFO: (5) /api/v1/namespaces/proxy-383/services/proxy-service-6lw44:portname2/proxy/: bar (200; 76.606806ms) +Apr 29 11:34:38.408: INFO: (6) /api/v1/namespaces/proxy-383/pods/proxy-service-6lw44-6r54z:160/proxy/: foo (200; 22.286588ms) +Apr 29 11:34:38.409: INFO: (6) /api/v1/namespaces/proxy-383/pods/http:proxy-service-6lw44-6r54z:160/proxy/: foo (200; 22.729637ms) +Apr 29 11:34:38.411: INFO: (6) /api/v1/namespaces/proxy-383/pods/proxy-service-6lw44-6r54z/proxy/: test (200; 25.752703ms) +Apr 29 11:34:38.413: INFO: (6) /api/v1/namespaces/proxy-383/pods/http:proxy-service-6lw44-6r54z:162/proxy/: bar (200; 27.445563ms) +Apr 29 11:34:38.434: INFO: (6) /api/v1/namespaces/proxy-383/pods/http:proxy-service-6lw44-6r54z:1080/proxy/: t... (200; 47.423548ms) +Apr 29 11:34:38.434: INFO: (6) /api/v1/namespaces/proxy-383/pods/proxy-service-6lw44-6r54z:162/proxy/: bar (200; 46.801193ms) +Apr 29 11:34:38.443: INFO: (6) /api/v1/namespaces/proxy-383/services/http:proxy-service-6lw44:portname2/proxy/: bar (200; 58.527124ms) +Apr 29 11:34:38.446: INFO: (6) /api/v1/namespaces/proxy-383/services/https:proxy-service-6lw44:tlsportname1/proxy/: tls baz (200; 59.951761ms) +Apr 29 11:34:38.448: INFO: (6) /api/v1/namespaces/proxy-383/pods/https:proxy-service-6lw44-6r54z:443/proxy/: testt... (200; 37.728524ms) +Apr 29 11:34:38.513: INFO: (7) /api/v1/namespaces/proxy-383/pods/proxy-service-6lw44-6r54z/proxy/: test (200; 41.100243ms) +Apr 29 11:34:38.513: INFO: (7) /api/v1/namespaces/proxy-383/pods/https:proxy-service-6lw44-6r54z:460/proxy/: tls baz (200; 41.798556ms) +Apr 29 11:34:38.513: INFO: (7) /api/v1/namespaces/proxy-383/pods/http:proxy-service-6lw44-6r54z:160/proxy/: foo (200; 40.795622ms) +Apr 29 11:34:38.513: INFO: (7) /api/v1/namespaces/proxy-383/pods/http:proxy-service-6lw44-6r54z:162/proxy/: bar (200; 40.854986ms) +Apr 29 11:34:38.513: INFO: (7) /api/v1/namespaces/proxy-383/pods/proxy-service-6lw44-6r54z:162/proxy/: bar (200; 40.832003ms) +Apr 29 11:34:38.513: INFO: (7) /api/v1/namespaces/proxy-383/pods/https:proxy-service-6lw44-6r54z:443/proxy/: testtest (200; 18.028756ms) +Apr 29 11:34:38.558: INFO: (8) /api/v1/namespaces/proxy-383/pods/https:proxy-service-6lw44-6r54z:460/proxy/: tls baz (200; 22.322684ms) +Apr 29 11:34:38.564: INFO: (8) /api/v1/namespaces/proxy-383/pods/https:proxy-service-6lw44-6r54z:443/proxy/: testt... (200; 40.092186ms) +Apr 29 11:34:38.576: INFO: (8) /api/v1/namespaces/proxy-383/pods/proxy-service-6lw44-6r54z:160/proxy/: foo (200; 40.697801ms) +Apr 29 11:34:38.578: INFO: (8) /api/v1/namespaces/proxy-383/services/proxy-service-6lw44:portname1/proxy/: foo (200; 42.702159ms) +Apr 29 11:34:38.579: INFO: (8) /api/v1/namespaces/proxy-383/services/https:proxy-service-6lw44:tlsportname1/proxy/: tls baz (200; 43.675985ms) +Apr 29 11:34:38.579: INFO: (8) /api/v1/namespaces/proxy-383/services/http:proxy-service-6lw44:portname1/proxy/: foo (200; 43.921619ms) +Apr 29 11:34:38.580: INFO: (8) /api/v1/namespaces/proxy-383/services/proxy-service-6lw44:portname2/proxy/: bar (200; 44.037861ms) +Apr 29 11:34:38.580: INFO: (8) /api/v1/namespaces/proxy-383/services/http:proxy-service-6lw44:portname2/proxy/: bar (200; 44.455051ms) +Apr 29 11:34:38.581: INFO: (8) /api/v1/namespaces/proxy-383/services/https:proxy-service-6lw44:tlsportname2/proxy/: tls qux (200; 45.496514ms) +Apr 29 11:34:38.596: INFO: (9) /api/v1/namespaces/proxy-383/pods/https:proxy-service-6lw44-6r54z:460/proxy/: tls baz (200; 15.115824ms) +Apr 29 11:34:38.603: INFO: (9) /api/v1/namespaces/proxy-383/pods/https:proxy-service-6lw44-6r54z:443/proxy/: testtest (200; 22.478898ms) +Apr 29 11:34:38.610: INFO: (9) /api/v1/namespaces/proxy-383/services/proxy-service-6lw44:portname2/proxy/: bar (200; 28.699419ms) +Apr 29 11:34:38.610: INFO: (9) /api/v1/namespaces/proxy-383/services/proxy-service-6lw44:portname1/proxy/: foo (200; 25.052714ms) +Apr 29 11:34:38.611: INFO: (9) /api/v1/namespaces/proxy-383/services/http:proxy-service-6lw44:portname2/proxy/: bar (200; 28.089616ms) +Apr 29 11:34:38.611: INFO: (9) /api/v1/namespaces/proxy-383/services/https:proxy-service-6lw44:tlsportname2/proxy/: tls qux (200; 27.953395ms) +Apr 29 11:34:38.612: INFO: (9) /api/v1/namespaces/proxy-383/pods/proxy-service-6lw44-6r54z:160/proxy/: foo (200; 29.474939ms) +Apr 29 11:34:38.613: INFO: (9) /api/v1/namespaces/proxy-383/services/https:proxy-service-6lw44:tlsportname1/proxy/: tls baz (200; 28.655904ms) +Apr 29 11:34:38.615: INFO: (9) /api/v1/namespaces/proxy-383/pods/http:proxy-service-6lw44-6r54z:162/proxy/: bar (200; 29.921482ms) +Apr 29 11:34:38.615: INFO: (9) /api/v1/namespaces/proxy-383/pods/http:proxy-service-6lw44-6r54z:1080/proxy/: t... (200; 31.250499ms) +Apr 29 11:34:38.616: INFO: (9) /api/v1/namespaces/proxy-383/pods/https:proxy-service-6lw44-6r54z:462/proxy/: tls qux (200; 31.921823ms) +Apr 29 11:34:38.617: INFO: (9) /api/v1/namespaces/proxy-383/services/http:proxy-service-6lw44:portname1/proxy/: foo (200; 34.061031ms) +Apr 29 11:34:38.617: INFO: (9) /api/v1/namespaces/proxy-383/pods/http:proxy-service-6lw44-6r54z:160/proxy/: foo (200; 32.452053ms) +Apr 29 11:34:38.617: INFO: (9) /api/v1/namespaces/proxy-383/pods/proxy-service-6lw44-6r54z:162/proxy/: bar (200; 31.806486ms) +Apr 29 11:34:38.695: INFO: (10) /api/v1/namespaces/proxy-383/pods/http:proxy-service-6lw44-6r54z:162/proxy/: bar (200; 75.592493ms) +Apr 29 11:34:38.695: INFO: (10) /api/v1/namespaces/proxy-383/pods/https:proxy-service-6lw44-6r54z:460/proxy/: tls baz (200; 75.53152ms) +Apr 29 11:34:38.695: INFO: (10) /api/v1/namespaces/proxy-383/pods/proxy-service-6lw44-6r54z:1080/proxy/: testtest (200; 78.809097ms) +Apr 29 11:34:38.697: INFO: (10) /api/v1/namespaces/proxy-383/pods/proxy-service-6lw44-6r54z:162/proxy/: bar (200; 76.66028ms) +Apr 29 11:34:38.697: INFO: (10) /api/v1/namespaces/proxy-383/pods/http:proxy-service-6lw44-6r54z:160/proxy/: foo (200; 77.404324ms) +Apr 29 11:34:38.697: INFO: (10) /api/v1/namespaces/proxy-383/pods/http:proxy-service-6lw44-6r54z:1080/proxy/: t... (200; 78.880952ms) +Apr 29 11:34:38.698: INFO: (10) /api/v1/namespaces/proxy-383/pods/https:proxy-service-6lw44-6r54z:462/proxy/: tls qux (200; 79.344851ms) +Apr 29 11:34:38.700: INFO: (10) /api/v1/namespaces/proxy-383/services/http:proxy-service-6lw44:portname1/proxy/: foo (200; 81.959273ms) +Apr 29 11:34:38.702: INFO: (10) /api/v1/namespaces/proxy-383/services/https:proxy-service-6lw44:tlsportname1/proxy/: tls baz (200; 82.496515ms) +Apr 29 11:34:38.702: INFO: (10) /api/v1/namespaces/proxy-383/pods/proxy-service-6lw44-6r54z:160/proxy/: foo (200; 82.030863ms) +Apr 29 11:34:38.703: INFO: (10) /api/v1/namespaces/proxy-383/services/https:proxy-service-6lw44:tlsportname2/proxy/: tls qux (200; 85.034371ms) +Apr 29 11:34:38.703: INFO: (10) /api/v1/namespaces/proxy-383/services/proxy-service-6lw44:portname2/proxy/: bar (200; 82.938046ms) +Apr 29 11:34:38.704: INFO: (10) /api/v1/namespaces/proxy-383/services/proxy-service-6lw44:portname1/proxy/: foo (200; 83.632996ms) +Apr 29 11:34:38.704: INFO: (10) /api/v1/namespaces/proxy-383/services/http:proxy-service-6lw44:portname2/proxy/: bar (200; 85.596573ms) +Apr 29 11:34:38.738: INFO: (11) /api/v1/namespaces/proxy-383/pods/https:proxy-service-6lw44-6r54z:462/proxy/: tls qux (200; 32.947028ms) +Apr 29 11:34:38.745: INFO: (11) /api/v1/namespaces/proxy-383/pods/https:proxy-service-6lw44-6r54z:460/proxy/: tls baz (200; 40.824947ms) +Apr 29 11:34:38.749: INFO: (11) /api/v1/namespaces/proxy-383/services/http:proxy-service-6lw44:portname1/proxy/: foo (200; 43.106469ms) +Apr 29 11:34:38.752: INFO: (11) /api/v1/namespaces/proxy-383/services/https:proxy-service-6lw44:tlsportname1/proxy/: tls baz (200; 46.566367ms) +Apr 29 11:34:38.752: INFO: (11) /api/v1/namespaces/proxy-383/pods/http:proxy-service-6lw44-6r54z:1080/proxy/: t... (200; 47.36989ms) +Apr 29 11:34:38.752: INFO: (11) /api/v1/namespaces/proxy-383/pods/proxy-service-6lw44-6r54z/proxy/: test (200; 45.488093ms) +Apr 29 11:34:38.752: INFO: (11) /api/v1/namespaces/proxy-383/services/http:proxy-service-6lw44:portname2/proxy/: bar (200; 45.354017ms) +Apr 29 11:34:38.752: INFO: (11) /api/v1/namespaces/proxy-383/services/https:proxy-service-6lw44:tlsportname2/proxy/: tls qux (200; 45.506774ms) +Apr 29 11:34:38.752: INFO: (11) /api/v1/namespaces/proxy-383/services/proxy-service-6lw44:portname2/proxy/: bar (200; 46.221132ms) +Apr 29 11:34:38.752: INFO: (11) /api/v1/namespaces/proxy-383/pods/proxy-service-6lw44-6r54z:1080/proxy/: testtest (200; 45.938688ms) +Apr 29 11:34:38.805: INFO: (12) /api/v1/namespaces/proxy-383/pods/http:proxy-service-6lw44-6r54z:1080/proxy/: t... (200; 45.029792ms) +Apr 29 11:34:38.808: INFO: (12) /api/v1/namespaces/proxy-383/pods/proxy-service-6lw44-6r54z:160/proxy/: foo (200; 48.793195ms) +Apr 29 11:34:38.813: INFO: (12) /api/v1/namespaces/proxy-383/pods/proxy-service-6lw44-6r54z:1080/proxy/: testtesttest (200; 56.025333ms) +Apr 29 11:34:38.898: INFO: (13) /api/v1/namespaces/proxy-383/pods/http:proxy-service-6lw44-6r54z:1080/proxy/: t... (200; 55.148382ms) +Apr 29 11:34:38.905: INFO: (13) /api/v1/namespaces/proxy-383/services/http:proxy-service-6lw44:portname2/proxy/: bar (200; 62.684132ms) +Apr 29 11:34:38.905: INFO: (13) /api/v1/namespaces/proxy-383/pods/proxy-service-6lw44-6r54z:160/proxy/: foo (200; 63.090909ms) +Apr 29 11:34:38.905: INFO: (13) /api/v1/namespaces/proxy-383/services/https:proxy-service-6lw44:tlsportname1/proxy/: tls baz (200; 62.180512ms) +Apr 29 11:34:38.905: INFO: (13) /api/v1/namespaces/proxy-383/pods/http:proxy-service-6lw44-6r54z:160/proxy/: foo (200; 62.064921ms) +Apr 29 11:34:38.905: INFO: (13) /api/v1/namespaces/proxy-383/services/https:proxy-service-6lw44:tlsportname2/proxy/: tls qux (200; 62.885541ms) +Apr 29 11:34:38.908: INFO: (13) /api/v1/namespaces/proxy-383/services/http:proxy-service-6lw44:portname1/proxy/: foo (200; 65.85456ms) +Apr 29 11:34:38.928: INFO: (14) /api/v1/namespaces/proxy-383/pods/http:proxy-service-6lw44-6r54z:162/proxy/: bar (200; 19.214313ms) +Apr 29 11:34:38.929: INFO: (14) /api/v1/namespaces/proxy-383/pods/http:proxy-service-6lw44-6r54z:1080/proxy/: t... (200; 20.925171ms) +Apr 29 11:34:38.934: INFO: (14) /api/v1/namespaces/proxy-383/pods/proxy-service-6lw44-6r54z/proxy/: test (200; 25.400906ms) +Apr 29 11:34:38.950: INFO: (14) /api/v1/namespaces/proxy-383/services/https:proxy-service-6lw44:tlsportname1/proxy/: tls baz (200; 41.249706ms) +Apr 29 11:34:38.953: INFO: (14) /api/v1/namespaces/proxy-383/pods/https:proxy-service-6lw44-6r54z:443/proxy/: testtestt... (200; 16.658862ms) +Apr 29 11:34:39.013: INFO: (15) /api/v1/namespaces/proxy-383/pods/https:proxy-service-6lw44-6r54z:462/proxy/: tls qux (200; 19.167995ms) +Apr 29 11:34:39.017: INFO: (15) /api/v1/namespaces/proxy-383/pods/http:proxy-service-6lw44-6r54z:162/proxy/: bar (200; 22.290635ms) +Apr 29 11:34:39.030: INFO: (15) /api/v1/namespaces/proxy-383/services/https:proxy-service-6lw44:tlsportname2/proxy/: tls qux (200; 35.608711ms) +Apr 29 11:34:39.032: INFO: (15) /api/v1/namespaces/proxy-383/services/https:proxy-service-6lw44:tlsportname1/proxy/: tls baz (200; 37.018069ms) +Apr 29 11:34:39.058: INFO: (15) /api/v1/namespaces/proxy-383/pods/https:proxy-service-6lw44-6r54z:460/proxy/: tls baz (200; 62.845544ms) +Apr 29 11:34:39.059: INFO: (15) /api/v1/namespaces/proxy-383/pods/http:proxy-service-6lw44-6r54z:160/proxy/: foo (200; 64.792259ms) +Apr 29 11:34:39.059: INFO: (15) /api/v1/namespaces/proxy-383/services/http:proxy-service-6lw44:portname1/proxy/: foo (200; 64.705595ms) +Apr 29 11:34:39.059: INFO: (15) /api/v1/namespaces/proxy-383/pods/proxy-service-6lw44-6r54z:162/proxy/: bar (200; 63.871948ms) +Apr 29 11:34:39.059: INFO: (15) /api/v1/namespaces/proxy-383/pods/proxy-service-6lw44-6r54z:160/proxy/: foo (200; 63.787846ms) +Apr 29 11:34:39.060: INFO: (15) /api/v1/namespaces/proxy-383/pods/proxy-service-6lw44-6r54z/proxy/: test (200; 64.332136ms) +Apr 29 11:34:39.060: INFO: (15) /api/v1/namespaces/proxy-383/services/proxy-service-6lw44:portname2/proxy/: bar (200; 64.863798ms) +Apr 29 11:34:39.061: INFO: (15) /api/v1/namespaces/proxy-383/services/http:proxy-service-6lw44:portname2/proxy/: bar (200; 66.63422ms) +Apr 29 11:34:39.061: INFO: (15) /api/v1/namespaces/proxy-383/pods/https:proxy-service-6lw44-6r54z:443/proxy/: t... (200; 32.219905ms) +Apr 29 11:34:39.100: INFO: (16) /api/v1/namespaces/proxy-383/services/https:proxy-service-6lw44:tlsportname1/proxy/: tls baz (200; 37.399212ms) +Apr 29 11:34:39.105: INFO: (16) /api/v1/namespaces/proxy-383/services/https:proxy-service-6lw44:tlsportname2/proxy/: tls qux (200; 42.036369ms) +Apr 29 11:34:39.106: INFO: (16) /api/v1/namespaces/proxy-383/pods/https:proxy-service-6lw44-6r54z:462/proxy/: tls qux (200; 43.225776ms) +Apr 29 11:34:39.130: INFO: (16) /api/v1/namespaces/proxy-383/pods/http:proxy-service-6lw44-6r54z:162/proxy/: bar (200; 67.16714ms) +Apr 29 11:34:39.130: INFO: (16) /api/v1/namespaces/proxy-383/pods/https:proxy-service-6lw44-6r54z:443/proxy/: testtest (200; 82.343405ms) +Apr 29 11:34:39.147: INFO: (16) /api/v1/namespaces/proxy-383/pods/proxy-service-6lw44-6r54z:162/proxy/: bar (200; 84.365274ms) +Apr 29 11:34:39.148: INFO: (16) /api/v1/namespaces/proxy-383/services/http:proxy-service-6lw44:portname1/proxy/: foo (200; 85.007124ms) +Apr 29 11:34:39.148: INFO: (16) /api/v1/namespaces/proxy-383/services/proxy-service-6lw44:portname2/proxy/: bar (200; 85.306138ms) +Apr 29 11:34:39.149: INFO: (16) /api/v1/namespaces/proxy-383/pods/http:proxy-service-6lw44-6r54z:160/proxy/: foo (200; 86.400286ms) +Apr 29 11:34:39.176: INFO: (17) /api/v1/namespaces/proxy-383/services/proxy-service-6lw44:portname1/proxy/: foo (200; 26.785974ms) +Apr 29 11:34:39.176: INFO: (17) /api/v1/namespaces/proxy-383/pods/http:proxy-service-6lw44-6r54z:1080/proxy/: t... (200; 26.413647ms) +Apr 29 11:34:39.176: INFO: (17) /api/v1/namespaces/proxy-383/pods/https:proxy-service-6lw44-6r54z:460/proxy/: tls baz (200; 26.258874ms) +Apr 29 11:34:39.177: INFO: (17) /api/v1/namespaces/proxy-383/pods/https:proxy-service-6lw44-6r54z:462/proxy/: tls qux (200; 27.047903ms) +Apr 29 11:34:39.177: INFO: (17) /api/v1/namespaces/proxy-383/pods/https:proxy-service-6lw44-6r54z:443/proxy/: testtest (200; 32.124908ms) +Apr 29 11:34:39.183: INFO: (17) /api/v1/namespaces/proxy-383/services/http:proxy-service-6lw44:portname2/proxy/: bar (200; 33.992959ms) +Apr 29 11:34:39.185: INFO: (17) /api/v1/namespaces/proxy-383/services/http:proxy-service-6lw44:portname1/proxy/: foo (200; 34.606761ms) +Apr 29 11:34:39.185: INFO: (17) /api/v1/namespaces/proxy-383/services/https:proxy-service-6lw44:tlsportname2/proxy/: tls qux (200; 35.234658ms) +Apr 29 11:34:39.185: INFO: (17) /api/v1/namespaces/proxy-383/services/https:proxy-service-6lw44:tlsportname1/proxy/: tls baz (200; 35.194883ms) +Apr 29 11:34:39.186: INFO: (17) /api/v1/namespaces/proxy-383/services/proxy-service-6lw44:portname2/proxy/: bar (200; 35.798903ms) +Apr 29 11:34:39.201: INFO: (18) /api/v1/namespaces/proxy-383/pods/proxy-service-6lw44-6r54z:160/proxy/: foo (200; 15.245844ms) +Apr 29 11:34:39.205: INFO: (18) /api/v1/namespaces/proxy-383/pods/proxy-service-6lw44-6r54z:1080/proxy/: testt... (200; 18.302834ms) +Apr 29 11:34:39.206: INFO: (18) /api/v1/namespaces/proxy-383/pods/proxy-service-6lw44-6r54z/proxy/: test (200; 19.287256ms) +Apr 29 11:34:39.206: INFO: (18) /api/v1/namespaces/proxy-383/pods/proxy-service-6lw44-6r54z:162/proxy/: bar (200; 19.37181ms) +Apr 29 11:34:39.206: INFO: (18) /api/v1/namespaces/proxy-383/pods/http:proxy-service-6lw44-6r54z:162/proxy/: bar (200; 19.4433ms) +Apr 29 11:34:39.207: INFO: (18) /api/v1/namespaces/proxy-383/pods/https:proxy-service-6lw44-6r54z:460/proxy/: tls baz (200; 20.198895ms) +Apr 29 11:34:39.207: INFO: (18) /api/v1/namespaces/proxy-383/pods/https:proxy-service-6lw44-6r54z:443/proxy/: test (200; 18.455412ms) +Apr 29 11:34:39.237: INFO: (19) /api/v1/namespaces/proxy-383/pods/https:proxy-service-6lw44-6r54z:462/proxy/: tls qux (200; 19.867524ms) +Apr 29 11:34:39.237: INFO: (19) /api/v1/namespaces/proxy-383/pods/https:proxy-service-6lw44-6r54z:460/proxy/: tls baz (200; 19.782581ms) +Apr 29 11:34:39.237: INFO: (19) /api/v1/namespaces/proxy-383/pods/https:proxy-service-6lw44-6r54z:443/proxy/: t... (200; 22.66222ms) +Apr 29 11:34:39.240: INFO: (19) /api/v1/namespaces/proxy-383/pods/proxy-service-6lw44-6r54z:1080/proxy/: test>> kubeConfig: /tmp/kubeconfig-185508174 +STEP: Building a namespace api object, basename pod-network-test +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in pod-network-test-4454 +STEP: Waiting for a default service account to be provisioned in namespace +[It] should function for node-pod communication: http [LinuxOnly] [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: Performing setup for networking test in namespace pod-network-test-4454 +STEP: creating a selector +STEP: Creating the service pods in kubernetes +Apr 29 11:34:50.853: INFO: Waiting up to 10m0s for all (but 0) nodes to be schedulable +STEP: Creating test pods +Apr 29 11:35:21.109: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s --max-time 15 --connect-timeout 1 http://192.168.148.77:8080/hostName | grep -v '^\s*$'] Namespace:pod-network-test-4454 PodName:host-test-container-pod ContainerName:hostexec Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} +Apr 29 11:35:21.110: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174 +Apr 29 11:35:21.319: INFO: Found all expected endpoints: [netserver-0] +Apr 29 11:35:21.323: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s --max-time 15 --connect-timeout 1 http://192.168.205.78:8080/hostName | grep -v '^\s*$'] Namespace:pod-network-test-4454 PodName:host-test-container-pod ContainerName:hostexec Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} +Apr 29 11:35:21.323: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174 +Apr 29 11:35:21.496: INFO: Found all expected endpoints: [netserver-1] +Apr 29 11:35:21.507: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s --max-time 15 --connect-timeout 1 http://192.168.14.8:8080/hostName | grep -v '^\s*$'] Namespace:pod-network-test-4454 PodName:host-test-container-pod ContainerName:hostexec Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} +Apr 29 11:35:21.507: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174 +Apr 29 11:35:21.698: INFO: Found all expected endpoints: [netserver-2] +[AfterEach] [sig-network] Networking + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Apr 29 11:35:21.699: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "pod-network-test-4454" for this suite. +Apr 29 11:35:45.728: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Apr 29 11:35:45.975: INFO: namespace pod-network-test-4454 deletion completed in 24.266258635s + +• [SLOW TEST:55.341 seconds] +[sig-network] Networking +/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:25 + Granular Checks: Pods + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:28 + should function for node-pod communication: http [LinuxOnly] [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +------------------------------ +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +------------------------------ +[sig-storage] ConfigMap + should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [sig-storage] ConfigMap + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +STEP: Creating a kubernetes client +Apr 29 11:35:45.975: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174 +STEP: Building a namespace api object, basename configmap +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in configmap-6593 +STEP: Waiting for a default service account to be provisioned in namespace +[It] should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: Creating configMap with name configmap-test-volume-ede7bb19-6a72-11e9-b6ef-0e71f8c7e888 +STEP: Creating a pod to test consume configMaps +Apr 29 11:35:46.197: INFO: Waiting up to 5m0s for pod "pod-configmaps-ede93592-6a72-11e9-b6ef-0e71f8c7e888" in namespace "configmap-6593" to be "success or failure" +Apr 29 11:35:46.203: INFO: Pod "pod-configmaps-ede93592-6a72-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 5.426306ms +Apr 29 11:35:48.216: INFO: Pod "pod-configmaps-ede93592-6a72-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 2.019229447s +Apr 29 11:35:50.223: INFO: Pod "pod-configmaps-ede93592-6a72-11e9-b6ef-0e71f8c7e888": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.025865609s +STEP: Saw pod success +Apr 29 11:35:50.223: INFO: Pod "pod-configmaps-ede93592-6a72-11e9-b6ef-0e71f8c7e888" satisfied condition "success or failure" +Apr 29 11:35:50.229: INFO: Trying to get logs from node worker-4av2h-6665d7b5ff-9nc42 pod pod-configmaps-ede93592-6a72-11e9-b6ef-0e71f8c7e888 container configmap-volume-test: +STEP: delete the pod +Apr 29 11:35:50.262: INFO: Waiting for pod pod-configmaps-ede93592-6a72-11e9-b6ef-0e71f8c7e888 to disappear +Apr 29 11:35:50.268: INFO: Pod pod-configmaps-ede93592-6a72-11e9-b6ef-0e71f8c7e888 no longer exists +[AfterEach] [sig-storage] ConfigMap + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Apr 29 11:35:50.270: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "configmap-6593" for this suite. +Apr 29 11:35:56.307: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Apr 29 11:35:56.562: INFO: namespace configmap-6593 deletion completed in 6.284748642s + +• [SLOW TEST:10.587 seconds] +[sig-storage] ConfigMap +/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:32 + should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +------------------------------ +SSS +------------------------------ +[k8s.io] Pods + should support retrieving logs from the container over websockets [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [k8s.io] Pods + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +STEP: Creating a kubernetes client +Apr 29 11:35:56.563: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174 +STEP: Building a namespace api object, basename pods +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in pods-9117 +STEP: Waiting for a default service account to be provisioned in namespace +[BeforeEach] [k8s.io] Pods + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:135 +[It] should support retrieving logs from the container over websockets [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +Apr 29 11:35:56.747: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174 +STEP: creating the pod +STEP: submitting the pod to kubernetes +[AfterEach] [k8s.io] Pods + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Apr 29 11:36:02.822: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "pods-9117" for this suite. +Apr 29 11:36:42.856: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Apr 29 11:36:43.009: INFO: namespace pods-9117 deletion completed in 40.177014678s + +• [SLOW TEST:46.446 seconds] +[k8s.io] Pods +/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687 + should support retrieving logs from the container over websockets [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +------------------------------ +SSSSSSSSSSSSSSSSSSSSSSSS +------------------------------ +[k8s.io] Pods + should be updated [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [k8s.io] Pods + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +STEP: Creating a kubernetes client +Apr 29 11:36:43.021: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174 +STEP: Building a namespace api object, basename pods +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in pods-5315 +STEP: Waiting for a default service account to be provisioned in namespace +[BeforeEach] [k8s.io] Pods + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:135 +[It] should be updated [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: creating the pod +STEP: submitting the pod to kubernetes +STEP: verifying the pod is in kubernetes +STEP: updating the pod +Apr 29 11:36:51.836: INFO: Successfully updated pod "pod-update-0fed3004-6a73-11e9-b6ef-0e71f8c7e888" +STEP: verifying the updated pod is in kubernetes +Apr 29 11:36:51.857: INFO: Pod update OK +[AfterEach] [k8s.io] Pods + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Apr 29 11:36:51.857: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "pods-5315" for this suite. +Apr 29 11:37:15.889: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Apr 29 11:37:16.144: INFO: namespace pods-5315 deletion completed in 24.277156453s + +• [SLOW TEST:33.124 seconds] +[k8s.io] Pods +/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687 + should be updated [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +------------------------------ +SSSSSSSSSSSSSSSSSSSSSSSS +------------------------------ +[sig-cli] Kubectl client [k8s.io] Guestbook application + should create and stop a working application [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [sig-cli] Kubectl client + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +STEP: Creating a kubernetes client +Apr 29 11:37:16.161: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174 +STEP: Building a namespace api object, basename kubectl +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-6570 +STEP: Waiting for a default service account to be provisioned in namespace +[BeforeEach] [sig-cli] Kubectl client + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:213 +[It] should create and stop a working application [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: creating all guestbook components +Apr 29 11:37:16.407: INFO: apiVersion: v1 +kind: Service +metadata: + name: redis-slave + labels: + app: redis + role: slave + tier: backend +spec: + ports: + - port: 6379 + selector: + app: redis + role: slave + tier: backend + +Apr 29 11:37:16.407: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 create -f - --namespace=kubectl-6570' +Apr 29 11:37:18.654: INFO: stderr: "" +Apr 29 11:37:18.654: INFO: stdout: "service/redis-slave created\n" +Apr 29 11:37:18.654: INFO: apiVersion: v1 +kind: Service +metadata: + name: redis-master + labels: + app: redis + role: master + tier: backend +spec: + ports: + - port: 6379 + targetPort: 6379 + selector: + app: redis + role: master + tier: backend + +Apr 29 11:37:18.654: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 create -f - --namespace=kubectl-6570' +Apr 29 11:37:19.315: INFO: stderr: "" +Apr 29 11:37:19.315: INFO: stdout: "service/redis-master created\n" +Apr 29 11:37:19.316: INFO: apiVersion: v1 +kind: Service +metadata: + name: frontend + labels: + app: guestbook + tier: frontend +spec: + # if your cluster supports it, uncomment the following to automatically create + # an external load-balanced IP for the frontend service. + # type: LoadBalancer + ports: + - port: 80 + selector: + app: guestbook + tier: frontend + +Apr 29 11:37:19.316: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 create -f - --namespace=kubectl-6570' +Apr 29 11:37:19.655: INFO: stderr: "" +Apr 29 11:37:19.655: INFO: stdout: "service/frontend created\n" +Apr 29 11:37:19.655: INFO: apiVersion: apps/v1 +kind: Deployment +metadata: + name: frontend +spec: + replicas: 3 + selector: + matchLabels: + app: guestbook + tier: frontend + template: + metadata: + labels: + app: guestbook + tier: frontend + spec: + containers: + - name: php-redis + image: gcr.io/google-samples/gb-frontend:v6 + resources: + requests: + cpu: 100m + memory: 100Mi + env: + - name: GET_HOSTS_FROM + value: dns + # If your cluster config does not include a dns service, then to + # instead access environment variables to find service host + # info, comment out the 'value: dns' line above, and uncomment the + # line below: + # value: env + ports: + - containerPort: 80 + +Apr 29 11:37:19.655: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 create -f - --namespace=kubectl-6570' +Apr 29 11:37:19.996: INFO: stderr: "" +Apr 29 11:37:19.996: INFO: stdout: "deployment.apps/frontend created\n" +Apr 29 11:37:19.997: INFO: apiVersion: apps/v1 +kind: Deployment +metadata: + name: redis-master +spec: + replicas: 1 + selector: + matchLabels: + app: redis + role: master + tier: backend + template: + metadata: + labels: + app: redis + role: master + tier: backend + spec: + containers: + - name: master + image: gcr.io/kubernetes-e2e-test-images/redis:1.0 + resources: + requests: + cpu: 100m + memory: 100Mi + ports: + - containerPort: 6379 + +Apr 29 11:37:19.997: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 create -f - --namespace=kubectl-6570' +Apr 29 11:37:20.515: INFO: stderr: "" +Apr 29 11:37:20.516: INFO: stdout: "deployment.apps/redis-master created\n" +Apr 29 11:37:20.516: INFO: apiVersion: apps/v1 +kind: Deployment +metadata: + name: redis-slave +spec: + replicas: 2 + selector: + matchLabels: + app: redis + role: slave + tier: backend + template: + metadata: + labels: + app: redis + role: slave + tier: backend + spec: + containers: + - name: slave + image: gcr.io/google-samples/gb-redisslave:v3 + resources: + requests: + cpu: 100m + memory: 100Mi + env: + - name: GET_HOSTS_FROM + value: dns + # If your cluster config does not include a dns service, then to + # instead access an environment variable to find the master + # service's host, comment out the 'value: dns' line above, and + # uncomment the line below: + # value: env + ports: + - containerPort: 6379 + +Apr 29 11:37:20.516: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 create -f - --namespace=kubectl-6570' +Apr 29 11:37:20.893: INFO: stderr: "" +Apr 29 11:37:20.893: INFO: stdout: "deployment.apps/redis-slave created\n" +STEP: validating guestbook app +Apr 29 11:37:20.893: INFO: Waiting for all frontend pods to be Running. +Apr 29 11:37:55.946: INFO: Waiting for frontend to serve content. +Apr 29 11:37:56.011: INFO: Trying to add a new entry to the guestbook. +Apr 29 11:37:56.054: INFO: Verifying that added entry can be retrieved. +STEP: using delete to clean up resources +Apr 29 11:37:56.077: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 delete --grace-period=0 --force -f - --namespace=kubectl-6570' +Apr 29 11:37:56.280: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" +Apr 29 11:37:56.280: INFO: stdout: "service \"redis-slave\" force deleted\n" +STEP: using delete to clean up resources +Apr 29 11:37:56.280: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 delete --grace-period=0 --force -f - --namespace=kubectl-6570' +Apr 29 11:37:56.469: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" +Apr 29 11:37:56.470: INFO: stdout: "service \"redis-master\" force deleted\n" +STEP: using delete to clean up resources +Apr 29 11:37:56.470: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 delete --grace-period=0 --force -f - --namespace=kubectl-6570' +Apr 29 11:37:56.647: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" +Apr 29 11:37:56.647: INFO: stdout: "service \"frontend\" force deleted\n" +STEP: using delete to clean up resources +Apr 29 11:37:56.648: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 delete --grace-period=0 --force -f - --namespace=kubectl-6570' +Apr 29 11:37:56.799: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" +Apr 29 11:37:56.799: INFO: stdout: "deployment.apps \"frontend\" force deleted\n" +STEP: using delete to clean up resources +Apr 29 11:37:56.799: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 delete --grace-period=0 --force -f - --namespace=kubectl-6570' +Apr 29 11:37:56.978: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" +Apr 29 11:37:56.978: INFO: stdout: "deployment.apps \"redis-master\" force deleted\n" +STEP: using delete to clean up resources +Apr 29 11:37:56.978: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 delete --grace-period=0 --force -f - --namespace=kubectl-6570' +Apr 29 11:37:57.173: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" +Apr 29 11:37:57.174: INFO: stdout: "deployment.apps \"redis-slave\" force deleted\n" +[AfterEach] [sig-cli] Kubectl client + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Apr 29 11:37:57.174: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "kubectl-6570" for this suite. +Apr 29 11:38:37.314: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Apr 29 11:38:37.516: INFO: namespace kubectl-6570 deletion completed in 40.292223475s + +• [SLOW TEST:81.356 seconds] +[sig-cli] Kubectl client +/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 + [k8s.io] Guestbook application + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687 + should create and stop a working application [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +------------------------------ +SSSSSSSSSSSSSSSSSSSSSSS +------------------------------ +[k8s.io] Probing container + should have monotonically increasing restart count [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [k8s.io] Probing container + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +STEP: Creating a kubernetes client +Apr 29 11:38:37.520: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174 +STEP: Building a namespace api object, basename container-probe +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in container-probe-6989 +STEP: Waiting for a default service account to be provisioned in namespace +[BeforeEach] [k8s.io] Probing container + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/container_probe.go:51 +[It] should have monotonically increasing restart count [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: Creating pod liveness-http in namespace container-probe-6989 +Apr 29 11:38:43.735: INFO: Started pod liveness-http in namespace container-probe-6989 +STEP: checking the pod's current state and verifying that restartCount is present +Apr 29 11:38:43.740: INFO: Initial restart count of pod liveness-http is 0 +Apr 29 11:39:01.809: INFO: Restart count of pod container-probe-6989/liveness-http is now 1 (18.068782706s elapsed) +Apr 29 11:39:21.904: INFO: Restart count of pod container-probe-6989/liveness-http is now 2 (38.16417124s elapsed) +Apr 29 11:39:42.018: INFO: Restart count of pod container-probe-6989/liveness-http is now 3 (58.278655644s elapsed) +Apr 29 11:40:00.107: INFO: Restart count of pod container-probe-6989/liveness-http is now 4 (1m16.366795594s elapsed) +Apr 29 11:41:10.370: INFO: Restart count of pod container-probe-6989/liveness-http is now 5 (2m26.630431396s elapsed) +STEP: deleting the pod +[AfterEach] [k8s.io] Probing container + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Apr 29 11:41:10.393: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "container-probe-6989" for this suite. +Apr 29 11:41:16.426: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Apr 29 11:41:16.617: INFO: namespace container-probe-6989 deletion completed in 6.211864905s + +• [SLOW TEST:159.097 seconds] +[k8s.io] Probing container +/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687 + should have monotonically increasing restart count [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +------------------------------ +SSSSSSS +------------------------------ +[sig-storage] Projected configMap + should be consumable from pods in volume as non-root [LinuxOnly] [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [sig-storage] Projected configMap + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +STEP: Creating a kubernetes client +Apr 29 11:41:16.619: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174 +STEP: Building a namespace api object, basename projected +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-6863 +STEP: Waiting for a default service account to be provisioned in namespace +[It] should be consumable from pods in volume as non-root [LinuxOnly] [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: Creating configMap with name projected-configmap-test-volume-b2fcbd1f-6a73-11e9-b6ef-0e71f8c7e888 +STEP: Creating a pod to test consume configMaps +Apr 29 11:41:16.854: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-b2fe63bc-6a73-11e9-b6ef-0e71f8c7e888" in namespace "projected-6863" to be "success or failure" +Apr 29 11:41:16.882: INFO: Pod "pod-projected-configmaps-b2fe63bc-6a73-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 27.594772ms +Apr 29 11:41:18.888: INFO: Pod "pod-projected-configmaps-b2fe63bc-6a73-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 2.034204153s +Apr 29 11:41:20.897: INFO: Pod "pod-projected-configmaps-b2fe63bc-6a73-11e9-b6ef-0e71f8c7e888": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.043206204s +STEP: Saw pod success +Apr 29 11:41:20.898: INFO: Pod "pod-projected-configmaps-b2fe63bc-6a73-11e9-b6ef-0e71f8c7e888" satisfied condition "success or failure" +Apr 29 11:41:20.905: INFO: Trying to get logs from node worker-4av2h-6665d7b5ff-9nc42 pod pod-projected-configmaps-b2fe63bc-6a73-11e9-b6ef-0e71f8c7e888 container projected-configmap-volume-test: +STEP: delete the pod +Apr 29 11:41:20.953: INFO: Waiting for pod pod-projected-configmaps-b2fe63bc-6a73-11e9-b6ef-0e71f8c7e888 to disappear +Apr 29 11:41:20.960: INFO: Pod pod-projected-configmaps-b2fe63bc-6a73-11e9-b6ef-0e71f8c7e888 no longer exists +[AfterEach] [sig-storage] Projected configMap + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Apr 29 11:41:20.961: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "projected-6863" for this suite. +Apr 29 11:41:26.997: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Apr 29 11:41:27.257: INFO: namespace projected-6863 deletion completed in 6.282381659s + +• [SLOW TEST:10.638 seconds] +[sig-storage] Projected configMap +/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_configmap.go:33 + should be consumable from pods in volume as non-root [LinuxOnly] [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +------------------------------ +SSSSSSSSSS +------------------------------ +[sig-apps] Daemon set [Serial] + should update pod when spec was updated and update strategy is RollingUpdate [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [sig-apps] Daemon set [Serial] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +STEP: Creating a kubernetes client +Apr 29 11:41:27.261: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174 +STEP: Building a namespace api object, basename daemonsets +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in daemonsets-5459 +STEP: Waiting for a default service account to be provisioned in namespace +[BeforeEach] [sig-apps] Daemon set [Serial] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:102 +[It] should update pod when spec was updated and update strategy is RollingUpdate [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +Apr 29 11:41:27.518: INFO: Creating simple daemon set daemon-set +STEP: Check that daemon pods launch on every node of the cluster. +Apr 29 11:41:27.542: INFO: DaemonSet pods can't tolerate node master-h2h3g-655686fb8c-8kgdn with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node +Apr 29 11:41:27.551: INFO: Number of nodes with available pods: 0 +Apr 29 11:41:27.551: INFO: Node worker-3oo4n-86cbf586b7-9q98g is running more than one daemon pod +Apr 29 11:41:28.563: INFO: DaemonSet pods can't tolerate node master-h2h3g-655686fb8c-8kgdn with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node +Apr 29 11:41:28.569: INFO: Number of nodes with available pods: 0 +Apr 29 11:41:28.569: INFO: Node worker-3oo4n-86cbf586b7-9q98g is running more than one daemon pod +Apr 29 11:41:29.563: INFO: DaemonSet pods can't tolerate node master-h2h3g-655686fb8c-8kgdn with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node +Apr 29 11:41:29.574: INFO: Number of nodes with available pods: 0 +Apr 29 11:41:29.575: INFO: Node worker-3oo4n-86cbf586b7-9q98g is running more than one daemon pod +Apr 29 11:41:30.564: INFO: DaemonSet pods can't tolerate node master-h2h3g-655686fb8c-8kgdn with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node +Apr 29 11:41:30.573: INFO: Number of nodes with available pods: 1 +Apr 29 11:41:30.573: INFO: Node worker-3oo4n-86cbf586b7-9q98g is running more than one daemon pod +Apr 29 11:41:31.564: INFO: DaemonSet pods can't tolerate node master-h2h3g-655686fb8c-8kgdn with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node +Apr 29 11:41:31.573: INFO: Number of nodes with available pods: 1 +Apr 29 11:41:31.573: INFO: Node worker-3oo4n-86cbf586b7-9q98g is running more than one daemon pod +Apr 29 11:41:32.559: INFO: DaemonSet pods can't tolerate node master-h2h3g-655686fb8c-8kgdn with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node +Apr 29 11:41:32.564: INFO: Number of nodes with available pods: 1 +Apr 29 11:41:32.564: INFO: Node worker-3oo4n-86cbf586b7-9q98g is running more than one daemon pod +Apr 29 11:41:33.559: INFO: DaemonSet pods can't tolerate node master-h2h3g-655686fb8c-8kgdn with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node +Apr 29 11:41:33.564: INFO: Number of nodes with available pods: 1 +Apr 29 11:41:33.564: INFO: Node worker-3oo4n-86cbf586b7-9q98g is running more than one daemon pod +Apr 29 11:41:34.560: INFO: DaemonSet pods can't tolerate node master-h2h3g-655686fb8c-8kgdn with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node +Apr 29 11:41:34.567: INFO: Number of nodes with available pods: 1 +Apr 29 11:41:34.568: INFO: Node worker-3oo4n-86cbf586b7-9q98g is running more than one daemon pod +Apr 29 11:41:35.562: INFO: DaemonSet pods can't tolerate node master-h2h3g-655686fb8c-8kgdn with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node +Apr 29 11:41:35.571: INFO: Number of nodes with available pods: 3 +Apr 29 11:41:35.571: INFO: Number of running nodes: 3, number of available pods: 3 +STEP: Update daemon pods image. +STEP: Check that daemon pods images are updated. +Apr 29 11:41:35.630: INFO: Wrong image for pod: daemon-set-dj57v. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Apr 29 11:41:35.630: INFO: Wrong image for pod: daemon-set-ld6w2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Apr 29 11:41:35.630: INFO: Wrong image for pod: daemon-set-nhkqz. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Apr 29 11:41:35.646: INFO: DaemonSet pods can't tolerate node master-h2h3g-655686fb8c-8kgdn with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node +Apr 29 11:41:36.653: INFO: Wrong image for pod: daemon-set-dj57v. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Apr 29 11:41:36.653: INFO: Wrong image for pod: daemon-set-ld6w2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Apr 29 11:41:36.653: INFO: Wrong image for pod: daemon-set-nhkqz. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Apr 29 11:41:36.663: INFO: DaemonSet pods can't tolerate node master-h2h3g-655686fb8c-8kgdn with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node +Apr 29 11:41:37.653: INFO: Wrong image for pod: daemon-set-dj57v. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Apr 29 11:41:37.653: INFO: Wrong image for pod: daemon-set-ld6w2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Apr 29 11:41:37.653: INFO: Wrong image for pod: daemon-set-nhkqz. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Apr 29 11:41:37.659: INFO: DaemonSet pods can't tolerate node master-h2h3g-655686fb8c-8kgdn with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node +Apr 29 11:41:38.657: INFO: Wrong image for pod: daemon-set-dj57v. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Apr 29 11:41:38.657: INFO: Wrong image for pod: daemon-set-ld6w2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Apr 29 11:41:38.657: INFO: Pod daemon-set-ld6w2 is not available +Apr 29 11:41:38.657: INFO: Wrong image for pod: daemon-set-nhkqz. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Apr 29 11:41:38.662: INFO: DaemonSet pods can't tolerate node master-h2h3g-655686fb8c-8kgdn with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node +Apr 29 11:41:39.655: INFO: Wrong image for pod: daemon-set-dj57v. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Apr 29 11:41:39.655: INFO: Wrong image for pod: daemon-set-ld6w2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Apr 29 11:41:39.656: INFO: Pod daemon-set-ld6w2 is not available +Apr 29 11:41:39.656: INFO: Wrong image for pod: daemon-set-nhkqz. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Apr 29 11:41:39.664: INFO: DaemonSet pods can't tolerate node master-h2h3g-655686fb8c-8kgdn with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node +Apr 29 11:41:40.654: INFO: Wrong image for pod: daemon-set-dj57v. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Apr 29 11:41:40.654: INFO: Wrong image for pod: daemon-set-ld6w2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Apr 29 11:41:40.655: INFO: Pod daemon-set-ld6w2 is not available +Apr 29 11:41:40.655: INFO: Wrong image for pod: daemon-set-nhkqz. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Apr 29 11:41:40.661: INFO: DaemonSet pods can't tolerate node master-h2h3g-655686fb8c-8kgdn with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node +Apr 29 11:41:41.654: INFO: Wrong image for pod: daemon-set-dj57v. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Apr 29 11:41:41.654: INFO: Wrong image for pod: daemon-set-ld6w2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Apr 29 11:41:41.654: INFO: Pod daemon-set-ld6w2 is not available +Apr 29 11:41:41.654: INFO: Wrong image for pod: daemon-set-nhkqz. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Apr 29 11:41:41.661: INFO: DaemonSet pods can't tolerate node master-h2h3g-655686fb8c-8kgdn with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node +Apr 29 11:41:42.653: INFO: Wrong image for pod: daemon-set-dj57v. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Apr 29 11:41:42.653: INFO: Wrong image for pod: daemon-set-ld6w2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Apr 29 11:41:42.653: INFO: Pod daemon-set-ld6w2 is not available +Apr 29 11:41:42.653: INFO: Wrong image for pod: daemon-set-nhkqz. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Apr 29 11:41:42.659: INFO: DaemonSet pods can't tolerate node master-h2h3g-655686fb8c-8kgdn with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node +Apr 29 11:41:43.652: INFO: Wrong image for pod: daemon-set-dj57v. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Apr 29 11:41:43.652: INFO: Wrong image for pod: daemon-set-ld6w2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Apr 29 11:41:43.652: INFO: Pod daemon-set-ld6w2 is not available +Apr 29 11:41:43.652: INFO: Wrong image for pod: daemon-set-nhkqz. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Apr 29 11:41:43.658: INFO: DaemonSet pods can't tolerate node master-h2h3g-655686fb8c-8kgdn with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node +Apr 29 11:41:44.655: INFO: Wrong image for pod: daemon-set-dj57v. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Apr 29 11:41:44.655: INFO: Wrong image for pod: daemon-set-ld6w2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Apr 29 11:41:44.655: INFO: Pod daemon-set-ld6w2 is not available +Apr 29 11:41:44.655: INFO: Wrong image for pod: daemon-set-nhkqz. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Apr 29 11:41:44.661: INFO: DaemonSet pods can't tolerate node master-h2h3g-655686fb8c-8kgdn with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node +Apr 29 11:41:45.655: INFO: Wrong image for pod: daemon-set-dj57v. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Apr 29 11:41:45.655: INFO: Wrong image for pod: daemon-set-ld6w2. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Apr 29 11:41:45.655: INFO: Pod daemon-set-ld6w2 is not available +Apr 29 11:41:45.655: INFO: Wrong image for pod: daemon-set-nhkqz. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Apr 29 11:41:45.660: INFO: DaemonSet pods can't tolerate node master-h2h3g-655686fb8c-8kgdn with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node +Apr 29 11:41:46.654: INFO: Pod daemon-set-bthkv is not available +Apr 29 11:41:46.654: INFO: Wrong image for pod: daemon-set-dj57v. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Apr 29 11:41:46.655: INFO: Wrong image for pod: daemon-set-nhkqz. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Apr 29 11:41:46.662: INFO: DaemonSet pods can't tolerate node master-h2h3g-655686fb8c-8kgdn with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node +Apr 29 11:41:47.654: INFO: Pod daemon-set-bthkv is not available +Apr 29 11:41:47.654: INFO: Wrong image for pod: daemon-set-dj57v. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Apr 29 11:41:47.654: INFO: Wrong image for pod: daemon-set-nhkqz. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Apr 29 11:41:47.660: INFO: DaemonSet pods can't tolerate node master-h2h3g-655686fb8c-8kgdn with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node +Apr 29 11:41:48.653: INFO: Pod daemon-set-bthkv is not available +Apr 29 11:41:48.653: INFO: Wrong image for pod: daemon-set-dj57v. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Apr 29 11:41:48.653: INFO: Wrong image for pod: daemon-set-nhkqz. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Apr 29 11:41:48.660: INFO: DaemonSet pods can't tolerate node master-h2h3g-655686fb8c-8kgdn with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node +Apr 29 11:41:49.653: INFO: Pod daemon-set-bthkv is not available +Apr 29 11:41:49.653: INFO: Wrong image for pod: daemon-set-dj57v. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Apr 29 11:41:49.653: INFO: Wrong image for pod: daemon-set-nhkqz. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Apr 29 11:41:49.663: INFO: DaemonSet pods can't tolerate node master-h2h3g-655686fb8c-8kgdn with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node +Apr 29 11:41:50.654: INFO: Wrong image for pod: daemon-set-dj57v. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Apr 29 11:41:50.654: INFO: Wrong image for pod: daemon-set-nhkqz. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Apr 29 11:41:50.661: INFO: DaemonSet pods can't tolerate node master-h2h3g-655686fb8c-8kgdn with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node +Apr 29 11:41:51.654: INFO: Wrong image for pod: daemon-set-dj57v. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Apr 29 11:41:51.654: INFO: Pod daemon-set-dj57v is not available +Apr 29 11:41:51.654: INFO: Wrong image for pod: daemon-set-nhkqz. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Apr 29 11:41:51.660: INFO: DaemonSet pods can't tolerate node master-h2h3g-655686fb8c-8kgdn with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node +Apr 29 11:41:52.653: INFO: Wrong image for pod: daemon-set-nhkqz. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Apr 29 11:41:52.653: INFO: Pod daemon-set-pr5ls is not available +Apr 29 11:41:52.660: INFO: DaemonSet pods can't tolerate node master-h2h3g-655686fb8c-8kgdn with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node +Apr 29 11:41:53.654: INFO: Wrong image for pod: daemon-set-nhkqz. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Apr 29 11:41:53.654: INFO: Pod daemon-set-pr5ls is not available +Apr 29 11:41:53.661: INFO: DaemonSet pods can't tolerate node master-h2h3g-655686fb8c-8kgdn with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node +Apr 29 11:41:54.654: INFO: Wrong image for pod: daemon-set-nhkqz. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Apr 29 11:41:54.654: INFO: Pod daemon-set-pr5ls is not available +Apr 29 11:41:54.660: INFO: DaemonSet pods can't tolerate node master-h2h3g-655686fb8c-8kgdn with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node +Apr 29 11:41:55.654: INFO: Wrong image for pod: daemon-set-nhkqz. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Apr 29 11:41:55.654: INFO: Pod daemon-set-pr5ls is not available +Apr 29 11:41:55.659: INFO: DaemonSet pods can't tolerate node master-h2h3g-655686fb8c-8kgdn with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node +Apr 29 11:41:56.707: INFO: Wrong image for pod: daemon-set-nhkqz. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Apr 29 11:41:56.707: INFO: Pod daemon-set-pr5ls is not available +Apr 29 11:41:56.720: INFO: DaemonSet pods can't tolerate node master-h2h3g-655686fb8c-8kgdn with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node +Apr 29 11:41:57.654: INFO: Wrong image for pod: daemon-set-nhkqz. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Apr 29 11:41:57.661: INFO: DaemonSet pods can't tolerate node master-h2h3g-655686fb8c-8kgdn with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node +Apr 29 11:41:58.653: INFO: Wrong image for pod: daemon-set-nhkqz. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Apr 29 11:41:58.662: INFO: DaemonSet pods can't tolerate node master-h2h3g-655686fb8c-8kgdn with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node +Apr 29 11:41:59.654: INFO: Wrong image for pod: daemon-set-nhkqz. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Apr 29 11:41:59.654: INFO: Pod daemon-set-nhkqz is not available +Apr 29 11:41:59.662: INFO: DaemonSet pods can't tolerate node master-h2h3g-655686fb8c-8kgdn with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node +Apr 29 11:42:00.656: INFO: Wrong image for pod: daemon-set-nhkqz. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Apr 29 11:42:00.656: INFO: Pod daemon-set-nhkqz is not available +Apr 29 11:42:00.663: INFO: DaemonSet pods can't tolerate node master-h2h3g-655686fb8c-8kgdn with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node +Apr 29 11:42:01.654: INFO: Wrong image for pod: daemon-set-nhkqz. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Apr 29 11:42:01.654: INFO: Pod daemon-set-nhkqz is not available +Apr 29 11:42:01.660: INFO: DaemonSet pods can't tolerate node master-h2h3g-655686fb8c-8kgdn with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node +Apr 29 11:42:02.652: INFO: Wrong image for pod: daemon-set-nhkqz. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Apr 29 11:42:02.653: INFO: Pod daemon-set-nhkqz is not available +Apr 29 11:42:02.657: INFO: DaemonSet pods can't tolerate node master-h2h3g-655686fb8c-8kgdn with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node +Apr 29 11:42:03.653: INFO: Wrong image for pod: daemon-set-nhkqz. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. +Apr 29 11:42:03.653: INFO: Pod daemon-set-nhkqz is not available +Apr 29 11:42:03.658: INFO: DaemonSet pods can't tolerate node master-h2h3g-655686fb8c-8kgdn with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node +Apr 29 11:42:04.655: INFO: Pod daemon-set-4qzxr is not available +Apr 29 11:42:04.660: INFO: DaemonSet pods can't tolerate node master-h2h3g-655686fb8c-8kgdn with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node +STEP: Check that daemon pods are still running on every node of the cluster. +Apr 29 11:42:04.666: INFO: DaemonSet pods can't tolerate node master-h2h3g-655686fb8c-8kgdn with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node +Apr 29 11:42:04.671: INFO: Number of nodes with available pods: 2 +Apr 29 11:42:04.671: INFO: Node worker-4av2h-6665d7b5ff-9nc42 is running more than one daemon pod +Apr 29 11:42:05.682: INFO: DaemonSet pods can't tolerate node master-h2h3g-655686fb8c-8kgdn with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node +Apr 29 11:42:05.688: INFO: Number of nodes with available pods: 2 +Apr 29 11:42:05.688: INFO: Node worker-4av2h-6665d7b5ff-9nc42 is running more than one daemon pod +Apr 29 11:42:06.684: INFO: DaemonSet pods can't tolerate node master-h2h3g-655686fb8c-8kgdn with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node +Apr 29 11:42:06.690: INFO: Number of nodes with available pods: 3 +Apr 29 11:42:06.690: INFO: Number of running nodes: 3, number of available pods: 3 +[AfterEach] [sig-apps] Daemon set [Serial] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:68 +STEP: Deleting DaemonSet "daemon-set" +STEP: deleting DaemonSet.extensions daemon-set in namespace daemonsets-5459, will wait for the garbage collector to delete the pods +Apr 29 11:42:06.787: INFO: Deleting DaemonSet.extensions daemon-set took: 11.472272ms +Apr 29 11:42:06.888: INFO: Terminating DaemonSet.extensions daemon-set pods took: 100.540949ms +Apr 29 11:42:14.394: INFO: Number of nodes with available pods: 0 +Apr 29 11:42:14.394: INFO: Number of running nodes: 0, number of available pods: 0 +Apr 29 11:42:14.399: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"selfLink":"/apis/apps/v1/namespaces/daemonsets-5459/daemonsets","resourceVersion":"5888"},"items":null} + +Apr 29 11:42:14.404: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"selfLink":"/api/v1/namespaces/daemonsets-5459/pods","resourceVersion":"5888"},"items":null} + +[AfterEach] [sig-apps] Daemon set [Serial] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Apr 29 11:42:14.432: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "daemonsets-5459" for this suite. +Apr 29 11:42:22.470: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Apr 29 11:42:22.674: INFO: namespace daemonsets-5459 deletion completed in 8.237059845s + +• [SLOW TEST:55.414 seconds] +[sig-apps] Daemon set [Serial] +/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22 + should update pod when spec was updated and update strategy is RollingUpdate [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +------------------------------ +SSSS +------------------------------ +[sig-node] ConfigMap + should be consumable via the environment [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [sig-node] ConfigMap + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +STEP: Creating a kubernetes client +Apr 29 11:42:22.676: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174 +STEP: Building a namespace api object, basename configmap +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in configmap-9954 +STEP: Waiting for a default service account to be provisioned in namespace +[It] should be consumable via the environment [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: Creating configMap configmap-9954/configmap-test-da58ab7d-6a73-11e9-b6ef-0e71f8c7e888 +STEP: Creating a pod to test consume configMaps +Apr 29 11:42:22.876: INFO: Waiting up to 5m0s for pod "pod-configmaps-da598a99-6a73-11e9-b6ef-0e71f8c7e888" in namespace "configmap-9954" to be "success or failure" +Apr 29 11:42:22.890: INFO: Pod "pod-configmaps-da598a99-6a73-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 13.949412ms +Apr 29 11:42:24.897: INFO: Pod "pod-configmaps-da598a99-6a73-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 2.020740107s +Apr 29 11:42:26.903: INFO: Pod "pod-configmaps-da598a99-6a73-11e9-b6ef-0e71f8c7e888": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.027193804s +STEP: Saw pod success +Apr 29 11:42:26.903: INFO: Pod "pod-configmaps-da598a99-6a73-11e9-b6ef-0e71f8c7e888" satisfied condition "success or failure" +Apr 29 11:42:26.911: INFO: Trying to get logs from node worker-4av2h-6665d7b5ff-9nc42 pod pod-configmaps-da598a99-6a73-11e9-b6ef-0e71f8c7e888 container env-test: +STEP: delete the pod +Apr 29 11:42:26.948: INFO: Waiting for pod pod-configmaps-da598a99-6a73-11e9-b6ef-0e71f8c7e888 to disappear +Apr 29 11:42:26.953: INFO: Pod pod-configmaps-da598a99-6a73-11e9-b6ef-0e71f8c7e888 no longer exists +[AfterEach] [sig-node] ConfigMap + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Apr 29 11:42:26.953: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "configmap-9954" for this suite. +Apr 29 11:42:32.987: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Apr 29 11:42:33.224: INFO: namespace configmap-9954 deletion completed in 6.262206827s + +• [SLOW TEST:10.548 seconds] +[sig-node] ConfigMap +/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/configmap.go:32 + should be consumable via the environment [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +------------------------------ +SSSSSS +------------------------------ +[sig-apps] Daemon set [Serial] + should run and stop complex daemon [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [sig-apps] Daemon set [Serial] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +STEP: Creating a kubernetes client +Apr 29 11:42:33.226: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174 +STEP: Building a namespace api object, basename daemonsets +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in daemonsets-6765 +STEP: Waiting for a default service account to be provisioned in namespace +[BeforeEach] [sig-apps] Daemon set [Serial] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:102 +[It] should run and stop complex daemon [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +Apr 29 11:42:33.464: INFO: Creating daemon "daemon-set" with a node selector +STEP: Initially, daemon pods should not be running on any nodes. +Apr 29 11:42:33.476: INFO: Number of nodes with available pods: 0 +Apr 29 11:42:33.476: INFO: Number of running nodes: 0, number of available pods: 0 +STEP: Change node label to blue, check that daemon pod is launched. +Apr 29 11:42:33.505: INFO: Number of nodes with available pods: 0 +Apr 29 11:42:33.505: INFO: Node worker-3oo4n-86cbf586b7-9q98g is running more than one daemon pod +Apr 29 11:42:34.511: INFO: Number of nodes with available pods: 0 +Apr 29 11:42:34.511: INFO: Node worker-3oo4n-86cbf586b7-9q98g is running more than one daemon pod +Apr 29 11:42:35.511: INFO: Number of nodes with available pods: 0 +Apr 29 11:42:35.511: INFO: Node worker-3oo4n-86cbf586b7-9q98g is running more than one daemon pod +Apr 29 11:42:36.513: INFO: Number of nodes with available pods: 1 +Apr 29 11:42:36.513: INFO: Number of running nodes: 1, number of available pods: 1 +STEP: Update the node label to green, and wait for daemons to be unscheduled +Apr 29 11:42:36.584: INFO: Number of nodes with available pods: 1 +Apr 29 11:42:36.584: INFO: Number of running nodes: 0, number of available pods: 1 +Apr 29 11:42:37.592: INFO: Number of nodes with available pods: 0 +Apr 29 11:42:37.592: INFO: Number of running nodes: 0, number of available pods: 0 +STEP: Update DaemonSet node selector to green, and change its update strategy to RollingUpdate +Apr 29 11:42:37.613: INFO: Number of nodes with available pods: 0 +Apr 29 11:42:37.613: INFO: Node worker-3oo4n-86cbf586b7-9q98g is running more than one daemon pod +Apr 29 11:42:38.621: INFO: Number of nodes with available pods: 0 +Apr 29 11:42:38.621: INFO: Node worker-3oo4n-86cbf586b7-9q98g is running more than one daemon pod +Apr 29 11:42:39.623: INFO: Number of nodes with available pods: 0 +Apr 29 11:42:39.623: INFO: Node worker-3oo4n-86cbf586b7-9q98g is running more than one daemon pod +Apr 29 11:42:40.623: INFO: Number of nodes with available pods: 0 +Apr 29 11:42:40.623: INFO: Node worker-3oo4n-86cbf586b7-9q98g is running more than one daemon pod +Apr 29 11:42:41.619: INFO: Number of nodes with available pods: 0 +Apr 29 11:42:41.619: INFO: Node worker-3oo4n-86cbf586b7-9q98g is running more than one daemon pod +Apr 29 11:42:42.620: INFO: Number of nodes with available pods: 0 +Apr 29 11:42:42.620: INFO: Node worker-3oo4n-86cbf586b7-9q98g is running more than one daemon pod +Apr 29 11:42:43.618: INFO: Number of nodes with available pods: 0 +Apr 29 11:42:43.618: INFO: Node worker-3oo4n-86cbf586b7-9q98g is running more than one daemon pod +Apr 29 11:42:44.623: INFO: Number of nodes with available pods: 0 +Apr 29 11:42:44.623: INFO: Node worker-3oo4n-86cbf586b7-9q98g is running more than one daemon pod +Apr 29 11:42:45.619: INFO: Number of nodes with available pods: 0 +Apr 29 11:42:45.619: INFO: Node worker-3oo4n-86cbf586b7-9q98g is running more than one daemon pod +Apr 29 11:42:46.620: INFO: Number of nodes with available pods: 0 +Apr 29 11:42:46.620: INFO: Node worker-3oo4n-86cbf586b7-9q98g is running more than one daemon pod +Apr 29 11:42:47.633: INFO: Number of nodes with available pods: 0 +Apr 29 11:42:47.633: INFO: Node worker-3oo4n-86cbf586b7-9q98g is running more than one daemon pod +Apr 29 11:42:48.623: INFO: Number of nodes with available pods: 0 +Apr 29 11:42:48.623: INFO: Node worker-3oo4n-86cbf586b7-9q98g is running more than one daemon pod +Apr 29 11:42:49.620: INFO: Number of nodes with available pods: 0 +Apr 29 11:42:49.620: INFO: Node worker-3oo4n-86cbf586b7-9q98g is running more than one daemon pod +Apr 29 11:42:50.621: INFO: Number of nodes with available pods: 1 +Apr 29 11:42:50.622: INFO: Number of running nodes: 1, number of available pods: 1 +[AfterEach] [sig-apps] Daemon set [Serial] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:68 +STEP: Deleting DaemonSet "daemon-set" +STEP: deleting DaemonSet.extensions daemon-set in namespace daemonsets-6765, will wait for the garbage collector to delete the pods +Apr 29 11:42:50.736: INFO: Deleting DaemonSet.extensions daemon-set took: 47.166822ms +Apr 29 11:42:50.837: INFO: Terminating DaemonSet.extensions daemon-set pods took: 100.461782ms +Apr 29 11:43:05.745: INFO: Number of nodes with available pods: 0 +Apr 29 11:43:05.745: INFO: Number of running nodes: 0, number of available pods: 0 +Apr 29 11:43:05.750: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"selfLink":"/apis/apps/v1/namespaces/daemonsets-6765/daemonsets","resourceVersion":"6132"},"items":null} + +Apr 29 11:43:05.754: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"selfLink":"/api/v1/namespaces/daemonsets-6765/pods","resourceVersion":"6132"},"items":null} + +[AfterEach] [sig-apps] Daemon set [Serial] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Apr 29 11:43:05.792: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "daemonsets-6765" for this suite. +Apr 29 11:43:11.821: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Apr 29 11:43:12.096: INFO: namespace daemonsets-6765 deletion completed in 6.297001692s + +• [SLOW TEST:38.871 seconds] +[sig-apps] Daemon set [Serial] +/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22 + should run and stop complex daemon [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +------------------------------ +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +------------------------------ +[k8s.io] Variable Expansion + should allow substituting values in a container's command [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [k8s.io] Variable Expansion + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +STEP: Creating a kubernetes client +Apr 29 11:43:12.115: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174 +STEP: Building a namespace api object, basename var-expansion +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in var-expansion-6768 +STEP: Waiting for a default service account to be provisioned in namespace +[It] should allow substituting values in a container's command [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: Creating a pod to test substitution in container's command +Apr 29 11:43:12.318: INFO: Waiting up to 5m0s for pod "var-expansion-f7d12e48-6a73-11e9-b6ef-0e71f8c7e888" in namespace "var-expansion-6768" to be "success or failure" +Apr 29 11:43:12.338: INFO: Pod "var-expansion-f7d12e48-6a73-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 19.672685ms +Apr 29 11:43:14.345: INFO: Pod "var-expansion-f7d12e48-6a73-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 2.02632642s +Apr 29 11:43:16.352: INFO: Pod "var-expansion-f7d12e48-6a73-11e9-b6ef-0e71f8c7e888": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.033878887s +STEP: Saw pod success +Apr 29 11:43:16.352: INFO: Pod "var-expansion-f7d12e48-6a73-11e9-b6ef-0e71f8c7e888" satisfied condition "success or failure" +Apr 29 11:43:16.356: INFO: Trying to get logs from node worker-4av2h-6665d7b5ff-9nc42 pod var-expansion-f7d12e48-6a73-11e9-b6ef-0e71f8c7e888 container dapi-container: +STEP: delete the pod +Apr 29 11:43:16.386: INFO: Waiting for pod var-expansion-f7d12e48-6a73-11e9-b6ef-0e71f8c7e888 to disappear +Apr 29 11:43:16.389: INFO: Pod var-expansion-f7d12e48-6a73-11e9-b6ef-0e71f8c7e888 no longer exists +[AfterEach] [k8s.io] Variable Expansion + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Apr 29 11:43:16.389: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "var-expansion-6768" for this suite. +Apr 29 11:43:22.418: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Apr 29 11:43:22.627: INFO: namespace var-expansion-6768 deletion completed in 6.232091583s + +• [SLOW TEST:10.513 seconds] +[k8s.io] Variable Expansion +/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687 + should allow substituting values in a container's command [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +------------------------------ +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +------------------------------ +[k8s.io] InitContainer [NodeConformance] + should invoke init containers on a RestartAlways pod [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [k8s.io] InitContainer [NodeConformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +STEP: Creating a kubernetes client +Apr 29 11:43:22.637: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174 +STEP: Building a namespace api object, basename init-container +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in init-container-867 +STEP: Waiting for a default service account to be provisioned in namespace +[BeforeEach] [k8s.io] InitContainer [NodeConformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/init_container.go:43 +[It] should invoke init containers on a RestartAlways pod [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: creating the pod +Apr 29 11:43:22.832: INFO: PodSpec: initContainers in spec.initContainers +[AfterEach] [k8s.io] InitContainer [NodeConformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Apr 29 11:43:27.722: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "init-container-867" for this suite. +Apr 29 11:43:51.746: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Apr 29 11:43:51.964: INFO: namespace init-container-867 deletion completed in 24.235175225s + +• [SLOW TEST:29.327 seconds] +[k8s.io] InitContainer [NodeConformance] +/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687 + should invoke init containers on a RestartAlways pod [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +------------------------------ +SS +------------------------------ +[sig-apps] Deployment + deployment should support proportional scaling [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [sig-apps] Deployment + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +STEP: Creating a kubernetes client +Apr 29 11:43:51.965: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174 +STEP: Building a namespace api object, basename deployment +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in deployment-6291 +STEP: Waiting for a default service account to be provisioned in namespace +[BeforeEach] [sig-apps] Deployment + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:65 +[It] deployment should support proportional scaling [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +Apr 29 11:43:52.157: INFO: Creating deployment "nginx-deployment" +Apr 29 11:43:52.174: INFO: Waiting for observed generation 1 +Apr 29 11:43:54.193: INFO: Waiting for all required pods to come up +Apr 29 11:43:54.213: INFO: Pod name nginx: Found 10 pods out of 10 +STEP: ensuring each pod is running +Apr 29 11:43:58.289: INFO: Waiting for deployment "nginx-deployment" to complete +Apr 29 11:43:58.305: INFO: Updating deployment "nginx-deployment" with a non-existent image +Apr 29 11:43:58.336: INFO: Updating deployment nginx-deployment +Apr 29 11:43:58.336: INFO: Waiting for observed generation 2 +Apr 29 11:44:00.374: INFO: Waiting for the first rollout's replicaset to have .status.availableReplicas = 8 +Apr 29 11:44:00.396: INFO: Waiting for the first rollout's replicaset to have .spec.replicas = 8 +Apr 29 11:44:00.407: INFO: Waiting for the first rollout's replicaset of deployment "nginx-deployment" to have desired number of replicas +Apr 29 11:44:00.420: INFO: Verifying that the second rollout's replicaset has .status.availableReplicas = 0 +Apr 29 11:44:00.421: INFO: Waiting for the second rollout's replicaset to have .spec.replicas = 5 +Apr 29 11:44:00.425: INFO: Waiting for the second rollout's replicaset of deployment "nginx-deployment" to have desired number of replicas +Apr 29 11:44:00.440: INFO: Verifying that deployment "nginx-deployment" has minimum required number of available replicas +Apr 29 11:44:00.440: INFO: Scaling up the deployment "nginx-deployment" from 10 to 30 +Apr 29 11:44:00.465: INFO: Updating deployment nginx-deployment +Apr 29 11:44:00.465: INFO: Waiting for the replicasets of deployment "nginx-deployment" to have desired number of replicas +Apr 29 11:44:00.486: INFO: Verifying that first rollout's replicaset has .spec.replicas = 20 +Apr 29 11:44:02.534: INFO: Verifying that second rollout's replicaset has .spec.replicas = 13 +[AfterEach] [sig-apps] Deployment + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:59 +Apr 29 11:44:02.544: INFO: Deployment "nginx-deployment": +&Deployment{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment,GenerateName:,Namespace:deployment-6291,SelfLink:/apis/apps/v1/namespaces/deployment-6291/deployments/nginx-deployment,UID:0f6c50ad-6a74-11e9-92a7-deadbef3c36d,ResourceVersion:6632,Generation:3,CreationTimestamp:2019-04-29 11:43:51 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,},Annotations:map[string]string{deployment.kubernetes.io/revision: 2,},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:DeploymentSpec{Replicas:*30,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: nginx,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[],Containers:[{nginx nginx:404 [] [] [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,},},Strategy:DeploymentStrategy{Type:RollingUpdate,RollingUpdate:&RollingUpdateDeployment{MaxUnavailable:2,MaxSurge:3,},},MinReadySeconds:0,RevisionHistoryLimit:*10,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:3,Replicas:33,UpdatedReplicas:13,AvailableReplicas:8,UnavailableReplicas:25,Conditions:[{Available False 2019-04-29 11:44:00 +0000 UTC 2019-04-29 11:44:00 +0000 UTC MinimumReplicasUnavailable Deployment does not have minimum availability.} {Progressing True 2019-04-29 11:44:01 +0000 UTC 2019-04-29 11:43:51 +0000 UTC ReplicaSetUpdated ReplicaSet "nginx-deployment-5f9595f595" is progressing.}],ReadyReplicas:8,CollisionCount:nil,},} + +Apr 29 11:44:02.549: INFO: New ReplicaSet "nginx-deployment-5f9595f595" of Deployment "nginx-deployment": +&ReplicaSet{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-5f9595f595,GenerateName:,Namespace:deployment-6291,SelfLink:/apis/apps/v1/namespaces/deployment-6291/replicasets/nginx-deployment-5f9595f595,UID:1318c440-6a74-11e9-92a7-deadbef3c36d,ResourceVersion:6630,Generation:3,CreationTimestamp:2019-04-29 11:43:58 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 5f9595f595,},Annotations:map[string]string{deployment.kubernetes.io/desired-replicas: 30,deployment.kubernetes.io/max-replicas: 33,deployment.kubernetes.io/revision: 2,},OwnerReferences:[{apps/v1 Deployment nginx-deployment 0f6c50ad-6a74-11e9-92a7-deadbef3c36d 0xc002471d77 0xc002471d78}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:ReplicaSetSpec{Replicas:*13,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: nginx,pod-template-hash: 5f9595f595,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 5f9595f595,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[],Containers:[{nginx nginx:404 [] [] [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,},},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:13,FullyLabeledReplicas:13,ObservedGeneration:3,ReadyReplicas:0,AvailableReplicas:0,Conditions:[],},} +Apr 29 11:44:02.549: INFO: All old ReplicaSets of Deployment "nginx-deployment": +Apr 29 11:44:02.550: INFO: &ReplicaSet{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-6f478d8d8,GenerateName:,Namespace:deployment-6291,SelfLink:/apis/apps/v1/namespaces/deployment-6291/replicasets/nginx-deployment-6f478d8d8,UID:0f6f5a34-6a74-11e9-92a7-deadbef3c36d,ResourceVersion:6618,Generation:3,CreationTimestamp:2019-04-29 11:43:51 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 6f478d8d8,},Annotations:map[string]string{deployment.kubernetes.io/desired-replicas: 30,deployment.kubernetes.io/max-replicas: 33,deployment.kubernetes.io/revision: 1,},OwnerReferences:[{apps/v1 Deployment nginx-deployment 0f6c50ad-6a74-11e9-92a7-deadbef3c36d 0xc002471ec7 0xc002471ec8}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:ReplicaSetSpec{Replicas:*20,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: nginx,pod-template-hash: 6f478d8d8,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 6f478d8d8,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,},},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:20,FullyLabeledReplicas:20,ObservedGeneration:3,ReadyReplicas:8,AvailableReplicas:8,Conditions:[],},} +Apr 29 11:44:02.564: INFO: Pod "nginx-deployment-5f9595f595-6kt4s" is not available: +&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-5f9595f595-6kt4s,GenerateName:nginx-deployment-5f9595f595-,Namespace:deployment-6291,SelfLink:/api/v1/namespaces/deployment-6291/pods/nginx-deployment-5f9595f595-6kt4s,UID:14b3c201-6a74-11e9-92a7-deadbef3c36d,ResourceVersion:6614,Generation:0,CreationTimestamp:2019-04-29 11:44:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 5f9595f595,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-5f9595f595 1318c440-6a74-11e9-92a7-deadbef3c36d 0xc002a5f4f0 0xc002a5f4f1}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-dhs49 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-dhs49,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] [] [] [] [] {map[] map[]} [{default-token-dhs49 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:worker-4av2h-6665d7b5ff-9nc42,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc002a5f560} {node.kubernetes.io/unreachable Exists NoExecute 0xc002a5f580}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:44:00 +0000 UTC }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} +Apr 29 11:44:02.565: INFO: Pod "nginx-deployment-5f9595f595-6vlxg" is not available: +&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-5f9595f595-6vlxg,GenerateName:nginx-deployment-5f9595f595-,Namespace:deployment-6291,SelfLink:/api/v1/namespaces/deployment-6291/pods/nginx-deployment-5f9595f595-6vlxg,UID:14d8b421-6a74-11e9-92a7-deadbef3c36d,ResourceVersion:6626,Generation:0,CreationTimestamp:2019-04-29 11:44:01 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 5f9595f595,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-5f9595f595 1318c440-6a74-11e9-92a7-deadbef3c36d 0xc002a5f600 0xc002a5f601}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-dhs49 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-dhs49,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] [] [] [] [] {map[] map[]} [{default-token-dhs49 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:worker-3oo4n-86cbf586b7-9q98g,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc002a5f670} {node.kubernetes.io/unreachable Exists NoExecute 0xc002a5f690}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:44:01 +0000 UTC }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} +Apr 29 11:44:02.565: INFO: Pod "nginx-deployment-5f9595f595-7vkzj" is not available: +&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-5f9595f595-7vkzj,GenerateName:nginx-deployment-5f9595f595-,Namespace:deployment-6291,SelfLink:/api/v1/namespaces/deployment-6291/pods/nginx-deployment-5f9595f595-7vkzj,UID:146a6abf-6a74-11e9-92a7-deadbef3c36d,ResourceVersion:6629,Generation:0,CreationTimestamp:2019-04-29 11:44:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 5f9595f595,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-5f9595f595 1318c440-6a74-11e9-92a7-deadbef3c36d 0xc002a5f710 0xc002a5f711}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-dhs49 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-dhs49,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] [] [] [] [] {map[] map[]} [{default-token-dhs49 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:worker-p95fb-54c845dd45-crdqs,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc002a5f780} {node.kubernetes.io/unreachable Exists NoExecute 0xc002a5f7a0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:44:00 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:44:00 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:44:00 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:44:00 +0000 UTC }],Message:,Reason:,HostIP:172.23.0.166,PodIP:,StartTime:2019-04-29 11:44:00 +0000 UTC,ContainerStatuses:[{nginx {ContainerStateWaiting{Reason:ContainerCreating,Message:,} nil nil} {nil nil nil} false 0 nginx:404 }],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} +Apr 29 11:44:02.566: INFO: Pod "nginx-deployment-5f9595f595-gsfnr" is not available: +&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-5f9595f595-gsfnr,GenerateName:nginx-deployment-5f9595f595-,Namespace:deployment-6291,SelfLink:/api/v1/namespaces/deployment-6291/pods/nginx-deployment-5f9595f595-gsfnr,UID:1329ad40-6a74-11e9-92a7-deadbef3c36d,ResourceVersion:6516,Generation:0,CreationTimestamp:2019-04-29 11:43:58 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 5f9595f595,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-5f9595f595 1318c440-6a74-11e9-92a7-deadbef3c36d 0xc002a5f870 0xc002a5f871}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-dhs49 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-dhs49,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] [] [] [] [] {map[] map[]} [{default-token-dhs49 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:worker-p95fb-54c845dd45-crdqs,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc002a5f8e0} {node.kubernetes.io/unreachable Exists NoExecute 0xc002a5f900}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:43:58 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:43:58 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:43:58 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:43:58 +0000 UTC }],Message:,Reason:,HostIP:172.23.0.166,PodIP:,StartTime:2019-04-29 11:43:58 +0000 UTC,ContainerStatuses:[{nginx {ContainerStateWaiting{Reason:ContainerCreating,Message:,} nil nil} {nil nil nil} false 0 nginx:404 }],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} +Apr 29 11:44:02.567: INFO: Pod "nginx-deployment-5f9595f595-h6ggv" is not available: +&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-5f9595f595-h6ggv,GenerateName:nginx-deployment-5f9595f595-,Namespace:deployment-6291,SelfLink:/api/v1/namespaces/deployment-6291/pods/nginx-deployment-5f9595f595-h6ggv,UID:14b6dd43-6a74-11e9-92a7-deadbef3c36d,ResourceVersion:6619,Generation:0,CreationTimestamp:2019-04-29 11:44:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 5f9595f595,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-5f9595f595 1318c440-6a74-11e9-92a7-deadbef3c36d 0xc002a5f9d0 0xc002a5f9d1}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-dhs49 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-dhs49,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] [] [] [] [] {map[] map[]} [{default-token-dhs49 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:worker-p95fb-54c845dd45-crdqs,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc002a5fa40} {node.kubernetes.io/unreachable Exists NoExecute 0xc002a5fa60}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:44:00 +0000 UTC }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} +Apr 29 11:44:02.568: INFO: Pod "nginx-deployment-5f9595f595-jjfs4" is not available: +&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-5f9595f595-jjfs4,GenerateName:nginx-deployment-5f9595f595-,Namespace:deployment-6291,SelfLink:/api/v1/namespaces/deployment-6291/pods/nginx-deployment-5f9595f595-jjfs4,UID:13543f28-6a74-11e9-92a7-deadbef3c36d,ResourceVersion:6538,Generation:0,CreationTimestamp:2019-04-29 11:43:58 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 5f9595f595,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-5f9595f595 1318c440-6a74-11e9-92a7-deadbef3c36d 0xc002a5fae0 0xc002a5fae1}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-dhs49 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-dhs49,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] [] [] [] [] {map[] map[]} [{default-token-dhs49 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:worker-4av2h-6665d7b5ff-9nc42,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc002a5fb50} {node.kubernetes.io/unreachable Exists NoExecute 0xc002a5fb70}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:43:58 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:43:58 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:43:58 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:43:58 +0000 UTC }],Message:,Reason:,HostIP:172.23.0.238,PodIP:,StartTime:2019-04-29 11:43:58 +0000 UTC,ContainerStatuses:[{nginx {ContainerStateWaiting{Reason:ContainerCreating,Message:,} nil nil} {nil nil nil} false 0 nginx:404 }],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} +Apr 29 11:44:02.568: INFO: Pod "nginx-deployment-5f9595f595-mfwdx" is not available: +&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-5f9595f595-mfwdx,GenerateName:nginx-deployment-5f9595f595-,Namespace:deployment-6291,SelfLink:/api/v1/namespaces/deployment-6291/pods/nginx-deployment-5f9595f595-mfwdx,UID:1476144e-6a74-11e9-92a7-deadbef3c36d,ResourceVersion:6628,Generation:0,CreationTimestamp:2019-04-29 11:44:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 5f9595f595,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-5f9595f595 1318c440-6a74-11e9-92a7-deadbef3c36d 0xc002a5fc40 0xc002a5fc41}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-dhs49 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-dhs49,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] [] [] [] [] {map[] map[]} [{default-token-dhs49 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:worker-4av2h-6665d7b5ff-9nc42,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc002a5fcb0} {node.kubernetes.io/unreachable Exists NoExecute 0xc002a5fcd0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:44:01 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:44:01 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:44:01 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:44:00 +0000 UTC }],Message:,Reason:,HostIP:172.23.0.238,PodIP:,StartTime:2019-04-29 11:44:01 +0000 UTC,ContainerStatuses:[{nginx {ContainerStateWaiting{Reason:ContainerCreating,Message:,} nil nil} {nil nil nil} false 0 nginx:404 }],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} +Apr 29 11:44:02.569: INFO: Pod "nginx-deployment-5f9595f595-n5pj7" is not available: +&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-5f9595f595-n5pj7,GenerateName:nginx-deployment-5f9595f595-,Namespace:deployment-6291,SelfLink:/api/v1/namespaces/deployment-6291/pods/nginx-deployment-5f9595f595-n5pj7,UID:13216339-6a74-11e9-92a7-deadbef3c36d,ResourceVersion:6515,Generation:0,CreationTimestamp:2019-04-29 11:43:58 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 5f9595f595,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-5f9595f595 1318c440-6a74-11e9-92a7-deadbef3c36d 0xc002a5fda0 0xc002a5fda1}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-dhs49 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-dhs49,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] [] [] [] [] {map[] map[]} [{default-token-dhs49 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:worker-4av2h-6665d7b5ff-9nc42,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc002a5fe10} {node.kubernetes.io/unreachable Exists NoExecute 0xc002a5fe30}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:43:58 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:43:58 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:43:58 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:43:58 +0000 UTC }],Message:,Reason:,HostIP:172.23.0.238,PodIP:,StartTime:2019-04-29 11:43:58 +0000 UTC,ContainerStatuses:[{nginx {ContainerStateWaiting{Reason:ContainerCreating,Message:,} nil nil} {nil nil nil} false 0 nginx:404 }],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} +Apr 29 11:44:02.570: INFO: Pod "nginx-deployment-5f9595f595-qfvd8" is not available: +&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-5f9595f595-qfvd8,GenerateName:nginx-deployment-5f9595f595-,Namespace:deployment-6291,SelfLink:/api/v1/namespaces/deployment-6291/pods/nginx-deployment-5f9595f595-qfvd8,UID:13628fec-6a74-11e9-92a7-deadbef3c36d,ResourceVersion:6537,Generation:0,CreationTimestamp:2019-04-29 11:43:58 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 5f9595f595,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-5f9595f595 1318c440-6a74-11e9-92a7-deadbef3c36d 0xc002a5ff00 0xc002a5ff01}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-dhs49 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-dhs49,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] [] [] [] [] {map[] map[]} [{default-token-dhs49 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:worker-3oo4n-86cbf586b7-9q98g,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc002a5ff70} {node.kubernetes.io/unreachable Exists NoExecute 0xc002a5ff90}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:43:58 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:43:58 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:43:58 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:43:58 +0000 UTC }],Message:,Reason:,HostIP:172.23.0.214,PodIP:,StartTime:2019-04-29 11:43:58 +0000 UTC,ContainerStatuses:[{nginx {ContainerStateWaiting{Reason:ContainerCreating,Message:,} nil nil} {nil nil nil} false 0 nginx:404 }],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} +Apr 29 11:44:02.571: INFO: Pod "nginx-deployment-5f9595f595-rjn6g" is not available: +&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-5f9595f595-rjn6g,GenerateName:nginx-deployment-5f9595f595-,Namespace:deployment-6291,SelfLink:/api/v1/namespaces/deployment-6291/pods/nginx-deployment-5f9595f595-rjn6g,UID:14b565df-6a74-11e9-92a7-deadbef3c36d,ResourceVersion:6616,Generation:0,CreationTimestamp:2019-04-29 11:44:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 5f9595f595,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-5f9595f595 1318c440-6a74-11e9-92a7-deadbef3c36d 0xc00296a080 0xc00296a081}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-dhs49 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-dhs49,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] [] [] [] [] {map[] map[]} [{default-token-dhs49 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:worker-3oo4n-86cbf586b7-9q98g,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc00296a160} {node.kubernetes.io/unreachable Exists NoExecute 0xc00296a190}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:44:00 +0000 UTC }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} +Apr 29 11:44:02.573: INFO: Pod "nginx-deployment-5f9595f595-tl9rc" is not available: +&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-5f9595f595-tl9rc,GenerateName:nginx-deployment-5f9595f595-,Namespace:deployment-6291,SelfLink:/api/v1/namespaces/deployment-6291/pods/nginx-deployment-5f9595f595-tl9rc,UID:14ad642c-6a74-11e9-92a7-deadbef3c36d,ResourceVersion:6612,Generation:0,CreationTimestamp:2019-04-29 11:44:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 5f9595f595,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-5f9595f595 1318c440-6a74-11e9-92a7-deadbef3c36d 0xc00296a270 0xc00296a271}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-dhs49 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-dhs49,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] [] [] [] [] {map[] map[]} [{default-token-dhs49 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:worker-p95fb-54c845dd45-crdqs,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc00296a310} {node.kubernetes.io/unreachable Exists NoExecute 0xc00296a360}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:44:00 +0000 UTC }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} +Apr 29 11:44:02.573: INFO: Pod "nginx-deployment-5f9595f595-tll98" is not available: +&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-5f9595f595-tll98,GenerateName:nginx-deployment-5f9595f595-,Namespace:deployment-6291,SelfLink:/api/v1/namespaces/deployment-6291/pods/nginx-deployment-5f9595f595-tll98,UID:131c169c-6a74-11e9-92a7-deadbef3c36d,ResourceVersion:6671,Generation:0,CreationTimestamp:2019-04-29 11:43:58 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 5f9595f595,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-5f9595f595 1318c440-6a74-11e9-92a7-deadbef3c36d 0xc00296a440 0xc00296a441}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-dhs49 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-dhs49,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] [] [] [] [] {map[] map[]} [{default-token-dhs49 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:worker-3oo4n-86cbf586b7-9q98g,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc00296a510} {node.kubernetes.io/unreachable Exists NoExecute 0xc00296a530}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:43:58 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:43:58 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:43:58 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:43:58 +0000 UTC }],Message:,Reason:,HostIP:172.23.0.214,PodIP:192.168.14.18,StartTime:2019-04-29 11:43:58 +0000 UTC,ContainerStatuses:[{nginx {ContainerStateWaiting{Reason:ErrImagePull,Message:rpc error: code = Unknown desc = Error response from daemon: manifest for nginx:404 not found,} nil nil} {nil nil nil} false 0 nginx:404 }],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} +Apr 29 11:44:02.574: INFO: Pod "nginx-deployment-5f9595f595-zh67l" is not available: +&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-5f9595f595-zh67l,GenerateName:nginx-deployment-5f9595f595-,Namespace:deployment-6291,SelfLink:/api/v1/namespaces/deployment-6291/pods/nginx-deployment-5f9595f595-zh67l,UID:1479aa52-6a74-11e9-92a7-deadbef3c36d,ResourceVersion:6633,Generation:0,CreationTimestamp:2019-04-29 11:44:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 5f9595f595,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-5f9595f595 1318c440-6a74-11e9-92a7-deadbef3c36d 0xc00296a6f0 0xc00296a6f1}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-dhs49 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-dhs49,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] [] [] [] [] {map[] map[]} [{default-token-dhs49 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:worker-3oo4n-86cbf586b7-9q98g,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc00296a7a0} {node.kubernetes.io/unreachable Exists NoExecute 0xc00296a7d0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:44:01 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:44:01 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:44:01 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:44:00 +0000 UTC }],Message:,Reason:,HostIP:172.23.0.214,PodIP:,StartTime:2019-04-29 11:44:01 +0000 UTC,ContainerStatuses:[{nginx {ContainerStateWaiting{Reason:ContainerCreating,Message:,} nil nil} {nil nil nil} false 0 nginx:404 }],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} +Apr 29 11:44:02.575: INFO: Pod "nginx-deployment-6f478d8d8-5ps29" is not available: +&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-6f478d8d8-5ps29,GenerateName:nginx-deployment-6f478d8d8-,Namespace:deployment-6291,SelfLink:/api/v1/namespaces/deployment-6291/pods/nginx-deployment-6f478d8d8-5ps29,UID:14688888-6a74-11e9-92a7-deadbef3c36d,ResourceVersion:6611,Generation:0,CreationTimestamp:2019-04-29 11:44:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 6f478d8d8,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-6f478d8d8 0f6f5a34-6a74-11e9-92a7-deadbef3c36d 0xc00296a8e0 0xc00296a8e1}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-dhs49 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-dhs49,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-dhs49 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:worker-p95fb-54c845dd45-crdqs,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc00296a970} {node.kubernetes.io/unreachable Exists NoExecute 0xc00296aa30}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:44:00 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:44:00 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:44:00 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:44:00 +0000 UTC }],Message:,Reason:,HostIP:172.23.0.166,PodIP:,StartTime:2019-04-29 11:44:00 +0000 UTC,ContainerStatuses:[{nginx {ContainerStateWaiting{Reason:ContainerCreating,Message:,} nil nil} {nil nil nil} false 0 docker.io/library/nginx:1.14-alpine }],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} +Apr 29 11:44:02.576: INFO: Pod "nginx-deployment-6f478d8d8-7jpjr" is not available: +&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-6f478d8d8-7jpjr,GenerateName:nginx-deployment-6f478d8d8-,Namespace:deployment-6291,SelfLink:/api/v1/namespaces/deployment-6291/pods/nginx-deployment-6f478d8d8-7jpjr,UID:1490ba02-6a74-11e9-92a7-deadbef3c36d,ResourceVersion:6607,Generation:0,CreationTimestamp:2019-04-29 11:44:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 6f478d8d8,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-6f478d8d8 0f6f5a34-6a74-11e9-92a7-deadbef3c36d 0xc00296ab40 0xc00296ab41}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-dhs49 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-dhs49,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-dhs49 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:worker-4av2h-6665d7b5ff-9nc42,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc00296abf0} {node.kubernetes.io/unreachable Exists NoExecute 0xc00296ac10}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:44:00 +0000 UTC }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} +Apr 29 11:44:02.576: INFO: Pod "nginx-deployment-6f478d8d8-7k5k7" is available: +&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-6f478d8d8-7k5k7,GenerateName:nginx-deployment-6f478d8d8-,Namespace:deployment-6291,SelfLink:/api/v1/namespaces/deployment-6291/pods/nginx-deployment-6f478d8d8-7k5k7,UID:0f763548-6a74-11e9-92a7-deadbef3c36d,ResourceVersion:6452,Generation:0,CreationTimestamp:2019-04-29 11:43:51 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 6f478d8d8,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-6f478d8d8 0f6f5a34-6a74-11e9-92a7-deadbef3c36d 0xc00296ac90 0xc00296ac91}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-dhs49 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-dhs49,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-dhs49 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:worker-p95fb-54c845dd45-crdqs,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc00296acf0} {node.kubernetes.io/unreachable Exists NoExecute 0xc00296ad10}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Running,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:43:52 +0000 UTC } {Ready True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:43:56 +0000 UTC } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:43:56 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:43:51 +0000 UTC }],Message:,Reason:,HostIP:172.23.0.166,PodIP:192.168.148.81,StartTime:2019-04-29 11:43:52 +0000 UTC,ContainerStatuses:[{nginx {nil ContainerStateRunning{StartedAt:2019-04-29 11:43:55 +0000 UTC,} nil} {nil nil nil} true 0 nginx:1.14-alpine docker-pullable://nginx@sha256:485b610fefec7ff6c463ced9623314a04ed67e3945b9c08d7e53a47f6d108dc7 docker://9e4a95607aee14b3e11b50f546db36be41c4049a90e9dc2691f590588e56dc0b}],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} +Apr 29 11:44:02.578: INFO: Pod "nginx-deployment-6f478d8d8-7pdl2" is available: +&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-6f478d8d8-7pdl2,GenerateName:nginx-deployment-6f478d8d8-,Namespace:deployment-6291,SelfLink:/api/v1/namespaces/deployment-6291/pods/nginx-deployment-6f478d8d8-7pdl2,UID:0f8f25d2-6a74-11e9-92a7-deadbef3c36d,ResourceVersion:6480,Generation:0,CreationTimestamp:2019-04-29 11:43:52 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 6f478d8d8,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-6f478d8d8 0f6f5a34-6a74-11e9-92a7-deadbef3c36d 0xc00296ade0 0xc00296ade1}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-dhs49 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-dhs49,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-dhs49 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:worker-4av2h-6665d7b5ff-9nc42,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc00296ae40} {node.kubernetes.io/unreachable Exists NoExecute 0xc00296ae60}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Running,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:43:52 +0000 UTC } {Ready True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:43:57 +0000 UTC } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:43:57 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:43:52 +0000 UTC }],Message:,Reason:,HostIP:172.23.0.238,PodIP:192.168.205.96,StartTime:2019-04-29 11:43:52 +0000 UTC,ContainerStatuses:[{nginx {nil ContainerStateRunning{StartedAt:2019-04-29 11:43:56 +0000 UTC,} nil} {nil nil nil} true 0 nginx:1.14-alpine docker-pullable://nginx@sha256:485b610fefec7ff6c463ced9623314a04ed67e3945b9c08d7e53a47f6d108dc7 docker://c00ac098866cf6b2761ca9633388b5ecd808398f5dfbd7bd6dd9d34e02db66e3}],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} +Apr 29 11:44:02.578: INFO: Pod "nginx-deployment-6f478d8d8-865zt" is available: +&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-6f478d8d8-865zt,GenerateName:nginx-deployment-6f478d8d8-,Namespace:deployment-6291,SelfLink:/api/v1/namespaces/deployment-6291/pods/nginx-deployment-6f478d8d8-865zt,UID:0f83b927-6a74-11e9-92a7-deadbef3c36d,ResourceVersion:6478,Generation:0,CreationTimestamp:2019-04-29 11:43:52 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 6f478d8d8,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-6f478d8d8 0f6f5a34-6a74-11e9-92a7-deadbef3c36d 0xc00296af30 0xc00296af31}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-dhs49 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-dhs49,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-dhs49 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:worker-4av2h-6665d7b5ff-9nc42,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc00296af90} {node.kubernetes.io/unreachable Exists NoExecute 0xc00296afb0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Running,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:43:52 +0000 UTC } {Ready True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:43:57 +0000 UTC } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:43:57 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:43:52 +0000 UTC }],Message:,Reason:,HostIP:172.23.0.238,PodIP:192.168.205.94,StartTime:2019-04-29 11:43:52 +0000 UTC,ContainerStatuses:[{nginx {nil ContainerStateRunning{StartedAt:2019-04-29 11:43:56 +0000 UTC,} nil} {nil nil nil} true 0 nginx:1.14-alpine docker-pullable://nginx@sha256:485b610fefec7ff6c463ced9623314a04ed67e3945b9c08d7e53a47f6d108dc7 docker://f6f1517e09c258502ed56ca9fbe9fa1c5f69ef8b410fd2b02e3642dd22970667}],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} +Apr 29 11:44:02.579: INFO: Pod "nginx-deployment-6f478d8d8-8lkd2" is not available: +&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-6f478d8d8-8lkd2,GenerateName:nginx-deployment-6f478d8d8-,Namespace:deployment-6291,SelfLink:/api/v1/namespaces/deployment-6291/pods/nginx-deployment-6f478d8d8-8lkd2,UID:1474f3ba-6a74-11e9-92a7-deadbef3c36d,ResourceVersion:6636,Generation:0,CreationTimestamp:2019-04-29 11:44:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 6f478d8d8,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-6f478d8d8 0f6f5a34-6a74-11e9-92a7-deadbef3c36d 0xc00296b080 0xc00296b081}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-dhs49 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-dhs49,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-dhs49 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:worker-p95fb-54c845dd45-crdqs,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc00296b0e0} {node.kubernetes.io/unreachable Exists NoExecute 0xc00296b100}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:44:01 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:44:01 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:44:01 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:44:00 +0000 UTC }],Message:,Reason:,HostIP:172.23.0.166,PodIP:,StartTime:2019-04-29 11:44:01 +0000 UTC,ContainerStatuses:[{nginx {ContainerStateWaiting{Reason:ContainerCreating,Message:,} nil nil} {nil nil nil} false 0 docker.io/library/nginx:1.14-alpine }],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} +Apr 29 11:44:02.580: INFO: Pod "nginx-deployment-6f478d8d8-9b5t4" is available: +&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-6f478d8d8-9b5t4,GenerateName:nginx-deployment-6f478d8d8-,Namespace:deployment-6291,SelfLink:/api/v1/namespaces/deployment-6291/pods/nginx-deployment-6f478d8d8-9b5t4,UID:0f8edbbf-6a74-11e9-92a7-deadbef3c36d,ResourceVersion:6468,Generation:0,CreationTimestamp:2019-04-29 11:43:52 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 6f478d8d8,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-6f478d8d8 0f6f5a34-6a74-11e9-92a7-deadbef3c36d 0xc00296b1c0 0xc00296b1c1}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-dhs49 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-dhs49,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-dhs49 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:worker-3oo4n-86cbf586b7-9q98g,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc00296b220} {node.kubernetes.io/unreachable Exists NoExecute 0xc00296b240}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Running,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:43:52 +0000 UTC } {Ready True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:43:56 +0000 UTC } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:43:56 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:43:52 +0000 UTC }],Message:,Reason:,HostIP:172.23.0.214,PodIP:192.168.14.16,StartTime:2019-04-29 11:43:52 +0000 UTC,ContainerStatuses:[{nginx {nil ContainerStateRunning{StartedAt:2019-04-29 11:43:55 +0000 UTC,} nil} {nil nil nil} true 0 nginx:1.14-alpine docker-pullable://nginx@sha256:485b610fefec7ff6c463ced9623314a04ed67e3945b9c08d7e53a47f6d108dc7 docker://c1f9e74b5453d28a5287c502f820ca8bc818a4d3c112287bf30e563d229f4c43}],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} +Apr 29 11:44:02.580: INFO: Pod "nginx-deployment-6f478d8d8-9fgf4" is not available: +&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-6f478d8d8-9fgf4,GenerateName:nginx-deployment-6f478d8d8-,Namespace:deployment-6291,SelfLink:/api/v1/namespaces/deployment-6291/pods/nginx-deployment-6f478d8d8-9fgf4,UID:1467f33a-6a74-11e9-92a7-deadbef3c36d,ResourceVersion:6583,Generation:0,CreationTimestamp:2019-04-29 11:44:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 6f478d8d8,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-6f478d8d8 0f6f5a34-6a74-11e9-92a7-deadbef3c36d 0xc00296b310 0xc00296b311}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-dhs49 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-dhs49,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-dhs49 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:worker-4av2h-6665d7b5ff-9nc42,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc00296b420} {node.kubernetes.io/unreachable Exists NoExecute 0xc00296b440}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:44:00 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:44:00 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:44:00 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:44:00 +0000 UTC }],Message:,Reason:,HostIP:172.23.0.238,PodIP:,StartTime:2019-04-29 11:44:00 +0000 UTC,ContainerStatuses:[{nginx {ContainerStateWaiting{Reason:ContainerCreating,Message:,} nil nil} {nil nil nil} false 0 docker.io/library/nginx:1.14-alpine }],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} +Apr 29 11:44:02.581: INFO: Pod "nginx-deployment-6f478d8d8-b2dpf" is not available: +&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-6f478d8d8-b2dpf,GenerateName:nginx-deployment-6f478d8d8-,Namespace:deployment-6291,SelfLink:/api/v1/namespaces/deployment-6291/pods/nginx-deployment-6f478d8d8-b2dpf,UID:1491f7e3-6a74-11e9-92a7-deadbef3c36d,ResourceVersion:6609,Generation:0,CreationTimestamp:2019-04-29 11:44:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 6f478d8d8,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-6f478d8d8 0f6f5a34-6a74-11e9-92a7-deadbef3c36d 0xc00296b580 0xc00296b581}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-dhs49 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-dhs49,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-dhs49 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:worker-p95fb-54c845dd45-crdqs,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc00296b600} {node.kubernetes.io/unreachable Exists NoExecute 0xc00296b670}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:44:00 +0000 UTC }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} +Apr 29 11:44:02.582: INFO: Pod "nginx-deployment-6f478d8d8-bzhdv" is not available: +&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-6f478d8d8-bzhdv,GenerateName:nginx-deployment-6f478d8d8-,Namespace:deployment-6291,SelfLink:/api/v1/namespaces/deployment-6291/pods/nginx-deployment-6f478d8d8-bzhdv,UID:14910c71-6a74-11e9-92a7-deadbef3c36d,ResourceVersion:6605,Generation:0,CreationTimestamp:2019-04-29 11:44:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 6f478d8d8,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-6f478d8d8 0f6f5a34-6a74-11e9-92a7-deadbef3c36d 0xc00296b730 0xc00296b731}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-dhs49 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-dhs49,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-dhs49 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:worker-p95fb-54c845dd45-crdqs,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc00296b7b0} {node.kubernetes.io/unreachable Exists NoExecute 0xc00296b7d0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:44:00 +0000 UTC }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} +Apr 29 11:44:02.583: INFO: Pod "nginx-deployment-6f478d8d8-lxpbp" is not available: +&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-6f478d8d8-lxpbp,GenerateName:nginx-deployment-6f478d8d8-,Namespace:deployment-6291,SelfLink:/api/v1/namespaces/deployment-6291/pods/nginx-deployment-6f478d8d8-lxpbp,UID:147464d1-6a74-11e9-92a7-deadbef3c36d,ResourceVersion:6600,Generation:0,CreationTimestamp:2019-04-29 11:44:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 6f478d8d8,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-6f478d8d8 0f6f5a34-6a74-11e9-92a7-deadbef3c36d 0xc00296b880 0xc00296b881}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-dhs49 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-dhs49,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-dhs49 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:worker-3oo4n-86cbf586b7-9q98g,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc00296b950} {node.kubernetes.io/unreachable Exists NoExecute 0xc00296b9a0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:44:00 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:44:00 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:44:00 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:44:00 +0000 UTC }],Message:,Reason:,HostIP:172.23.0.214,PodIP:,StartTime:2019-04-29 11:44:00 +0000 UTC,ContainerStatuses:[{nginx {ContainerStateWaiting{Reason:ContainerCreating,Message:,} nil nil} {nil nil nil} false 0 docker.io/library/nginx:1.14-alpine }],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} +Apr 29 11:44:02.583: INFO: Pod "nginx-deployment-6f478d8d8-mhthb" is available: +&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-6f478d8d8-mhthb,GenerateName:nginx-deployment-6f478d8d8-,Namespace:deployment-6291,SelfLink:/api/v1/namespaces/deployment-6291/pods/nginx-deployment-6f478d8d8-mhthb,UID:0f82a4f8-6a74-11e9-92a7-deadbef3c36d,ResourceVersion:6472,Generation:0,CreationTimestamp:2019-04-29 11:43:52 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 6f478d8d8,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-6f478d8d8 0f6f5a34-6a74-11e9-92a7-deadbef3c36d 0xc00296bb30 0xc00296bb31}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-dhs49 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-dhs49,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-dhs49 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:worker-3oo4n-86cbf586b7-9q98g,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc00296bb90} {node.kubernetes.io/unreachable Exists NoExecute 0xc00296bc00}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Running,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:43:52 +0000 UTC } {Ready True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:43:56 +0000 UTC } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:43:56 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:43:52 +0000 UTC }],Message:,Reason:,HostIP:172.23.0.214,PodIP:192.168.14.17,StartTime:2019-04-29 11:43:52 +0000 UTC,ContainerStatuses:[{nginx {nil ContainerStateRunning{StartedAt:2019-04-29 11:43:55 +0000 UTC,} nil} {nil nil nil} true 0 nginx:1.14-alpine docker-pullable://nginx@sha256:485b610fefec7ff6c463ced9623314a04ed67e3945b9c08d7e53a47f6d108dc7 docker://1f8d095bd8092fbd5bacae965949bcf9ed0f2f60306e710bac3c51f5cce14a87}],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} +Apr 29 11:44:02.584: INFO: Pod "nginx-deployment-6f478d8d8-pw6z5" is available: +&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-6f478d8d8-pw6z5,GenerateName:nginx-deployment-6f478d8d8-,Namespace:deployment-6291,SelfLink:/api/v1/namespaces/deployment-6291/pods/nginx-deployment-6f478d8d8-pw6z5,UID:0f732799-6a74-11e9-92a7-deadbef3c36d,ResourceVersion:6462,Generation:0,CreationTimestamp:2019-04-29 11:43:51 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 6f478d8d8,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-6f478d8d8 0f6f5a34-6a74-11e9-92a7-deadbef3c36d 0xc00296bd00 0xc00296bd01}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-dhs49 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-dhs49,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-dhs49 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:worker-4av2h-6665d7b5ff-9nc42,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc00296bd90} {node.kubernetes.io/unreachable Exists NoExecute 0xc00296be10}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Running,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:43:52 +0000 UTC } {Ready True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:43:56 +0000 UTC } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:43:56 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:43:51 +0000 UTC }],Message:,Reason:,HostIP:172.23.0.238,PodIP:192.168.205.93,StartTime:2019-04-29 11:43:52 +0000 UTC,ContainerStatuses:[{nginx {nil ContainerStateRunning{StartedAt:2019-04-29 11:43:55 +0000 UTC,} nil} {nil nil nil} true 0 nginx:1.14-alpine docker-pullable://nginx@sha256:485b610fefec7ff6c463ced9623314a04ed67e3945b9c08d7e53a47f6d108dc7 docker://c2160289af610d9e448fcfd94b1fd7e5b12d3c6ab17950dd0fdf71777bf9c4cd}],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} +Apr 29 11:44:02.585: INFO: Pod "nginx-deployment-6f478d8d8-qdqs2" is not available: +&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-6f478d8d8-qdqs2,GenerateName:nginx-deployment-6f478d8d8-,Namespace:deployment-6291,SelfLink:/api/v1/namespaces/deployment-6291/pods/nginx-deployment-6f478d8d8-qdqs2,UID:1491b882-6a74-11e9-92a7-deadbef3c36d,ResourceVersion:6606,Generation:0,CreationTimestamp:2019-04-29 11:44:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 6f478d8d8,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-6f478d8d8 0f6f5a34-6a74-11e9-92a7-deadbef3c36d 0xc00296bf90 0xc00296bf91}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-dhs49 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-dhs49,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-dhs49 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:worker-3oo4n-86cbf586b7-9q98g,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc002d72020} {node.kubernetes.io/unreachable Exists NoExecute 0xc002d72040}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:44:00 +0000 UTC }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} +Apr 29 11:44:02.585: INFO: Pod "nginx-deployment-6f478d8d8-qgkwz" is not available: +&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-6f478d8d8-qgkwz,GenerateName:nginx-deployment-6f478d8d8-,Namespace:deployment-6291,SelfLink:/api/v1/namespaces/deployment-6291/pods/nginx-deployment-6f478d8d8-qgkwz,UID:147790fb-6a74-11e9-92a7-deadbef3c36d,ResourceVersion:6621,Generation:0,CreationTimestamp:2019-04-29 11:44:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 6f478d8d8,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-6f478d8d8 0f6f5a34-6a74-11e9-92a7-deadbef3c36d 0xc002d720c0 0xc002d720c1}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-dhs49 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-dhs49,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-dhs49 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:worker-3oo4n-86cbf586b7-9q98g,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc002d72120} {node.kubernetes.io/unreachable Exists NoExecute 0xc002d72140}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:44:00 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:44:00 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:44:00 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:44:00 +0000 UTC }],Message:,Reason:,HostIP:172.23.0.214,PodIP:,StartTime:2019-04-29 11:44:00 +0000 UTC,ContainerStatuses:[{nginx {ContainerStateWaiting{Reason:ContainerCreating,Message:,} nil nil} {nil nil nil} false 0 docker.io/library/nginx:1.14-alpine }],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} +Apr 29 11:44:02.586: INFO: Pod "nginx-deployment-6f478d8d8-qjhcp" is not available: +&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-6f478d8d8-qjhcp,GenerateName:nginx-deployment-6f478d8d8-,Namespace:deployment-6291,SelfLink:/api/v1/namespaces/deployment-6291/pods/nginx-deployment-6f478d8d8-qjhcp,UID:1474aceb-6a74-11e9-92a7-deadbef3c36d,ResourceVersion:6615,Generation:0,CreationTimestamp:2019-04-29 11:44:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 6f478d8d8,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-6f478d8d8 0f6f5a34-6a74-11e9-92a7-deadbef3c36d 0xc002d72370 0xc002d72371}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-dhs49 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-dhs49,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-dhs49 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:worker-4av2h-6665d7b5ff-9nc42,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc002d723d0} {node.kubernetes.io/unreachable Exists NoExecute 0xc002d72470}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:44:00 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:44:00 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:44:00 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:44:00 +0000 UTC }],Message:,Reason:,HostIP:172.23.0.238,PodIP:,StartTime:2019-04-29 11:44:00 +0000 UTC,ContainerStatuses:[{nginx {ContainerStateWaiting{Reason:ContainerCreating,Message:,} nil nil} {nil nil nil} false 0 docker.io/library/nginx:1.14-alpine }],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} +Apr 29 11:44:02.587: INFO: Pod "nginx-deployment-6f478d8d8-rnw7k" is available: +&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-6f478d8d8-rnw7k,GenerateName:nginx-deployment-6f478d8d8-,Namespace:deployment-6291,SelfLink:/api/v1/namespaces/deployment-6291/pods/nginx-deployment-6f478d8d8-rnw7k,UID:0f799fc5-6a74-11e9-92a7-deadbef3c36d,ResourceVersion:6465,Generation:0,CreationTimestamp:2019-04-29 11:43:51 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 6f478d8d8,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-6f478d8d8 0f6f5a34-6a74-11e9-92a7-deadbef3c36d 0xc002d725a0 0xc002d725a1}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-dhs49 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-dhs49,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-dhs49 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:worker-3oo4n-86cbf586b7-9q98g,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc002d72620} {node.kubernetes.io/unreachable Exists NoExecute 0xc002d72640}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Running,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:43:52 +0000 UTC } {Ready True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:43:56 +0000 UTC } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:43:56 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:43:52 +0000 UTC }],Message:,Reason:,HostIP:172.23.0.214,PodIP:192.168.14.15,StartTime:2019-04-29 11:43:52 +0000 UTC,ContainerStatuses:[{nginx {nil ContainerStateRunning{StartedAt:2019-04-29 11:43:55 +0000 UTC,} nil} {nil nil nil} true 0 nginx:1.14-alpine docker-pullable://nginx@sha256:485b610fefec7ff6c463ced9623314a04ed67e3945b9c08d7e53a47f6d108dc7 docker://e31055911f87d5ef52ebea4fd9831a8c3abb80b5b839037915df7bdb2eecc10b}],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} +Apr 29 11:44:02.588: INFO: Pod "nginx-deployment-6f478d8d8-t8mpl" is not available: +&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-6f478d8d8-t8mpl,GenerateName:nginx-deployment-6f478d8d8-,Namespace:deployment-6291,SelfLink:/api/v1/namespaces/deployment-6291/pods/nginx-deployment-6f478d8d8-t8mpl,UID:1460a859-6a74-11e9-92a7-deadbef3c36d,ResourceVersion:6590,Generation:0,CreationTimestamp:2019-04-29 11:44:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 6f478d8d8,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-6f478d8d8 0f6f5a34-6a74-11e9-92a7-deadbef3c36d 0xc002d727b0 0xc002d727b1}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-dhs49 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-dhs49,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-dhs49 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:worker-p95fb-54c845dd45-crdqs,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc002d72890} {node.kubernetes.io/unreachable Exists NoExecute 0xc002d728b0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:44:00 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:44:00 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:44:00 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:44:00 +0000 UTC }],Message:,Reason:,HostIP:172.23.0.166,PodIP:,StartTime:2019-04-29 11:44:00 +0000 UTC,ContainerStatuses:[{nginx {ContainerStateWaiting{Reason:ContainerCreating,Message:,} nil nil} {nil nil nil} false 0 docker.io/library/nginx:1.14-alpine }],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} +Apr 29 11:44:02.588: INFO: Pod "nginx-deployment-6f478d8d8-xps6j" is available: +&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-6f478d8d8-xps6j,GenerateName:nginx-deployment-6f478d8d8-,Namespace:deployment-6291,SelfLink:/api/v1/namespaces/deployment-6291/pods/nginx-deployment-6f478d8d8-xps6j,UID:0f8b9880-6a74-11e9-92a7-deadbef3c36d,ResourceVersion:6455,Generation:0,CreationTimestamp:2019-04-29 11:43:52 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 6f478d8d8,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-6f478d8d8 0f6f5a34-6a74-11e9-92a7-deadbef3c36d 0xc002d729a0 0xc002d729a1}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-dhs49 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-dhs49,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-dhs49 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:worker-p95fb-54c845dd45-crdqs,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc002d72a00} {node.kubernetes.io/unreachable Exists NoExecute 0xc002d72a20}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Running,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:43:52 +0000 UTC } {Ready True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:43:56 +0000 UTC } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:43:56 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:43:52 +0000 UTC }],Message:,Reason:,HostIP:172.23.0.166,PodIP:192.168.148.83,StartTime:2019-04-29 11:43:52 +0000 UTC,ContainerStatuses:[{nginx {nil ContainerStateRunning{StartedAt:2019-04-29 11:43:56 +0000 UTC,} nil} {nil nil nil} true 0 nginx:1.14-alpine docker-pullable://nginx@sha256:485b610fefec7ff6c463ced9623314a04ed67e3945b9c08d7e53a47f6d108dc7 docker://15b4f76caace19f6afc88eb1712c43d7d3200d3a5c652889b88c7bf98b9eb875}],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} +Apr 29 11:44:02.589: INFO: Pod "nginx-deployment-6f478d8d8-zznr7" is not available: +&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-6f478d8d8-zznr7,GenerateName:nginx-deployment-6f478d8d8-,Namespace:deployment-6291,SelfLink:/api/v1/namespaces/deployment-6291/pods/nginx-deployment-6f478d8d8-zznr7,UID:1492b1ee-6a74-11e9-92a7-deadbef3c36d,ResourceVersion:6608,Generation:0,CreationTimestamp:2019-04-29 11:44:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 6f478d8d8,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-6f478d8d8 0f6f5a34-6a74-11e9-92a7-deadbef3c36d 0xc002d72af0 0xc002d72af1}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-dhs49 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-dhs49,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-dhs49 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:worker-4av2h-6665d7b5ff-9nc42,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc002d72bc0} {node.kubernetes.io/unreachable Exists NoExecute 0xc002d72be0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:44:00 +0000 UTC }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} +[AfterEach] [sig-apps] Deployment + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Apr 29 11:44:02.590: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "deployment-6291" for this suite. +Apr 29 11:44:10.617: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Apr 29 11:44:10.804: INFO: namespace deployment-6291 deletion completed in 8.207051149s + +• [SLOW TEST:18.839 seconds] +[sig-apps] Deployment +/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22 + deployment should support proportional scaling [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +------------------------------ +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +------------------------------ +[sig-storage] Projected secret + optional updates should be reflected in volume [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [sig-storage] Projected secret + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +STEP: Creating a kubernetes client +Apr 29 11:44:10.805: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174 +STEP: Building a namespace api object, basename projected +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-6923 +STEP: Waiting for a default service account to be provisioned in namespace +[It] optional updates should be reflected in volume [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: Creating secret with name s-test-opt-del-1ad15ef7-6a74-11e9-b6ef-0e71f8c7e888 +STEP: Creating secret with name s-test-opt-upd-1ad15f50-6a74-11e9-b6ef-0e71f8c7e888 +STEP: Creating the pod +STEP: Deleting secret s-test-opt-del-1ad15ef7-6a74-11e9-b6ef-0e71f8c7e888 +STEP: Updating secret s-test-opt-upd-1ad15f50-6a74-11e9-b6ef-0e71f8c7e888 +STEP: Creating secret with name s-test-opt-create-1ad15f72-6a74-11e9-b6ef-0e71f8c7e888 +STEP: waiting to observe update in volume +[AfterEach] [sig-storage] Projected secret + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Apr 29 11:45:50.077: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "projected-6923" for this suite. +Apr 29 11:46:14.114: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Apr 29 11:46:14.366: INFO: namespace projected-6923 deletion completed in 24.275977481s + +• [SLOW TEST:123.561 seconds] +[sig-storage] Projected secret +/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_secret.go:33 + optional updates should be reflected in volume [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +------------------------------ +SS +------------------------------ +[k8s.io] Pods + should support remote command execution over websockets [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [k8s.io] Pods + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +STEP: Creating a kubernetes client +Apr 29 11:46:14.374: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174 +STEP: Building a namespace api object, basename pods +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in pods-4309 +STEP: Waiting for a default service account to be provisioned in namespace +[BeforeEach] [k8s.io] Pods + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:135 +[It] should support remote command execution over websockets [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +Apr 29 11:46:14.615: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174 +STEP: creating the pod +STEP: submitting the pod to kubernetes +[AfterEach] [k8s.io] Pods + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Apr 29 11:46:18.829: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "pods-4309" for this suite. +Apr 29 11:46:58.855: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Apr 29 11:46:59.077: INFO: namespace pods-4309 deletion completed in 40.242725967s + +• [SLOW TEST:44.704 seconds] +[k8s.io] Pods +/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687 + should support remote command execution over websockets [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +------------------------------ +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +------------------------------ +[sig-storage] EmptyDir volumes + should support (non-root,0777,default) [LinuxOnly] [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [sig-storage] EmptyDir volumes + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +STEP: Creating a kubernetes client +Apr 29 11:46:59.079: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174 +STEP: Building a namespace api object, basename emptydir +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in emptydir-6641 +STEP: Waiting for a default service account to be provisioned in namespace +[It] should support (non-root,0777,default) [LinuxOnly] [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: Creating a pod to test emptydir 0777 on node default medium +Apr 29 11:46:59.287: INFO: Waiting up to 5m0s for pod "pod-7f194694-6a74-11e9-b6ef-0e71f8c7e888" in namespace "emptydir-6641" to be "success or failure" +Apr 29 11:46:59.298: INFO: Pod "pod-7f194694-6a74-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 10.887371ms +Apr 29 11:47:01.307: INFO: Pod "pod-7f194694-6a74-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 2.020481519s +Apr 29 11:47:03.316: INFO: Pod "pod-7f194694-6a74-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 4.028899917s +Apr 29 11:47:05.322: INFO: Pod "pod-7f194694-6a74-11e9-b6ef-0e71f8c7e888": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.035442955s +STEP: Saw pod success +Apr 29 11:47:05.322: INFO: Pod "pod-7f194694-6a74-11e9-b6ef-0e71f8c7e888" satisfied condition "success or failure" +Apr 29 11:47:05.327: INFO: Trying to get logs from node worker-4av2h-6665d7b5ff-9nc42 pod pod-7f194694-6a74-11e9-b6ef-0e71f8c7e888 container test-container: +STEP: delete the pod +Apr 29 11:47:05.355: INFO: Waiting for pod pod-7f194694-6a74-11e9-b6ef-0e71f8c7e888 to disappear +Apr 29 11:47:05.358: INFO: Pod pod-7f194694-6a74-11e9-b6ef-0e71f8c7e888 no longer exists +[AfterEach] [sig-storage] EmptyDir volumes + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Apr 29 11:47:05.359: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "emptydir-6641" for this suite. +Apr 29 11:47:11.384: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Apr 29 11:47:11.630: INFO: namespace emptydir-6641 deletion completed in 6.263466521s + +• [SLOW TEST:12.551 seconds] +[sig-storage] EmptyDir volumes +/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:41 + should support (non-root,0777,default) [LinuxOnly] [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +------------------------------ +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +------------------------------ +[k8s.io] Docker Containers + should be able to override the image's default command and arguments [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [k8s.io] Docker Containers + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +STEP: Creating a kubernetes client +Apr 29 11:47:11.640: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174 +STEP: Building a namespace api object, basename containers +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in containers-8544 +STEP: Waiting for a default service account to be provisioned in namespace +[It] should be able to override the image's default command and arguments [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: Creating a pod to test override all +Apr 29 11:47:11.871: INFO: Waiting up to 5m0s for pod "client-containers-869a16a0-6a74-11e9-b6ef-0e71f8c7e888" in namespace "containers-8544" to be "success or failure" +Apr 29 11:47:11.888: INFO: Pod "client-containers-869a16a0-6a74-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 16.771187ms +Apr 29 11:47:13.905: INFO: Pod "client-containers-869a16a0-6a74-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 2.033663042s +Apr 29 11:47:15.916: INFO: Pod "client-containers-869a16a0-6a74-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 4.044498868s +Apr 29 11:47:17.926: INFO: Pod "client-containers-869a16a0-6a74-11e9-b6ef-0e71f8c7e888": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.054055987s +STEP: Saw pod success +Apr 29 11:47:17.926: INFO: Pod "client-containers-869a16a0-6a74-11e9-b6ef-0e71f8c7e888" satisfied condition "success or failure" +Apr 29 11:47:17.931: INFO: Trying to get logs from node worker-4av2h-6665d7b5ff-9nc42 pod client-containers-869a16a0-6a74-11e9-b6ef-0e71f8c7e888 container test-container: +STEP: delete the pod +Apr 29 11:47:17.972: INFO: Waiting for pod client-containers-869a16a0-6a74-11e9-b6ef-0e71f8c7e888 to disappear +Apr 29 11:47:17.982: INFO: Pod client-containers-869a16a0-6a74-11e9-b6ef-0e71f8c7e888 no longer exists +[AfterEach] [k8s.io] Docker Containers + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Apr 29 11:47:17.982: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "containers-8544" for this suite. +Apr 29 11:47:24.020: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Apr 29 11:47:24.260: INFO: namespace containers-8544 deletion completed in 6.272017162s + +• [SLOW TEST:12.620 seconds] +[k8s.io] Docker Containers +/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687 + should be able to override the image's default command and arguments [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +------------------------------ +[sig-api-machinery] Garbage collector + should not delete dependents that have both valid owner and owner that's waiting for dependents to be deleted [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [sig-api-machinery] Garbage collector + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +STEP: Creating a kubernetes client +Apr 29 11:47:24.264: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174 +STEP: Building a namespace api object, basename gc +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in gc-4069 +STEP: Waiting for a default service account to be provisioned in namespace +[It] should not delete dependents that have both valid owner and owner that's waiting for dependents to be deleted [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: create the rc1 +STEP: create the rc2 +STEP: set half of pods created by rc simpletest-rc-to-be-deleted to have rc simpletest-rc-to-stay as owner as well +STEP: delete the rc simpletest-rc-to-be-deleted +STEP: wait for the rc to be deleted +STEP: Gathering metrics +W0429 11:47:34.687158 15 metrics_grabber.go:79] Master node is not registered. Grabbing metrics from Scheduler, ControllerManager and ClusterAutoscaler is disabled. +Apr 29 11:47:34.687: INFO: For apiserver_request_total: +For apiserver_request_latencies_summary: +For apiserver_init_events_total: +For garbage_collector_attempt_to_delete_queue_latency: +For garbage_collector_attempt_to_delete_work_duration: +For garbage_collector_attempt_to_orphan_queue_latency: +For garbage_collector_attempt_to_orphan_work_duration: +For garbage_collector_dirty_processing_latency_microseconds: +For garbage_collector_event_processing_latency_microseconds: +For garbage_collector_graph_changes_queue_latency: +For garbage_collector_graph_changes_work_duration: +For garbage_collector_orphan_processing_latency_microseconds: +For namespace_queue_latency: +For namespace_queue_latency_sum: +For namespace_queue_latency_count: +For namespace_retries: +For namespace_work_duration: +For namespace_work_duration_sum: +For namespace_work_duration_count: +For function_duration_seconds: +For errors_total: +For evicted_pods_total: + +[AfterEach] [sig-api-machinery] Garbage collector + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Apr 29 11:47:34.687: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "gc-4069" for this suite. +Apr 29 11:47:42.711: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Apr 29 11:47:42.974: INFO: namespace gc-4069 deletion completed in 8.28124769s + +• [SLOW TEST:18.710 seconds] +[sig-api-machinery] Garbage collector +/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:22 + should not delete dependents that have both valid owner and owner that's waiting for dependents to be deleted [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +------------------------------ +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +------------------------------ +[sig-apps] Deployment + deployment should delete old replica sets [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [sig-apps] Deployment + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +STEP: Creating a kubernetes client +Apr 29 11:47:43.000: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174 +STEP: Building a namespace api object, basename deployment +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in deployment-3168 +STEP: Waiting for a default service account to be provisioned in namespace +[BeforeEach] [sig-apps] Deployment + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:65 +[It] deployment should delete old replica sets [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +Apr 29 11:47:43.278: INFO: Pod name cleanup-pod: Found 0 pods out of 1 +Apr 29 11:47:48.284: INFO: Pod name cleanup-pod: Found 1 pods out of 1 +STEP: ensuring each pod is running +Apr 29 11:47:48.285: INFO: Creating deployment test-cleanup-deployment +STEP: Waiting for deployment test-cleanup-deployment history to be cleaned up +[AfterEach] [sig-apps] Deployment + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:59 +Apr 29 11:47:48.330: INFO: Deployment "test-cleanup-deployment": +&Deployment{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-cleanup-deployment,GenerateName:,Namespace:deployment-3168,SelfLink:/apis/apps/v1/namespaces/deployment-3168/deployments/test-cleanup-deployment,UID:9c2996e1-6a74-11e9-92a7-deadbef3c36d,ResourceVersion:7768,Generation:1,CreationTimestamp:2019-04-29 11:47:48 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: cleanup-pod,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:DeploymentSpec{Replicas:*1,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: cleanup-pod,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: cleanup-pod,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[],Containers:[{redis gcr.io/kubernetes-e2e-test-images/redis:1.0 [] [] [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,},},Strategy:DeploymentStrategy{Type:RollingUpdate,RollingUpdate:&RollingUpdateDeployment{MaxUnavailable:25%!,(MISSING)MaxSurge:25%!,(MISSING)},},MinReadySeconds:0,RevisionHistoryLimit:*0,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:0,Replicas:0,UpdatedReplicas:0,AvailableReplicas:0,UnavailableReplicas:0,Conditions:[],ReadyReplicas:0,CollisionCount:nil,},} + +Apr 29 11:47:48.338: INFO: New ReplicaSet of Deployment "test-cleanup-deployment" is nil. +Apr 29 11:47:48.338: INFO: All old ReplicaSets of Deployment "test-cleanup-deployment": +Apr 29 11:47:48.339: INFO: &ReplicaSet{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-cleanup-controller,GenerateName:,Namespace:deployment-3168,SelfLink:/apis/apps/v1/namespaces/deployment-3168/replicasets/test-cleanup-controller,UID:992a0238-6a74-11e9-92a7-deadbef3c36d,ResourceVersion:7769,Generation:1,CreationTimestamp:2019-04-29 11:47:42 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: cleanup-pod,pod: nginx,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 Deployment test-cleanup-deployment 9c2996e1-6a74-11e9-92a7-deadbef3c36d 0xc00214cdff 0xc00214ce10}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:ReplicaSetSpec{Replicas:*1,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: cleanup-pod,pod: nginx,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: cleanup-pod,pod: nginx,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,},},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:1,FullyLabeledReplicas:1,ObservedGeneration:1,ReadyReplicas:1,AvailableReplicas:1,Conditions:[],},} +Apr 29 11:47:48.350: INFO: Pod "test-cleanup-controller-lx4zt" is available: +&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-cleanup-controller-lx4zt,GenerateName:test-cleanup-controller-,Namespace:deployment-3168,SelfLink:/api/v1/namespaces/deployment-3168/pods/test-cleanup-controller-lx4zt,UID:992c8625-6a74-11e9-92a7-deadbef3c36d,ResourceVersion:7761,Generation:0,CreationTimestamp:2019-04-29 11:47:43 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: cleanup-pod,pod: nginx,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet test-cleanup-controller 992a0238-6a74-11e9-92a7-deadbef3c36d 0xc00214d36f 0xc00214d380}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-lc62d {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-lc62d,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-lc62d true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:worker-4av2h-6665d7b5ff-9nc42,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc00214d3e0} {node.kubernetes.io/unreachable Exists NoExecute 0xc00214d400}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Running,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:47:43 +0000 UTC } {Ready True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:47:45 +0000 UTC } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:47:45 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 11:47:43 +0000 UTC }],Message:,Reason:,HostIP:172.23.0.238,PodIP:192.168.205.106,StartTime:2019-04-29 11:47:43 +0000 UTC,ContainerStatuses:[{nginx {nil ContainerStateRunning{StartedAt:2019-04-29 11:47:45 +0000 UTC,} nil} {nil nil nil} true 0 nginx:1.14-alpine docker-pullable://nginx@sha256:485b610fefec7ff6c463ced9623314a04ed67e3945b9c08d7e53a47f6d108dc7 docker://eb1d2da057d619d0ee3cb08af1dd40acd54b8b6d30331254b6532d0542160479}],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} +[AfterEach] [sig-apps] Deployment + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Apr 29 11:47:48.350: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "deployment-3168" for this suite. +Apr 29 11:47:56.409: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Apr 29 11:47:56.661: INFO: namespace deployment-3168 deletion completed in 8.293627973s + +• [SLOW TEST:13.661 seconds] +[sig-apps] Deployment +/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22 + deployment should delete old replica sets [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +------------------------------ +SSSSSSSSSSSSSSSSS +------------------------------ +[sig-storage] Projected downwardAPI + should provide container's cpu request [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [sig-storage] Projected downwardAPI + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +STEP: Creating a kubernetes client +Apr 29 11:47:56.670: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174 +STEP: Building a namespace api object, basename projected +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-9797 +STEP: Waiting for a default service account to be provisioned in namespace +[BeforeEach] [sig-storage] Projected downwardAPI + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:39 +[It] should provide container's cpu request [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: Creating a pod to test downward API volume plugin +Apr 29 11:47:56.867: INFO: Waiting up to 5m0s for pod "downwardapi-volume-a16d5422-6a74-11e9-b6ef-0e71f8c7e888" in namespace "projected-9797" to be "success or failure" +Apr 29 11:47:56.874: INFO: Pod "downwardapi-volume-a16d5422-6a74-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 6.835594ms +Apr 29 11:47:58.882: INFO: Pod "downwardapi-volume-a16d5422-6a74-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 2.014874655s +Apr 29 11:48:00.889: INFO: Pod "downwardapi-volume-a16d5422-6a74-11e9-b6ef-0e71f8c7e888": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.022637574s +STEP: Saw pod success +Apr 29 11:48:00.889: INFO: Pod "downwardapi-volume-a16d5422-6a74-11e9-b6ef-0e71f8c7e888" satisfied condition "success or failure" +Apr 29 11:48:00.894: INFO: Trying to get logs from node worker-4av2h-6665d7b5ff-9nc42 pod downwardapi-volume-a16d5422-6a74-11e9-b6ef-0e71f8c7e888 container client-container: +STEP: delete the pod +Apr 29 11:48:00.931: INFO: Waiting for pod downwardapi-volume-a16d5422-6a74-11e9-b6ef-0e71f8c7e888 to disappear +Apr 29 11:48:00.934: INFO: Pod downwardapi-volume-a16d5422-6a74-11e9-b6ef-0e71f8c7e888 no longer exists +[AfterEach] [sig-storage] Projected downwardAPI + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Apr 29 11:48:00.934: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "projected-9797" for this suite. +Apr 29 11:48:06.966: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Apr 29 11:48:07.181: INFO: namespace projected-9797 deletion completed in 6.239820737s + +• [SLOW TEST:10.512 seconds] +[sig-storage] Projected downwardAPI +/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:33 + should provide container's cpu request [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +------------------------------ +SSSSSSSSSSSSSSSSSSS +------------------------------ +[k8s.io] Kubelet when scheduling a busybox command that always fails in a pod + should have an terminated reason [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [k8s.io] Kubelet + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +STEP: Creating a kubernetes client +Apr 29 11:48:07.185: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174 +STEP: Building a namespace api object, basename kubelet-test +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubelet-test-8469 +STEP: Waiting for a default service account to be provisioned in namespace +[BeforeEach] [k8s.io] Kubelet + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:37 +[BeforeEach] when scheduling a busybox command that always fails in a pod + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:81 +[It] should have an terminated reason [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[AfterEach] [k8s.io] Kubelet + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Apr 29 11:48:11.398: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "kubelet-test-8469" for this suite. +Apr 29 11:48:17.431: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Apr 29 11:48:17.684: INFO: namespace kubelet-test-8469 deletion completed in 6.278029854s + +• [SLOW TEST:10.499 seconds] +[k8s.io] Kubelet +/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687 + when scheduling a busybox command that always fails in a pod + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:78 + should have an terminated reason [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +------------------------------ +SSSSSSSS +------------------------------ +[k8s.io] KubeletManagedEtcHosts + should test kubelet managed /etc/hosts file [LinuxOnly] [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [k8s.io] KubeletManagedEtcHosts + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +STEP: Creating a kubernetes client +Apr 29 11:48:17.696: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174 +STEP: Building a namespace api object, basename e2e-kubelet-etc-hosts +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-kubelet-etc-hosts-1478 +STEP: Waiting for a default service account to be provisioned in namespace +[It] should test kubelet managed /etc/hosts file [LinuxOnly] [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: Setting up the test +STEP: Creating hostNetwork=false pod +STEP: Creating hostNetwork=true pod +STEP: Running the test +STEP: Verifying /etc/hosts of container is kubelet-managed for pod with hostNetwork=false +Apr 29 11:48:28.000: INFO: ExecWithOptions {Command:[cat /etc/hosts] Namespace:e2e-kubelet-etc-hosts-1478 PodName:test-pod ContainerName:busybox-1 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} +Apr 29 11:48:28.000: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174 +Apr 29 11:48:28.175: INFO: Exec stderr: "" +Apr 29 11:48:28.175: INFO: ExecWithOptions {Command:[cat /etc/hosts-original] Namespace:e2e-kubelet-etc-hosts-1478 PodName:test-pod ContainerName:busybox-1 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} +Apr 29 11:48:28.175: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174 +Apr 29 11:48:28.381: INFO: Exec stderr: "" +Apr 29 11:48:28.381: INFO: ExecWithOptions {Command:[cat /etc/hosts] Namespace:e2e-kubelet-etc-hosts-1478 PodName:test-pod ContainerName:busybox-2 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} +Apr 29 11:48:28.382: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174 +Apr 29 11:48:28.531: INFO: Exec stderr: "" +Apr 29 11:48:28.531: INFO: ExecWithOptions {Command:[cat /etc/hosts-original] Namespace:e2e-kubelet-etc-hosts-1478 PodName:test-pod ContainerName:busybox-2 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} +Apr 29 11:48:28.531: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174 +Apr 29 11:48:28.692: INFO: Exec stderr: "" +STEP: Verifying /etc/hosts of container is not kubelet-managed since container specifies /etc/hosts mount +Apr 29 11:48:28.692: INFO: ExecWithOptions {Command:[cat /etc/hosts] Namespace:e2e-kubelet-etc-hosts-1478 PodName:test-pod ContainerName:busybox-3 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} +Apr 29 11:48:28.692: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174 +Apr 29 11:48:28.857: INFO: Exec stderr: "" +Apr 29 11:48:28.857: INFO: ExecWithOptions {Command:[cat /etc/hosts-original] Namespace:e2e-kubelet-etc-hosts-1478 PodName:test-pod ContainerName:busybox-3 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} +Apr 29 11:48:28.857: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174 +Apr 29 11:48:29.071: INFO: Exec stderr: "" +STEP: Verifying /etc/hosts content of container is not kubelet-managed for pod with hostNetwork=true +Apr 29 11:48:29.072: INFO: ExecWithOptions {Command:[cat /etc/hosts] Namespace:e2e-kubelet-etc-hosts-1478 PodName:test-host-network-pod ContainerName:busybox-1 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} +Apr 29 11:48:29.072: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174 +Apr 29 11:48:29.246: INFO: Exec stderr: "" +Apr 29 11:48:29.247: INFO: ExecWithOptions {Command:[cat /etc/hosts-original] Namespace:e2e-kubelet-etc-hosts-1478 PodName:test-host-network-pod ContainerName:busybox-1 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} +Apr 29 11:48:29.247: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174 +Apr 29 11:48:29.423: INFO: Exec stderr: "" +Apr 29 11:48:29.424: INFO: ExecWithOptions {Command:[cat /etc/hosts] Namespace:e2e-kubelet-etc-hosts-1478 PodName:test-host-network-pod ContainerName:busybox-2 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} +Apr 29 11:48:29.424: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174 +Apr 29 11:48:29.624: INFO: Exec stderr: "" +Apr 29 11:48:29.624: INFO: ExecWithOptions {Command:[cat /etc/hosts-original] Namespace:e2e-kubelet-etc-hosts-1478 PodName:test-host-network-pod ContainerName:busybox-2 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} +Apr 29 11:48:29.624: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174 +Apr 29 11:48:29.832: INFO: Exec stderr: "" +[AfterEach] [k8s.io] KubeletManagedEtcHosts + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Apr 29 11:48:29.832: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "e2e-kubelet-etc-hosts-1478" for this suite. +Apr 29 11:49:25.860: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Apr 29 11:49:26.068: INFO: namespace e2e-kubelet-etc-hosts-1478 deletion completed in 56.226285543s + +• [SLOW TEST:68.374 seconds] +[k8s.io] KubeletManagedEtcHosts +/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687 + should test kubelet managed /etc/hosts file [LinuxOnly] [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +------------------------------ +SSSSSSSSSSSSSSSSSS +------------------------------ +[sig-cli] Kubectl client [k8s.io] Kubectl version + should check is all data is printed [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [sig-cli] Kubectl client + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +STEP: Creating a kubernetes client +Apr 29 11:49:26.070: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174 +STEP: Building a namespace api object, basename kubectl +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-3571 +STEP: Waiting for a default service account to be provisioned in namespace +[BeforeEach] [sig-cli] Kubectl client + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:213 +[It] should check is all data is printed [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +Apr 29 11:49:26.267: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 version' +Apr 29 11:49:26.392: INFO: stderr: "" +Apr 29 11:49:26.392: INFO: stdout: "Client Version: version.Info{Major:\"1\", Minor:\"14\", GitVersion:\"v1.14.1\", GitCommit:\"b7394102d6ef778017f2ca4046abbaa23b88c290\", GitTreeState:\"clean\", BuildDate:\"2019-04-08T17:11:31Z\", GoVersion:\"go1.12.1\", Compiler:\"gc\", Platform:\"linux/amd64\"}\nServer Version: version.Info{Major:\"1\", Minor:\"14\", GitVersion:\"v1.14.1\", GitCommit:\"b7394102d6ef778017f2ca4046abbaa23b88c290\", GitTreeState:\"clean\", BuildDate:\"2019-04-08T17:02:58Z\", GoVersion:\"go1.12.1\", Compiler:\"gc\", Platform:\"linux/amd64\"}\n" +[AfterEach] [sig-cli] Kubectl client + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Apr 29 11:49:26.393: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "kubectl-3571" for this suite. +Apr 29 11:49:32.423: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Apr 29 11:49:32.657: INFO: namespace kubectl-3571 deletion completed in 6.255841768s + +• [SLOW TEST:6.589 seconds] +[sig-cli] Kubectl client +/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 + [k8s.io] Kubectl version + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687 + should check is all data is printed [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +------------------------------ +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +------------------------------ +[sig-storage] ConfigMap + should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [sig-storage] ConfigMap + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +STEP: Creating a kubernetes client +Apr 29 11:49:32.659: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174 +STEP: Building a namespace api object, basename configmap +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in configmap-669 +STEP: Waiting for a default service account to be provisioned in namespace +[It] should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: Creating configMap with name configmap-test-volume-daaba56e-6a74-11e9-b6ef-0e71f8c7e888 +STEP: Creating a pod to test consume configMaps +Apr 29 11:49:32.920: INFO: Waiting up to 5m0s for pod "pod-configmaps-daad7ab4-6a74-11e9-b6ef-0e71f8c7e888" in namespace "configmap-669" to be "success or failure" +Apr 29 11:49:32.924: INFO: Pod "pod-configmaps-daad7ab4-6a74-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 4.416354ms +Apr 29 11:49:34.932: INFO: Pod "pod-configmaps-daad7ab4-6a74-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 2.012282653s +Apr 29 11:49:36.947: INFO: Pod "pod-configmaps-daad7ab4-6a74-11e9-b6ef-0e71f8c7e888": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.02726236s +STEP: Saw pod success +Apr 29 11:49:36.947: INFO: Pod "pod-configmaps-daad7ab4-6a74-11e9-b6ef-0e71f8c7e888" satisfied condition "success or failure" +Apr 29 11:49:36.954: INFO: Trying to get logs from node worker-4av2h-6665d7b5ff-9nc42 pod pod-configmaps-daad7ab4-6a74-11e9-b6ef-0e71f8c7e888 container configmap-volume-test: +STEP: delete the pod +Apr 29 11:49:37.008: INFO: Waiting for pod pod-configmaps-daad7ab4-6a74-11e9-b6ef-0e71f8c7e888 to disappear +Apr 29 11:49:37.020: INFO: Pod pod-configmaps-daad7ab4-6a74-11e9-b6ef-0e71f8c7e888 no longer exists +[AfterEach] [sig-storage] ConfigMap + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Apr 29 11:49:37.020: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "configmap-669" for this suite. +Apr 29 11:49:43.079: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Apr 29 11:49:43.343: INFO: namespace configmap-669 deletion completed in 6.300437598s + +• [SLOW TEST:10.685 seconds] +[sig-storage] ConfigMap +/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:32 + should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +------------------------------ +SSSSSSS +------------------------------ +[sig-storage] Downward API volume + should update labels on modification [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [sig-storage] Downward API volume + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +STEP: Creating a kubernetes client +Apr 29 11:49:43.350: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174 +STEP: Building a namespace api object, basename downward-api +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in downward-api-9664 +STEP: Waiting for a default service account to be provisioned in namespace +[BeforeEach] [sig-storage] Downward API volume + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:39 +[It] should update labels on modification [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: Creating the pod +Apr 29 11:49:48.138: INFO: Successfully updated pod "labelsupdatee1061aea-6a74-11e9-b6ef-0e71f8c7e888" +[AfterEach] [sig-storage] Downward API volume + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Apr 29 11:49:50.181: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "downward-api-9664" for this suite. +Apr 29 11:50:14.210: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Apr 29 11:50:14.488: INFO: namespace downward-api-9664 deletion completed in 24.298860921s + +• [SLOW TEST:31.140 seconds] +[sig-storage] Downward API volume +/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:34 + should update labels on modification [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +------------------------------ +SSSSSSSSSSSS +------------------------------ +[sig-apps] ReplicationController + should adopt matching pods on creation [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [sig-apps] ReplicationController + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +STEP: Creating a kubernetes client +Apr 29 11:50:14.489: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174 +STEP: Building a namespace api object, basename replication-controller +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in replication-controller-9615 +STEP: Waiting for a default service account to be provisioned in namespace +[It] should adopt matching pods on creation [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: Given a Pod with a 'name' label pod-adoption is created +STEP: When a replication controller with a matching selector is created +STEP: Then the orphan pod is adopted +[AfterEach] [sig-apps] ReplicationController + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Apr 29 11:50:19.761: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "replication-controller-9615" for this suite. +Apr 29 11:50:43.793: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Apr 29 11:50:44.061: INFO: namespace replication-controller-9615 deletion completed in 24.29056331s + +• [SLOW TEST:29.573 seconds] +[sig-apps] ReplicationController +/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22 + should adopt matching pods on creation [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +------------------------------ +SSSSSSSSS +------------------------------ +[sig-storage] Projected downwardAPI + should update labels on modification [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [sig-storage] Projected downwardAPI + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +STEP: Creating a kubernetes client +Apr 29 11:50:44.065: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174 +STEP: Building a namespace api object, basename projected +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-7569 +STEP: Waiting for a default service account to be provisioned in namespace +[BeforeEach] [sig-storage] Projected downwardAPI + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:39 +[It] should update labels on modification [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: Creating the pod +Apr 29 11:50:48.860: INFO: Successfully updated pod "labelsupdate05377e54-6a75-11e9-b6ef-0e71f8c7e888" +[AfterEach] [sig-storage] Projected downwardAPI + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Apr 29 11:50:50.902: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "projected-7569" for this suite. +Apr 29 11:51:14.929: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Apr 29 11:51:15.104: INFO: namespace projected-7569 deletion completed in 24.194870246s + +• [SLOW TEST:31.039 seconds] +[sig-storage] Projected downwardAPI +/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:33 + should update labels on modification [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +------------------------------ +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +------------------------------ +[sig-storage] Downward API volume + should provide container's cpu request [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [sig-storage] Downward API volume + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +STEP: Creating a kubernetes client +Apr 29 11:51:15.105: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174 +STEP: Building a namespace api object, basename downward-api +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in downward-api-6051 +STEP: Waiting for a default service account to be provisioned in namespace +[BeforeEach] [sig-storage] Downward API volume + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:39 +[It] should provide container's cpu request [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: Creating a pod to test downward API volume plugin +Apr 29 11:51:15.309: INFO: Waiting up to 5m0s for pod "downwardapi-volume-17b25d84-6a75-11e9-b6ef-0e71f8c7e888" in namespace "downward-api-6051" to be "success or failure" +Apr 29 11:51:15.329: INFO: Pod "downwardapi-volume-17b25d84-6a75-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 19.684098ms +Apr 29 11:51:17.337: INFO: Pod "downwardapi-volume-17b25d84-6a75-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 2.027525088s +Apr 29 11:51:19.344: INFO: Pod "downwardapi-volume-17b25d84-6a75-11e9-b6ef-0e71f8c7e888": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.034480988s +STEP: Saw pod success +Apr 29 11:51:19.344: INFO: Pod "downwardapi-volume-17b25d84-6a75-11e9-b6ef-0e71f8c7e888" satisfied condition "success or failure" +Apr 29 11:51:19.348: INFO: Trying to get logs from node worker-4av2h-6665d7b5ff-9nc42 pod downwardapi-volume-17b25d84-6a75-11e9-b6ef-0e71f8c7e888 container client-container: +STEP: delete the pod +Apr 29 11:51:19.385: INFO: Waiting for pod downwardapi-volume-17b25d84-6a75-11e9-b6ef-0e71f8c7e888 to disappear +Apr 29 11:51:19.389: INFO: Pod downwardapi-volume-17b25d84-6a75-11e9-b6ef-0e71f8c7e888 no longer exists +[AfterEach] [sig-storage] Downward API volume + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Apr 29 11:51:19.389: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "downward-api-6051" for this suite. +Apr 29 11:51:25.415: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Apr 29 11:51:25.628: INFO: namespace downward-api-6051 deletion completed in 6.231930535s + +• [SLOW TEST:10.523 seconds] +[sig-storage] Downward API volume +/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:34 + should provide container's cpu request [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +------------------------------ +SSS +------------------------------ +[sig-storage] EmptyDir volumes + should support (non-root,0666,default) [LinuxOnly] [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [sig-storage] EmptyDir volumes + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +STEP: Creating a kubernetes client +Apr 29 11:51:25.629: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174 +STEP: Building a namespace api object, basename emptydir +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in emptydir-1769 +STEP: Waiting for a default service account to be provisioned in namespace +[It] should support (non-root,0666,default) [LinuxOnly] [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: Creating a pod to test emptydir 0666 on node default medium +Apr 29 11:51:25.860: INFO: Waiting up to 5m0s for pod "pod-1dfd93ea-6a75-11e9-b6ef-0e71f8c7e888" in namespace "emptydir-1769" to be "success or failure" +Apr 29 11:51:25.865: INFO: Pod "pod-1dfd93ea-6a75-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 4.736641ms +Apr 29 11:51:27.875: INFO: Pod "pod-1dfd93ea-6a75-11e9-b6ef-0e71f8c7e888": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.015362095s +STEP: Saw pod success +Apr 29 11:51:27.876: INFO: Pod "pod-1dfd93ea-6a75-11e9-b6ef-0e71f8c7e888" satisfied condition "success or failure" +Apr 29 11:51:27.887: INFO: Trying to get logs from node worker-4av2h-6665d7b5ff-9nc42 pod pod-1dfd93ea-6a75-11e9-b6ef-0e71f8c7e888 container test-container: +STEP: delete the pod +Apr 29 11:51:27.939: INFO: Waiting for pod pod-1dfd93ea-6a75-11e9-b6ef-0e71f8c7e888 to disappear +Apr 29 11:51:27.951: INFO: Pod pod-1dfd93ea-6a75-11e9-b6ef-0e71f8c7e888 no longer exists +[AfterEach] [sig-storage] EmptyDir volumes + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Apr 29 11:51:27.951: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "emptydir-1769" for this suite. +Apr 29 11:51:34.091: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Apr 29 11:51:34.290: INFO: namespace emptydir-1769 deletion completed in 6.30348349s + +• [SLOW TEST:8.661 seconds] +[sig-storage] EmptyDir volumes +/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:41 + should support (non-root,0666,default) [LinuxOnly] [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +------------------------------ +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +------------------------------ +[k8s.io] Variable Expansion + should allow substituting values in a container's args [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [k8s.io] Variable Expansion + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +STEP: Creating a kubernetes client +Apr 29 11:51:34.306: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174 +STEP: Building a namespace api object, basename var-expansion +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in var-expansion-6854 +STEP: Waiting for a default service account to be provisioned in namespace +[It] should allow substituting values in a container's args [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: Creating a pod to test substitution in container's args +Apr 29 11:51:34.528: INFO: Waiting up to 5m0s for pod "var-expansion-2329846e-6a75-11e9-b6ef-0e71f8c7e888" in namespace "var-expansion-6854" to be "success or failure" +Apr 29 11:51:34.538: INFO: Pod "var-expansion-2329846e-6a75-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 8.913218ms +Apr 29 11:51:36.547: INFO: Pod "var-expansion-2329846e-6a75-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 2.01855565s +Apr 29 11:51:38.556: INFO: Pod "var-expansion-2329846e-6a75-11e9-b6ef-0e71f8c7e888": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.027454186s +STEP: Saw pod success +Apr 29 11:51:38.556: INFO: Pod "var-expansion-2329846e-6a75-11e9-b6ef-0e71f8c7e888" satisfied condition "success or failure" +Apr 29 11:51:38.562: INFO: Trying to get logs from node worker-4av2h-6665d7b5ff-9nc42 pod var-expansion-2329846e-6a75-11e9-b6ef-0e71f8c7e888 container dapi-container: +STEP: delete the pod +Apr 29 11:51:38.601: INFO: Waiting for pod var-expansion-2329846e-6a75-11e9-b6ef-0e71f8c7e888 to disappear +Apr 29 11:51:38.608: INFO: Pod var-expansion-2329846e-6a75-11e9-b6ef-0e71f8c7e888 no longer exists +[AfterEach] [k8s.io] Variable Expansion + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Apr 29 11:51:38.609: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "var-expansion-6854" for this suite. +Apr 29 11:51:44.642: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Apr 29 11:51:44.912: INFO: namespace var-expansion-6854 deletion completed in 6.291215209s + +• [SLOW TEST:10.607 seconds] +[k8s.io] Variable Expansion +/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687 + should allow substituting values in a container's args [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +------------------------------ +SSSSSSSSSSS +------------------------------ +[sig-api-machinery] Watchers + should observe an object deletion if it stops meeting the requirements of the selector [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [sig-api-machinery] Watchers + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +STEP: Creating a kubernetes client +Apr 29 11:51:44.914: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174 +STEP: Building a namespace api object, basename watch +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in watch-7054 +STEP: Waiting for a default service account to be provisioned in namespace +[It] should observe an object deletion if it stops meeting the requirements of the selector [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: creating a watch on configmaps with a certain label +STEP: creating a new configmap +STEP: modifying the configmap once +STEP: changing the label value of the configmap +STEP: Expecting to observe a delete notification for the watched object +Apr 29 11:51:45.143: INFO: Got : ADDED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-label-changed,GenerateName:,Namespace:watch-7054,SelfLink:/api/v1/namespaces/watch-7054/configmaps/e2e-watch-test-label-changed,UID:294efd45-6a75-11e9-92a7-deadbef3c36d,ResourceVersion:8643,Generation:0,CreationTimestamp:2019-04-29 11:51:44 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: label-changed-and-restored,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{},BinaryData:map[string][]byte{},} +Apr 29 11:51:45.143: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-label-changed,GenerateName:,Namespace:watch-7054,SelfLink:/api/v1/namespaces/watch-7054/configmaps/e2e-watch-test-label-changed,UID:294efd45-6a75-11e9-92a7-deadbef3c36d,ResourceVersion:8644,Generation:0,CreationTimestamp:2019-04-29 11:51:44 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: label-changed-and-restored,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{mutation: 1,},BinaryData:map[string][]byte{},} +Apr 29 11:51:45.143: INFO: Got : DELETED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-label-changed,GenerateName:,Namespace:watch-7054,SelfLink:/api/v1/namespaces/watch-7054/configmaps/e2e-watch-test-label-changed,UID:294efd45-6a75-11e9-92a7-deadbef3c36d,ResourceVersion:8645,Generation:0,CreationTimestamp:2019-04-29 11:51:44 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: label-changed-and-restored,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{mutation: 1,},BinaryData:map[string][]byte{},} +STEP: modifying the configmap a second time +STEP: Expecting not to observe a notification because the object no longer meets the selector's requirements +STEP: changing the label value of the configmap back +STEP: modifying the configmap a third time +STEP: deleting the configmap +STEP: Expecting to observe an add notification for the watched object when the label value was restored +Apr 29 11:51:55.201: INFO: Got : ADDED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-label-changed,GenerateName:,Namespace:watch-7054,SelfLink:/api/v1/namespaces/watch-7054/configmaps/e2e-watch-test-label-changed,UID:294efd45-6a75-11e9-92a7-deadbef3c36d,ResourceVersion:8665,Generation:0,CreationTimestamp:2019-04-29 11:51:44 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: label-changed-and-restored,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},} +Apr 29 11:51:55.201: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-label-changed,GenerateName:,Namespace:watch-7054,SelfLink:/api/v1/namespaces/watch-7054/configmaps/e2e-watch-test-label-changed,UID:294efd45-6a75-11e9-92a7-deadbef3c36d,ResourceVersion:8666,Generation:0,CreationTimestamp:2019-04-29 11:51:44 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: label-changed-and-restored,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{mutation: 3,},BinaryData:map[string][]byte{},} +Apr 29 11:51:55.201: INFO: Got : DELETED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-label-changed,GenerateName:,Namespace:watch-7054,SelfLink:/api/v1/namespaces/watch-7054/configmaps/e2e-watch-test-label-changed,UID:294efd45-6a75-11e9-92a7-deadbef3c36d,ResourceVersion:8667,Generation:0,CreationTimestamp:2019-04-29 11:51:44 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: label-changed-and-restored,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{mutation: 3,},BinaryData:map[string][]byte{},} +[AfterEach] [sig-api-machinery] Watchers + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Apr 29 11:51:55.201: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "watch-7054" for this suite. +Apr 29 11:52:01.241: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Apr 29 11:52:01.482: INFO: namespace watch-7054 deletion completed in 6.273265444s + +• [SLOW TEST:16.568 seconds] +[sig-api-machinery] Watchers +/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:22 + should observe an object deletion if it stops meeting the requirements of the selector [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +------------------------------ +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +------------------------------ +[sig-storage] Projected downwardAPI + should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [sig-storage] Projected downwardAPI + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +STEP: Creating a kubernetes client +Apr 29 11:52:01.486: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174 +STEP: Building a namespace api object, basename projected +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-6083 +STEP: Waiting for a default service account to be provisioned in namespace +[BeforeEach] [sig-storage] Projected downwardAPI + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:39 +[It] should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: Creating a pod to test downward API volume plugin +Apr 29 11:52:01.715: INFO: Waiting up to 5m0s for pod "downwardapi-volume-335d4794-6a75-11e9-b6ef-0e71f8c7e888" in namespace "projected-6083" to be "success or failure" +Apr 29 11:52:01.721: INFO: Pod "downwardapi-volume-335d4794-6a75-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 6.869942ms +Apr 29 11:52:03.739: INFO: Pod "downwardapi-volume-335d4794-6a75-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 2.024563943s +Apr 29 11:52:05.749: INFO: Pod "downwardapi-volume-335d4794-6a75-11e9-b6ef-0e71f8c7e888": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.034733408s +STEP: Saw pod success +Apr 29 11:52:05.749: INFO: Pod "downwardapi-volume-335d4794-6a75-11e9-b6ef-0e71f8c7e888" satisfied condition "success or failure" +Apr 29 11:52:05.756: INFO: Trying to get logs from node worker-4av2h-6665d7b5ff-9nc42 pod downwardapi-volume-335d4794-6a75-11e9-b6ef-0e71f8c7e888 container client-container: +STEP: delete the pod +Apr 29 11:52:05.786: INFO: Waiting for pod downwardapi-volume-335d4794-6a75-11e9-b6ef-0e71f8c7e888 to disappear +Apr 29 11:52:05.791: INFO: Pod downwardapi-volume-335d4794-6a75-11e9-b6ef-0e71f8c7e888 no longer exists +[AfterEach] [sig-storage] Projected downwardAPI + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Apr 29 11:52:05.791: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "projected-6083" for this suite. +Apr 29 11:52:11.827: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Apr 29 11:52:12.003: INFO: namespace projected-6083 deletion completed in 6.20575908s + +• [SLOW TEST:10.517 seconds] +[sig-storage] Projected downwardAPI +/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:33 + should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +------------------------------ +SS +------------------------------ +[sig-api-machinery] Garbage collector + should orphan pods created by rc if delete options say so [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [sig-api-machinery] Garbage collector + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +STEP: Creating a kubernetes client +Apr 29 11:52:12.005: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174 +STEP: Building a namespace api object, basename gc +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in gc-1892 +STEP: Waiting for a default service account to be provisioned in namespace +[It] should orphan pods created by rc if delete options say so [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: create the rc +STEP: delete the rc +STEP: wait for the rc to be deleted +STEP: wait for 30 seconds to see if the garbage collector mistakenly deletes the pods +STEP: Gathering metrics +W0429 11:52:52.268177 15 metrics_grabber.go:79] Master node is not registered. Grabbing metrics from Scheduler, ControllerManager and ClusterAutoscaler is disabled. +Apr 29 11:52:52.268: INFO: For apiserver_request_total: +For apiserver_request_latencies_summary: +For apiserver_init_events_total: +For garbage_collector_attempt_to_delete_queue_latency: +For garbage_collector_attempt_to_delete_work_duration: +For garbage_collector_attempt_to_orphan_queue_latency: +For garbage_collector_attempt_to_orphan_work_duration: +For garbage_collector_dirty_processing_latency_microseconds: +For garbage_collector_event_processing_latency_microseconds: +For garbage_collector_graph_changes_queue_latency: +For garbage_collector_graph_changes_work_duration: +For garbage_collector_orphan_processing_latency_microseconds: +For namespace_queue_latency: +For namespace_queue_latency_sum: +For namespace_queue_latency_count: +For namespace_retries: +For namespace_work_duration: +For namespace_work_duration_sum: +For namespace_work_duration_count: +For function_duration_seconds: +For errors_total: +For evicted_pods_total: + +[AfterEach] [sig-api-machinery] Garbage collector + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Apr 29 11:52:52.268: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "gc-1892" for this suite. +Apr 29 11:53:00.300: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Apr 29 11:53:00.599: INFO: namespace gc-1892 deletion completed in 8.325773841s + +• [SLOW TEST:48.594 seconds] +[sig-api-machinery] Garbage collector +/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:22 + should orphan pods created by rc if delete options say so [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +------------------------------ +S +------------------------------ +[sig-storage] Projected downwardAPI + should set mode on item file [LinuxOnly] [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [sig-storage] Projected downwardAPI + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +STEP: Creating a kubernetes client +Apr 29 11:53:00.602: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174 +STEP: Building a namespace api object, basename projected +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-3850 +STEP: Waiting for a default service account to be provisioned in namespace +[BeforeEach] [sig-storage] Projected downwardAPI + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:39 +[It] should set mode on item file [LinuxOnly] [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: Creating a pod to test downward API volume plugin +Apr 29 11:53:00.865: INFO: Waiting up to 5m0s for pod "downwardapi-volume-569e7978-6a75-11e9-b6ef-0e71f8c7e888" in namespace "projected-3850" to be "success or failure" +Apr 29 11:53:00.873: INFO: Pod "downwardapi-volume-569e7978-6a75-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 8.073524ms +Apr 29 11:53:02.884: INFO: Pod "downwardapi-volume-569e7978-6a75-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 2.019121656s +Apr 29 11:53:04.892: INFO: Pod "downwardapi-volume-569e7978-6a75-11e9-b6ef-0e71f8c7e888": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.027272452s +STEP: Saw pod success +Apr 29 11:53:04.893: INFO: Pod "downwardapi-volume-569e7978-6a75-11e9-b6ef-0e71f8c7e888" satisfied condition "success or failure" +Apr 29 11:53:04.898: INFO: Trying to get logs from node worker-4av2h-6665d7b5ff-9nc42 pod downwardapi-volume-569e7978-6a75-11e9-b6ef-0e71f8c7e888 container client-container: +STEP: delete the pod +Apr 29 11:53:04.946: INFO: Waiting for pod downwardapi-volume-569e7978-6a75-11e9-b6ef-0e71f8c7e888 to disappear +Apr 29 11:53:04.963: INFO: Pod downwardapi-volume-569e7978-6a75-11e9-b6ef-0e71f8c7e888 no longer exists +[AfterEach] [sig-storage] Projected downwardAPI + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Apr 29 11:53:04.963: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "projected-3850" for this suite. +Apr 29 11:53:10.993: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Apr 29 11:53:11.292: INFO: namespace projected-3850 deletion completed in 6.320219456s + +• [SLOW TEST:10.690 seconds] +[sig-storage] Projected downwardAPI +/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:33 + should set mode on item file [LinuxOnly] [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +------------------------------ +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +------------------------------ +[sig-node] ConfigMap + should fail to create ConfigMap with empty key [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [sig-node] ConfigMap + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +STEP: Creating a kubernetes client +Apr 29 11:53:11.293: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174 +STEP: Building a namespace api object, basename configmap +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in configmap-7163 +STEP: Waiting for a default service account to be provisioned in namespace +[It] should fail to create ConfigMap with empty key [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: Creating configMap that has name configmap-test-emptyKey-5cfe689b-6a75-11e9-b6ef-0e71f8c7e888 +[AfterEach] [sig-node] ConfigMap + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Apr 29 11:53:11.551: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "configmap-7163" for this suite. +Apr 29 11:53:17.589: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Apr 29 11:53:17.837: INFO: namespace configmap-7163 deletion completed in 6.27362046s + +• [SLOW TEST:6.544 seconds] +[sig-node] ConfigMap +/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/configmap.go:32 + should fail to create ConfigMap with empty key [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +------------------------------ +SSSSSSS +------------------------------ +[k8s.io] InitContainer [NodeConformance] + should invoke init containers on a RestartNever pod [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [k8s.io] InitContainer [NodeConformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +STEP: Creating a kubernetes client +Apr 29 11:53:17.840: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174 +STEP: Building a namespace api object, basename init-container +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in init-container-4931 +STEP: Waiting for a default service account to be provisioned in namespace +[BeforeEach] [k8s.io] InitContainer [NodeConformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/init_container.go:43 +[It] should invoke init containers on a RestartNever pod [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: creating the pod +Apr 29 11:53:18.038: INFO: PodSpec: initContainers in spec.initContainers +[AfterEach] [k8s.io] InitContainer [NodeConformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Apr 29 11:53:23.545: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "init-container-4931" for this suite. +Apr 29 11:53:29.585: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Apr 29 11:53:29.828: INFO: namespace init-container-4931 deletion completed in 6.268322275s + +• [SLOW TEST:11.989 seconds] +[k8s.io] InitContainer [NodeConformance] +/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687 + should invoke init containers on a RestartNever pod [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +------------------------------ +SSSSSSSSSSSSSS +------------------------------ +[sig-storage] Projected secret + should be consumable from pods in volume [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [sig-storage] Projected secret + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +STEP: Creating a kubernetes client +Apr 29 11:53:29.831: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174 +STEP: Building a namespace api object, basename projected +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-3425 +STEP: Waiting for a default service account to be provisioned in namespace +[It] should be consumable from pods in volume [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: Creating projection with secret that has name projected-secret-test-6806cc06-6a75-11e9-b6ef-0e71f8c7e888 +STEP: Creating a pod to test consume secrets +Apr 29 11:53:30.089: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-6809bbe2-6a75-11e9-b6ef-0e71f8c7e888" in namespace "projected-3425" to be "success or failure" +Apr 29 11:53:30.108: INFO: Pod "pod-projected-secrets-6809bbe2-6a75-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 18.879309ms +Apr 29 11:53:32.118: INFO: Pod "pod-projected-secrets-6809bbe2-6a75-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 2.028358085s +Apr 29 11:53:34.124: INFO: Pod "pod-projected-secrets-6809bbe2-6a75-11e9-b6ef-0e71f8c7e888": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.034743965s +STEP: Saw pod success +Apr 29 11:53:34.124: INFO: Pod "pod-projected-secrets-6809bbe2-6a75-11e9-b6ef-0e71f8c7e888" satisfied condition "success or failure" +Apr 29 11:53:34.129: INFO: Trying to get logs from node worker-4av2h-6665d7b5ff-9nc42 pod pod-projected-secrets-6809bbe2-6a75-11e9-b6ef-0e71f8c7e888 container projected-secret-volume-test: +STEP: delete the pod +Apr 29 11:53:34.163: INFO: Waiting for pod pod-projected-secrets-6809bbe2-6a75-11e9-b6ef-0e71f8c7e888 to disappear +Apr 29 11:53:34.167: INFO: Pod pod-projected-secrets-6809bbe2-6a75-11e9-b6ef-0e71f8c7e888 no longer exists +[AfterEach] [sig-storage] Projected secret + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Apr 29 11:53:34.167: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "projected-3425" for this suite. +Apr 29 11:53:40.191: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Apr 29 11:53:40.432: INFO: namespace projected-3425 deletion completed in 6.258214242s + +• [SLOW TEST:10.602 seconds] +[sig-storage] Projected secret +/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_secret.go:33 + should be consumable from pods in volume [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +------------------------------ +SSSSS +------------------------------ +[sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] + should perform canary updates and phased rolling updates of template modifications [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [sig-apps] StatefulSet + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +STEP: Creating a kubernetes client +Apr 29 11:53:40.438: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174 +STEP: Building a namespace api object, basename statefulset +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in statefulset-8071 +STEP: Waiting for a default service account to be provisioned in namespace +[BeforeEach] [sig-apps] StatefulSet + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:59 +[BeforeEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:74 +STEP: Creating service test in namespace statefulset-8071 +[It] should perform canary updates and phased rolling updates of template modifications [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: Creating a new StatefulSet +Apr 29 11:53:40.664: INFO: Found 0 stateful pods, waiting for 3 +Apr 29 11:53:50.672: INFO: Waiting for pod ss2-0 to enter Running - Ready=true, currently Running - Ready=true +Apr 29 11:53:50.673: INFO: Waiting for pod ss2-1 to enter Running - Ready=true, currently Running - Ready=true +Apr 29 11:53:50.673: INFO: Waiting for pod ss2-2 to enter Running - Ready=true, currently Running - Ready=false +Apr 29 11:54:00.671: INFO: Waiting for pod ss2-0 to enter Running - Ready=true, currently Running - Ready=true +Apr 29 11:54:00.671: INFO: Waiting for pod ss2-1 to enter Running - Ready=true, currently Running - Ready=true +Apr 29 11:54:00.671: INFO: Waiting for pod ss2-2 to enter Running - Ready=true, currently Running - Ready=true +STEP: Updating stateful set template: update image from docker.io/library/nginx:1.14-alpine to docker.io/library/nginx:1.15-alpine +Apr 29 11:54:00.706: INFO: Updating stateful set ss2 +STEP: Creating a new revision +STEP: Not applying an update when the partition is greater than the number of replicas +STEP: Performing a canary update +Apr 29 11:54:10.767: INFO: Updating stateful set ss2 +Apr 29 11:54:10.805: INFO: Waiting for Pod statefulset-8071/ss2-2 to have revision ss2-c79899b9 update revision ss2-787997d666 +STEP: Restoring Pods to the correct revision when they are deleted +Apr 29 11:54:20.983: INFO: Found 2 stateful pods, waiting for 3 +Apr 29 11:54:30.995: INFO: Waiting for pod ss2-0 to enter Running - Ready=true, currently Running - Ready=true +Apr 29 11:54:30.996: INFO: Waiting for pod ss2-1 to enter Running - Ready=true, currently Running - Ready=true +Apr 29 11:54:30.997: INFO: Waiting for pod ss2-2 to enter Running - Ready=true, currently Running - Ready=true +STEP: Performing a phased rolling update +Apr 29 11:54:31.045: INFO: Updating stateful set ss2 +Apr 29 11:54:31.082: INFO: Waiting for Pod statefulset-8071/ss2-1 to have revision ss2-c79899b9 update revision ss2-787997d666 +Apr 29 11:54:41.122: INFO: Updating stateful set ss2 +Apr 29 11:54:41.154: INFO: Waiting for StatefulSet statefulset-8071/ss2 to complete update +Apr 29 11:54:41.154: INFO: Waiting for Pod statefulset-8071/ss2-0 to have revision ss2-c79899b9 update revision ss2-787997d666 +Apr 29 11:54:51.174: INFO: Waiting for StatefulSet statefulset-8071/ss2 to complete update +Apr 29 11:54:51.174: INFO: Waiting for Pod statefulset-8071/ss2-0 to have revision ss2-c79899b9 update revision ss2-787997d666 +Apr 29 11:55:01.170: INFO: Waiting for StatefulSet statefulset-8071/ss2 to complete update +[AfterEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:85 +Apr 29 11:55:11.180: INFO: Deleting all statefulset in ns statefulset-8071 +Apr 29 11:55:11.185: INFO: Scaling statefulset ss2 to 0 +Apr 29 11:55:41.238: INFO: Waiting for statefulset status.replicas updated to 0 +Apr 29 11:55:41.245: INFO: Deleting statefulset ss2 +[AfterEach] [sig-apps] StatefulSet + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Apr 29 11:55:41.280: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "statefulset-8071" for this suite. +Apr 29 11:55:49.313: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Apr 29 11:55:49.544: INFO: namespace statefulset-8071 deletion completed in 8.252873015s + +• [SLOW TEST:129.106 seconds] +[sig-apps] StatefulSet +/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22 + [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687 + should perform canary updates and phased rolling updates of template modifications [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +------------------------------ +SSSSSSSSSSSSSSSSS +------------------------------ +[sig-network] Networking Granular Checks: Pods + should function for node-pod communication: udp [LinuxOnly] [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [sig-network] Networking + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +STEP: Creating a kubernetes client +Apr 29 11:55:49.544: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174 +STEP: Building a namespace api object, basename pod-network-test +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in pod-network-test-1969 +STEP: Waiting for a default service account to be provisioned in namespace +[It] should function for node-pod communication: udp [LinuxOnly] [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: Performing setup for networking test in namespace pod-network-test-1969 +STEP: creating a selector +STEP: Creating the service pods in kubernetes +Apr 29 11:55:49.749: INFO: Waiting up to 10m0s for all (but 0) nodes to be schedulable +STEP: Creating test pods +Apr 29 11:56:31.993: INFO: ExecWithOptions {Command:[/bin/sh -c echo hostName | nc -w 1 -u 192.168.14.29 8081 | grep -v '^\s*$'] Namespace:pod-network-test-1969 PodName:host-test-container-pod ContainerName:hostexec Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} +Apr 29 11:56:31.993: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174 +Apr 29 11:56:33.158: INFO: Found all expected endpoints: [netserver-0] +Apr 29 11:56:33.164: INFO: ExecWithOptions {Command:[/bin/sh -c echo hostName | nc -w 1 -u 192.168.205.66 8081 | grep -v '^\s*$'] Namespace:pod-network-test-1969 PodName:host-test-container-pod ContainerName:hostexec Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} +Apr 29 11:56:33.164: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174 +Apr 29 11:56:34.350: INFO: Found all expected endpoints: [netserver-1] +Apr 29 11:56:34.356: INFO: ExecWithOptions {Command:[/bin/sh -c echo hostName | nc -w 1 -u 192.168.148.94 8081 | grep -v '^\s*$'] Namespace:pod-network-test-1969 PodName:host-test-container-pod ContainerName:hostexec Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} +Apr 29 11:56:34.356: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174 +Apr 29 11:56:35.583: INFO: Found all expected endpoints: [netserver-2] +[AfterEach] [sig-network] Networking + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Apr 29 11:56:35.583: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "pod-network-test-1969" for this suite. +Apr 29 11:56:59.625: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Apr 29 11:56:59.878: INFO: namespace pod-network-test-1969 deletion completed in 24.286534107s + +• [SLOW TEST:70.334 seconds] +[sig-network] Networking +/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:25 + Granular Checks: Pods + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:28 + should function for node-pod communication: udp [LinuxOnly] [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +------------------------------ +S +------------------------------ +[sig-storage] ConfigMap + optional updates should be reflected in volume [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [sig-storage] ConfigMap + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +STEP: Creating a kubernetes client +Apr 29 11:56:59.878: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174 +STEP: Building a namespace api object, basename configmap +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in configmap-3906 +STEP: Waiting for a default service account to be provisioned in namespace +[It] optional updates should be reflected in volume [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: Creating configMap with name cm-test-opt-del-e53add79-6a75-11e9-b6ef-0e71f8c7e888 +STEP: Creating configMap with name cm-test-opt-upd-e53added-6a75-11e9-b6ef-0e71f8c7e888 +STEP: Creating the pod +STEP: Deleting configmap cm-test-opt-del-e53add79-6a75-11e9-b6ef-0e71f8c7e888 +STEP: Updating configmap cm-test-opt-upd-e53added-6a75-11e9-b6ef-0e71f8c7e888 +STEP: Creating configMap with name cm-test-opt-create-e53adeab-6a75-11e9-b6ef-0e71f8c7e888 +STEP: waiting to observe update in volume +[AfterEach] [sig-storage] ConfigMap + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Apr 29 11:57:08.311: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "configmap-3906" for this suite. +Apr 29 11:57:32.344: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Apr 29 11:57:32.556: INFO: namespace configmap-3906 deletion completed in 24.238727162s + +• [SLOW TEST:32.678 seconds] +[sig-storage] ConfigMap +/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:32 + optional updates should be reflected in volume [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +------------------------------ +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +------------------------------ +[sig-storage] Projected downwardAPI + should update annotations on modification [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [sig-storage] Projected downwardAPI + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +STEP: Creating a kubernetes client +Apr 29 11:57:32.570: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174 +STEP: Building a namespace api object, basename projected +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-9180 +STEP: Waiting for a default service account to be provisioned in namespace +[BeforeEach] [sig-storage] Projected downwardAPI + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:39 +[It] should update annotations on modification [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: Creating the pod +Apr 29 11:57:37.315: INFO: Successfully updated pod "annotationupdatef8add3e1-6a75-11e9-b6ef-0e71f8c7e888" +[AfterEach] [sig-storage] Projected downwardAPI + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Apr 29 11:57:39.348: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "projected-9180" for this suite. +Apr 29 11:57:59.380: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Apr 29 11:57:59.625: INFO: namespace projected-9180 deletion completed in 20.266635791s + +• [SLOW TEST:27.055 seconds] +[sig-storage] Projected downwardAPI +/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:33 + should update annotations on modification [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +------------------------------ +SSSSSSS +------------------------------ +[sig-cli] Kubectl client [k8s.io] Update Demo + should do a rolling update of a replication controller [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [sig-cli] Kubectl client + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +STEP: Creating a kubernetes client +Apr 29 11:57:59.632: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174 +STEP: Building a namespace api object, basename kubectl +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-7204 +STEP: Waiting for a default service account to be provisioned in namespace +[BeforeEach] [sig-cli] Kubectl client + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:213 +[BeforeEach] [k8s.io] Update Demo + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:265 +[It] should do a rolling update of a replication controller [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: creating the initial replication controller +Apr 29 11:57:59.832: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 create -f - --namespace=kubectl-7204' +Apr 29 11:58:02.635: INFO: stderr: "" +Apr 29 11:58:02.635: INFO: stdout: "replicationcontroller/update-demo-nautilus created\n" +STEP: waiting for all containers in name=update-demo pods to come up. +Apr 29 11:58:02.635: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-7204' +Apr 29 11:58:02.755: INFO: stderr: "" +Apr 29 11:58:02.755: INFO: stdout: "update-demo-nautilus-dxfsd update-demo-nautilus-pxcfg " +Apr 29 11:58:02.755: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 get pods update-demo-nautilus-dxfsd -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-7204' +Apr 29 11:58:02.871: INFO: stderr: "" +Apr 29 11:58:02.871: INFO: stdout: "" +Apr 29 11:58:02.871: INFO: update-demo-nautilus-dxfsd is created but not running +Apr 29 11:58:07.872: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-7204' +Apr 29 11:58:08.002: INFO: stderr: "" +Apr 29 11:58:08.002: INFO: stdout: "update-demo-nautilus-dxfsd update-demo-nautilus-pxcfg " +Apr 29 11:58:08.003: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 get pods update-demo-nautilus-dxfsd -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-7204' +Apr 29 11:58:08.128: INFO: stderr: "" +Apr 29 11:58:08.128: INFO: stdout: "" +Apr 29 11:58:08.128: INFO: update-demo-nautilus-dxfsd is created but not running +Apr 29 11:58:13.128: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-7204' +Apr 29 11:58:13.248: INFO: stderr: "" +Apr 29 11:58:13.248: INFO: stdout: "update-demo-nautilus-dxfsd update-demo-nautilus-pxcfg " +Apr 29 11:58:13.248: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 get pods update-demo-nautilus-dxfsd -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-7204' +Apr 29 11:58:13.382: INFO: stderr: "" +Apr 29 11:58:13.382: INFO: stdout: "true" +Apr 29 11:58:13.382: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 get pods update-demo-nautilus-dxfsd -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-7204' +Apr 29 11:58:13.523: INFO: stderr: "" +Apr 29 11:58:13.523: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0" +Apr 29 11:58:13.523: INFO: validating pod update-demo-nautilus-dxfsd +Apr 29 11:58:13.532: INFO: got data: { + "image": "nautilus.jpg" +} + +Apr 29 11:58:13.532: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . +Apr 29 11:58:13.532: INFO: update-demo-nautilus-dxfsd is verified up and running +Apr 29 11:58:13.532: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 get pods update-demo-nautilus-pxcfg -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-7204' +Apr 29 11:58:13.654: INFO: stderr: "" +Apr 29 11:58:13.654: INFO: stdout: "true" +Apr 29 11:58:13.655: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 get pods update-demo-nautilus-pxcfg -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-7204' +Apr 29 11:58:13.779: INFO: stderr: "" +Apr 29 11:58:13.779: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0" +Apr 29 11:58:13.779: INFO: validating pod update-demo-nautilus-pxcfg +Apr 29 11:58:13.792: INFO: got data: { + "image": "nautilus.jpg" +} + +Apr 29 11:58:13.792: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . +Apr 29 11:58:13.792: INFO: update-demo-nautilus-pxcfg is verified up and running +STEP: rolling-update to new replication controller +Apr 29 11:58:13.798: INFO: scanned /root for discovery docs: +Apr 29 11:58:13.798: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 rolling-update update-demo-nautilus --update-period=1s -f - --namespace=kubectl-7204' +Apr 29 11:58:38.606: INFO: stderr: "Command \"rolling-update\" is deprecated, use \"rollout\" instead\n" +Apr 29 11:58:38.606: INFO: stdout: "Created update-demo-kitten\nScaling up update-demo-kitten from 0 to 2, scaling down update-demo-nautilus from 2 to 0 (keep 2 pods available, don't exceed 3 pods)\nScaling update-demo-kitten up to 1\nScaling update-demo-nautilus down to 1\nScaling update-demo-kitten up to 2\nScaling update-demo-nautilus down to 0\nUpdate succeeded. Deleting old controller: update-demo-nautilus\nRenaming update-demo-kitten to update-demo-nautilus\nreplicationcontroller/update-demo-nautilus rolling updated\n" +STEP: waiting for all containers in name=update-demo pods to come up. +Apr 29 11:58:38.606: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-7204' +Apr 29 11:58:38.730: INFO: stderr: "" +Apr 29 11:58:38.730: INFO: stdout: "update-demo-kitten-qp7nh update-demo-kitten-zt7mx " +Apr 29 11:58:38.730: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 get pods update-demo-kitten-qp7nh -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-7204' +Apr 29 11:58:38.853: INFO: stderr: "" +Apr 29 11:58:38.853: INFO: stdout: "true" +Apr 29 11:58:38.853: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 get pods update-demo-kitten-qp7nh -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-7204' +Apr 29 11:58:38.966: INFO: stderr: "" +Apr 29 11:58:38.966: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/kitten:1.0" +Apr 29 11:58:38.966: INFO: validating pod update-demo-kitten-qp7nh +Apr 29 11:58:38.982: INFO: got data: { + "image": "kitten.jpg" +} + +Apr 29 11:58:38.982: INFO: Unmarshalled json jpg/img => {kitten.jpg} , expecting kitten.jpg . +Apr 29 11:58:38.982: INFO: update-demo-kitten-qp7nh is verified up and running +Apr 29 11:58:38.982: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 get pods update-demo-kitten-zt7mx -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-7204' +Apr 29 11:58:39.102: INFO: stderr: "" +Apr 29 11:58:39.102: INFO: stdout: "true" +Apr 29 11:58:39.102: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 get pods update-demo-kitten-zt7mx -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-7204' +Apr 29 11:58:39.220: INFO: stderr: "" +Apr 29 11:58:39.220: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/kitten:1.0" +Apr 29 11:58:39.220: INFO: validating pod update-demo-kitten-zt7mx +Apr 29 11:58:39.230: INFO: got data: { + "image": "kitten.jpg" +} + +Apr 29 11:58:39.230: INFO: Unmarshalled json jpg/img => {kitten.jpg} , expecting kitten.jpg . +Apr 29 11:58:39.230: INFO: update-demo-kitten-zt7mx is verified up and running +[AfterEach] [sig-cli] Kubectl client + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Apr 29 11:58:39.230: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "kubectl-7204" for this suite. +Apr 29 11:59:03.256: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Apr 29 11:59:03.514: INFO: namespace kubectl-7204 deletion completed in 24.276427488s + +• [SLOW TEST:63.883 seconds] +[sig-cli] Kubectl client +/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 + [k8s.io] Update Demo + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687 + should do a rolling update of a replication controller [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +------------------------------ +SSSSS +------------------------------ +[sig-storage] EmptyDir volumes + should support (non-root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [sig-storage] EmptyDir volumes + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +STEP: Creating a kubernetes client +Apr 29 11:59:03.515: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174 +STEP: Building a namespace api object, basename emptydir +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in emptydir-1278 +STEP: Waiting for a default service account to be provisioned in namespace +[It] should support (non-root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: Creating a pod to test emptydir 0777 on tmpfs +Apr 29 11:59:03.812: INFO: Waiting up to 5m0s for pod "pod-2ef09a56-6a76-11e9-b6ef-0e71f8c7e888" in namespace "emptydir-1278" to be "success or failure" +Apr 29 11:59:03.820: INFO: Pod "pod-2ef09a56-6a76-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 7.831492ms +Apr 29 11:59:05.828: INFO: Pod "pod-2ef09a56-6a76-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 2.01659204s +Apr 29 11:59:07.836: INFO: Pod "pod-2ef09a56-6a76-11e9-b6ef-0e71f8c7e888": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.023996666s +STEP: Saw pod success +Apr 29 11:59:07.836: INFO: Pod "pod-2ef09a56-6a76-11e9-b6ef-0e71f8c7e888" satisfied condition "success or failure" +Apr 29 11:59:07.840: INFO: Trying to get logs from node worker-4av2h-6665d7b5ff-9nc42 pod pod-2ef09a56-6a76-11e9-b6ef-0e71f8c7e888 container test-container: +STEP: delete the pod +Apr 29 11:59:07.872: INFO: Waiting for pod pod-2ef09a56-6a76-11e9-b6ef-0e71f8c7e888 to disappear +Apr 29 11:59:07.878: INFO: Pod pod-2ef09a56-6a76-11e9-b6ef-0e71f8c7e888 no longer exists +[AfterEach] [sig-storage] EmptyDir volumes + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Apr 29 11:59:07.879: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "emptydir-1278" for this suite. +Apr 29 11:59:13.922: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Apr 29 11:59:14.137: INFO: namespace emptydir-1278 deletion completed in 6.244404653s + +• [SLOW TEST:10.623 seconds] +[sig-storage] EmptyDir volumes +/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:41 + should support (non-root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +------------------------------ +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +------------------------------ +[sig-storage] Projected secret + should be consumable from pods in volume with mappings and Item Mode set [LinuxOnly] [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [sig-storage] Projected secret + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +STEP: Creating a kubernetes client +Apr 29 11:59:14.151: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174 +STEP: Building a namespace api object, basename projected +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-3637 +STEP: Waiting for a default service account to be provisioned in namespace +[It] should be consumable from pods in volume with mappings and Item Mode set [LinuxOnly] [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: Creating projection with secret that has name projected-secret-test-map-35403537-6a76-11e9-b6ef-0e71f8c7e888 +STEP: Creating a pod to test consume secrets +Apr 29 11:59:14.397: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-35415654-6a76-11e9-b6ef-0e71f8c7e888" in namespace "projected-3637" to be "success or failure" +Apr 29 11:59:14.411: INFO: Pod "pod-projected-secrets-35415654-6a76-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 13.230733ms +Apr 29 11:59:16.418: INFO: Pod "pod-projected-secrets-35415654-6a76-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 2.020325639s +Apr 29 11:59:18.426: INFO: Pod "pod-projected-secrets-35415654-6a76-11e9-b6ef-0e71f8c7e888": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.028337812s +STEP: Saw pod success +Apr 29 11:59:18.426: INFO: Pod "pod-projected-secrets-35415654-6a76-11e9-b6ef-0e71f8c7e888" satisfied condition "success or failure" +Apr 29 11:59:18.431: INFO: Trying to get logs from node worker-4av2h-6665d7b5ff-9nc42 pod pod-projected-secrets-35415654-6a76-11e9-b6ef-0e71f8c7e888 container projected-secret-volume-test: +STEP: delete the pod +Apr 29 11:59:18.459: INFO: Waiting for pod pod-projected-secrets-35415654-6a76-11e9-b6ef-0e71f8c7e888 to disappear +Apr 29 11:59:18.472: INFO: Pod pod-projected-secrets-35415654-6a76-11e9-b6ef-0e71f8c7e888 no longer exists +[AfterEach] [sig-storage] Projected secret + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Apr 29 11:59:18.473: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "projected-3637" for this suite. +Apr 29 11:59:24.514: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Apr 29 11:59:24.739: INFO: namespace projected-3637 deletion completed in 6.253908215s + +• [SLOW TEST:10.589 seconds] +[sig-storage] Projected secret +/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_secret.go:33 + should be consumable from pods in volume with mappings and Item Mode set [LinuxOnly] [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +------------------------------ +[sig-storage] Projected secret + should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [sig-storage] Projected secret + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +STEP: Creating a kubernetes client +Apr 29 11:59:24.740: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174 +STEP: Building a namespace api object, basename projected +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-5550 +STEP: Waiting for a default service account to be provisioned in namespace +[It] should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: Creating projection with secret that has name projected-secret-test-3b8def02-6a76-11e9-b6ef-0e71f8c7e888 +STEP: Creating a pod to test consume secrets +Apr 29 11:59:24.967: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-3b8f69ab-6a76-11e9-b6ef-0e71f8c7e888" in namespace "projected-5550" to be "success or failure" +Apr 29 11:59:24.980: INFO: Pod "pod-projected-secrets-3b8f69ab-6a76-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 12.930682ms +Apr 29 11:59:26.996: INFO: Pod "pod-projected-secrets-3b8f69ab-6a76-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 2.028926791s +Apr 29 11:59:29.003: INFO: Pod "pod-projected-secrets-3b8f69ab-6a76-11e9-b6ef-0e71f8c7e888": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.036233483s +STEP: Saw pod success +Apr 29 11:59:29.004: INFO: Pod "pod-projected-secrets-3b8f69ab-6a76-11e9-b6ef-0e71f8c7e888" satisfied condition "success or failure" +Apr 29 11:59:29.009: INFO: Trying to get logs from node worker-4av2h-6665d7b5ff-9nc42 pod pod-projected-secrets-3b8f69ab-6a76-11e9-b6ef-0e71f8c7e888 container projected-secret-volume-test: +STEP: delete the pod +Apr 29 11:59:29.050: INFO: Waiting for pod pod-projected-secrets-3b8f69ab-6a76-11e9-b6ef-0e71f8c7e888 to disappear +Apr 29 11:59:29.065: INFO: Pod pod-projected-secrets-3b8f69ab-6a76-11e9-b6ef-0e71f8c7e888 no longer exists +[AfterEach] [sig-storage] Projected secret + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Apr 29 11:59:29.066: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "projected-5550" for this suite. +Apr 29 11:59:35.124: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Apr 29 11:59:35.416: INFO: namespace projected-5550 deletion completed in 6.334020868s + +• [SLOW TEST:10.677 seconds] +[sig-storage] Projected secret +/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_secret.go:33 + should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +------------------------------ +SSSSSSSSSSSSSSSSS +------------------------------ +[sig-scheduling] SchedulerPredicates [Serial] + validates that NodeSelector is respected if matching [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [sig-scheduling] SchedulerPredicates [Serial] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +STEP: Creating a kubernetes client +Apr 29 11:59:35.428: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174 +STEP: Building a namespace api object, basename sched-pred +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in sched-pred-4679 +STEP: Waiting for a default service account to be provisioned in namespace +[BeforeEach] [sig-scheduling] SchedulerPredicates [Serial] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:79 +Apr 29 11:59:35.632: INFO: Waiting up to 1m0s for all (but 0) nodes to be ready +Apr 29 11:59:35.648: INFO: Waiting for terminating namespaces to be deleted... +Apr 29 11:59:35.653: INFO: +Logging pods the kubelet thinks is on node worker-3oo4n-86cbf586b7-9q98g before test +Apr 29 11:59:35.678: INFO: kube-state-metrics-5fdb649879-lqkfk from kube-system started at 2019-04-29 11:17:01 +0000 UTC (2 container statuses recorded) +Apr 29 11:59:35.679: INFO: Container addon-resizer ready: true, restart count 0 +Apr 29 11:59:35.679: INFO: Container kube-state-metrics ready: true, restart count 0 +Apr 29 11:59:35.679: INFO: coredns-58f7d854b4-xlcnp from kube-system started at 2019-04-29 11:16:23 +0000 UTC (1 container statuses recorded) +Apr 29 11:59:35.679: INFO: Container coredns ready: true, restart count 0 +Apr 29 11:59:35.680: INFO: cert-exporter-kgwcd from kube-system started at 2019-04-29 11:16:36 +0000 UTC (1 container statuses recorded) +Apr 29 11:59:35.680: INFO: Container cert-exporter ready: true, restart count 0 +Apr 29 11:59:35.680: INFO: calico-node-fmqgw from kube-system started at 2019-04-29 11:13:35 +0000 UTC (1 container statuses recorded) +Apr 29 11:59:35.680: INFO: Container calico-node ready: true, restart count 0 +Apr 29 11:59:35.680: INFO: kube-proxy-dw2qf from kube-system started at 2019-04-29 11:13:55 +0000 UTC (1 container statuses recorded) +Apr 29 11:59:35.680: INFO: Container kube-proxy ready: true, restart count 0 +Apr 29 11:59:35.680: INFO: node-exporter-j2kbr from kube-system started at 2019-04-29 11:21:42 +0000 UTC (1 container statuses recorded) +Apr 29 11:59:35.680: INFO: Container node-exporter ready: true, restart count 0 +Apr 29 11:59:35.680: INFO: nginx-ingress-controller-6cffb97fb5-5fbv2 from kube-system started at 2019-04-29 11:23:55 +0000 UTC (1 container statuses recorded) +Apr 29 11:59:35.680: INFO: Container nginx-ingress-controller ready: true, restart count 0 +Apr 29 11:59:35.681: INFO: net-exporter-wbqq5 from kube-system started at 2019-04-29 11:21:44 +0000 UTC (1 container statuses recorded) +Apr 29 11:59:35.681: INFO: Container net-exporter ready: true, restart count 0 +Apr 29 11:59:35.681: INFO: sonobuoy-systemd-logs-daemon-set-8191105fdead4396-6mc85 from heptio-sonobuoy started at 2019-04-29 11:30:45 +0000 UTC (2 container statuses recorded) +Apr 29 11:59:35.681: INFO: Container sonobuoy-worker ready: true, restart count 0 +Apr 29 11:59:35.681: INFO: Container systemd-logs ready: true, restart count 0 +Apr 29 11:59:35.681: INFO: +Logging pods the kubelet thinks is on node worker-4av2h-6665d7b5ff-9nc42 before test +Apr 29 11:59:35.706: INFO: cert-exporter-pn7vb from kube-system started at 2019-04-29 11:16:35 +0000 UTC (1 container statuses recorded) +Apr 29 11:59:35.706: INFO: Container cert-exporter ready: true, restart count 0 +Apr 29 11:59:35.706: INFO: node-exporter-pcjs5 from kube-system started at 2019-04-29 11:21:43 +0000 UTC (1 container statuses recorded) +Apr 29 11:59:35.706: INFO: Container node-exporter ready: true, restart count 0 +Apr 29 11:59:35.706: INFO: nginx-ingress-controller-6cffb97fb5-mfcvr from kube-system started at 2019-04-29 11:23:04 +0000 UTC (1 container statuses recorded) +Apr 29 11:59:35.707: INFO: Container nginx-ingress-controller ready: true, restart count 0 +Apr 29 11:59:35.707: INFO: tiller-deploy-54494c4fb6-k7mrg from giantswarm started at 2019-04-29 11:15:23 +0000 UTC (1 container statuses recorded) +Apr 29 11:59:35.707: INFO: Container tiller ready: true, restart count 0 +Apr 29 11:59:35.707: INFO: calico-node-k64gm from kube-system started at 2019-04-29 11:13:35 +0000 UTC (1 container statuses recorded) +Apr 29 11:59:35.707: INFO: Container calico-node ready: true, restart count 0 +Apr 29 11:59:35.707: INFO: kube-proxy-mf5pp from kube-system started at 2019-04-29 11:13:55 +0000 UTC (1 container statuses recorded) +Apr 29 11:59:35.708: INFO: Container kube-proxy ready: true, restart count 0 +Apr 29 11:59:35.708: INFO: net-exporter-vhll2 from kube-system started at 2019-04-29 11:21:44 +0000 UTC (1 container statuses recorded) +Apr 29 11:59:35.708: INFO: Container net-exporter ready: true, restart count 0 +Apr 29 11:59:35.708: INFO: default-http-backend-77b6545877-zd4s2 from kube-system started at 2019-04-29 11:21:34 +0000 UTC (1 container statuses recorded) +Apr 29 11:59:35.708: INFO: Container default-http-backend ready: true, restart count 0 +Apr 29 11:59:35.708: INFO: coredns-58f7d854b4-c4lkh from kube-system started at 2019-04-29 11:16:23 +0000 UTC (1 container statuses recorded) +Apr 29 11:59:35.708: INFO: Container coredns ready: true, restart count 0 +Apr 29 11:59:35.708: INFO: sonobuoy-systemd-logs-daemon-set-8191105fdead4396-gcldz from heptio-sonobuoy started at 2019-04-29 11:30:45 +0000 UTC (2 container statuses recorded) +Apr 29 11:59:35.708: INFO: Container sonobuoy-worker ready: true, restart count 0 +Apr 29 11:59:35.708: INFO: Container systemd-logs ready: true, restart count 0 +Apr 29 11:59:35.708: INFO: +Logging pods the kubelet thinks is on node worker-p95fb-54c845dd45-crdqs before test +Apr 29 11:59:35.729: INFO: cert-exporter-jkxfh from kube-system started at 2019-04-29 11:16:35 +0000 UTC (1 container statuses recorded) +Apr 29 11:59:35.729: INFO: Container cert-exporter ready: true, restart count 0 +Apr 29 11:59:35.730: INFO: kube-proxy-tppgw from kube-system started at 2019-04-29 11:13:55 +0000 UTC (1 container statuses recorded) +Apr 29 11:59:35.730: INFO: Container kube-proxy ready: true, restart count 0 +Apr 29 11:59:35.730: INFO: chart-operator-58979ccd8d-wvbzg from giantswarm started at 2019-04-29 11:16:10 +0000 UTC (1 container statuses recorded) +Apr 29 11:59:35.730: INFO: Container chart-operator ready: true, restart count 0 +Apr 29 11:59:35.730: INFO: node-exporter-98xpc from kube-system started at 2019-04-29 11:21:42 +0000 UTC (1 container statuses recorded) +Apr 29 11:59:35.730: INFO: Container node-exporter ready: true, restart count 0 +Apr 29 11:59:35.730: INFO: sonobuoy from heptio-sonobuoy started at 2019-04-29 11:30:35 +0000 UTC (1 container statuses recorded) +Apr 29 11:59:35.730: INFO: Container kube-sonobuoy ready: true, restart count 0 +Apr 29 11:59:35.730: INFO: sonobuoy-e2e-job-9cf815dd5d0c4d66 from heptio-sonobuoy started at 2019-04-29 11:30:45 +0000 UTC (2 container statuses recorded) +Apr 29 11:59:35.730: INFO: Container e2e ready: true, restart count 0 +Apr 29 11:59:35.730: INFO: Container sonobuoy-worker ready: true, restart count 0 +Apr 29 11:59:35.730: INFO: nginx-ingress-controller-6cffb97fb5-vqksl from kube-system started at 2019-04-29 11:21:47 +0000 UTC (1 container statuses recorded) +Apr 29 11:59:35.730: INFO: Container nginx-ingress-controller ready: true, restart count 0 +Apr 29 11:59:35.730: INFO: net-exporter-xcjtp from kube-system started at 2019-04-29 11:21:44 +0000 UTC (1 container statuses recorded) +Apr 29 11:59:35.731: INFO: Container net-exporter ready: true, restart count 0 +Apr 29 11:59:35.731: INFO: default-http-backend-77b6545877-jwz64 from kube-system started at 2019-04-29 11:22:25 +0000 UTC (1 container statuses recorded) +Apr 29 11:59:35.731: INFO: Container default-http-backend ready: true, restart count 0 +Apr 29 11:59:35.731: INFO: calico-node-hwth9 from kube-system started at 2019-04-29 11:13:35 +0000 UTC (1 container statuses recorded) +Apr 29 11:59:35.731: INFO: Container calico-node ready: true, restart count 0 +Apr 29 11:59:35.731: INFO: metrics-server-b94b95fb4-ncg25 from kube-system started at 2019-04-29 11:16:39 +0000 UTC (1 container statuses recorded) +Apr 29 11:59:35.731: INFO: Container metrics-server ready: true, restart count 0 +Apr 29 11:59:35.731: INFO: sonobuoy-systemd-logs-daemon-set-8191105fdead4396-w5l9x from heptio-sonobuoy started at 2019-04-29 11:30:45 +0000 UTC (2 container statuses recorded) +Apr 29 11:59:35.731: INFO: Container sonobuoy-worker ready: true, restart count 0 +Apr 29 11:59:35.731: INFO: Container systemd-logs ready: true, restart count 0 +[It] validates that NodeSelector is respected if matching [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: Trying to launch a pod without a label to get a node which can launch it. +STEP: Explicitly delete pod here to free the resource it takes. +STEP: Trying to apply a random label on the found node. +STEP: verifying the node has the label kubernetes.io/e2e-44704cb3-6a76-11e9-b6ef-0e71f8c7e888 42 +STEP: Trying to relaunch the pod, now with labels. +STEP: removing the label kubernetes.io/e2e-44704cb3-6a76-11e9-b6ef-0e71f8c7e888 off the node worker-4av2h-6665d7b5ff-9nc42 +STEP: verifying the node doesn't have the label kubernetes.io/e2e-44704cb3-6a76-11e9-b6ef-0e71f8c7e888 +[AfterEach] [sig-scheduling] SchedulerPredicates [Serial] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Apr 29 11:59:43.973: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "sched-pred-4679" for this suite. +Apr 29 11:59:54.029: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Apr 29 11:59:54.242: INFO: namespace sched-pred-4679 deletion completed in 10.243223168s +[AfterEach] [sig-scheduling] SchedulerPredicates [Serial] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:70 + +• [SLOW TEST:18.815 seconds] +[sig-scheduling] SchedulerPredicates [Serial] +/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/framework.go:22 + validates that NodeSelector is respected if matching [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +------------------------------ +SSSSSSSSSSSSSSSSSSSSSS +------------------------------ +[sig-api-machinery] Garbage collector + should delete RS created by deployment when not orphaning [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [sig-api-machinery] Garbage collector + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +STEP: Creating a kubernetes client +Apr 29 11:59:54.243: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174 +STEP: Building a namespace api object, basename gc +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in gc-1709 +STEP: Waiting for a default service account to be provisioned in namespace +[It] should delete RS created by deployment when not orphaning [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: create the deployment +STEP: Wait for the Deployment to create new ReplicaSet +STEP: delete the deployment +STEP: wait for all rs to be garbage collected +STEP: expected 0 rs, got 1 rs +STEP: expected 0 pods, got 2 pods +STEP: expected 0 rs, got 1 rs +STEP: expected 0 pods, got 2 pods +STEP: Gathering metrics +W0429 11:59:56.009585 15 metrics_grabber.go:79] Master node is not registered. Grabbing metrics from Scheduler, ControllerManager and ClusterAutoscaler is disabled. +Apr 29 11:59:56.009: INFO: For apiserver_request_total: +For apiserver_request_latencies_summary: +For apiserver_init_events_total: +For garbage_collector_attempt_to_delete_queue_latency: +For garbage_collector_attempt_to_delete_work_duration: +For garbage_collector_attempt_to_orphan_queue_latency: +For garbage_collector_attempt_to_orphan_work_duration: +For garbage_collector_dirty_processing_latency_microseconds: +For garbage_collector_event_processing_latency_microseconds: +For garbage_collector_graph_changes_queue_latency: +For garbage_collector_graph_changes_work_duration: +For garbage_collector_orphan_processing_latency_microseconds: +For namespace_queue_latency: +For namespace_queue_latency_sum: +For namespace_queue_latency_count: +For namespace_retries: +For namespace_work_duration: +For namespace_work_duration_sum: +For namespace_work_duration_count: +For function_duration_seconds: +For errors_total: +For evicted_pods_total: + +[AfterEach] [sig-api-machinery] Garbage collector + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Apr 29 11:59:56.010: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "gc-1709" for this suite. +Apr 29 12:00:02.054: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Apr 29 12:00:02.269: INFO: namespace gc-1709 deletion completed in 6.242280917s + +• [SLOW TEST:8.026 seconds] +[sig-api-machinery] Garbage collector +/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:22 + should delete RS created by deployment when not orphaning [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +------------------------------ +SSSSSSSSSSSSSSSSSSSSSSS +------------------------------ +[sig-storage] EmptyDir volumes + should support (root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [sig-storage] EmptyDir volumes + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +STEP: Creating a kubernetes client +Apr 29 12:00:02.270: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174 +STEP: Building a namespace api object, basename emptydir +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in emptydir-9243 +STEP: Waiting for a default service account to be provisioned in namespace +[It] should support (root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: Creating a pod to test emptydir 0777 on tmpfs +Apr 29 12:00:02.478: INFO: Waiting up to 5m0s for pod "pod-51eafb7e-6a76-11e9-b6ef-0e71f8c7e888" in namespace "emptydir-9243" to be "success or failure" +Apr 29 12:00:02.488: INFO: Pod "pod-51eafb7e-6a76-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 10.558319ms +Apr 29 12:00:04.497: INFO: Pod "pod-51eafb7e-6a76-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 2.019109896s +Apr 29 12:00:06.506: INFO: Pod "pod-51eafb7e-6a76-11e9-b6ef-0e71f8c7e888": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.028545763s +STEP: Saw pod success +Apr 29 12:00:06.507: INFO: Pod "pod-51eafb7e-6a76-11e9-b6ef-0e71f8c7e888" satisfied condition "success or failure" +Apr 29 12:00:06.512: INFO: Trying to get logs from node worker-4av2h-6665d7b5ff-9nc42 pod pod-51eafb7e-6a76-11e9-b6ef-0e71f8c7e888 container test-container: +STEP: delete the pod +Apr 29 12:00:06.561: INFO: Waiting for pod pod-51eafb7e-6a76-11e9-b6ef-0e71f8c7e888 to disappear +Apr 29 12:00:06.565: INFO: Pod pod-51eafb7e-6a76-11e9-b6ef-0e71f8c7e888 no longer exists +[AfterEach] [sig-storage] EmptyDir volumes + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Apr 29 12:00:06.566: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "emptydir-9243" for this suite. +Apr 29 12:00:12.599: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Apr 29 12:00:12.759: INFO: namespace emptydir-9243 deletion completed in 6.18162386s + +• [SLOW TEST:10.489 seconds] +[sig-storage] EmptyDir volumes +/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:41 + should support (root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +------------------------------ +SSSSSSSSSSSSSS +------------------------------ +[k8s.io] Docker Containers + should use the image defaults if command and args are blank [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [k8s.io] Docker Containers + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +STEP: Creating a kubernetes client +Apr 29 12:00:12.765: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174 +STEP: Building a namespace api object, basename containers +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in containers-5759 +STEP: Waiting for a default service account to be provisioned in namespace +[It] should use the image defaults if command and args are blank [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: Creating a pod to test use defaults +Apr 29 12:00:12.972: INFO: Waiting up to 5m0s for pod "client-containers-582de0f8-6a76-11e9-b6ef-0e71f8c7e888" in namespace "containers-5759" to be "success or failure" +Apr 29 12:00:12.976: INFO: Pod "client-containers-582de0f8-6a76-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 4.061561ms +Apr 29 12:00:14.983: INFO: Pod "client-containers-582de0f8-6a76-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 2.011346849s +Apr 29 12:00:16.992: INFO: Pod "client-containers-582de0f8-6a76-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 4.019997776s +Apr 29 12:00:18.998: INFO: Pod "client-containers-582de0f8-6a76-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 6.025558599s +Apr 29 12:00:21.004: INFO: Pod "client-containers-582de0f8-6a76-11e9-b6ef-0e71f8c7e888": Phase="Succeeded", Reason="", readiness=false. Elapsed: 8.032327942s +STEP: Saw pod success +Apr 29 12:00:21.004: INFO: Pod "client-containers-582de0f8-6a76-11e9-b6ef-0e71f8c7e888" satisfied condition "success or failure" +Apr 29 12:00:21.010: INFO: Trying to get logs from node worker-4av2h-6665d7b5ff-9nc42 pod client-containers-582de0f8-6a76-11e9-b6ef-0e71f8c7e888 container test-container: +STEP: delete the pod +Apr 29 12:00:21.041: INFO: Waiting for pod client-containers-582de0f8-6a76-11e9-b6ef-0e71f8c7e888 to disappear +Apr 29 12:00:21.047: INFO: Pod client-containers-582de0f8-6a76-11e9-b6ef-0e71f8c7e888 no longer exists +[AfterEach] [k8s.io] Docker Containers + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Apr 29 12:00:21.048: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "containers-5759" for this suite. +Apr 29 12:00:27.096: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Apr 29 12:00:27.296: INFO: namespace containers-5759 deletion completed in 6.231186687s + +• [SLOW TEST:14.531 seconds] +[k8s.io] Docker Containers +/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687 + should use the image defaults if command and args are blank [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +------------------------------ +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +------------------------------ +[k8s.io] InitContainer [NodeConformance] + should not start app containers if init containers fail on a RestartAlways pod [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [k8s.io] InitContainer [NodeConformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +STEP: Creating a kubernetes client +Apr 29 12:00:27.296: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174 +STEP: Building a namespace api object, basename init-container +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in init-container-3096 +STEP: Waiting for a default service account to be provisioned in namespace +[BeforeEach] [k8s.io] InitContainer [NodeConformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/init_container.go:43 +[It] should not start app containers if init containers fail on a RestartAlways pod [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: creating the pod +Apr 29 12:00:27.506: INFO: PodSpec: initContainers in spec.initContainers +Apr 29 12:01:11.048: INFO: init container has failed twice: &v1.Pod{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"pod-init-60d985e8-6a76-11e9-b6ef-0e71f8c7e888", GenerateName:"", Namespace:"init-container-3096", SelfLink:"/api/v1/namespaces/init-container-3096/pods/pod-init-60d985e8-6a76-11e9-b6ef-0e71f8c7e888", UID:"60adeeb7-6a76-11e9-92a7-deadbef3c36d", ResourceVersion:"11081", Generation:0, CreationTimestamp:v1.Time{Time:time.Time{wall:0x0, ext:63692136027, loc:(*time.Location)(0x8a060e0)}}, DeletionTimestamp:(*v1.Time)(nil), DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"name":"foo", "time":"506435375"}, Annotations:map[string]string{"kubernetes.io/psp":"cert-exporter-psp"}, OwnerReferences:[]v1.OwnerReference(nil), Initializers:(*v1.Initializers)(nil), Finalizers:[]string(nil), ClusterName:"", ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v1.PodSpec{Volumes:[]v1.Volume{v1.Volume{Name:"default-token-4mjlt", VolumeSource:v1.VolumeSource{HostPath:(*v1.HostPathVolumeSource)(nil), EmptyDir:(*v1.EmptyDirVolumeSource)(nil), GCEPersistentDisk:(*v1.GCEPersistentDiskVolumeSource)(nil), AWSElasticBlockStore:(*v1.AWSElasticBlockStoreVolumeSource)(nil), GitRepo:(*v1.GitRepoVolumeSource)(nil), Secret:(*v1.SecretVolumeSource)(0xc001e78180), NFS:(*v1.NFSVolumeSource)(nil), ISCSI:(*v1.ISCSIVolumeSource)(nil), Glusterfs:(*v1.GlusterfsVolumeSource)(nil), PersistentVolumeClaim:(*v1.PersistentVolumeClaimVolumeSource)(nil), RBD:(*v1.RBDVolumeSource)(nil), FlexVolume:(*v1.FlexVolumeSource)(nil), Cinder:(*v1.CinderVolumeSource)(nil), CephFS:(*v1.CephFSVolumeSource)(nil), Flocker:(*v1.FlockerVolumeSource)(nil), DownwardAPI:(*v1.DownwardAPIVolumeSource)(nil), FC:(*v1.FCVolumeSource)(nil), AzureFile:(*v1.AzureFileVolumeSource)(nil), ConfigMap:(*v1.ConfigMapVolumeSource)(nil), VsphereVolume:(*v1.VsphereVirtualDiskVolumeSource)(nil), Quobyte:(*v1.QuobyteVolumeSource)(nil), AzureDisk:(*v1.AzureDiskVolumeSource)(nil), PhotonPersistentDisk:(*v1.PhotonPersistentDiskVolumeSource)(nil), Projected:(*v1.ProjectedVolumeSource)(nil), PortworxVolume:(*v1.PortworxVolumeSource)(nil), ScaleIO:(*v1.ScaleIOVolumeSource)(nil), StorageOS:(*v1.StorageOSVolumeSource)(nil), CSI:(*v1.CSIVolumeSource)(nil)}}}, InitContainers:[]v1.Container{v1.Container{Name:"init1", Image:"docker.io/library/busybox:1.29", Command:[]string{"/bin/false"}, Args:[]string(nil), WorkingDir:"", Ports:[]v1.ContainerPort(nil), EnvFrom:[]v1.EnvFromSource(nil), Env:[]v1.EnvVar(nil), Resources:v1.ResourceRequirements{Limits:v1.ResourceList(nil), Requests:v1.ResourceList(nil)}, VolumeMounts:[]v1.VolumeMount{v1.VolumeMount{Name:"default-token-4mjlt", ReadOnly:true, MountPath:"/var/run/secrets/kubernetes.io/serviceaccount", SubPath:"", MountPropagation:(*v1.MountPropagationMode)(nil), SubPathExpr:""}}, VolumeDevices:[]v1.VolumeDevice(nil), LivenessProbe:(*v1.Probe)(nil), ReadinessProbe:(*v1.Probe)(nil), Lifecycle:(*v1.Lifecycle)(nil), TerminationMessagePath:"/dev/termination-log", TerminationMessagePolicy:"File", ImagePullPolicy:"IfNotPresent", SecurityContext:(*v1.SecurityContext)(nil), Stdin:false, StdinOnce:false, TTY:false}, v1.Container{Name:"init2", Image:"docker.io/library/busybox:1.29", Command:[]string{"/bin/true"}, Args:[]string(nil), WorkingDir:"", Ports:[]v1.ContainerPort(nil), EnvFrom:[]v1.EnvFromSource(nil), Env:[]v1.EnvVar(nil), Resources:v1.ResourceRequirements{Limits:v1.ResourceList(nil), Requests:v1.ResourceList(nil)}, VolumeMounts:[]v1.VolumeMount{v1.VolumeMount{Name:"default-token-4mjlt", ReadOnly:true, MountPath:"/var/run/secrets/kubernetes.io/serviceaccount", SubPath:"", MountPropagation:(*v1.MountPropagationMode)(nil), SubPathExpr:""}}, VolumeDevices:[]v1.VolumeDevice(nil), LivenessProbe:(*v1.Probe)(nil), ReadinessProbe:(*v1.Probe)(nil), Lifecycle:(*v1.Lifecycle)(nil), TerminationMessagePath:"/dev/termination-log", TerminationMessagePolicy:"File", ImagePullPolicy:"IfNotPresent", SecurityContext:(*v1.SecurityContext)(nil), Stdin:false, StdinOnce:false, TTY:false}}, Containers:[]v1.Container{v1.Container{Name:"run1", Image:"k8s.gcr.io/pause:3.1", Command:[]string(nil), Args:[]string(nil), WorkingDir:"", Ports:[]v1.ContainerPort(nil), EnvFrom:[]v1.EnvFromSource(nil), Env:[]v1.EnvVar(nil), Resources:v1.ResourceRequirements{Limits:v1.ResourceList{"cpu":resource.Quantity{i:resource.int64Amount{value:100, scale:-3}, d:resource.infDecAmount{Dec:(*inf.Dec)(nil)}, s:"100m", Format:"DecimalSI"}, "memory":resource.Quantity{i:resource.int64Amount{value:52428800, scale:0}, d:resource.infDecAmount{Dec:(*inf.Dec)(nil)}, s:"52428800", Format:"DecimalSI"}}, Requests:v1.ResourceList{"cpu":resource.Quantity{i:resource.int64Amount{value:100, scale:-3}, d:resource.infDecAmount{Dec:(*inf.Dec)(nil)}, s:"100m", Format:"DecimalSI"}, "memory":resource.Quantity{i:resource.int64Amount{value:52428800, scale:0}, d:resource.infDecAmount{Dec:(*inf.Dec)(nil)}, s:"52428800", Format:"DecimalSI"}}}, VolumeMounts:[]v1.VolumeMount{v1.VolumeMount{Name:"default-token-4mjlt", ReadOnly:true, MountPath:"/var/run/secrets/kubernetes.io/serviceaccount", SubPath:"", MountPropagation:(*v1.MountPropagationMode)(nil), SubPathExpr:""}}, VolumeDevices:[]v1.VolumeDevice(nil), LivenessProbe:(*v1.Probe)(nil), ReadinessProbe:(*v1.Probe)(nil), Lifecycle:(*v1.Lifecycle)(nil), TerminationMessagePath:"/dev/termination-log", TerminationMessagePolicy:"File", ImagePullPolicy:"IfNotPresent", SecurityContext:(*v1.SecurityContext)(nil), Stdin:false, StdinOnce:false, TTY:false}}, RestartPolicy:"Always", TerminationGracePeriodSeconds:(*int64)(0xc001d6a288), ActiveDeadlineSeconds:(*int64)(nil), DNSPolicy:"ClusterFirst", NodeSelector:map[string]string(nil), ServiceAccountName:"default", DeprecatedServiceAccount:"default", AutomountServiceAccountToken:(*bool)(nil), NodeName:"worker-4av2h-6665d7b5ff-9nc42", HostNetwork:false, HostPID:false, HostIPC:false, ShareProcessNamespace:(*bool)(nil), SecurityContext:(*v1.PodSecurityContext)(0xc002a0e060), ImagePullSecrets:[]v1.LocalObjectReference(nil), Hostname:"", Subdomain:"", Affinity:(*v1.Affinity)(nil), SchedulerName:"default-scheduler", Tolerations:[]v1.Toleration{v1.Toleration{Key:"node.kubernetes.io/not-ready", Operator:"Exists", Value:"", Effect:"NoExecute", TolerationSeconds:(*int64)(0xc001d6a300)}, v1.Toleration{Key:"node.kubernetes.io/unreachable", Operator:"Exists", Value:"", Effect:"NoExecute", TolerationSeconds:(*int64)(0xc001d6a320)}}, HostAliases:[]v1.HostAlias(nil), PriorityClassName:"", Priority:(*int32)(0xc001d6a328), DNSConfig:(*v1.PodDNSConfig)(nil), ReadinessGates:[]v1.PodReadinessGate(nil), RuntimeClassName:(*string)(nil), EnableServiceLinks:(*bool)(0xc001d6a32c)}, Status:v1.PodStatus{Phase:"Pending", Conditions:[]v1.PodCondition{v1.PodCondition{Type:"Initialized", Status:"False", LastProbeTime:v1.Time{Time:time.Time{wall:0x0, ext:0, loc:(*time.Location)(nil)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63692136027, loc:(*time.Location)(0x8a060e0)}}, Reason:"ContainersNotInitialized", Message:"containers with incomplete status: [init1 init2]"}, v1.PodCondition{Type:"Ready", Status:"False", LastProbeTime:v1.Time{Time:time.Time{wall:0x0, ext:0, loc:(*time.Location)(nil)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63692136027, loc:(*time.Location)(0x8a060e0)}}, Reason:"ContainersNotReady", Message:"containers with unready status: [run1]"}, v1.PodCondition{Type:"ContainersReady", Status:"False", LastProbeTime:v1.Time{Time:time.Time{wall:0x0, ext:0, loc:(*time.Location)(nil)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63692136027, loc:(*time.Location)(0x8a060e0)}}, Reason:"ContainersNotReady", Message:"containers with unready status: [run1]"}, v1.PodCondition{Type:"PodScheduled", Status:"True", LastProbeTime:v1.Time{Time:time.Time{wall:0x0, ext:0, loc:(*time.Location)(nil)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63692136027, loc:(*time.Location)(0x8a060e0)}}, Reason:"", Message:""}}, Message:"", Reason:"", NominatedNodeName:"", HostIP:"172.23.0.238", PodIP:"192.168.205.84", StartTime:(*v1.Time)(0xc00292a1c0), InitContainerStatuses:[]v1.ContainerStatus{v1.ContainerStatus{Name:"init1", State:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(nil), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(0xc000312e00)}, LastTerminationState:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(nil), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(0xc000312e70)}, Ready:false, RestartCount:3, Image:"busybox:1.29", ImageID:"docker-pullable://busybox@sha256:8ccbac733d19c0dd4d70b4f0c1e12245b5fa3ad24758a11035ee505c629c0796", ContainerID:"docker://ec05eb21922392357ca4361728a090b61325363d910ff538fb2d503406d8ceba"}, v1.ContainerStatus{Name:"init2", State:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(0xc00292a200), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(nil)}, LastTerminationState:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(nil), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(nil)}, Ready:false, RestartCount:0, Image:"docker.io/library/busybox:1.29", ImageID:"", ContainerID:""}}, ContainerStatuses:[]v1.ContainerStatus{v1.ContainerStatus{Name:"run1", State:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(0xc00292a1e0), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(nil)}, LastTerminationState:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(nil), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(nil)}, Ready:false, RestartCount:0, Image:"k8s.gcr.io/pause:3.1", ImageID:"", ContainerID:""}}, QOSClass:"Guaranteed"}} +[AfterEach] [k8s.io] InitContainer [NodeConformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Apr 29 12:01:11.049: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "init-container-3096" for this suite. +Apr 29 12:01:51.112: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Apr 29 12:01:51.322: INFO: namespace init-container-3096 deletion completed in 40.249634931s + +• [SLOW TEST:84.026 seconds] +[k8s.io] InitContainer [NodeConformance] +/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687 + should not start app containers if init containers fail on a RestartAlways pod [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +------------------------------ +SSSSSSSSS +------------------------------ +[k8s.io] Probing container + should be restarted with a /healthz http liveness probe [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [k8s.io] Probing container + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +STEP: Creating a kubernetes client +Apr 29 12:01:51.325: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174 +STEP: Building a namespace api object, basename container-probe +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in container-probe-8730 +STEP: Waiting for a default service account to be provisioned in namespace +[BeforeEach] [k8s.io] Probing container + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/container_probe.go:51 +[It] should be restarted with a /healthz http liveness probe [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: Creating pod liveness-http in namespace container-probe-8730 +Apr 29 12:01:55.580: INFO: Started pod liveness-http in namespace container-probe-8730 +STEP: checking the pod's current state and verifying that restartCount is present +Apr 29 12:01:55.586: INFO: Initial restart count of pod liveness-http is 0 +Apr 29 12:02:13.665: INFO: Restart count of pod container-probe-8730/liveness-http is now 1 (18.078563585s elapsed) +STEP: deleting the pod +[AfterEach] [k8s.io] Probing container + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Apr 29 12:02:13.679: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "container-probe-8730" for this suite. +Apr 29 12:02:19.731: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Apr 29 12:02:19.925: INFO: namespace container-probe-8730 deletion completed in 6.228497123s + +• [SLOW TEST:28.603 seconds] +[k8s.io] Probing container +/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687 + should be restarted with a /healthz http liveness probe [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +------------------------------ +S +------------------------------ +[sig-storage] Downward API volume + should provide podname only [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [sig-storage] Downward API volume + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +STEP: Creating a kubernetes client +Apr 29 12:02:19.931: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174 +STEP: Building a namespace api object, basename downward-api +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in downward-api-7449 +STEP: Waiting for a default service account to be provisioned in namespace +[BeforeEach] [sig-storage] Downward API volume + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:39 +[It] should provide podname only [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: Creating a pod to test downward API volume plugin +Apr 29 12:02:20.188: INFO: Waiting up to 5m0s for pod "downwardapi-volume-a401064e-6a76-11e9-b6ef-0e71f8c7e888" in namespace "downward-api-7449" to be "success or failure" +Apr 29 12:02:20.205: INFO: Pod "downwardapi-volume-a401064e-6a76-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 16.79821ms +Apr 29 12:02:22.212: INFO: Pod "downwardapi-volume-a401064e-6a76-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 2.024620543s +Apr 29 12:02:24.220: INFO: Pod "downwardapi-volume-a401064e-6a76-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 4.03270125s +Apr 29 12:02:26.230: INFO: Pod "downwardapi-volume-a401064e-6a76-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 6.042214733s +Apr 29 12:02:28.237: INFO: Pod "downwardapi-volume-a401064e-6a76-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 8.049145152s +Apr 29 12:02:30.256: INFO: Pod "downwardapi-volume-a401064e-6a76-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 10.068190681s +Apr 29 12:02:32.262: INFO: Pod "downwardapi-volume-a401064e-6a76-11e9-b6ef-0e71f8c7e888": Phase="Succeeded", Reason="", readiness=false. Elapsed: 12.074459921s +STEP: Saw pod success +Apr 29 12:02:32.262: INFO: Pod "downwardapi-volume-a401064e-6a76-11e9-b6ef-0e71f8c7e888" satisfied condition "success or failure" +Apr 29 12:02:32.265: INFO: Trying to get logs from node worker-4av2h-6665d7b5ff-9nc42 pod downwardapi-volume-a401064e-6a76-11e9-b6ef-0e71f8c7e888 container client-container: +STEP: delete the pod +Apr 29 12:02:32.295: INFO: Waiting for pod downwardapi-volume-a401064e-6a76-11e9-b6ef-0e71f8c7e888 to disappear +Apr 29 12:02:32.303: INFO: Pod downwardapi-volume-a401064e-6a76-11e9-b6ef-0e71f8c7e888 no longer exists +[AfterEach] [sig-storage] Downward API volume + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Apr 29 12:02:32.303: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "downward-api-7449" for this suite. +Apr 29 12:02:38.333: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Apr 29 12:02:38.512: INFO: namespace downward-api-7449 deletion completed in 6.20236996s + +• [SLOW TEST:18.582 seconds] +[sig-storage] Downward API volume +/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:34 + should provide podname only [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +------------------------------ +SSSSSSSS +------------------------------ +[k8s.io] Pods + should contain environment variables for services [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [k8s.io] Pods + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +STEP: Creating a kubernetes client +Apr 29 12:02:38.513: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174 +STEP: Building a namespace api object, basename pods +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in pods-6304 +STEP: Waiting for a default service account to be provisioned in namespace +[BeforeEach] [k8s.io] Pods + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:135 +[It] should contain environment variables for services [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +Apr 29 12:02:44.788: INFO: Waiting up to 5m0s for pod "client-envvars-b2ab282e-6a76-11e9-b6ef-0e71f8c7e888" in namespace "pods-6304" to be "success or failure" +Apr 29 12:02:44.798: INFO: Pod "client-envvars-b2ab282e-6a76-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 10.312567ms +Apr 29 12:02:46.804: INFO: Pod "client-envvars-b2ab282e-6a76-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 2.01632131s +Apr 29 12:02:48.811: INFO: Pod "client-envvars-b2ab282e-6a76-11e9-b6ef-0e71f8c7e888": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.023388755s +STEP: Saw pod success +Apr 29 12:02:48.811: INFO: Pod "client-envvars-b2ab282e-6a76-11e9-b6ef-0e71f8c7e888" satisfied condition "success or failure" +Apr 29 12:02:48.816: INFO: Trying to get logs from node worker-4av2h-6665d7b5ff-9nc42 pod client-envvars-b2ab282e-6a76-11e9-b6ef-0e71f8c7e888 container env3cont: +STEP: delete the pod +Apr 29 12:02:48.849: INFO: Waiting for pod client-envvars-b2ab282e-6a76-11e9-b6ef-0e71f8c7e888 to disappear +Apr 29 12:02:48.860: INFO: Pod client-envvars-b2ab282e-6a76-11e9-b6ef-0e71f8c7e888 no longer exists +[AfterEach] [k8s.io] Pods + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Apr 29 12:02:48.860: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "pods-6304" for this suite. +Apr 29 12:03:36.891: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Apr 29 12:03:37.120: INFO: namespace pods-6304 deletion completed in 48.250914809s + +• [SLOW TEST:58.608 seconds] +[k8s.io] Pods +/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687 + should contain environment variables for services [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +------------------------------ +SSSSS +------------------------------ +[sig-cli] Kubectl client [k8s.io] Kubectl rolling-update + should support rolling-update to same image [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [sig-cli] Kubectl client + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +STEP: Creating a kubernetes client +Apr 29 12:03:37.120: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174 +STEP: Building a namespace api object, basename kubectl +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-6955 +STEP: Waiting for a default service account to be provisioned in namespace +[BeforeEach] [sig-cli] Kubectl client + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:213 +[BeforeEach] [k8s.io] Kubectl rolling-update + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1414 +[It] should support rolling-update to same image [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: running the image docker.io/library/nginx:1.14-alpine +Apr 29 12:03:37.342: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 run e2e-test-nginx-rc --image=docker.io/library/nginx:1.14-alpine --generator=run/v1 --namespace=kubectl-6955' +Apr 29 12:03:37.462: INFO: stderr: "kubectl run --generator=run/v1 is DEPRECATED and will be removed in a future version. Use kubectl run --generator=run-pod/v1 or kubectl create instead.\n" +Apr 29 12:03:37.462: INFO: stdout: "replicationcontroller/e2e-test-nginx-rc created\n" +STEP: verifying the rc e2e-test-nginx-rc was created +Apr 29 12:03:37.471: INFO: Waiting for rc e2e-test-nginx-rc to stabilize, generation 1 observed generation 0 spec.replicas 1 status.replicas 0 +Apr 29 12:03:37.474: INFO: Waiting for rc e2e-test-nginx-rc to stabilize, generation 1 observed generation 1 spec.replicas 1 status.replicas 0 +STEP: rolling-update to same image controller +Apr 29 12:03:37.540: INFO: scanned /root for discovery docs: +Apr 29 12:03:37.540: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 rolling-update e2e-test-nginx-rc --update-period=1s --image=docker.io/library/nginx:1.14-alpine --image-pull-policy=IfNotPresent --namespace=kubectl-6955' +Apr 29 12:03:53.481: INFO: stderr: "Command \"rolling-update\" is deprecated, use \"rollout\" instead\n" +Apr 29 12:03:53.481: INFO: stdout: "Created e2e-test-nginx-rc-b93d2bfd66fe5cbfc1af898fcad2c4b5\nScaling up e2e-test-nginx-rc-b93d2bfd66fe5cbfc1af898fcad2c4b5 from 0 to 1, scaling down e2e-test-nginx-rc from 1 to 0 (keep 1 pods available, don't exceed 2 pods)\nScaling e2e-test-nginx-rc-b93d2bfd66fe5cbfc1af898fcad2c4b5 up to 1\nScaling e2e-test-nginx-rc down to 0\nUpdate succeeded. Deleting old controller: e2e-test-nginx-rc\nRenaming e2e-test-nginx-rc-b93d2bfd66fe5cbfc1af898fcad2c4b5 to e2e-test-nginx-rc\nreplicationcontroller/e2e-test-nginx-rc rolling updated\n" +Apr 29 12:03:53.481: INFO: stdout: "Created e2e-test-nginx-rc-b93d2bfd66fe5cbfc1af898fcad2c4b5\nScaling up e2e-test-nginx-rc-b93d2bfd66fe5cbfc1af898fcad2c4b5 from 0 to 1, scaling down e2e-test-nginx-rc from 1 to 0 (keep 1 pods available, don't exceed 2 pods)\nScaling e2e-test-nginx-rc-b93d2bfd66fe5cbfc1af898fcad2c4b5 up to 1\nScaling e2e-test-nginx-rc down to 0\nUpdate succeeded. Deleting old controller: e2e-test-nginx-rc\nRenaming e2e-test-nginx-rc-b93d2bfd66fe5cbfc1af898fcad2c4b5 to e2e-test-nginx-rc\nreplicationcontroller/e2e-test-nginx-rc rolling updated\n" +STEP: waiting for all containers in run=e2e-test-nginx-rc pods to come up. +Apr 29 12:03:53.482: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l run=e2e-test-nginx-rc --namespace=kubectl-6955' +Apr 29 12:03:53.597: INFO: stderr: "" +Apr 29 12:03:53.597: INFO: stdout: "e2e-test-nginx-rc-b93d2bfd66fe5cbfc1af898fcad2c4b5-zkh87 e2e-test-nginx-rc-qrnjz " +STEP: Replicas for run=e2e-test-nginx-rc: expected=1 actual=2 +Apr 29 12:03:58.598: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l run=e2e-test-nginx-rc --namespace=kubectl-6955' +Apr 29 12:03:58.708: INFO: stderr: "" +Apr 29 12:03:58.708: INFO: stdout: "e2e-test-nginx-rc-b93d2bfd66fe5cbfc1af898fcad2c4b5-zkh87 " +Apr 29 12:03:58.708: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 get pods e2e-test-nginx-rc-b93d2bfd66fe5cbfc1af898fcad2c4b5-zkh87 -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "e2e-test-nginx-rc") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-6955' +Apr 29 12:03:58.814: INFO: stderr: "" +Apr 29 12:03:58.814: INFO: stdout: "true" +Apr 29 12:03:58.814: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 get pods e2e-test-nginx-rc-b93d2bfd66fe5cbfc1af898fcad2c4b5-zkh87 -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "e2e-test-nginx-rc"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-6955' +Apr 29 12:03:58.940: INFO: stderr: "" +Apr 29 12:03:58.940: INFO: stdout: "docker.io/library/nginx:1.14-alpine" +Apr 29 12:03:58.940: INFO: e2e-test-nginx-rc-b93d2bfd66fe5cbfc1af898fcad2c4b5-zkh87 is verified up and running +[AfterEach] [k8s.io] Kubectl rolling-update + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1420 +Apr 29 12:03:58.940: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 delete rc e2e-test-nginx-rc --namespace=kubectl-6955' +Apr 29 12:03:59.064: INFO: stderr: "" +Apr 29 12:03:59.064: INFO: stdout: "replicationcontroller \"e2e-test-nginx-rc\" deleted\n" +[AfterEach] [sig-cli] Kubectl client + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Apr 29 12:03:59.064: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "kubectl-6955" for this suite. +Apr 29 12:04:05.109: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Apr 29 12:04:05.367: INFO: namespace kubectl-6955 deletion completed in 6.294103224s + +• [SLOW TEST:28.246 seconds] +[sig-cli] Kubectl client +/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 + [k8s.io] Kubectl rolling-update + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687 + should support rolling-update to same image [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +------------------------------ +SSSSSSSSSSSS +------------------------------ +[sig-storage] EmptyDir volumes + should support (root,0644,default) [LinuxOnly] [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [sig-storage] EmptyDir volumes + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +STEP: Creating a kubernetes client +Apr 29 12:04:05.371: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174 +STEP: Building a namespace api object, basename emptydir +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in emptydir-1918 +STEP: Waiting for a default service account to be provisioned in namespace +[It] should support (root,0644,default) [LinuxOnly] [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: Creating a pod to test emptydir 0644 on node default medium +Apr 29 12:04:05.594: INFO: Waiting up to 5m0s for pod "pod-e2d5165c-6a76-11e9-b6ef-0e71f8c7e888" in namespace "emptydir-1918" to be "success or failure" +Apr 29 12:04:05.600: INFO: Pod "pod-e2d5165c-6a76-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 5.714138ms +Apr 29 12:04:07.610: INFO: Pod "pod-e2d5165c-6a76-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 2.015243838s +Apr 29 12:04:09.617: INFO: Pod "pod-e2d5165c-6a76-11e9-b6ef-0e71f8c7e888": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.022576326s +STEP: Saw pod success +Apr 29 12:04:09.617: INFO: Pod "pod-e2d5165c-6a76-11e9-b6ef-0e71f8c7e888" satisfied condition "success or failure" +Apr 29 12:04:09.622: INFO: Trying to get logs from node worker-4av2h-6665d7b5ff-9nc42 pod pod-e2d5165c-6a76-11e9-b6ef-0e71f8c7e888 container test-container: +STEP: delete the pod +Apr 29 12:04:09.662: INFO: Waiting for pod pod-e2d5165c-6a76-11e9-b6ef-0e71f8c7e888 to disappear +Apr 29 12:04:09.670: INFO: Pod pod-e2d5165c-6a76-11e9-b6ef-0e71f8c7e888 no longer exists +[AfterEach] [sig-storage] EmptyDir volumes + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Apr 29 12:04:09.670: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "emptydir-1918" for this suite. +Apr 29 12:04:15.697: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Apr 29 12:04:15.863: INFO: namespace emptydir-1918 deletion completed in 6.186014259s + +• [SLOW TEST:10.492 seconds] +[sig-storage] EmptyDir volumes +/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:41 + should support (root,0644,default) [LinuxOnly] [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +------------------------------ +SSSSSSSSSSSSSSSSSSSSSSSSSSS +------------------------------ +[sig-storage] EmptyDir wrapper volumes + should not conflict [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [sig-storage] EmptyDir wrapper volumes + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +STEP: Creating a kubernetes client +Apr 29 12:04:15.865: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174 +STEP: Building a namespace api object, basename emptydir-wrapper +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in emptydir-wrapper-5424 +STEP: Waiting for a default service account to be provisioned in namespace +[It] should not conflict [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: Cleaning up the secret +STEP: Cleaning up the configmap +STEP: Cleaning up the pod +[AfterEach] [sig-storage] EmptyDir wrapper volumes + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Apr 29 12:04:28.160: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "emptydir-wrapper-5424" for this suite. +Apr 29 12:04:34.187: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Apr 29 12:04:34.427: INFO: namespace emptydir-wrapper-5424 deletion completed in 6.258807615s + +• [SLOW TEST:18.563 seconds] +[sig-storage] EmptyDir wrapper volumes +/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:22 + should not conflict [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +------------------------------ +SSSSSSSSSSSSSSSSSSSSSSSSSSS +------------------------------ +[sig-cli] Kubectl client [k8s.io] Kubectl describe + should check if kubectl describe prints relevant information for rc and pods [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [sig-cli] Kubectl client + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +STEP: Creating a kubernetes client +Apr 29 12:04:34.432: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174 +STEP: Building a namespace api object, basename kubectl +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-5616 +STEP: Waiting for a default service account to be provisioned in namespace +[BeforeEach] [sig-cli] Kubectl client + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:213 +[It] should check if kubectl describe prints relevant information for rc and pods [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +Apr 29 12:04:34.649: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 version --client' +Apr 29 12:04:34.719: INFO: stderr: "" +Apr 29 12:04:34.720: INFO: stdout: "Client Version: version.Info{Major:\"1\", Minor:\"14\", GitVersion:\"v1.14.1\", GitCommit:\"b7394102d6ef778017f2ca4046abbaa23b88c290\", GitTreeState:\"clean\", BuildDate:\"2019-04-08T17:11:31Z\", GoVersion:\"go1.12.1\", Compiler:\"gc\", Platform:\"linux/amd64\"}\n" +Apr 29 12:04:34.723: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 create -f - --namespace=kubectl-5616' +Apr 29 12:04:35.240: INFO: stderr: "" +Apr 29 12:04:35.240: INFO: stdout: "replicationcontroller/redis-master created\n" +Apr 29 12:04:35.240: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 create -f - --namespace=kubectl-5616' +Apr 29 12:04:35.544: INFO: stderr: "" +Apr 29 12:04:35.544: INFO: stdout: "service/redis-master created\n" +STEP: Waiting for Redis master to start. +Apr 29 12:04:36.550: INFO: Selector matched 1 pods for map[app:redis] +Apr 29 12:04:36.550: INFO: Found 0 / 1 +Apr 29 12:04:37.551: INFO: Selector matched 1 pods for map[app:redis] +Apr 29 12:04:37.551: INFO: Found 0 / 1 +Apr 29 12:04:38.550: INFO: Selector matched 1 pods for map[app:redis] +Apr 29 12:04:38.550: INFO: Found 0 / 1 +Apr 29 12:04:39.560: INFO: Selector matched 1 pods for map[app:redis] +Apr 29 12:04:39.560: INFO: Found 1 / 1 +Apr 29 12:04:39.560: INFO: WaitFor completed with timeout 5m0s. Pods found = 1 out of 1 +Apr 29 12:04:39.566: INFO: Selector matched 1 pods for map[app:redis] +Apr 29 12:04:39.566: INFO: ForEach: Found 1 pods from the filter. Now looping through them. +Apr 29 12:04:39.566: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 describe pod redis-master-wknw4 --namespace=kubectl-5616' +Apr 29 12:04:39.686: INFO: stderr: "" +Apr 29 12:04:39.686: INFO: stdout: "Name: redis-master-wknw4\nNamespace: kubectl-5616\nPriority: 0\nPriorityClassName: \nNode: worker-4av2h-6665d7b5ff-9nc42/172.23.0.238\nStart Time: Mon, 29 Apr 2019 12:04:35 +0000\nLabels: app=redis\n role=master\nAnnotations: kubernetes.io/psp: e2e-test-privileged-psp\nStatus: Running\nIP: 192.168.205.95\nControlled By: ReplicationController/redis-master\nContainers:\n redis-master:\n Container ID: docker://f45ee6f624735bcf57b6c5a45f1e0a5dcc0dc8dabd4eddac23e70569a22c4d27\n Image: gcr.io/kubernetes-e2e-test-images/redis:1.0\n Image ID: docker-pullable://gcr.io/kubernetes-e2e-test-images/redis@sha256:af4748d1655c08dc54d4be5182135395db9ce87aba2d4699b26b14ae197c5830\n Port: 6379/TCP\n Host Port: 0/TCP\n State: Running\n Started: Mon, 29 Apr 2019 12:04:37 +0000\n Ready: True\n Restart Count: 0\n Environment: \n Mounts:\n /var/run/secrets/kubernetes.io/serviceaccount from default-token-m6kb8 (ro)\nConditions:\n Type Status\n Initialized True \n Ready True \n ContainersReady True \n PodScheduled True \nVolumes:\n default-token-m6kb8:\n Type: Secret (a volume populated by a Secret)\n SecretName: default-token-m6kb8\n Optional: false\nQoS Class: BestEffort\nNode-Selectors: \nTolerations: node.kubernetes.io/not-ready:NoExecute for 300s\n node.kubernetes.io/unreachable:NoExecute for 300s\nEvents:\n Type Reason Age From Message\n ---- ------ ---- ---- -------\n Normal Scheduled 5s default-scheduler Successfully assigned kubectl-5616/redis-master-wknw4 to worker-4av2h-6665d7b5ff-9nc42\n Normal Pulled 3s kubelet, worker-4av2h-6665d7b5ff-9nc42 Container image \"gcr.io/kubernetes-e2e-test-images/redis:1.0\" already present on machine\n Normal Created 2s kubelet, worker-4av2h-6665d7b5ff-9nc42 Created container redis-master\n Normal Started 2s kubelet, worker-4av2h-6665d7b5ff-9nc42 Started container redis-master\n" +Apr 29 12:04:39.686: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 describe rc redis-master --namespace=kubectl-5616' +Apr 29 12:04:39.904: INFO: stderr: "" +Apr 29 12:04:39.904: INFO: stdout: "Name: redis-master\nNamespace: kubectl-5616\nSelector: app=redis,role=master\nLabels: app=redis\n role=master\nAnnotations: \nReplicas: 1 current / 1 desired\nPods Status: 1 Running / 0 Waiting / 0 Succeeded / 0 Failed\nPod Template:\n Labels: app=redis\n role=master\n Containers:\n redis-master:\n Image: gcr.io/kubernetes-e2e-test-images/redis:1.0\n Port: 6379/TCP\n Host Port: 0/TCP\n Environment: \n Mounts: \n Volumes: \nEvents:\n Type Reason Age From Message\n ---- ------ ---- ---- -------\n Normal SuccessfulCreate 5s replication-controller Created pod: redis-master-wknw4\n" +Apr 29 12:04:39.904: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 describe service redis-master --namespace=kubectl-5616' +Apr 29 12:04:40.063: INFO: stderr: "" +Apr 29 12:04:40.063: INFO: stdout: "Name: redis-master\nNamespace: kubectl-5616\nLabels: app=redis\n role=master\nAnnotations: \nSelector: app=redis,role=master\nType: ClusterIP\nIP: 172.31.170.138\nPort: 6379/TCP\nTargetPort: redis-server/TCP\nEndpoints: 192.168.205.95:6379\nSession Affinity: None\nEvents: \n" +Apr 29 12:04:40.070: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 describe node master-h2h3g-655686fb8c-8kgdn' +Apr 29 12:04:40.316: INFO: stderr: "" +Apr 29 12:04:40.317: INFO: stdout: "Name: master-h2h3g-655686fb8c-8kgdn\nRoles: master\nLabels: beta.kubernetes.io/arch=amd64\n beta.kubernetes.io/os=linux\n giantswarm.io/provider=kvm\n ip=172.23.0.254\n kubernetes.io/arch=amd64\n kubernetes.io/hostname=master-h2h3g-655686fb8c-8kgdn\n kubernetes.io/os=linux\n kubernetes.io/role=master\n kvm-operator.giantswarm.io/version=3.5.0\n node-role.kubernetes.io/master=\n node.kubernetes.io/master=\nAnnotations: node.alpha.kubernetes.io/ttl: 0\n volumes.kubernetes.io/controller-managed-attach-detach: true\nCreationTimestamp: Mon, 29 Apr 2019 11:12:41 +0000\nTaints: node-role.kubernetes.io/master:NoSchedule\nUnschedulable: false\nConditions:\n Type Status LastHeartbeatTime LastTransitionTime Reason Message\n ---- ------ ----------------- ------------------ ------ -------\n NetworkUnavailable False Mon, 29 Apr 2019 11:14:07 +0000 Mon, 29 Apr 2019 11:14:07 +0000 CalicoIsUp Calico is running on this node\n MemoryPressure False Mon, 29 Apr 2019 12:03:48 +0000 Mon, 29 Apr 2019 11:12:35 +0000 KubeletHasSufficientMemory kubelet has sufficient memory available\n DiskPressure False Mon, 29 Apr 2019 12:03:48 +0000 Mon, 29 Apr 2019 11:12:35 +0000 KubeletHasNoDiskPressure kubelet has no disk pressure\n PIDPressure False Mon, 29 Apr 2019 12:03:48 +0000 Mon, 29 Apr 2019 11:12:35 +0000 KubeletHasSufficientPID kubelet has sufficient PID available\n Ready True Mon, 29 Apr 2019 12:03:48 +0000 Mon, 29 Apr 2019 11:14:52 +0000 KubeletReady kubelet is posting ready status\nAddresses:\n InternalIP: 172.23.0.254\n Hostname: master-h2h3g-655686fb8c-8kgdn\nCapacity:\n cpu: 2\n ephemeral-storage: 5095040Ki\n hugepages-1Gi: 0\n hugepages-2Mi: 0\n memory: 8170012Ki\n pods: 110\nAllocatable:\n cpu: 2\n ephemeral-storage: 5095040Ki\n hugepages-1Gi: 0\n hugepages-2Mi: 0\n memory: 7965212Ki\n pods: 110\nSystem Info:\n Machine ID: 0029049d43914765b661dcc2f063b62c\n System UUID: 311936e2303b034fe7ef70182235b8cb\n Boot ID: 280ff0a6-539c-4322-b425-ecf58cae3269\n Kernel Version: 4.19.25-coreos\n OS Image: Container Linux by CoreOS 2023.5.0 (Rhyolite)\n Operating System: linux\n Architecture: amd64\n Container Runtime Version: docker://18.6.1\n Kubelet Version: v1.14.1\n Kube-Proxy Version: v1.14.1\nNon-terminated Pods: (10 in total)\n Namespace Name CPU Requests CPU Limits Memory Requests Memory Limits AGE\n --------- ---- ------------ ---------- --------------- ------------- ---\n heptio-sonobuoy sonobuoy-systemd-logs-daemon-set-8191105fdead4396-b4lgk 0 (0%) 0 (0%) 0 (0%) 0 (0%) 33m\n kube-system calico-kube-controllers-689fccb766-9zj5q 250m (12%) 250m (12%) 100Mi (1%) 100Mi (1%) 51m\n kube-system calico-node-jllz6 250m (12%) 250m (12%) 150Mi (1%) 150Mi (1%) 51m\n kube-system cert-exporter-xb8gr 50m (2%) 50m (2%) 50Mi (0%) 50Mi (0%) 48m\n kube-system k8s-api-server-master-h2h3g-655686fb8c-8kgdn 300m (15%) 0 (0%) 300Mi (3%) 0 (0%) 50m\n kube-system k8s-controller-manager-master-h2h3g-655686fb8c-8kgdn 200m (10%) 0 (0%) 200Mi (2%) 0 (0%) 51m\n kube-system k8s-scheduler-master-h2h3g-655686fb8c-8kgdn 100m (5%) 0 (0%) 100Mi (1%) 0 (0%) 50m\n kube-system kube-proxy-jgxgc 75m (3%) 0 (0%) 80Mi (1%) 0 (0%) 50m\n kube-system net-exporter-plsfx 50m (2%) 50m (2%) 50Mi (0%) 50Mi (0%) 42m\n kube-system node-exporter-9cbwn 200m (10%) 200m (10%) 75Mi (0%) 75Mi (0%) 42m\nAllocated resources:\n (Total limits may be over 100 percent, i.e., overcommitted.)\n Resource Requests Limits\n -------- -------- ------\n cpu 1475m (73%) 800m (40%)\n memory 1105Mi (14%) 425Mi (5%)\n ephemeral-storage 0 (0%) 0 (0%)\nEvents:\n Type Reason Age From Message\n ---- ------ ---- ---- -------\n Normal NodeHasSufficientMemory 52m (x8 over 52m) kubelet, master-h2h3g-655686fb8c-8kgdn Node master-h2h3g-655686fb8c-8kgdn status is now: NodeHasSufficientMemory\n Normal NodeHasNoDiskPressure 52m (x8 over 52m) kubelet, master-h2h3g-655686fb8c-8kgdn Node master-h2h3g-655686fb8c-8kgdn status is now: NodeHasNoDiskPressure\n Normal NodeHasSufficientPID 52m (x7 over 52m) kubelet, master-h2h3g-655686fb8c-8kgdn Node master-h2h3g-655686fb8c-8kgdn status is now: NodeHasSufficientPID\n Normal Starting 50m kube-proxy, master-h2h3g-655686fb8c-8kgdn Starting kube-proxy.\n Normal Starting 49m kubelet, master-h2h3g-655686fb8c-8kgdn Starting kubelet.\n Normal NodeHasSufficientMemory 49m kubelet, master-h2h3g-655686fb8c-8kgdn Node master-h2h3g-655686fb8c-8kgdn status is now: NodeHasSufficientMemory\n Normal NodeHasNoDiskPressure 49m kubelet, master-h2h3g-655686fb8c-8kgdn Node master-h2h3g-655686fb8c-8kgdn status is now: NodeHasNoDiskPressure\n Normal NodeHasSufficientPID 49m kubelet, master-h2h3g-655686fb8c-8kgdn Node master-h2h3g-655686fb8c-8kgdn status is now: NodeHasSufficientPID\n Normal NodeNotReady 49m kubelet, master-h2h3g-655686fb8c-8kgdn Node master-h2h3g-655686fb8c-8kgdn status is now: NodeNotReady\n Normal NodeAllocatableEnforced 49m kubelet, master-h2h3g-655686fb8c-8kgdn Updated Node Allocatable limit across pods\n Normal NodeReady 49m kubelet, master-h2h3g-655686fb8c-8kgdn Node master-h2h3g-655686fb8c-8kgdn status is now: NodeReady\n" +Apr 29 12:04:40.317: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 describe namespace kubectl-5616' +Apr 29 12:04:40.484: INFO: stderr: "" +Apr 29 12:04:40.485: INFO: stdout: "Name: kubectl-5616\nLabels: e2e-framework=kubectl\n e2e-run=4c9f30da-6a72-11e9-b6ef-0e71f8c7e888\nAnnotations: \nStatus: Active\n\nNo resource quota.\n\nNo resource limits.\n" +[AfterEach] [sig-cli] Kubectl client + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Apr 29 12:04:40.485: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "kubectl-5616" for this suite. +Apr 29 12:05:04.523: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Apr 29 12:05:04.734: INFO: namespace kubectl-5616 deletion completed in 24.240443953s + +• [SLOW TEST:30.303 seconds] +[sig-cli] Kubectl client +/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 + [k8s.io] Kubectl describe + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687 + should check if kubectl describe prints relevant information for rc and pods [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +------------------------------ +SSSSSSSSSSSSSSS +------------------------------ +[sig-cli] Kubectl client [k8s.io] Proxy server + should support --unix-socket=/path [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [sig-cli] Kubectl client + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +STEP: Creating a kubernetes client +Apr 29 12:05:04.738: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174 +STEP: Building a namespace api object, basename kubectl +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-5005 +STEP: Waiting for a default service account to be provisioned in namespace +[BeforeEach] [sig-cli] Kubectl client + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:213 +[It] should support --unix-socket=/path [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: Starting the proxy +Apr 29 12:05:04.937: INFO: Asynchronously running '/usr/local/bin/kubectl kubectl --kubeconfig=/tmp/kubeconfig-185508174 proxy --unix-socket=/tmp/kubectl-proxy-unix836987477/test' +STEP: retrieving proxy /api/ output +[AfterEach] [sig-cli] Kubectl client + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Apr 29 12:05:05.014: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "kubectl-5005" for this suite. +Apr 29 12:05:11.050: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Apr 29 12:05:11.335: INFO: namespace kubectl-5005 deletion completed in 6.312630015s + +• [SLOW TEST:6.597 seconds] +[sig-cli] Kubectl client +/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 + [k8s.io] Proxy server + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687 + should support --unix-socket=/path [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +------------------------------ +SSSSSSSSSSSSSSSSSSSSSS +------------------------------ +[sig-storage] EmptyDir volumes + should support (non-root,0644,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [sig-storage] EmptyDir volumes + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +STEP: Creating a kubernetes client +Apr 29 12:05:11.338: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174 +STEP: Building a namespace api object, basename emptydir +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in emptydir-5693 +STEP: Waiting for a default service account to be provisioned in namespace +[It] should support (non-root,0644,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: Creating a pod to test emptydir 0644 on tmpfs +Apr 29 12:05:11.559: INFO: Waiting up to 5m0s for pod "pod-0a25ddbe-6a77-11e9-b6ef-0e71f8c7e888" in namespace "emptydir-5693" to be "success or failure" +Apr 29 12:05:11.572: INFO: Pod "pod-0a25ddbe-6a77-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 13.052449ms +Apr 29 12:05:13.582: INFO: Pod "pod-0a25ddbe-6a77-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 2.023201862s +Apr 29 12:05:15.588: INFO: Pod "pod-0a25ddbe-6a77-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 4.029018028s +Apr 29 12:05:17.599: INFO: Pod "pod-0a25ddbe-6a77-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 6.039799327s +Apr 29 12:05:19.614: INFO: Pod "pod-0a25ddbe-6a77-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 8.05502633s +Apr 29 12:05:21.620: INFO: Pod "pod-0a25ddbe-6a77-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 10.061090557s +Apr 29 12:05:23.628: INFO: Pod "pod-0a25ddbe-6a77-11e9-b6ef-0e71f8c7e888": Phase="Succeeded", Reason="", readiness=false. Elapsed: 12.069026921s +STEP: Saw pod success +Apr 29 12:05:23.629: INFO: Pod "pod-0a25ddbe-6a77-11e9-b6ef-0e71f8c7e888" satisfied condition "success or failure" +Apr 29 12:05:23.634: INFO: Trying to get logs from node worker-4av2h-6665d7b5ff-9nc42 pod pod-0a25ddbe-6a77-11e9-b6ef-0e71f8c7e888 container test-container: +STEP: delete the pod +Apr 29 12:05:23.675: INFO: Waiting for pod pod-0a25ddbe-6a77-11e9-b6ef-0e71f8c7e888 to disappear +Apr 29 12:05:23.683: INFO: Pod pod-0a25ddbe-6a77-11e9-b6ef-0e71f8c7e888 no longer exists +[AfterEach] [sig-storage] EmptyDir volumes + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Apr 29 12:05:23.684: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "emptydir-5693" for this suite. +Apr 29 12:05:29.723: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Apr 29 12:05:29.911: INFO: namespace emptydir-5693 deletion completed in 6.21478749s + +• [SLOW TEST:18.574 seconds] +[sig-storage] EmptyDir volumes +/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:41 + should support (non-root,0644,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +------------------------------ +SSSS +------------------------------ +[sig-storage] Projected configMap + should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [sig-storage] Projected configMap + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +STEP: Creating a kubernetes client +Apr 29 12:05:29.912: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174 +STEP: Building a namespace api object, basename projected +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-1507 +STEP: Waiting for a default service account to be provisioned in namespace +[It] should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: Creating configMap with name projected-configmap-test-volume-153770b1-6a77-11e9-b6ef-0e71f8c7e888 +STEP: Creating a pod to test consume configMaps +Apr 29 12:05:30.154: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-1539e0ae-6a77-11e9-b6ef-0e71f8c7e888" in namespace "projected-1507" to be "success or failure" +Apr 29 12:05:30.164: INFO: Pod "pod-projected-configmaps-1539e0ae-6a77-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 9.765479ms +Apr 29 12:05:32.175: INFO: Pod "pod-projected-configmaps-1539e0ae-6a77-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 2.021162159s +Apr 29 12:05:34.183: INFO: Pod "pod-projected-configmaps-1539e0ae-6a77-11e9-b6ef-0e71f8c7e888": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.028681478s +STEP: Saw pod success +Apr 29 12:05:34.183: INFO: Pod "pod-projected-configmaps-1539e0ae-6a77-11e9-b6ef-0e71f8c7e888" satisfied condition "success or failure" +Apr 29 12:05:34.187: INFO: Trying to get logs from node worker-4av2h-6665d7b5ff-9nc42 pod pod-projected-configmaps-1539e0ae-6a77-11e9-b6ef-0e71f8c7e888 container projected-configmap-volume-test: +STEP: delete the pod +Apr 29 12:05:34.224: INFO: Waiting for pod pod-projected-configmaps-1539e0ae-6a77-11e9-b6ef-0e71f8c7e888 to disappear +Apr 29 12:05:34.230: INFO: Pod pod-projected-configmaps-1539e0ae-6a77-11e9-b6ef-0e71f8c7e888 no longer exists +[AfterEach] [sig-storage] Projected configMap + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Apr 29 12:05:34.230: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "projected-1507" for this suite. +Apr 29 12:05:40.263: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Apr 29 12:05:40.547: INFO: namespace projected-1507 deletion completed in 6.303451471s + +• [SLOW TEST:10.635 seconds] +[sig-storage] Projected configMap +/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_configmap.go:33 + should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +------------------------------ +SSSS +------------------------------ +[sig-cli] Kubectl client [k8s.io] Kubectl run --rm job + should create a job from an image, then delete the job [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [sig-cli] Kubectl client + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +STEP: Creating a kubernetes client +Apr 29 12:05:40.548: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174 +STEP: Building a namespace api object, basename kubectl +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-7535 +STEP: Waiting for a default service account to be provisioned in namespace +[BeforeEach] [sig-cli] Kubectl client + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:213 +[It] should create a job from an image, then delete the job [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: executing a command with run --rm and attach with stdin +Apr 29 12:05:40.775: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 --namespace=kubectl-7535 run e2e-test-rm-busybox-job --image=docker.io/library/busybox:1.29 --rm=true --generator=job/v1 --restart=OnFailure --attach=true --stdin -- sh -c cat && echo 'stdin closed'' +Apr 29 12:05:43.555: INFO: stderr: "kubectl run --generator=job/v1 is DEPRECATED and will be removed in a future version. Use kubectl run --generator=run-pod/v1 or kubectl create instead.\nIf you don't see a command prompt, try pressing enter.\n" +Apr 29 12:05:43.555: INFO: stdout: "abcd1234stdin closed\njob.batch \"e2e-test-rm-busybox-job\" deleted\n" +STEP: verifying the job e2e-test-rm-busybox-job was deleted +[AfterEach] [sig-cli] Kubectl client + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Apr 29 12:05:45.566: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "kubectl-7535" for this suite. +Apr 29 12:05:55.597: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Apr 29 12:05:55.848: INFO: namespace kubectl-7535 deletion completed in 10.273994851s + +• [SLOW TEST:15.300 seconds] +[sig-cli] Kubectl client +/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 + [k8s.io] Kubectl run --rm job + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687 + should create a job from an image, then delete the job [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +------------------------------ +SSSSSSSSSSS +------------------------------ +[k8s.io] Pods + should be submitted and removed [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [k8s.io] Pods + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +STEP: Creating a kubernetes client +Apr 29 12:05:55.849: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174 +STEP: Building a namespace api object, basename pods +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in pods-5301 +STEP: Waiting for a default service account to be provisioned in namespace +[BeforeEach] [k8s.io] Pods + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:135 +[It] should be submitted and removed [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: creating the pod +STEP: setting up watch +STEP: submitting the pod to kubernetes +Apr 29 12:05:56.068: INFO: observed the pod list +STEP: verifying the pod is in kubernetes +STEP: verifying pod creation was observed +STEP: deleting the pod gracefully +STEP: verifying the kubelet observed the termination notice +Apr 29 12:06:05.160: INFO: no pod exists with the name we were looking for, assuming the termination request was observed and completed +STEP: verifying pod deletion was observed +[AfterEach] [k8s.io] Pods + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Apr 29 12:06:05.167: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "pods-5301" for this suite. +Apr 29 12:06:11.203: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Apr 29 12:06:11.436: INFO: namespace pods-5301 deletion completed in 6.263378007s + +• [SLOW TEST:15.587 seconds] +[k8s.io] Pods +/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687 + should be submitted and removed [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +------------------------------ +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +------------------------------ +[sig-api-machinery] Watchers + should be able to start watching from a specific resource version [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [sig-api-machinery] Watchers + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +STEP: Creating a kubernetes client +Apr 29 12:06:11.439: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174 +STEP: Building a namespace api object, basename watch +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in watch-70 +STEP: Waiting for a default service account to be provisioned in namespace +[It] should be able to start watching from a specific resource version [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: creating a new configmap +STEP: modifying the configmap once +STEP: modifying the configmap a second time +STEP: deleting the configmap +STEP: creating a watch on configmaps from the resource version returned by the first update +STEP: Expecting to observe notifications for all changes to the configmap after the first update +Apr 29 12:06:11.712: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-resource-version,GenerateName:,Namespace:watch-70,SelfLink:/api/v1/namespaces/watch-70/configmaps/e2e-watch-test-resource-version,UID:2dcb960e-6a77-11e9-92a7-deadbef3c36d,ResourceVersion:12171,Generation:0,CreationTimestamp:2019-04-29 12:06:11 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: from-resource-version,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},} +Apr 29 12:06:11.712: INFO: Got : DELETED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-resource-version,GenerateName:,Namespace:watch-70,SelfLink:/api/v1/namespaces/watch-70/configmaps/e2e-watch-test-resource-version,UID:2dcb960e-6a77-11e9-92a7-deadbef3c36d,ResourceVersion:12172,Generation:0,CreationTimestamp:2019-04-29 12:06:11 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: from-resource-version,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},} +[AfterEach] [sig-api-machinery] Watchers + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Apr 29 12:06:11.713: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "watch-70" for this suite. +Apr 29 12:06:17.747: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Apr 29 12:06:17.987: INFO: namespace watch-70 deletion completed in 6.266821437s + +• [SLOW TEST:6.549 seconds] +[sig-api-machinery] Watchers +/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:22 + should be able to start watching from a specific resource version [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +------------------------------ +[sig-storage] EmptyDir volumes + should support (non-root,0644,default) [LinuxOnly] [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [sig-storage] EmptyDir volumes + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +STEP: Creating a kubernetes client +Apr 29 12:06:17.989: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174 +STEP: Building a namespace api object, basename emptydir +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in emptydir-6934 +STEP: Waiting for a default service account to be provisioned in namespace +[It] should support (non-root,0644,default) [LinuxOnly] [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: Creating a pod to test emptydir 0644 on node default medium +Apr 29 12:06:18.205: INFO: Waiting up to 5m0s for pod "pod-31df5c2e-6a77-11e9-b6ef-0e71f8c7e888" in namespace "emptydir-6934" to be "success or failure" +Apr 29 12:06:18.217: INFO: Pod "pod-31df5c2e-6a77-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 12.158106ms +Apr 29 12:06:20.226: INFO: Pod "pod-31df5c2e-6a77-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 2.020944631s +Apr 29 12:06:22.233: INFO: Pod "pod-31df5c2e-6a77-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 4.02799826s +Apr 29 12:06:24.247: INFO: Pod "pod-31df5c2e-6a77-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 6.042197199s +Apr 29 12:06:26.254: INFO: Pod "pod-31df5c2e-6a77-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 8.048895162s +Apr 29 12:06:28.261: INFO: Pod "pod-31df5c2e-6a77-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 10.056010303s +Apr 29 12:06:30.285: INFO: Pod "pod-31df5c2e-6a77-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 12.080055966s +Apr 29 12:06:32.294: INFO: Pod "pod-31df5c2e-6a77-11e9-b6ef-0e71f8c7e888": Phase="Succeeded", Reason="", readiness=false. Elapsed: 14.089161238s +STEP: Saw pod success +Apr 29 12:06:32.295: INFO: Pod "pod-31df5c2e-6a77-11e9-b6ef-0e71f8c7e888" satisfied condition "success or failure" +Apr 29 12:06:32.304: INFO: Trying to get logs from node worker-4av2h-6665d7b5ff-9nc42 pod pod-31df5c2e-6a77-11e9-b6ef-0e71f8c7e888 container test-container: +STEP: delete the pod +Apr 29 12:06:32.335: INFO: Waiting for pod pod-31df5c2e-6a77-11e9-b6ef-0e71f8c7e888 to disappear +Apr 29 12:06:32.341: INFO: Pod pod-31df5c2e-6a77-11e9-b6ef-0e71f8c7e888 no longer exists +[AfterEach] [sig-storage] EmptyDir volumes + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Apr 29 12:06:32.341: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "emptydir-6934" for this suite. +Apr 29 12:06:38.375: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Apr 29 12:06:38.600: INFO: namespace emptydir-6934 deletion completed in 6.253031219s + +• [SLOW TEST:20.612 seconds] +[sig-storage] EmptyDir volumes +/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:41 + should support (non-root,0644,default) [LinuxOnly] [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +------------------------------ +SSSSS +------------------------------ +[sig-storage] Subpath Atomic writer volumes + should support subpaths with configmap pod [LinuxOnly] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [sig-storage] Subpath + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +STEP: Creating a kubernetes client +Apr 29 12:06:38.606: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174 +STEP: Building a namespace api object, basename subpath +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in subpath-6171 +STEP: Waiting for a default service account to be provisioned in namespace +[BeforeEach] Atomic writer volumes + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:38 +STEP: Setting up data +[It] should support subpaths with configmap pod [LinuxOnly] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: Creating pod pod-subpath-test-configmap-4hkj +STEP: Creating a pod to test atomic-volume-subpath +Apr 29 12:06:38.838: INFO: Waiting up to 5m0s for pod "pod-subpath-test-configmap-4hkj" in namespace "subpath-6171" to be "success or failure" +Apr 29 12:06:38.847: INFO: Pod "pod-subpath-test-configmap-4hkj": Phase="Pending", Reason="", readiness=false. Elapsed: 8.562262ms +Apr 29 12:06:40.856: INFO: Pod "pod-subpath-test-configmap-4hkj": Phase="Pending", Reason="", readiness=false. Elapsed: 2.017263969s +Apr 29 12:06:42.863: INFO: Pod "pod-subpath-test-configmap-4hkj": Phase="Running", Reason="", readiness=true. Elapsed: 4.024544788s +Apr 29 12:06:44.871: INFO: Pod "pod-subpath-test-configmap-4hkj": Phase="Running", Reason="", readiness=true. Elapsed: 6.032754408s +Apr 29 12:06:46.880: INFO: Pod "pod-subpath-test-configmap-4hkj": Phase="Running", Reason="", readiness=true. Elapsed: 8.041453683s +Apr 29 12:06:48.890: INFO: Pod "pod-subpath-test-configmap-4hkj": Phase="Running", Reason="", readiness=true. Elapsed: 10.051686946s +Apr 29 12:06:50.897: INFO: Pod "pod-subpath-test-configmap-4hkj": Phase="Running", Reason="", readiness=true. Elapsed: 12.058881752s +Apr 29 12:06:52.905: INFO: Pod "pod-subpath-test-configmap-4hkj": Phase="Running", Reason="", readiness=true. Elapsed: 14.06668165s +Apr 29 12:06:54.913: INFO: Pod "pod-subpath-test-configmap-4hkj": Phase="Running", Reason="", readiness=true. Elapsed: 16.074884994s +Apr 29 12:06:56.923: INFO: Pod "pod-subpath-test-configmap-4hkj": Phase="Running", Reason="", readiness=true. Elapsed: 18.085014185s +Apr 29 12:06:58.932: INFO: Pod "pod-subpath-test-configmap-4hkj": Phase="Running", Reason="", readiness=true. Elapsed: 20.093897655s +Apr 29 12:07:00.941: INFO: Pod "pod-subpath-test-configmap-4hkj": Phase="Running", Reason="", readiness=true. Elapsed: 22.102898567s +Apr 29 12:07:02.949: INFO: Pod "pod-subpath-test-configmap-4hkj": Phase="Succeeded", Reason="", readiness=false. Elapsed: 24.110969075s +STEP: Saw pod success +Apr 29 12:07:02.950: INFO: Pod "pod-subpath-test-configmap-4hkj" satisfied condition "success or failure" +Apr 29 12:07:02.957: INFO: Trying to get logs from node worker-4av2h-6665d7b5ff-9nc42 pod pod-subpath-test-configmap-4hkj container test-container-subpath-configmap-4hkj: +STEP: delete the pod +Apr 29 12:07:02.993: INFO: Waiting for pod pod-subpath-test-configmap-4hkj to disappear +Apr 29 12:07:03.000: INFO: Pod pod-subpath-test-configmap-4hkj no longer exists +STEP: Deleting pod pod-subpath-test-configmap-4hkj +Apr 29 12:07:03.000: INFO: Deleting pod "pod-subpath-test-configmap-4hkj" in namespace "subpath-6171" +[AfterEach] [sig-storage] Subpath + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Apr 29 12:07:03.008: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "subpath-6171" for this suite. +Apr 29 12:07:09.043: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Apr 29 12:07:09.274: INFO: namespace subpath-6171 deletion completed in 6.253726429s + +• [SLOW TEST:30.670 seconds] +[sig-storage] Subpath +/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:22 + Atomic writer volumes + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:34 + should support subpaths with configmap pod [LinuxOnly] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +------------------------------ +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +------------------------------ +[sig-auth] ServiceAccounts + should allow opting out of API token automount [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [sig-auth] ServiceAccounts + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +STEP: Creating a kubernetes client +Apr 29 12:07:09.292: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174 +STEP: Building a namespace api object, basename svcaccounts +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in svcaccounts-8640 +STEP: Waiting for a default service account to be provisioned in namespace +[It] should allow opting out of API token automount [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: getting the auto-created API token +Apr 29 12:07:10.063: INFO: created pod pod-service-account-defaultsa +Apr 29 12:07:10.063: INFO: pod pod-service-account-defaultsa service account token volume mount: true +Apr 29 12:07:10.104: INFO: created pod pod-service-account-mountsa +Apr 29 12:07:10.104: INFO: pod pod-service-account-mountsa service account token volume mount: true +Apr 29 12:07:10.128: INFO: created pod pod-service-account-nomountsa +Apr 29 12:07:10.128: INFO: pod pod-service-account-nomountsa service account token volume mount: false +Apr 29 12:07:10.147: INFO: created pod pod-service-account-defaultsa-mountspec +Apr 29 12:07:10.147: INFO: pod pod-service-account-defaultsa-mountspec service account token volume mount: true +Apr 29 12:07:10.171: INFO: created pod pod-service-account-mountsa-mountspec +Apr 29 12:07:10.171: INFO: pod pod-service-account-mountsa-mountspec service account token volume mount: true +Apr 29 12:07:10.200: INFO: created pod pod-service-account-nomountsa-mountspec +Apr 29 12:07:10.200: INFO: pod pod-service-account-nomountsa-mountspec service account token volume mount: true +Apr 29 12:07:10.212: INFO: created pod pod-service-account-defaultsa-nomountspec +Apr 29 12:07:10.212: INFO: pod pod-service-account-defaultsa-nomountspec service account token volume mount: false +Apr 29 12:07:10.228: INFO: created pod pod-service-account-mountsa-nomountspec +Apr 29 12:07:10.228: INFO: pod pod-service-account-mountsa-nomountspec service account token volume mount: false +Apr 29 12:07:10.258: INFO: created pod pod-service-account-nomountsa-nomountspec +Apr 29 12:07:10.258: INFO: pod pod-service-account-nomountsa-nomountspec service account token volume mount: false +[AfterEach] [sig-auth] ServiceAccounts + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Apr 29 12:07:10.259: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "svcaccounts-8640" for this suite. +Apr 29 12:07:34.314: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Apr 29 12:07:34.549: INFO: namespace svcaccounts-8640 deletion completed in 24.267344904s + +• [SLOW TEST:25.258 seconds] +[sig-auth] ServiceAccounts +/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/auth/framework.go:22 + should allow opting out of API token automount [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +------------------------------ +SSSSSSSSSSSS +------------------------------ +[sig-storage] Projected configMap + should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [sig-storage] Projected configMap + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +STEP: Creating a kubernetes client +Apr 29 12:07:34.552: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174 +STEP: Building a namespace api object, basename projected +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-1873 +STEP: Waiting for a default service account to be provisioned in namespace +[It] should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: Creating configMap with name projected-configmap-test-volume-5f871e31-6a77-11e9-b6ef-0e71f8c7e888 +STEP: Creating a pod to test consume configMaps +Apr 29 12:07:34.819: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-5f886e80-6a77-11e9-b6ef-0e71f8c7e888" in namespace "projected-1873" to be "success or failure" +Apr 29 12:07:34.823: INFO: Pod "pod-projected-configmaps-5f886e80-6a77-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 4.100476ms +Apr 29 12:07:36.831: INFO: Pod "pod-projected-configmaps-5f886e80-6a77-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 2.011739135s +Apr 29 12:07:38.838: INFO: Pod "pod-projected-configmaps-5f886e80-6a77-11e9-b6ef-0e71f8c7e888": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.019169191s +STEP: Saw pod success +Apr 29 12:07:38.839: INFO: Pod "pod-projected-configmaps-5f886e80-6a77-11e9-b6ef-0e71f8c7e888" satisfied condition "success or failure" +Apr 29 12:07:38.842: INFO: Trying to get logs from node worker-4av2h-6665d7b5ff-9nc42 pod pod-projected-configmaps-5f886e80-6a77-11e9-b6ef-0e71f8c7e888 container projected-configmap-volume-test: +STEP: delete the pod +Apr 29 12:07:38.887: INFO: Waiting for pod pod-projected-configmaps-5f886e80-6a77-11e9-b6ef-0e71f8c7e888 to disappear +Apr 29 12:07:38.892: INFO: Pod pod-projected-configmaps-5f886e80-6a77-11e9-b6ef-0e71f8c7e888 no longer exists +[AfterEach] [sig-storage] Projected configMap + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Apr 29 12:07:38.893: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "projected-1873" for this suite. +Apr 29 12:07:44.936: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Apr 29 12:07:45.167: INFO: namespace projected-1873 deletion completed in 6.262124112s + +• [SLOW TEST:10.615 seconds] +[sig-storage] Projected configMap +/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_configmap.go:33 + should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +------------------------------ +SSSSSSSSSSSSSSSSSSSSSSSSSS +------------------------------ +[k8s.io] InitContainer [NodeConformance] + should not start app containers and fail the pod if init containers fail on a RestartNever pod [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [k8s.io] InitContainer [NodeConformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +STEP: Creating a kubernetes client +Apr 29 12:07:45.168: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174 +STEP: Building a namespace api object, basename init-container +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in init-container-5748 +STEP: Waiting for a default service account to be provisioned in namespace +[BeforeEach] [k8s.io] InitContainer [NodeConformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/init_container.go:43 +[It] should not start app containers and fail the pod if init containers fail on a RestartNever pod [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: creating the pod +Apr 29 12:07:45.364: INFO: PodSpec: initContainers in spec.initContainers +[AfterEach] [k8s.io] InitContainer [NodeConformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Apr 29 12:07:49.540: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "init-container-5748" for this suite. +Apr 29 12:07:55.577: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Apr 29 12:07:55.850: INFO: namespace init-container-5748 deletion completed in 6.296710603s + +• [SLOW TEST:10.683 seconds] +[k8s.io] InitContainer [NodeConformance] +/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687 + should not start app containers and fail the pod if init containers fail on a RestartNever pod [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +------------------------------ +SSSSSSSSSSSSSSSSSSSSSS +------------------------------ +[k8s.io] Variable Expansion + should allow composing env vars into new env vars [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [k8s.io] Variable Expansion + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +STEP: Creating a kubernetes client +Apr 29 12:07:55.853: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174 +STEP: Building a namespace api object, basename var-expansion +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in var-expansion-2046 +STEP: Waiting for a default service account to be provisioned in namespace +[It] should allow composing env vars into new env vars [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: Creating a pod to test env composition +Apr 29 12:07:56.079: INFO: Waiting up to 5m0s for pod "var-expansion-6c35b215-6a77-11e9-b6ef-0e71f8c7e888" in namespace "var-expansion-2046" to be "success or failure" +Apr 29 12:07:56.097: INFO: Pod "var-expansion-6c35b215-6a77-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 17.951081ms +Apr 29 12:07:58.105: INFO: Pod "var-expansion-6c35b215-6a77-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 2.025810733s +Apr 29 12:08:00.113: INFO: Pod "var-expansion-6c35b215-6a77-11e9-b6ef-0e71f8c7e888": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.034281629s +STEP: Saw pod success +Apr 29 12:08:00.114: INFO: Pod "var-expansion-6c35b215-6a77-11e9-b6ef-0e71f8c7e888" satisfied condition "success or failure" +Apr 29 12:08:00.124: INFO: Trying to get logs from node worker-4av2h-6665d7b5ff-9nc42 pod var-expansion-6c35b215-6a77-11e9-b6ef-0e71f8c7e888 container dapi-container: +STEP: delete the pod +Apr 29 12:08:00.161: INFO: Waiting for pod var-expansion-6c35b215-6a77-11e9-b6ef-0e71f8c7e888 to disappear +Apr 29 12:08:00.170: INFO: Pod var-expansion-6c35b215-6a77-11e9-b6ef-0e71f8c7e888 no longer exists +[AfterEach] [k8s.io] Variable Expansion + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Apr 29 12:08:00.170: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "var-expansion-2046" for this suite. +Apr 29 12:08:06.209: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Apr 29 12:08:06.461: INFO: namespace var-expansion-2046 deletion completed in 6.275472448s + +• [SLOW TEST:10.608 seconds] +[k8s.io] Variable Expansion +/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687 + should allow composing env vars into new env vars [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +------------------------------ +SSSSSSSS +------------------------------ +[sig-cli] Kubectl client [k8s.io] Kubectl label + should update the label on a resource [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [sig-cli] Kubectl client + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +STEP: Creating a kubernetes client +Apr 29 12:08:06.462: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174 +STEP: Building a namespace api object, basename kubectl +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-6279 +STEP: Waiting for a default service account to be provisioned in namespace +[BeforeEach] [sig-cli] Kubectl client + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:213 +[BeforeEach] [k8s.io] Kubectl label + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1108 +STEP: creating the pod +Apr 29 12:08:06.676: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 create -f - --namespace=kubectl-6279' +Apr 29 12:08:09.469: INFO: stderr: "" +Apr 29 12:08:09.469: INFO: stdout: "pod/pause created\n" +Apr 29 12:08:09.469: INFO: Waiting up to 5m0s for 1 pods to be running and ready: [pause] +Apr 29 12:08:09.469: INFO: Waiting up to 5m0s for pod "pause" in namespace "kubectl-6279" to be "running and ready" +Apr 29 12:08:09.479: INFO: Pod "pause": Phase="Pending", Reason="", readiness=false. Elapsed: 9.882906ms +Apr 29 12:08:11.487: INFO: Pod "pause": Phase="Pending", Reason="", readiness=false. Elapsed: 2.017241427s +Apr 29 12:08:13.493: INFO: Pod "pause": Phase="Running", Reason="", readiness=true. Elapsed: 4.023992223s +Apr 29 12:08:13.493: INFO: Pod "pause" satisfied condition "running and ready" +Apr 29 12:08:13.493: INFO: Wanted all 1 pods to be running and ready. Result: true. Pods: [pause] +[It] should update the label on a resource [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: adding the label testing-label with value testing-label-value to a pod +Apr 29 12:08:13.494: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 label pods pause testing-label=testing-label-value --namespace=kubectl-6279' +Apr 29 12:08:13.658: INFO: stderr: "" +Apr 29 12:08:13.658: INFO: stdout: "pod/pause labeled\n" +STEP: verifying the pod has the label testing-label with the value testing-label-value +Apr 29 12:08:13.658: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 get pod pause -L testing-label --namespace=kubectl-6279' +Apr 29 12:08:13.758: INFO: stderr: "" +Apr 29 12:08:13.758: INFO: stdout: "NAME READY STATUS RESTARTS AGE TESTING-LABEL\npause 1/1 Running 0 4s testing-label-value\n" +STEP: removing the label testing-label of a pod +Apr 29 12:08:13.758: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 label pods pause testing-label- --namespace=kubectl-6279' +Apr 29 12:08:13.919: INFO: stderr: "" +Apr 29 12:08:13.919: INFO: stdout: "pod/pause labeled\n" +STEP: verifying the pod doesn't have the label testing-label +Apr 29 12:08:13.919: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 get pod pause -L testing-label --namespace=kubectl-6279' +Apr 29 12:08:14.031: INFO: stderr: "" +Apr 29 12:08:14.031: INFO: stdout: "NAME READY STATUS RESTARTS AGE TESTING-LABEL\npause 1/1 Running 0 4s \n" +[AfterEach] [k8s.io] Kubectl label + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1115 +STEP: using delete to clean up resources +Apr 29 12:08:14.031: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 delete --grace-period=0 --force -f - --namespace=kubectl-6279' +Apr 29 12:08:14.184: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" +Apr 29 12:08:14.184: INFO: stdout: "pod \"pause\" force deleted\n" +Apr 29 12:08:14.185: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 get rc,svc -l name=pause --no-headers --namespace=kubectl-6279' +Apr 29 12:08:14.307: INFO: stderr: "No resources found.\n" +Apr 29 12:08:14.307: INFO: stdout: "" +Apr 29 12:08:14.307: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 get pods -l name=pause --namespace=kubectl-6279 -o go-template={{ range .items }}{{ if not .metadata.deletionTimestamp }}{{ .metadata.name }}{{ "\n" }}{{ end }}{{ end }}' +Apr 29 12:08:14.403: INFO: stderr: "" +Apr 29 12:08:14.404: INFO: stdout: "" +[AfterEach] [sig-cli] Kubectl client + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Apr 29 12:08:14.404: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "kubectl-6279" for this suite. +Apr 29 12:08:20.426: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Apr 29 12:08:20.632: INFO: namespace kubectl-6279 deletion completed in 6.223241765s + +• [SLOW TEST:14.171 seconds] +[sig-cli] Kubectl client +/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 + [k8s.io] Kubectl label + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687 + should update the label on a resource [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +------------------------------ +SSSSSSSSSSSS +------------------------------ +[sig-storage] Projected configMap + optional updates should be reflected in volume [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] [sig-storage] Projected configMap + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +STEP: Creating a kubernetes client +Apr 29 12:08:20.635: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174 +STEP: Building a namespace api object, basename projected +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-798 +STEP: Waiting for a default service account to be provisioned in namespace +[It] optional updates should be reflected in volume [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +STEP: Creating configMap with name cm-test-opt-del-7af7e4f1-6a77-11e9-b6ef-0e71f8c7e888 +STEP: Creating configMap with name cm-test-opt-upd-7af7e540-6a77-11e9-b6ef-0e71f8c7e888 +STEP: Creating the pod +STEP: Deleting configmap cm-test-opt-del-7af7e4f1-6a77-11e9-b6ef-0e71f8c7e888 +STEP: Updating configmap cm-test-opt-upd-7af7e540-6a77-11e9-b6ef-0e71f8c7e888 +STEP: Creating configMap with name cm-test-opt-create-7af7e55e-6a77-11e9-b6ef-0e71f8c7e888 +STEP: waiting to observe update in volume +[AfterEach] [sig-storage] Projected configMap + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 +Apr 29 12:08:35.036: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "projected-798" for this suite. +Apr 29 12:08:59.073: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered +Apr 29 12:08:59.269: INFO: namespace projected-798 deletion completed in 24.225057206s + +• [SLOW TEST:38.634 seconds] +[sig-storage] Projected configMap +/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_configmap.go:33 + optional updates should be reflected in volume [NodeConformance] [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +------------------------------ +SSSSSSSSSSSSS +------------------------------ +[sig-network] Proxy version v1 + should proxy logs on node using proxy subresource [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +[BeforeEach] version v1 + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149 +STEP: Creating a kubernetes client +Apr 29 12:08:59.270: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174 +STEP: Building a namespace api object, basename proxy +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in proxy-1354 +STEP: Waiting for a default service account to be provisioned in namespace +[It] should proxy logs on node using proxy subresource [Conformance] + /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 +Apr 29 12:08:59.464: INFO: (0) /api/v1/nodes/worker-3oo4n-86cbf586b7-9q98g/proxy/logs/:
+btmp
+containers/
+faillog... (200; 7.024721ms)
+Apr 29 12:08:59.473: INFO: (1) /api/v1/nodes/worker-3oo4n-86cbf586b7-9q98g/proxy/logs/: 
+btmp
+containers/
+faillog... (200; 8.669638ms)
+Apr 29 12:08:59.483: INFO: (2) /api/v1/nodes/worker-3oo4n-86cbf586b7-9q98g/proxy/logs/: 
+btmp
+containers/
+faillog... (200; 9.489522ms)
+Apr 29 12:08:59.496: INFO: (3) /api/v1/nodes/worker-3oo4n-86cbf586b7-9q98g/proxy/logs/: 
+btmp
+containers/
+faillog... (200; 13.485665ms)
+Apr 29 12:08:59.511: INFO: (4) /api/v1/nodes/worker-3oo4n-86cbf586b7-9q98g/proxy/logs/: 
+btmp
+containers/
+faillog... (200; 14.059682ms)
+Apr 29 12:08:59.523: INFO: (5) /api/v1/nodes/worker-3oo4n-86cbf586b7-9q98g/proxy/logs/: 
+btmp
+containers/
+faillog... (200; 11.53283ms)
+Apr 29 12:08:59.533: INFO: (6) /api/v1/nodes/worker-3oo4n-86cbf586b7-9q98g/proxy/logs/: 
+btmp
+containers/
+faillog... (200; 9.804839ms)
+Apr 29 12:08:59.540: INFO: (7) /api/v1/nodes/worker-3oo4n-86cbf586b7-9q98g/proxy/logs/: 
+btmp
+containers/
+faillog... (200; 7.532564ms)
+Apr 29 12:08:59.553: INFO: (8) /api/v1/nodes/worker-3oo4n-86cbf586b7-9q98g/proxy/logs/: 
+btmp
+containers/
+faillog... (200; 12.283229ms)
+Apr 29 12:08:59.565: INFO: (9) /api/v1/nodes/worker-3oo4n-86cbf586b7-9q98g/proxy/logs/: 
+btmp
+containers/
+faillog... (200; 11.936675ms)
+Apr 29 12:08:59.573: INFO: (10) /api/v1/nodes/worker-3oo4n-86cbf586b7-9q98g/proxy/logs/: 
+btmp
+containers/
+faillog... (200; 7.679301ms)
+Apr 29 12:08:59.582: INFO: (11) /api/v1/nodes/worker-3oo4n-86cbf586b7-9q98g/proxy/logs/: 
+btmp
+containers/
+faillog... (200; 9.554572ms)
+Apr 29 12:08:59.588: INFO: (12) /api/v1/nodes/worker-3oo4n-86cbf586b7-9q98g/proxy/logs/: 
+btmp
+containers/
+faillog... (200; 6.055898ms)
+Apr 29 12:08:59.595: INFO: (13) /api/v1/nodes/worker-3oo4n-86cbf586b7-9q98g/proxy/logs/: 
+btmp
+containers/
+faillog... (200; 6.201414ms)
+Apr 29 12:08:59.600: INFO: (14) /api/v1/nodes/worker-3oo4n-86cbf586b7-9q98g/proxy/logs/: 
+btmp
+containers/
+faillog... (200; 5.771524ms)
+Apr 29 12:08:59.606: INFO: (15) /api/v1/nodes/worker-3oo4n-86cbf586b7-9q98g/proxy/logs/: 
+btmp
+containers/
+faillog... (200; 5.897002ms)
+Apr 29 12:08:59.618: INFO: (16) /api/v1/nodes/worker-3oo4n-86cbf586b7-9q98g/proxy/logs/: 
+btmp
+containers/
+faillog... (200; 11.974889ms)
+Apr 29 12:08:59.630: INFO: (17) /api/v1/nodes/worker-3oo4n-86cbf586b7-9q98g/proxy/logs/: 
+btmp
+containers/
+faillog... (200; 11.127311ms)
+Apr 29 12:08:59.639: INFO: (18) /api/v1/nodes/worker-3oo4n-86cbf586b7-9q98g/proxy/logs/: 
+btmp
+containers/
+faillog... (200; 8.854703ms)
+Apr 29 12:08:59.645: INFO: (19) /api/v1/nodes/worker-3oo4n-86cbf586b7-9q98g/proxy/logs/: 
+btmp
+containers/
+faillog... (200; 6.185019ms)
+[AfterEach] version v1
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 12:08:59.645: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "proxy-1354" for this suite.
+Apr 29 12:09:05.670: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 12:09:05.874: INFO: namespace proxy-1354 deletion completed in 6.222237579s
+
+• [SLOW TEST:6.604 seconds]
+[sig-network] Proxy
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:22
+  version v1
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/proxy.go:56
+    should proxy logs on node using proxy subresource  [Conformance]
+    /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSSSSSSS
+------------------------------
+[sig-cli] Kubectl client [k8s.io] Kubectl expose 
+  should create services for rc  [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-cli] Kubectl client
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 12:09:05.874: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename kubectl
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-1934
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [sig-cli] Kubectl client
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:213
+[It] should create services for rc  [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: creating Redis RC
+Apr 29 12:09:06.070: INFO: namespace kubectl-1934
+Apr 29 12:09:06.070: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 create -f - --namespace=kubectl-1934'
+Apr 29 12:09:06.408: INFO: stderr: ""
+Apr 29 12:09:06.408: INFO: stdout: "replicationcontroller/redis-master created\n"
+STEP: Waiting for Redis master to start.
+Apr 29 12:09:07.414: INFO: Selector matched 1 pods for map[app:redis]
+Apr 29 12:09:07.414: INFO: Found 0 / 1
+Apr 29 12:09:08.416: INFO: Selector matched 1 pods for map[app:redis]
+Apr 29 12:09:08.416: INFO: Found 0 / 1
+Apr 29 12:09:09.415: INFO: Selector matched 1 pods for map[app:redis]
+Apr 29 12:09:09.415: INFO: Found 1 / 1
+Apr 29 12:09:09.415: INFO: WaitFor completed with timeout 5m0s.  Pods found = 1 out of 1
+Apr 29 12:09:09.419: INFO: Selector matched 1 pods for map[app:redis]
+Apr 29 12:09:09.420: INFO: ForEach: Found 1 pods from the filter.  Now looping through them.
+Apr 29 12:09:09.420: INFO: wait on redis-master startup in kubectl-1934 
+Apr 29 12:09:09.420: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 logs redis-master-2r4k7 redis-master --namespace=kubectl-1934'
+Apr 29 12:09:09.661: INFO: stderr: ""
+Apr 29 12:09:09.661: INFO: stdout: "                _._                                                  \n           _.-``__ ''-._                                             \n      _.-``    `.  `_.  ''-._           Redis 3.2.12 (35a5711f/0) 64 bit\n  .-`` .-```.  ```\\/    _.,_ ''-._                                   \n (    '      ,       .-`  | `,    )     Running in standalone mode\n |`-._`-...-` __...-.``-._|'` _.-'|     Port: 6379\n |    `-._   `._    /     _.-'    |     PID: 1\n  `-._    `-._  `-./  _.-'    _.-'                                   \n |`-._`-._    `-.__.-'    _.-'_.-'|                                  \n |    `-._`-._        _.-'_.-'    |           http://redis.io        \n  `-._    `-._`-.__.-'_.-'    _.-'                                   \n |`-._`-._    `-.__.-'    _.-'_.-'|                                  \n |    `-._`-._        _.-'_.-'    |                                  \n  `-._    `-._`-.__.-'_.-'    _.-'                                   \n      `-._    `-.__.-'    _.-'                                       \n          `-._        _.-'                                           \n              `-.__.-'                                               \n\n1:M 29 Apr 12:09:08.471 # WARNING: The TCP backlog setting of 511 cannot be enforced because /proc/sys/net/core/somaxconn is set to the lower value of 128.\n1:M 29 Apr 12:09:08.471 # Server started, Redis version 3.2.12\n1:M 29 Apr 12:09:08.471 # WARNING you have Transparent Huge Pages (THP) support enabled in your kernel. This will create latency and memory usage issues with Redis. To fix this issue run the command 'echo never > /sys/kernel/mm/transparent_hugepage/enabled' as root, and add it to your /etc/rc.local in order to retain the setting after a reboot. Redis must be restarted after THP is disabled.\n1:M 29 Apr 12:09:08.471 * The server is now ready to accept connections on port 6379\n"
+STEP: exposing RC
+Apr 29 12:09:09.662: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 expose rc redis-master --name=rm2 --port=1234 --target-port=6379 --namespace=kubectl-1934'
+Apr 29 12:09:09.878: INFO: stderr: ""
+Apr 29 12:09:09.878: INFO: stdout: "service/rm2 exposed\n"
+Apr 29 12:09:09.883: INFO: Service rm2 in namespace kubectl-1934 found.
+STEP: exposing service
+Apr 29 12:09:11.916: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 expose service rm2 --name=rm3 --port=2345 --target-port=6379 --namespace=kubectl-1934'
+Apr 29 12:09:12.056: INFO: stderr: ""
+Apr 29 12:09:12.056: INFO: stdout: "service/rm3 exposed\n"
+Apr 29 12:09:12.061: INFO: Service rm3 in namespace kubectl-1934 found.
+[AfterEach] [sig-cli] Kubectl client
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 12:09:14.078: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "kubectl-1934" for this suite.
+Apr 29 12:09:38.138: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 12:09:38.381: INFO: namespace kubectl-1934 deletion completed in 24.295561998s
+
+• [SLOW TEST:32.507 seconds]
+[sig-cli] Kubectl client
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23
+  [k8s.io] Kubectl expose
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
+    should create services for rc  [Conformance]
+    /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSSSSSSSS
+------------------------------
+[sig-cli] Kubectl client [k8s.io] Update Demo 
+  should create and stop a replication controller  [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-cli] Kubectl client
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 12:09:38.386: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename kubectl
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-2818
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [sig-cli] Kubectl client
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:213
+[BeforeEach] [k8s.io] Update Demo
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:265
+[It] should create and stop a replication controller  [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: creating a replication controller
+Apr 29 12:09:38.577: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 create -f - --namespace=kubectl-2818'
+Apr 29 12:09:38.860: INFO: stderr: ""
+Apr 29 12:09:38.860: INFO: stdout: "replicationcontroller/update-demo-nautilus created\n"
+STEP: waiting for all containers in name=update-demo pods to come up.
+Apr 29 12:09:38.860: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-2818'
+Apr 29 12:09:39.011: INFO: stderr: ""
+Apr 29 12:09:39.011: INFO: stdout: "update-demo-nautilus-pqks6 update-demo-nautilus-r2vcq "
+Apr 29 12:09:39.012: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 get pods update-demo-nautilus-pqks6 -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-2818'
+Apr 29 12:09:39.120: INFO: stderr: ""
+Apr 29 12:09:39.120: INFO: stdout: ""
+Apr 29 12:09:39.120: INFO: update-demo-nautilus-pqks6 is created but not running
+Apr 29 12:09:44.120: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-2818'
+Apr 29 12:09:44.223: INFO: stderr: ""
+Apr 29 12:09:44.223: INFO: stdout: "update-demo-nautilus-pqks6 update-demo-nautilus-r2vcq "
+Apr 29 12:09:44.223: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 get pods update-demo-nautilus-pqks6 -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-2818'
+Apr 29 12:09:44.323: INFO: stderr: ""
+Apr 29 12:09:44.323: INFO: stdout: "true"
+Apr 29 12:09:44.323: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 get pods update-demo-nautilus-pqks6 -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-2818'
+Apr 29 12:09:44.420: INFO: stderr: ""
+Apr 29 12:09:44.420: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0"
+Apr 29 12:09:44.420: INFO: validating pod update-demo-nautilus-pqks6
+Apr 29 12:09:44.431: INFO: got data: {
+  "image": "nautilus.jpg"
+}
+
+Apr 29 12:09:44.431: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg .
+Apr 29 12:09:44.431: INFO: update-demo-nautilus-pqks6 is verified up and running
+Apr 29 12:09:44.431: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 get pods update-demo-nautilus-r2vcq -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-2818'
+Apr 29 12:09:44.564: INFO: stderr: ""
+Apr 29 12:09:44.565: INFO: stdout: "true"
+Apr 29 12:09:44.565: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 get pods update-demo-nautilus-r2vcq -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-2818'
+Apr 29 12:09:44.680: INFO: stderr: ""
+Apr 29 12:09:44.680: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0"
+Apr 29 12:09:44.680: INFO: validating pod update-demo-nautilus-r2vcq
+Apr 29 12:09:44.689: INFO: got data: {
+  "image": "nautilus.jpg"
+}
+
+Apr 29 12:09:44.689: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg .
+Apr 29 12:09:44.689: INFO: update-demo-nautilus-r2vcq is verified up and running
+STEP: using delete to clean up resources
+Apr 29 12:09:44.689: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 delete --grace-period=0 --force -f - --namespace=kubectl-2818'
+Apr 29 12:09:44.817: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n"
+Apr 29 12:09:44.817: INFO: stdout: "replicationcontroller \"update-demo-nautilus\" force deleted\n"
+Apr 29 12:09:44.817: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 get rc,svc -l name=update-demo --no-headers --namespace=kubectl-2818'
+Apr 29 12:09:44.978: INFO: stderr: "No resources found.\n"
+Apr 29 12:09:44.978: INFO: stdout: ""
+Apr 29 12:09:44.979: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 get pods -l name=update-demo --namespace=kubectl-2818 -o go-template={{ range .items }}{{ if not .metadata.deletionTimestamp }}{{ .metadata.name }}{{ "\n" }}{{ end }}{{ end }}'
+Apr 29 12:09:45.098: INFO: stderr: ""
+Apr 29 12:09:45.098: INFO: stdout: ""
+[AfterEach] [sig-cli] Kubectl client
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 12:09:45.098: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "kubectl-2818" for this suite.
+Apr 29 12:10:09.157: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 12:10:09.369: INFO: namespace kubectl-2818 deletion completed in 24.262951698s
+
+• [SLOW TEST:30.984 seconds]
+[sig-cli] Kubectl client
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23
+  [k8s.io] Update Demo
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
+    should create and stop a replication controller  [Conformance]
+    /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+------------------------------
+[sig-storage] Secrets 
+  should be consumable from pods in volume with mappings and Item Mode set [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-storage] Secrets
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 12:10:09.370: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename secrets
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in secrets-4541
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should be consumable from pods in volume with mappings and Item Mode set [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating secret with name secret-test-map-bbc60914-6a77-11e9-b6ef-0e71f8c7e888
+STEP: Creating a pod to test consume secrets
+Apr 29 12:10:09.573: INFO: Waiting up to 5m0s for pod "pod-secrets-bbc704bb-6a77-11e9-b6ef-0e71f8c7e888" in namespace "secrets-4541" to be "success or failure"
+Apr 29 12:10:09.579: INFO: Pod "pod-secrets-bbc704bb-6a77-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 5.374336ms
+Apr 29 12:10:11.586: INFO: Pod "pod-secrets-bbc704bb-6a77-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 2.012433454s
+Apr 29 12:10:13.594: INFO: Pod "pod-secrets-bbc704bb-6a77-11e9-b6ef-0e71f8c7e888": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.020701555s
+STEP: Saw pod success
+Apr 29 12:10:13.594: INFO: Pod "pod-secrets-bbc704bb-6a77-11e9-b6ef-0e71f8c7e888" satisfied condition "success or failure"
+Apr 29 12:10:13.600: INFO: Trying to get logs from node worker-4av2h-6665d7b5ff-9nc42 pod pod-secrets-bbc704bb-6a77-11e9-b6ef-0e71f8c7e888 container secret-volume-test: 
+STEP: delete the pod
+Apr 29 12:10:13.642: INFO: Waiting for pod pod-secrets-bbc704bb-6a77-11e9-b6ef-0e71f8c7e888 to disappear
+Apr 29 12:10:13.647: INFO: Pod pod-secrets-bbc704bb-6a77-11e9-b6ef-0e71f8c7e888 no longer exists
+[AfterEach] [sig-storage] Secrets
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 12:10:13.647: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "secrets-4541" for this suite.
+Apr 29 12:10:19.674: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 12:10:19.882: INFO: namespace secrets-4541 deletion completed in 6.227539484s
+
+• [SLOW TEST:10.512 seconds]
+[sig-storage] Secrets
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/secrets_volume.go:33
+  should be consumable from pods in volume with mappings and Item Mode set [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSSSSSSSS
+------------------------------
+[sig-cli] Kubectl client [k8s.io] Kubectl run job 
+  should create a job from an image when restart is OnFailure  [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-cli] Kubectl client
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 12:10:19.883: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename kubectl
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-2499
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [sig-cli] Kubectl client
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:213
+[BeforeEach] [k8s.io] Kubectl run job
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1510
+[It] should create a job from an image when restart is OnFailure  [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: running the image docker.io/library/nginx:1.14-alpine
+Apr 29 12:10:20.090: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 run e2e-test-nginx-job --restart=OnFailure --generator=job/v1 --image=docker.io/library/nginx:1.14-alpine --namespace=kubectl-2499'
+Apr 29 12:10:20.230: INFO: stderr: "kubectl run --generator=job/v1 is DEPRECATED and will be removed in a future version. Use kubectl run --generator=run-pod/v1 or kubectl create instead.\n"
+Apr 29 12:10:20.230: INFO: stdout: "job.batch/e2e-test-nginx-job created\n"
+STEP: verifying the job e2e-test-nginx-job was created
+[AfterEach] [k8s.io] Kubectl run job
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1515
+Apr 29 12:10:20.248: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 delete jobs e2e-test-nginx-job --namespace=kubectl-2499'
+Apr 29 12:10:20.415: INFO: stderr: ""
+Apr 29 12:10:20.415: INFO: stdout: "job.batch \"e2e-test-nginx-job\" deleted\n"
+[AfterEach] [sig-cli] Kubectl client
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 12:10:20.415: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "kubectl-2499" for this suite.
+Apr 29 12:10:26.449: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 12:10:26.670: INFO: namespace kubectl-2499 deletion completed in 6.243484203s
+
+• [SLOW TEST:6.787 seconds]
+[sig-cli] Kubectl client
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23
+  [k8s.io] Kubectl run job
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
+    should create a job from an image when restart is OnFailure  [Conformance]
+    /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSSSSSSSSSSS
+------------------------------
+[k8s.io] Probing container 
+  with readiness probe should not be ready before initial delay and never restart [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [k8s.io] Probing container
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 12:10:26.671: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename container-probe
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in container-probe-4816
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [k8s.io] Probing container
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/container_probe.go:51
+[It] with readiness probe should not be ready before initial delay and never restart [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+Apr 29 12:10:52.890: INFO: Container started at 2019-04-29 12:10:28 +0000 UTC, pod became ready at 2019-04-29 12:10:52 +0000 UTC
+[AfterEach] [k8s.io] Probing container
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 12:10:52.891: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "container-probe-4816" for this suite.
+Apr 29 12:11:16.925: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 12:11:17.100: INFO: namespace container-probe-4816 deletion completed in 24.197931501s
+
+• [SLOW TEST:50.430 seconds]
+[k8s.io] Probing container
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
+  with readiness probe should not be ready before initial delay and never restart [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSSSSSSSSS
+------------------------------
+[sig-api-machinery] Garbage collector 
+  should keep the rc around until all its pods are deleted if the deleteOptions says so [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-api-machinery] Garbage collector
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 12:11:17.109: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename gc
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in gc-3511
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should keep the rc around until all its pods are deleted if the deleteOptions says so [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: create the rc
+STEP: delete the rc
+STEP: wait for the rc to be deleted
+Apr 29 12:11:23.481: INFO: 0 pods remaining
+Apr 29 12:11:23.481: INFO: 0 pods has nil DeletionTimestamp
+Apr 29 12:11:23.481: INFO: 
+STEP: Gathering metrics
+W0429 12:11:24.391181      15 metrics_grabber.go:79] Master node is not registered. Grabbing metrics from Scheduler, ControllerManager and ClusterAutoscaler is disabled.
+Apr 29 12:11:24.391: INFO: For apiserver_request_total:
+For apiserver_request_latencies_summary:
+For apiserver_init_events_total:
+For garbage_collector_attempt_to_delete_queue_latency:
+For garbage_collector_attempt_to_delete_work_duration:
+For garbage_collector_attempt_to_orphan_queue_latency:
+For garbage_collector_attempt_to_orphan_work_duration:
+For garbage_collector_dirty_processing_latency_microseconds:
+For garbage_collector_event_processing_latency_microseconds:
+For garbage_collector_graph_changes_queue_latency:
+For garbage_collector_graph_changes_work_duration:
+For garbage_collector_orphan_processing_latency_microseconds:
+For namespace_queue_latency:
+For namespace_queue_latency_sum:
+For namespace_queue_latency_count:
+For namespace_retries:
+For namespace_work_duration:
+For namespace_work_duration_sum:
+For namespace_work_duration_count:
+For function_duration_seconds:
+For errors_total:
+For evicted_pods_total:
+
+[AfterEach] [sig-api-machinery] Garbage collector
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 12:11:24.391: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "gc-3511" for this suite.
+Apr 29 12:11:32.431: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 12:11:32.599: INFO: namespace gc-3511 deletion completed in 8.198536638s
+
+• [SLOW TEST:15.491 seconds]
+[sig-api-machinery] Garbage collector
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:22
+  should keep the rc around until all its pods are deleted if the deleteOptions says so [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+------------------------------
+[sig-api-machinery] Secrets 
+  should be consumable via the environment [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-api-machinery] Secrets
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 12:11:32.602: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename secrets
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in secrets-3107
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should be consumable via the environment [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: creating secret secrets-3107/secret-test-ed62f233-6a77-11e9-b6ef-0e71f8c7e888
+STEP: Creating a pod to test consume secrets
+Apr 29 12:11:32.804: INFO: Waiting up to 5m0s for pod "pod-configmaps-ed646261-6a77-11e9-b6ef-0e71f8c7e888" in namespace "secrets-3107" to be "success or failure"
+Apr 29 12:11:32.809: INFO: Pod "pod-configmaps-ed646261-6a77-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 4.880222ms
+Apr 29 12:11:34.818: INFO: Pod "pod-configmaps-ed646261-6a77-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 2.013500684s
+Apr 29 12:11:36.825: INFO: Pod "pod-configmaps-ed646261-6a77-11e9-b6ef-0e71f8c7e888": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.020519459s
+STEP: Saw pod success
+Apr 29 12:11:36.825: INFO: Pod "pod-configmaps-ed646261-6a77-11e9-b6ef-0e71f8c7e888" satisfied condition "success or failure"
+Apr 29 12:11:36.828: INFO: Trying to get logs from node worker-4av2h-6665d7b5ff-9nc42 pod pod-configmaps-ed646261-6a77-11e9-b6ef-0e71f8c7e888 container env-test: 
+STEP: delete the pod
+Apr 29 12:11:36.868: INFO: Waiting for pod pod-configmaps-ed646261-6a77-11e9-b6ef-0e71f8c7e888 to disappear
+Apr 29 12:11:36.873: INFO: Pod pod-configmaps-ed646261-6a77-11e9-b6ef-0e71f8c7e888 no longer exists
+[AfterEach] [sig-api-machinery] Secrets
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 12:11:36.873: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "secrets-3107" for this suite.
+Apr 29 12:11:42.896: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 12:11:43.136: INFO: namespace secrets-3107 deletion completed in 6.257297189s
+
+• [SLOW TEST:10.534 seconds]
+[sig-api-machinery] Secrets
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/secrets.go:32
+  should be consumable via the environment [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSSSSSS
+------------------------------
+[sig-storage] Downward API volume 
+  should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-storage] Downward API volume
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 12:11:43.138: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename downward-api
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in downward-api-9230
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [sig-storage] Downward API volume
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:39
+[It] should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating a pod to test downward API volume plugin
+Apr 29 12:11:43.410: INFO: Waiting up to 5m0s for pod "downwardapi-volume-f3b14219-6a77-11e9-b6ef-0e71f8c7e888" in namespace "downward-api-9230" to be "success or failure"
+Apr 29 12:11:43.436: INFO: Pod "downwardapi-volume-f3b14219-6a77-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 26.192446ms
+Apr 29 12:11:45.444: INFO: Pod "downwardapi-volume-f3b14219-6a77-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 2.034265933s
+Apr 29 12:11:47.450: INFO: Pod "downwardapi-volume-f3b14219-6a77-11e9-b6ef-0e71f8c7e888": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.040178989s
+STEP: Saw pod success
+Apr 29 12:11:47.450: INFO: Pod "downwardapi-volume-f3b14219-6a77-11e9-b6ef-0e71f8c7e888" satisfied condition "success or failure"
+Apr 29 12:11:47.460: INFO: Trying to get logs from node worker-4av2h-6665d7b5ff-9nc42 pod downwardapi-volume-f3b14219-6a77-11e9-b6ef-0e71f8c7e888 container client-container: 
+STEP: delete the pod
+Apr 29 12:11:47.512: INFO: Waiting for pod downwardapi-volume-f3b14219-6a77-11e9-b6ef-0e71f8c7e888 to disappear
+Apr 29 12:11:47.519: INFO: Pod downwardapi-volume-f3b14219-6a77-11e9-b6ef-0e71f8c7e888 no longer exists
+[AfterEach] [sig-storage] Downward API volume
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 12:11:47.519: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "downward-api-9230" for this suite.
+Apr 29 12:11:53.549: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 12:11:53.777: INFO: namespace downward-api-9230 deletion completed in 6.248009269s
+
+• [SLOW TEST:10.640 seconds]
+[sig-storage] Downward API volume
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:34
+  should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+------------------------------
+[sig-apps] Daemon set [Serial] 
+  should retry creating failed daemon pods [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-apps] Daemon set [Serial]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 12:11:53.779: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename daemonsets
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in daemonsets-8857
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [sig-apps] Daemon set [Serial]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:102
+[It] should retry creating failed daemon pods [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating a simple DaemonSet "daemon-set"
+STEP: Check that daemon pods launch on every node of the cluster.
+Apr 29 12:11:54.098: INFO: DaemonSet pods can't tolerate node master-h2h3g-655686fb8c-8kgdn with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node
+Apr 29 12:11:54.105: INFO: Number of nodes with available pods: 0
+Apr 29 12:11:54.105: INFO: Node worker-3oo4n-86cbf586b7-9q98g is running more than one daemon pod
+Apr 29 12:11:55.116: INFO: DaemonSet pods can't tolerate node master-h2h3g-655686fb8c-8kgdn with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node
+Apr 29 12:11:55.124: INFO: Number of nodes with available pods: 0
+Apr 29 12:11:55.124: INFO: Node worker-3oo4n-86cbf586b7-9q98g is running more than one daemon pod
+Apr 29 12:11:56.116: INFO: DaemonSet pods can't tolerate node master-h2h3g-655686fb8c-8kgdn with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node
+Apr 29 12:11:56.127: INFO: Number of nodes with available pods: 0
+Apr 29 12:11:56.127: INFO: Node worker-3oo4n-86cbf586b7-9q98g is running more than one daemon pod
+Apr 29 12:11:57.122: INFO: DaemonSet pods can't tolerate node master-h2h3g-655686fb8c-8kgdn with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node
+Apr 29 12:11:57.138: INFO: Number of nodes with available pods: 3
+Apr 29 12:11:57.138: INFO: Number of running nodes: 3, number of available pods: 3
+STEP: Set a daemon pod's phase to 'Failed', check that the daemon pod is revived.
+Apr 29 12:11:57.194: INFO: DaemonSet pods can't tolerate node master-h2h3g-655686fb8c-8kgdn with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node
+Apr 29 12:11:57.216: INFO: Number of nodes with available pods: 2
+Apr 29 12:11:57.218: INFO: Node worker-4av2h-6665d7b5ff-9nc42 is running more than one daemon pod
+Apr 29 12:11:58.230: INFO: DaemonSet pods can't tolerate node master-h2h3g-655686fb8c-8kgdn with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node
+Apr 29 12:11:58.238: INFO: Number of nodes with available pods: 2
+Apr 29 12:11:58.239: INFO: Node worker-4av2h-6665d7b5ff-9nc42 is running more than one daemon pod
+Apr 29 12:11:59.227: INFO: DaemonSet pods can't tolerate node master-h2h3g-655686fb8c-8kgdn with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node
+Apr 29 12:11:59.233: INFO: Number of nodes with available pods: 2
+Apr 29 12:11:59.234: INFO: Node worker-4av2h-6665d7b5ff-9nc42 is running more than one daemon pod
+Apr 29 12:12:00.231: INFO: DaemonSet pods can't tolerate node master-h2h3g-655686fb8c-8kgdn with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node
+Apr 29 12:12:00.240: INFO: Number of nodes with available pods: 3
+Apr 29 12:12:00.240: INFO: Number of running nodes: 3, number of available pods: 3
+STEP: Wait for the failed daemon pod to be completely deleted.
+[AfterEach] [sig-apps] Daemon set [Serial]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:68
+STEP: Deleting DaemonSet "daemon-set"
+STEP: deleting DaemonSet.extensions daemon-set in namespace daemonsets-8857, will wait for the garbage collector to delete the pods
+Apr 29 12:12:00.323: INFO: Deleting DaemonSet.extensions daemon-set took: 13.402813ms
+Apr 29 12:12:00.524: INFO: Terminating DaemonSet.extensions daemon-set pods took: 200.987836ms
+Apr 29 12:12:14.438: INFO: Number of nodes with available pods: 0
+Apr 29 12:12:14.438: INFO: Number of running nodes: 0, number of available pods: 0
+Apr 29 12:12:14.443: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"selfLink":"/apis/apps/v1/namespaces/daemonsets-8857/daemonsets","resourceVersion":"13951"},"items":null}
+
+Apr 29 12:12:14.447: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"selfLink":"/api/v1/namespaces/daemonsets-8857/pods","resourceVersion":"13951"},"items":null}
+
+[AfterEach] [sig-apps] Daemon set [Serial]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 12:12:14.478: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "daemonsets-8857" for this suite.
+Apr 29 12:12:20.523: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 12:12:20.742: INFO: namespace daemonsets-8857 deletion completed in 6.243490102s
+
+• [SLOW TEST:26.963 seconds]
+[sig-apps] Daemon set [Serial]
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22
+  should retry creating failed daemon pods [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSSSSSSSSSSS
+------------------------------
+[sig-storage] Projected secret 
+  should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-storage] Projected secret
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 12:12:20.742: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename projected
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-3786
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating projection with secret that has name projected-secret-test-0a1a633a-6a78-11e9-b6ef-0e71f8c7e888
+STEP: Creating a pod to test consume secrets
+Apr 29 12:12:20.996: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-0a1bf069-6a78-11e9-b6ef-0e71f8c7e888" in namespace "projected-3786" to be "success or failure"
+Apr 29 12:12:21.003: INFO: Pod "pod-projected-secrets-0a1bf069-6a78-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 6.945032ms
+Apr 29 12:12:23.013: INFO: Pod "pod-projected-secrets-0a1bf069-6a78-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 2.017138808s
+Apr 29 12:12:25.021: INFO: Pod "pod-projected-secrets-0a1bf069-6a78-11e9-b6ef-0e71f8c7e888": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.025484055s
+STEP: Saw pod success
+Apr 29 12:12:25.022: INFO: Pod "pod-projected-secrets-0a1bf069-6a78-11e9-b6ef-0e71f8c7e888" satisfied condition "success or failure"
+Apr 29 12:12:25.027: INFO: Trying to get logs from node worker-4av2h-6665d7b5ff-9nc42 pod pod-projected-secrets-0a1bf069-6a78-11e9-b6ef-0e71f8c7e888 container projected-secret-volume-test: 
+STEP: delete the pod
+Apr 29 12:12:25.067: INFO: Waiting for pod pod-projected-secrets-0a1bf069-6a78-11e9-b6ef-0e71f8c7e888 to disappear
+Apr 29 12:12:25.084: INFO: Pod pod-projected-secrets-0a1bf069-6a78-11e9-b6ef-0e71f8c7e888 no longer exists
+[AfterEach] [sig-storage] Projected secret
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 12:12:25.084: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "projected-3786" for this suite.
+Apr 29 12:12:31.137: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 12:12:31.374: INFO: namespace projected-3786 deletion completed in 6.278151223s
+
+• [SLOW TEST:10.633 seconds]
+[sig-storage] Projected secret
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_secret.go:33
+  should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSSSSSSSSSSSS
+------------------------------
+[sig-storage] Secrets 
+  should be consumable from pods in volume [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-storage] Secrets
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 12:12:31.387: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename secrets
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in secrets-7786
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should be consumable from pods in volume [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating secret with name secret-test-106ee013-6a78-11e9-b6ef-0e71f8c7e888
+STEP: Creating a pod to test consume secrets
+Apr 29 12:12:31.605: INFO: Waiting up to 5m0s for pod "pod-secrets-106fef8b-6a78-11e9-b6ef-0e71f8c7e888" in namespace "secrets-7786" to be "success or failure"
+Apr 29 12:12:31.614: INFO: Pod "pod-secrets-106fef8b-6a78-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 8.215547ms
+Apr 29 12:12:33.625: INFO: Pod "pod-secrets-106fef8b-6a78-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 2.01955101s
+Apr 29 12:12:35.638: INFO: Pod "pod-secrets-106fef8b-6a78-11e9-b6ef-0e71f8c7e888": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.032712614s
+STEP: Saw pod success
+Apr 29 12:12:35.638: INFO: Pod "pod-secrets-106fef8b-6a78-11e9-b6ef-0e71f8c7e888" satisfied condition "success or failure"
+Apr 29 12:12:35.645: INFO: Trying to get logs from node worker-4av2h-6665d7b5ff-9nc42 pod pod-secrets-106fef8b-6a78-11e9-b6ef-0e71f8c7e888 container secret-volume-test: 
+STEP: delete the pod
+Apr 29 12:12:35.675: INFO: Waiting for pod pod-secrets-106fef8b-6a78-11e9-b6ef-0e71f8c7e888 to disappear
+Apr 29 12:12:35.680: INFO: Pod pod-secrets-106fef8b-6a78-11e9-b6ef-0e71f8c7e888 no longer exists
+[AfterEach] [sig-storage] Secrets
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 12:12:35.688: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "secrets-7786" for this suite.
+Apr 29 12:12:41.720: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 12:12:41.967: INFO: namespace secrets-7786 deletion completed in 6.270760896s
+
+• [SLOW TEST:10.580 seconds]
+[sig-storage] Secrets
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/secrets_volume.go:33
+  should be consumable from pods in volume [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSSSSSSSSSSS
+------------------------------
+[k8s.io] Probing container 
+  should *not* be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [k8s.io] Probing container
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 12:12:41.967: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename container-probe
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in container-probe-5739
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [k8s.io] Probing container
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/container_probe.go:51
+[It] should *not* be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating pod liveness-exec in namespace container-probe-5739
+Apr 29 12:12:46.226: INFO: Started pod liveness-exec in namespace container-probe-5739
+STEP: checking the pod's current state and verifying that restartCount is present
+Apr 29 12:12:46.232: INFO: Initial restart count of pod liveness-exec is 0
+STEP: deleting the pod
+[AfterEach] [k8s.io] Probing container
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 12:16:47.200: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "container-probe-5739" for this suite.
+Apr 29 12:16:53.226: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 12:16:53.448: INFO: namespace container-probe-5739 deletion completed in 6.239941513s
+
+• [SLOW TEST:251.481 seconds]
+[k8s.io] Probing container
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
+  should *not* be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSS
+------------------------------
+[sig-scheduling] SchedulerPredicates [Serial] 
+  validates that NodeSelector is respected if not matching  [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-scheduling] SchedulerPredicates [Serial]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 12:16:53.449: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename sched-pred
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in sched-pred-4105
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [sig-scheduling] SchedulerPredicates [Serial]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:79
+Apr 29 12:16:53.631: INFO: Waiting up to 1m0s for all (but 0) nodes to be ready
+Apr 29 12:16:53.651: INFO: Waiting for terminating namespaces to be deleted...
+Apr 29 12:16:53.656: INFO: 
+Logging pods the kubelet thinks is on node worker-3oo4n-86cbf586b7-9q98g before test
+Apr 29 12:16:53.674: INFO: kube-state-metrics-5fdb649879-lqkfk from kube-system started at 2019-04-29 11:17:01 +0000 UTC (2 container statuses recorded)
+Apr 29 12:16:53.674: INFO: 	Container addon-resizer ready: true, restart count 0
+Apr 29 12:16:53.674: INFO: 	Container kube-state-metrics ready: true, restart count 0
+Apr 29 12:16:53.674: INFO: coredns-58f7d854b4-xlcnp from kube-system started at 2019-04-29 11:16:23 +0000 UTC (1 container statuses recorded)
+Apr 29 12:16:53.674: INFO: 	Container coredns ready: true, restart count 0
+Apr 29 12:16:53.674: INFO: cert-exporter-kgwcd from kube-system started at 2019-04-29 11:16:36 +0000 UTC (1 container statuses recorded)
+Apr 29 12:16:53.674: INFO: 	Container cert-exporter ready: true, restart count 0
+Apr 29 12:16:53.674: INFO: calico-node-fmqgw from kube-system started at 2019-04-29 11:13:35 +0000 UTC (1 container statuses recorded)
+Apr 29 12:16:53.674: INFO: 	Container calico-node ready: true, restart count 0
+Apr 29 12:16:53.674: INFO: kube-proxy-dw2qf from kube-system started at 2019-04-29 11:13:55 +0000 UTC (1 container statuses recorded)
+Apr 29 12:16:53.674: INFO: 	Container kube-proxy ready: true, restart count 0
+Apr 29 12:16:53.674: INFO: node-exporter-j2kbr from kube-system started at 2019-04-29 11:21:42 +0000 UTC (1 container statuses recorded)
+Apr 29 12:16:53.674: INFO: 	Container node-exporter ready: true, restart count 0
+Apr 29 12:16:53.674: INFO: nginx-ingress-controller-6cffb97fb5-5fbv2 from kube-system started at 2019-04-29 11:23:55 +0000 UTC (1 container statuses recorded)
+Apr 29 12:16:53.674: INFO: 	Container nginx-ingress-controller ready: true, restart count 0
+Apr 29 12:16:53.674: INFO: net-exporter-wbqq5 from kube-system started at 2019-04-29 11:21:44 +0000 UTC (1 container statuses recorded)
+Apr 29 12:16:53.674: INFO: 	Container net-exporter ready: true, restart count 0
+Apr 29 12:16:53.674: INFO: sonobuoy-systemd-logs-daemon-set-8191105fdead4396-6mc85 from heptio-sonobuoy started at 2019-04-29 11:30:45 +0000 UTC (2 container statuses recorded)
+Apr 29 12:16:53.674: INFO: 	Container sonobuoy-worker ready: true, restart count 0
+Apr 29 12:16:53.674: INFO: 	Container systemd-logs ready: true, restart count 0
+Apr 29 12:16:53.674: INFO: 
+Logging pods the kubelet thinks is on node worker-4av2h-6665d7b5ff-9nc42 before test
+Apr 29 12:16:53.694: INFO: kube-proxy-mf5pp from kube-system started at 2019-04-29 11:13:55 +0000 UTC (1 container statuses recorded)
+Apr 29 12:16:53.695: INFO: 	Container kube-proxy ready: true, restart count 0
+Apr 29 12:16:53.695: INFO: net-exporter-vhll2 from kube-system started at 2019-04-29 11:21:44 +0000 UTC (1 container statuses recorded)
+Apr 29 12:16:53.695: INFO: 	Container net-exporter ready: true, restart count 0
+Apr 29 12:16:53.695: INFO: default-http-backend-77b6545877-zd4s2 from kube-system started at 2019-04-29 11:21:34 +0000 UTC (1 container statuses recorded)
+Apr 29 12:16:53.695: INFO: 	Container default-http-backend ready: true, restart count 0
+Apr 29 12:16:53.696: INFO: calico-node-k64gm from kube-system started at 2019-04-29 11:13:35 +0000 UTC (1 container statuses recorded)
+Apr 29 12:16:53.696: INFO: 	Container calico-node ready: true, restart count 0
+Apr 29 12:16:53.696: INFO: sonobuoy-systemd-logs-daemon-set-8191105fdead4396-gcldz from heptio-sonobuoy started at 2019-04-29 11:30:45 +0000 UTC (2 container statuses recorded)
+Apr 29 12:16:53.697: INFO: 	Container sonobuoy-worker ready: true, restart count 0
+Apr 29 12:16:53.697: INFO: 	Container systemd-logs ready: true, restart count 0
+Apr 29 12:16:53.697: INFO: coredns-58f7d854b4-c4lkh from kube-system started at 2019-04-29 11:16:23 +0000 UTC (1 container statuses recorded)
+Apr 29 12:16:53.697: INFO: 	Container coredns ready: true, restart count 0
+Apr 29 12:16:53.698: INFO: node-exporter-pcjs5 from kube-system started at 2019-04-29 11:21:43 +0000 UTC (1 container statuses recorded)
+Apr 29 12:16:53.698: INFO: 	Container node-exporter ready: true, restart count 0
+Apr 29 12:16:53.698: INFO: nginx-ingress-controller-6cffb97fb5-mfcvr from kube-system started at 2019-04-29 11:23:04 +0000 UTC (1 container statuses recorded)
+Apr 29 12:16:53.699: INFO: 	Container nginx-ingress-controller ready: true, restart count 0
+Apr 29 12:16:53.699: INFO: cert-exporter-pn7vb from kube-system started at 2019-04-29 11:16:35 +0000 UTC (1 container statuses recorded)
+Apr 29 12:16:53.699: INFO: 	Container cert-exporter ready: true, restart count 0
+Apr 29 12:16:53.700: INFO: tiller-deploy-54494c4fb6-k7mrg from giantswarm started at 2019-04-29 11:15:23 +0000 UTC (1 container statuses recorded)
+Apr 29 12:16:53.700: INFO: 	Container tiller ready: true, restart count 0
+Apr 29 12:16:53.700: INFO: 
+Logging pods the kubelet thinks is on node worker-p95fb-54c845dd45-crdqs before test
+Apr 29 12:16:53.726: INFO: calico-node-hwth9 from kube-system started at 2019-04-29 11:13:35 +0000 UTC (1 container statuses recorded)
+Apr 29 12:16:53.726: INFO: 	Container calico-node ready: true, restart count 0
+Apr 29 12:16:53.726: INFO: metrics-server-b94b95fb4-ncg25 from kube-system started at 2019-04-29 11:16:39 +0000 UTC (1 container statuses recorded)
+Apr 29 12:16:53.726: INFO: 	Container metrics-server ready: true, restart count 0
+Apr 29 12:16:53.726: INFO: sonobuoy-systemd-logs-daemon-set-8191105fdead4396-w5l9x from heptio-sonobuoy started at 2019-04-29 11:30:45 +0000 UTC (2 container statuses recorded)
+Apr 29 12:16:53.726: INFO: 	Container sonobuoy-worker ready: true, restart count 0
+Apr 29 12:16:53.726: INFO: 	Container systemd-logs ready: true, restart count 0
+Apr 29 12:16:53.726: INFO: cert-exporter-jkxfh from kube-system started at 2019-04-29 11:16:35 +0000 UTC (1 container statuses recorded)
+Apr 29 12:16:53.727: INFO: 	Container cert-exporter ready: true, restart count 0
+Apr 29 12:16:53.727: INFO: kube-proxy-tppgw from kube-system started at 2019-04-29 11:13:55 +0000 UTC (1 container statuses recorded)
+Apr 29 12:16:53.727: INFO: 	Container kube-proxy ready: true, restart count 0
+Apr 29 12:16:53.727: INFO: chart-operator-58979ccd8d-wvbzg from giantswarm started at 2019-04-29 11:16:10 +0000 UTC (1 container statuses recorded)
+Apr 29 12:16:53.727: INFO: 	Container chart-operator ready: true, restart count 0
+Apr 29 12:16:53.727: INFO: node-exporter-98xpc from kube-system started at 2019-04-29 11:21:42 +0000 UTC (1 container statuses recorded)
+Apr 29 12:16:53.727: INFO: 	Container node-exporter ready: true, restart count 0
+Apr 29 12:16:53.727: INFO: sonobuoy-e2e-job-9cf815dd5d0c4d66 from heptio-sonobuoy started at 2019-04-29 11:30:45 +0000 UTC (2 container statuses recorded)
+Apr 29 12:16:53.727: INFO: 	Container e2e ready: true, restart count 0
+Apr 29 12:16:53.728: INFO: 	Container sonobuoy-worker ready: true, restart count 0
+Apr 29 12:16:53.728: INFO: nginx-ingress-controller-6cffb97fb5-vqksl from kube-system started at 2019-04-29 11:21:47 +0000 UTC (1 container statuses recorded)
+Apr 29 12:16:53.728: INFO: 	Container nginx-ingress-controller ready: true, restart count 0
+Apr 29 12:16:53.728: INFO: net-exporter-xcjtp from kube-system started at 2019-04-29 11:21:44 +0000 UTC (1 container statuses recorded)
+Apr 29 12:16:53.728: INFO: 	Container net-exporter ready: true, restart count 0
+Apr 29 12:16:53.728: INFO: default-http-backend-77b6545877-jwz64 from kube-system started at 2019-04-29 11:22:25 +0000 UTC (1 container statuses recorded)
+Apr 29 12:16:53.728: INFO: 	Container default-http-backend ready: true, restart count 0
+Apr 29 12:16:53.728: INFO: sonobuoy from heptio-sonobuoy started at 2019-04-29 11:30:35 +0000 UTC (1 container statuses recorded)
+Apr 29 12:16:53.728: INFO: 	Container kube-sonobuoy ready: true, restart count 0
+[It] validates that NodeSelector is respected if not matching  [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Trying to schedule Pod with nonempty NodeSelector.
+STEP: Considering event: 
+Type = [Warning], Name = [restricted-pod.1599f113c63cd063], Reason = [FailedScheduling], Message = [0/4 nodes are available: 4 node(s) didn't match node selector.]
+[AfterEach] [sig-scheduling] SchedulerPredicates [Serial]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 12:16:54.791: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "sched-pred-4105" for this suite.
+Apr 29 12:17:00.813: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 12:17:01.167: INFO: namespace sched-pred-4105 deletion completed in 6.369356346s
+[AfterEach] [sig-scheduling] SchedulerPredicates [Serial]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:70
+
+• [SLOW TEST:7.719 seconds]
+[sig-scheduling] SchedulerPredicates [Serial]
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/framework.go:22
+  validates that NodeSelector is respected if not matching  [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSS
+------------------------------
+[sig-storage] Projected secret 
+  should be consumable in multiple volumes in a pod [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-storage] Projected secret
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 12:17:01.172: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename projected
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-604
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should be consumable in multiple volumes in a pod [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating secret with name projected-secret-test-b140bc2d-6a78-11e9-b6ef-0e71f8c7e888
+STEP: Creating a pod to test consume secrets
+Apr 29 12:17:01.428: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-b1426a4b-6a78-11e9-b6ef-0e71f8c7e888" in namespace "projected-604" to be "success or failure"
+Apr 29 12:17:01.442: INFO: Pod "pod-projected-secrets-b1426a4b-6a78-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 13.789903ms
+Apr 29 12:17:03.452: INFO: Pod "pod-projected-secrets-b1426a4b-6a78-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 2.02336887s
+Apr 29 12:17:05.459: INFO: Pod "pod-projected-secrets-b1426a4b-6a78-11e9-b6ef-0e71f8c7e888": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.031327911s
+STEP: Saw pod success
+Apr 29 12:17:05.460: INFO: Pod "pod-projected-secrets-b1426a4b-6a78-11e9-b6ef-0e71f8c7e888" satisfied condition "success or failure"
+Apr 29 12:17:05.465: INFO: Trying to get logs from node worker-4av2h-6665d7b5ff-9nc42 pod pod-projected-secrets-b1426a4b-6a78-11e9-b6ef-0e71f8c7e888 container secret-volume-test: 
+STEP: delete the pod
+Apr 29 12:17:05.529: INFO: Waiting for pod pod-projected-secrets-b1426a4b-6a78-11e9-b6ef-0e71f8c7e888 to disappear
+Apr 29 12:17:05.536: INFO: Pod pod-projected-secrets-b1426a4b-6a78-11e9-b6ef-0e71f8c7e888 no longer exists
+[AfterEach] [sig-storage] Projected secret
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 12:17:05.536: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "projected-604" for this suite.
+Apr 29 12:17:13.569: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 12:17:13.804: INFO: namespace projected-604 deletion completed in 8.259082137s
+
+• [SLOW TEST:12.632 seconds]
+[sig-storage] Projected secret
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_secret.go:33
+  should be consumable in multiple volumes in a pod [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSS
+------------------------------
+[sig-node] Downward API 
+  should provide pod UID as env vars [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-node] Downward API
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 12:17:13.806: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename downward-api
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in downward-api-4195
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should provide pod UID as env vars [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating a pod to test downward api env vars
+Apr 29 12:17:13.992: INFO: Waiting up to 5m0s for pod "downward-api-b8c17426-6a78-11e9-b6ef-0e71f8c7e888" in namespace "downward-api-4195" to be "success or failure"
+Apr 29 12:17:13.998: INFO: Pod "downward-api-b8c17426-6a78-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 5.869588ms
+Apr 29 12:17:16.005: INFO: Pod "downward-api-b8c17426-6a78-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 2.012922891s
+Apr 29 12:17:18.016: INFO: Pod "downward-api-b8c17426-6a78-11e9-b6ef-0e71f8c7e888": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.024158755s
+STEP: Saw pod success
+Apr 29 12:17:18.017: INFO: Pod "downward-api-b8c17426-6a78-11e9-b6ef-0e71f8c7e888" satisfied condition "success or failure"
+Apr 29 12:17:18.022: INFO: Trying to get logs from node worker-4av2h-6665d7b5ff-9nc42 pod downward-api-b8c17426-6a78-11e9-b6ef-0e71f8c7e888 container dapi-container: 
+STEP: delete the pod
+Apr 29 12:17:18.068: INFO: Waiting for pod downward-api-b8c17426-6a78-11e9-b6ef-0e71f8c7e888 to disappear
+Apr 29 12:17:18.075: INFO: Pod downward-api-b8c17426-6a78-11e9-b6ef-0e71f8c7e888 no longer exists
+[AfterEach] [sig-node] Downward API
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 12:17:18.076: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "downward-api-4195" for this suite.
+Apr 29 12:17:24.102: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 12:17:24.372: INFO: namespace downward-api-4195 deletion completed in 6.289023117s
+
+• [SLOW TEST:10.566 seconds]
+[sig-node] Downward API
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downward_api.go:38
+  should provide pod UID as env vars [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SS
+------------------------------
+[sig-storage] Subpath Atomic writer volumes 
+  should support subpaths with projected pod [LinuxOnly] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-storage] Subpath
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 12:17:24.373: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename subpath
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in subpath-8548
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] Atomic writer volumes
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:38
+STEP: Setting up data
+[It] should support subpaths with projected pod [LinuxOnly] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating pod pod-subpath-test-projected-5dps
+STEP: Creating a pod to test atomic-volume-subpath
+Apr 29 12:17:24.613: INFO: Waiting up to 5m0s for pod "pod-subpath-test-projected-5dps" in namespace "subpath-8548" to be "success or failure"
+Apr 29 12:17:24.620: INFO: Pod "pod-subpath-test-projected-5dps": Phase="Pending", Reason="", readiness=false. Elapsed: 6.56978ms
+Apr 29 12:17:26.629: INFO: Pod "pod-subpath-test-projected-5dps": Phase="Pending", Reason="", readiness=false. Elapsed: 2.015850005s
+Apr 29 12:17:28.638: INFO: Pod "pod-subpath-test-projected-5dps": Phase="Running", Reason="", readiness=true. Elapsed: 4.024003729s
+Apr 29 12:17:30.648: INFO: Pod "pod-subpath-test-projected-5dps": Phase="Running", Reason="", readiness=true. Elapsed: 6.034234941s
+Apr 29 12:17:32.658: INFO: Pod "pod-subpath-test-projected-5dps": Phase="Running", Reason="", readiness=true. Elapsed: 8.044807936s
+Apr 29 12:17:34.679: INFO: Pod "pod-subpath-test-projected-5dps": Phase="Running", Reason="", readiness=true. Elapsed: 10.065015832s
+Apr 29 12:17:36.690: INFO: Pod "pod-subpath-test-projected-5dps": Phase="Running", Reason="", readiness=true. Elapsed: 12.076650203s
+Apr 29 12:17:38.698: INFO: Pod "pod-subpath-test-projected-5dps": Phase="Running", Reason="", readiness=true. Elapsed: 14.084353602s
+Apr 29 12:17:40.709: INFO: Pod "pod-subpath-test-projected-5dps": Phase="Running", Reason="", readiness=true. Elapsed: 16.094999682s
+Apr 29 12:17:42.715: INFO: Pod "pod-subpath-test-projected-5dps": Phase="Running", Reason="", readiness=true. Elapsed: 18.101369653s
+Apr 29 12:17:44.725: INFO: Pod "pod-subpath-test-projected-5dps": Phase="Running", Reason="", readiness=true. Elapsed: 20.111189763s
+Apr 29 12:17:46.732: INFO: Pod "pod-subpath-test-projected-5dps": Phase="Running", Reason="", readiness=true. Elapsed: 22.118196238s
+Apr 29 12:17:48.741: INFO: Pod "pod-subpath-test-projected-5dps": Phase="Succeeded", Reason="", readiness=false. Elapsed: 24.1274126s
+STEP: Saw pod success
+Apr 29 12:17:48.741: INFO: Pod "pod-subpath-test-projected-5dps" satisfied condition "success or failure"
+Apr 29 12:17:48.745: INFO: Trying to get logs from node worker-4av2h-6665d7b5ff-9nc42 pod pod-subpath-test-projected-5dps container test-container-subpath-projected-5dps: 
+STEP: delete the pod
+Apr 29 12:17:48.782: INFO: Waiting for pod pod-subpath-test-projected-5dps to disappear
+Apr 29 12:17:48.790: INFO: Pod pod-subpath-test-projected-5dps no longer exists
+STEP: Deleting pod pod-subpath-test-projected-5dps
+Apr 29 12:17:48.790: INFO: Deleting pod "pod-subpath-test-projected-5dps" in namespace "subpath-8548"
+[AfterEach] [sig-storage] Subpath
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 12:17:48.794: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "subpath-8548" for this suite.
+Apr 29 12:17:54.827: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 12:17:55.052: INFO: namespace subpath-8548 deletion completed in 6.250165117s
+
+• [SLOW TEST:30.679 seconds]
+[sig-storage] Subpath
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:22
+  Atomic writer volumes
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:34
+    should support subpaths with projected pod [LinuxOnly] [Conformance]
+    /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+[sig-cli] Kubectl client [k8s.io] Kubectl replace 
+  should update a single-container pod's image  [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-cli] Kubectl client
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 12:17:55.058: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename kubectl
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-5473
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [sig-cli] Kubectl client
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:213
+[BeforeEach] [k8s.io] Kubectl replace
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1619
+[It] should update a single-container pod's image  [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: running the image docker.io/library/nginx:1.14-alpine
+Apr 29 12:17:55.278: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 run e2e-test-nginx-pod --generator=run-pod/v1 --image=docker.io/library/nginx:1.14-alpine --labels=run=e2e-test-nginx-pod --namespace=kubectl-5473'
+Apr 29 12:17:55.428: INFO: stderr: ""
+Apr 29 12:17:55.428: INFO: stdout: "pod/e2e-test-nginx-pod created\n"
+STEP: verifying the pod e2e-test-nginx-pod is running
+STEP: verifying the pod e2e-test-nginx-pod was created
+Apr 29 12:18:00.479: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 get pod e2e-test-nginx-pod --namespace=kubectl-5473 -o json'
+Apr 29 12:18:00.603: INFO: stderr: ""
+Apr 29 12:18:00.603: INFO: stdout: "{\n    \"apiVersion\": \"v1\",\n    \"kind\": \"Pod\",\n    \"metadata\": {\n        \"annotations\": {\n            \"kubernetes.io/psp\": \"cert-exporter-psp\"\n        },\n        \"creationTimestamp\": \"2019-04-29T12:17:55Z\",\n        \"labels\": {\n            \"run\": \"e2e-test-nginx-pod\"\n        },\n        \"name\": \"e2e-test-nginx-pod\",\n        \"namespace\": \"kubectl-5473\",\n        \"resourceVersion\": \"14841\",\n        \"selfLink\": \"/api/v1/namespaces/kubectl-5473/pods/e2e-test-nginx-pod\",\n        \"uid\": \"d13ed497-6a78-11e9-92a7-deadbef3c36d\"\n    },\n    \"spec\": {\n        \"containers\": [\n            {\n                \"image\": \"docker.io/library/nginx:1.14-alpine\",\n                \"imagePullPolicy\": \"IfNotPresent\",\n                \"name\": \"e2e-test-nginx-pod\",\n                \"resources\": {},\n                \"terminationMessagePath\": \"/dev/termination-log\",\n                \"terminationMessagePolicy\": \"File\",\n                \"volumeMounts\": [\n                    {\n                        \"mountPath\": \"/var/run/secrets/kubernetes.io/serviceaccount\",\n                        \"name\": \"default-token-nk5fz\",\n                        \"readOnly\": true\n                    }\n                ]\n            }\n        ],\n        \"dnsPolicy\": \"ClusterFirst\",\n        \"enableServiceLinks\": true,\n        \"nodeName\": \"worker-4av2h-6665d7b5ff-9nc42\",\n        \"priority\": 0,\n        \"restartPolicy\": \"Always\",\n        \"schedulerName\": \"default-scheduler\",\n        \"securityContext\": {},\n        \"serviceAccount\": \"default\",\n        \"serviceAccountName\": \"default\",\n        \"terminationGracePeriodSeconds\": 30,\n        \"tolerations\": [\n            {\n                \"effect\": \"NoExecute\",\n                \"key\": \"node.kubernetes.io/not-ready\",\n                \"operator\": \"Exists\",\n                \"tolerationSeconds\": 300\n            },\n            {\n                \"effect\": \"NoExecute\",\n                \"key\": \"node.kubernetes.io/unreachable\",\n                \"operator\": \"Exists\",\n                \"tolerationSeconds\": 300\n            }\n        ],\n        \"volumes\": [\n            {\n                \"name\": \"default-token-nk5fz\",\n                \"secret\": {\n                    \"defaultMode\": 420,\n                    \"secretName\": \"default-token-nk5fz\"\n                }\n            }\n        ]\n    },\n    \"status\": {\n        \"conditions\": [\n            {\n                \"lastProbeTime\": null,\n                \"lastTransitionTime\": \"2019-04-29T12:17:55Z\",\n                \"status\": \"True\",\n                \"type\": \"Initialized\"\n            },\n            {\n                \"lastProbeTime\": null,\n                \"lastTransitionTime\": \"2019-04-29T12:17:58Z\",\n                \"status\": \"True\",\n                \"type\": \"Ready\"\n            },\n            {\n                \"lastProbeTime\": null,\n                \"lastTransitionTime\": \"2019-04-29T12:17:58Z\",\n                \"status\": \"True\",\n                \"type\": \"ContainersReady\"\n            },\n            {\n                \"lastProbeTime\": null,\n                \"lastTransitionTime\": \"2019-04-29T12:17:55Z\",\n                \"status\": \"True\",\n                \"type\": \"PodScheduled\"\n            }\n        ],\n        \"containerStatuses\": [\n            {\n                \"containerID\": \"docker://80b18f098fc16aaf04f3345c9ea6fffb02351901b47621da7d50ae81cc75ad7f\",\n                \"image\": \"nginx:1.14-alpine\",\n                \"imageID\": \"docker-pullable://nginx@sha256:485b610fefec7ff6c463ced9623314a04ed67e3945b9c08d7e53a47f6d108dc7\",\n                \"lastState\": {},\n                \"name\": \"e2e-test-nginx-pod\",\n                \"ready\": true,\n                \"restartCount\": 0,\n                \"state\": {\n                    \"running\": {\n                        \"startedAt\": \"2019-04-29T12:17:57Z\"\n                    }\n                }\n            }\n        ],\n        \"hostIP\": \"172.23.0.238\",\n        \"phase\": \"Running\",\n        \"podIP\": \"192.168.205.127\",\n        \"qosClass\": \"BestEffort\",\n        \"startTime\": \"2019-04-29T12:17:55Z\"\n    }\n}\n"
+STEP: replace the image in the pod
+Apr 29 12:18:00.604: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 replace -f - --namespace=kubectl-5473'
+Apr 29 12:18:00.910: INFO: stderr: ""
+Apr 29 12:18:00.911: INFO: stdout: "pod/e2e-test-nginx-pod replaced\n"
+STEP: verifying the pod e2e-test-nginx-pod has the right image docker.io/library/busybox:1.29
+[AfterEach] [k8s.io] Kubectl replace
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1624
+Apr 29 12:18:00.920: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 delete pods e2e-test-nginx-pod --namespace=kubectl-5473'
+Apr 29 12:18:04.458: INFO: stderr: ""
+Apr 29 12:18:04.458: INFO: stdout: "pod \"e2e-test-nginx-pod\" deleted\n"
+[AfterEach] [sig-cli] Kubectl client
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 12:18:04.458: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "kubectl-5473" for this suite.
+Apr 29 12:18:10.497: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 12:18:10.694: INFO: namespace kubectl-5473 deletion completed in 6.224871789s
+
+• [SLOW TEST:15.637 seconds]
+[sig-cli] Kubectl client
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23
+  [k8s.io] Kubectl replace
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
+    should update a single-container pod's image  [Conformance]
+    /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSSSSSSSSS
+------------------------------
+[sig-storage] ConfigMap 
+  should be consumable from pods in volume as non-root [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-storage] ConfigMap
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 12:18:10.694: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename configmap
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in configmap-5438
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should be consumable from pods in volume as non-root [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating configMap with name configmap-test-volume-daacc19f-6a78-11e9-b6ef-0e71f8c7e888
+STEP: Creating a pod to test consume configMaps
+Apr 29 12:18:10.928: INFO: Waiting up to 5m0s for pod "pod-configmaps-daaeca6f-6a78-11e9-b6ef-0e71f8c7e888" in namespace "configmap-5438" to be "success or failure"
+Apr 29 12:18:10.940: INFO: Pod "pod-configmaps-daaeca6f-6a78-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 11.981349ms
+Apr 29 12:18:12.948: INFO: Pod "pod-configmaps-daaeca6f-6a78-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 2.019624023s
+Apr 29 12:18:14.959: INFO: Pod "pod-configmaps-daaeca6f-6a78-11e9-b6ef-0e71f8c7e888": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.03096753s
+STEP: Saw pod success
+Apr 29 12:18:14.959: INFO: Pod "pod-configmaps-daaeca6f-6a78-11e9-b6ef-0e71f8c7e888" satisfied condition "success or failure"
+Apr 29 12:18:14.963: INFO: Trying to get logs from node worker-4av2h-6665d7b5ff-9nc42 pod pod-configmaps-daaeca6f-6a78-11e9-b6ef-0e71f8c7e888 container configmap-volume-test: 
+STEP: delete the pod
+Apr 29 12:18:14.998: INFO: Waiting for pod pod-configmaps-daaeca6f-6a78-11e9-b6ef-0e71f8c7e888 to disappear
+Apr 29 12:18:15.008: INFO: Pod pod-configmaps-daaeca6f-6a78-11e9-b6ef-0e71f8c7e888 no longer exists
+[AfterEach] [sig-storage] ConfigMap
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 12:18:15.008: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "configmap-5438" for this suite.
+Apr 29 12:18:21.044: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 12:18:21.288: INFO: namespace configmap-5438 deletion completed in 6.264566176s
+
+• [SLOW TEST:10.594 seconds]
+[sig-storage] ConfigMap
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:32
+  should be consumable from pods in volume as non-root [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSS
+------------------------------
+[sig-storage] Projected downwardAPI 
+  should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-storage] Projected downwardAPI
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 12:18:21.293: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename projected
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-7184
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [sig-storage] Projected downwardAPI
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:39
+[It] should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating a pod to test downward API volume plugin
+Apr 29 12:18:21.511: INFO: Waiting up to 5m0s for pod "downwardapi-volume-e0fe589b-6a78-11e9-b6ef-0e71f8c7e888" in namespace "projected-7184" to be "success or failure"
+Apr 29 12:18:21.517: INFO: Pod "downwardapi-volume-e0fe589b-6a78-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 6.163422ms
+Apr 29 12:18:23.526: INFO: Pod "downwardapi-volume-e0fe589b-6a78-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 2.01528137s
+Apr 29 12:18:25.534: INFO: Pod "downwardapi-volume-e0fe589b-6a78-11e9-b6ef-0e71f8c7e888": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.022504592s
+STEP: Saw pod success
+Apr 29 12:18:25.534: INFO: Pod "downwardapi-volume-e0fe589b-6a78-11e9-b6ef-0e71f8c7e888" satisfied condition "success or failure"
+Apr 29 12:18:25.539: INFO: Trying to get logs from node worker-4av2h-6665d7b5ff-9nc42 pod downwardapi-volume-e0fe589b-6a78-11e9-b6ef-0e71f8c7e888 container client-container: 
+STEP: delete the pod
+Apr 29 12:18:25.574: INFO: Waiting for pod downwardapi-volume-e0fe589b-6a78-11e9-b6ef-0e71f8c7e888 to disappear
+Apr 29 12:18:25.578: INFO: Pod downwardapi-volume-e0fe589b-6a78-11e9-b6ef-0e71f8c7e888 no longer exists
+[AfterEach] [sig-storage] Projected downwardAPI
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 12:18:25.578: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "projected-7184" for this suite.
+Apr 29 12:18:31.602: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 12:18:31.800: INFO: namespace projected-7184 deletion completed in 6.216322939s
+
+• [SLOW TEST:10.508 seconds]
+[sig-storage] Projected downwardAPI
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:33
+  should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+------------------------------
+[k8s.io] Container Lifecycle Hook when create a pod with lifecycle hook 
+  should execute prestop http hook properly [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [k8s.io] Container Lifecycle Hook
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 12:18:31.801: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename container-lifecycle-hook
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in container-lifecycle-hook-6337
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] when create a pod with lifecycle hook
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:61
+STEP: create the container to handle the HTTPGet hook request.
+[It] should execute prestop http hook properly [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: create the pod with lifecycle hook
+STEP: delete the pod with lifecycle hook
+Apr 29 12:18:38.110: INFO: Waiting for pod pod-with-prestop-http-hook to disappear
+Apr 29 12:18:38.120: INFO: Pod pod-with-prestop-http-hook still exists
+Apr 29 12:18:40.121: INFO: Waiting for pod pod-with-prestop-http-hook to disappear
+Apr 29 12:18:40.135: INFO: Pod pod-with-prestop-http-hook still exists
+Apr 29 12:18:42.121: INFO: Waiting for pod pod-with-prestop-http-hook to disappear
+Apr 29 12:18:42.128: INFO: Pod pod-with-prestop-http-hook still exists
+Apr 29 12:18:44.121: INFO: Waiting for pod pod-with-prestop-http-hook to disappear
+Apr 29 12:18:44.127: INFO: Pod pod-with-prestop-http-hook still exists
+Apr 29 12:18:46.121: INFO: Waiting for pod pod-with-prestop-http-hook to disappear
+Apr 29 12:18:46.146: INFO: Pod pod-with-prestop-http-hook still exists
+Apr 29 12:18:48.121: INFO: Waiting for pod pod-with-prestop-http-hook to disappear
+Apr 29 12:18:48.127: INFO: Pod pod-with-prestop-http-hook still exists
+Apr 29 12:18:50.121: INFO: Waiting for pod pod-with-prestop-http-hook to disappear
+Apr 29 12:18:50.131: INFO: Pod pod-with-prestop-http-hook still exists
+Apr 29 12:18:52.121: INFO: Waiting for pod pod-with-prestop-http-hook to disappear
+Apr 29 12:18:52.130: INFO: Pod pod-with-prestop-http-hook still exists
+Apr 29 12:18:54.121: INFO: Waiting for pod pod-with-prestop-http-hook to disappear
+Apr 29 12:18:54.131: INFO: Pod pod-with-prestop-http-hook still exists
+Apr 29 12:18:56.123: INFO: Waiting for pod pod-with-prestop-http-hook to disappear
+Apr 29 12:18:56.131: INFO: Pod pod-with-prestop-http-hook no longer exists
+STEP: check prestop hook
+[AfterEach] [k8s.io] Container Lifecycle Hook
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 12:18:56.149: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "container-lifecycle-hook-6337" for this suite.
+Apr 29 12:19:20.184: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 12:19:20.456: INFO: namespace container-lifecycle-hook-6337 deletion completed in 24.297371821s
+
+• [SLOW TEST:48.656 seconds]
+[k8s.io] Container Lifecycle Hook
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
+  when create a pod with lifecycle hook
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:40
+    should execute prestop http hook properly [NodeConformance] [Conformance]
+    /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+------------------------------
+[sig-cli] Kubectl client [k8s.io] Update Demo 
+  should scale a replication controller  [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-cli] Kubectl client
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 12:19:20.466: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename kubectl
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-5588
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [sig-cli] Kubectl client
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:213
+[BeforeEach] [k8s.io] Update Demo
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:265
+[It] should scale a replication controller  [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: creating a replication controller
+Apr 29 12:19:20.641: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 create -f - --namespace=kubectl-5588'
+Apr 29 12:19:22.984: INFO: stderr: ""
+Apr 29 12:19:22.984: INFO: stdout: "replicationcontroller/update-demo-nautilus created\n"
+STEP: waiting for all containers in name=update-demo pods to come up.
+Apr 29 12:19:22.985: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-5588'
+Apr 29 12:19:23.144: INFO: stderr: ""
+Apr 29 12:19:23.144: INFO: stdout: "update-demo-nautilus-cmmt4 update-demo-nautilus-ddntb "
+Apr 29 12:19:23.145: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 get pods update-demo-nautilus-cmmt4 -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-5588'
+Apr 29 12:19:23.271: INFO: stderr: ""
+Apr 29 12:19:23.271: INFO: stdout: ""
+Apr 29 12:19:23.271: INFO: update-demo-nautilus-cmmt4 is created but not running
+Apr 29 12:19:28.272: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-5588'
+Apr 29 12:19:28.391: INFO: stderr: ""
+Apr 29 12:19:28.391: INFO: stdout: "update-demo-nautilus-cmmt4 update-demo-nautilus-ddntb "
+Apr 29 12:19:28.391: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 get pods update-demo-nautilus-cmmt4 -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-5588'
+Apr 29 12:19:28.505: INFO: stderr: ""
+Apr 29 12:19:28.505: INFO: stdout: "true"
+Apr 29 12:19:28.505: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 get pods update-demo-nautilus-cmmt4 -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-5588'
+Apr 29 12:19:28.615: INFO: stderr: ""
+Apr 29 12:19:28.615: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0"
+Apr 29 12:19:28.615: INFO: validating pod update-demo-nautilus-cmmt4
+Apr 29 12:19:28.624: INFO: got data: {
+  "image": "nautilus.jpg"
+}
+
+Apr 29 12:19:28.624: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg .
+Apr 29 12:19:28.624: INFO: update-demo-nautilus-cmmt4 is verified up and running
+Apr 29 12:19:28.624: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 get pods update-demo-nautilus-ddntb -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-5588'
+Apr 29 12:19:28.730: INFO: stderr: ""
+Apr 29 12:19:28.730: INFO: stdout: "true"
+Apr 29 12:19:28.731: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 get pods update-demo-nautilus-ddntb -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-5588'
+Apr 29 12:19:28.839: INFO: stderr: ""
+Apr 29 12:19:28.840: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0"
+Apr 29 12:19:28.840: INFO: validating pod update-demo-nautilus-ddntb
+Apr 29 12:19:28.848: INFO: got data: {
+  "image": "nautilus.jpg"
+}
+
+Apr 29 12:19:28.848: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg .
+Apr 29 12:19:28.848: INFO: update-demo-nautilus-ddntb is verified up and running
+STEP: scaling down the replication controller
+Apr 29 12:19:28.852: INFO: scanned /root for discovery docs: 
+Apr 29 12:19:28.852: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 scale rc update-demo-nautilus --replicas=1 --timeout=5m --namespace=kubectl-5588'
+Apr 29 12:19:30.029: INFO: stderr: ""
+Apr 29 12:19:30.029: INFO: stdout: "replicationcontroller/update-demo-nautilus scaled\n"
+STEP: waiting for all containers in name=update-demo pods to come up.
+Apr 29 12:19:30.029: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-5588'
+Apr 29 12:19:30.164: INFO: stderr: ""
+Apr 29 12:19:30.164: INFO: stdout: "update-demo-nautilus-cmmt4 update-demo-nautilus-ddntb "
+STEP: Replicas for name=update-demo: expected=1 actual=2
+Apr 29 12:19:35.164: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-5588'
+Apr 29 12:19:35.267: INFO: stderr: ""
+Apr 29 12:19:35.267: INFO: stdout: "update-demo-nautilus-ddntb "
+Apr 29 12:19:35.267: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 get pods update-demo-nautilus-ddntb -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-5588'
+Apr 29 12:19:35.368: INFO: stderr: ""
+Apr 29 12:19:35.368: INFO: stdout: "true"
+Apr 29 12:19:35.368: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 get pods update-demo-nautilus-ddntb -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-5588'
+Apr 29 12:19:35.479: INFO: stderr: ""
+Apr 29 12:19:35.479: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0"
+Apr 29 12:19:35.479: INFO: validating pod update-demo-nautilus-ddntb
+Apr 29 12:19:35.490: INFO: got data: {
+  "image": "nautilus.jpg"
+}
+
+Apr 29 12:19:35.490: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg .
+Apr 29 12:19:35.490: INFO: update-demo-nautilus-ddntb is verified up and running
+STEP: scaling up the replication controller
+Apr 29 12:19:35.493: INFO: scanned /root for discovery docs: 
+Apr 29 12:19:35.494: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 scale rc update-demo-nautilus --replicas=2 --timeout=5m --namespace=kubectl-5588'
+Apr 29 12:19:36.687: INFO: stderr: ""
+Apr 29 12:19:36.687: INFO: stdout: "replicationcontroller/update-demo-nautilus scaled\n"
+STEP: waiting for all containers in name=update-demo pods to come up.
+Apr 29 12:19:36.687: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-5588'
+Apr 29 12:19:36.799: INFO: stderr: ""
+Apr 29 12:19:36.799: INFO: stdout: "update-demo-nautilus-ddntb update-demo-nautilus-fnbbq "
+Apr 29 12:19:36.799: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 get pods update-demo-nautilus-ddntb -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-5588'
+Apr 29 12:19:36.912: INFO: stderr: ""
+Apr 29 12:19:36.912: INFO: stdout: "true"
+Apr 29 12:19:36.912: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 get pods update-demo-nautilus-ddntb -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-5588'
+Apr 29 12:19:37.028: INFO: stderr: ""
+Apr 29 12:19:37.028: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0"
+Apr 29 12:19:37.028: INFO: validating pod update-demo-nautilus-ddntb
+Apr 29 12:19:37.038: INFO: got data: {
+  "image": "nautilus.jpg"
+}
+
+Apr 29 12:19:37.038: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg .
+Apr 29 12:19:37.039: INFO: update-demo-nautilus-ddntb is verified up and running
+Apr 29 12:19:37.039: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 get pods update-demo-nautilus-fnbbq -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-5588'
+Apr 29 12:19:37.174: INFO: stderr: ""
+Apr 29 12:19:37.174: INFO: stdout: ""
+Apr 29 12:19:37.174: INFO: update-demo-nautilus-fnbbq is created but not running
+Apr 29 12:19:42.174: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-5588'
+Apr 29 12:19:42.275: INFO: stderr: ""
+Apr 29 12:19:42.276: INFO: stdout: "update-demo-nautilus-ddntb update-demo-nautilus-fnbbq "
+Apr 29 12:19:42.276: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 get pods update-demo-nautilus-ddntb -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-5588'
+Apr 29 12:19:42.390: INFO: stderr: ""
+Apr 29 12:19:42.390: INFO: stdout: "true"
+Apr 29 12:19:42.390: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 get pods update-demo-nautilus-ddntb -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-5588'
+Apr 29 12:19:42.497: INFO: stderr: ""
+Apr 29 12:19:42.497: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0"
+Apr 29 12:19:42.497: INFO: validating pod update-demo-nautilus-ddntb
+Apr 29 12:19:42.508: INFO: got data: {
+  "image": "nautilus.jpg"
+}
+
+Apr 29 12:19:42.508: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg .
+Apr 29 12:19:42.508: INFO: update-demo-nautilus-ddntb is verified up and running
+Apr 29 12:19:42.508: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 get pods update-demo-nautilus-fnbbq -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-5588'
+Apr 29 12:19:42.613: INFO: stderr: ""
+Apr 29 12:19:42.613: INFO: stdout: "true"
+Apr 29 12:19:42.613: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 get pods update-demo-nautilus-fnbbq -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-5588'
+Apr 29 12:19:42.722: INFO: stderr: ""
+Apr 29 12:19:42.722: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0"
+Apr 29 12:19:42.722: INFO: validating pod update-demo-nautilus-fnbbq
+Apr 29 12:19:42.729: INFO: got data: {
+  "image": "nautilus.jpg"
+}
+
+Apr 29 12:19:42.729: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg .
+Apr 29 12:19:42.729: INFO: update-demo-nautilus-fnbbq is verified up and running
+STEP: using delete to clean up resources
+Apr 29 12:19:42.729: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 delete --grace-period=0 --force -f - --namespace=kubectl-5588'
+Apr 29 12:19:42.846: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n"
+Apr 29 12:19:42.846: INFO: stdout: "replicationcontroller \"update-demo-nautilus\" force deleted\n"
+Apr 29 12:19:42.846: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 get rc,svc -l name=update-demo --no-headers --namespace=kubectl-5588'
+Apr 29 12:19:43.059: INFO: stderr: "No resources found.\n"
+Apr 29 12:19:43.060: INFO: stdout: ""
+Apr 29 12:19:43.060: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 get pods -l name=update-demo --namespace=kubectl-5588 -o go-template={{ range .items }}{{ if not .metadata.deletionTimestamp }}{{ .metadata.name }}{{ "\n" }}{{ end }}{{ end }}'
+Apr 29 12:19:43.208: INFO: stderr: ""
+Apr 29 12:19:43.208: INFO: stdout: ""
+[AfterEach] [sig-cli] Kubectl client
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 12:19:43.208: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "kubectl-5588" for this suite.
+Apr 29 12:20:07.311: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 12:20:07.547: INFO: namespace kubectl-5588 deletion completed in 24.320424892s
+
+• [SLOW TEST:47.081 seconds]
+[sig-cli] Kubectl client
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23
+  [k8s.io] Update Demo
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
+    should scale a replication controller  [Conformance]
+    /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSS
+------------------------------
+[sig-storage] Projected configMap 
+  should be consumable from pods in volume [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-storage] Projected configMap
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 12:20:07.548: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename projected
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-7409
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should be consumable from pods in volume [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating configMap with name projected-configmap-test-volume-205436ce-6a79-11e9-b6ef-0e71f8c7e888
+STEP: Creating a pod to test consume configMaps
+Apr 29 12:20:07.770: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-2055b2a5-6a79-11e9-b6ef-0e71f8c7e888" in namespace "projected-7409" to be "success or failure"
+Apr 29 12:20:07.778: INFO: Pod "pod-projected-configmaps-2055b2a5-6a79-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 7.266334ms
+Apr 29 12:20:09.786: INFO: Pod "pod-projected-configmaps-2055b2a5-6a79-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 2.014820486s
+Apr 29 12:20:11.792: INFO: Pod "pod-projected-configmaps-2055b2a5-6a79-11e9-b6ef-0e71f8c7e888": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.021443948s
+STEP: Saw pod success
+Apr 29 12:20:11.793: INFO: Pod "pod-projected-configmaps-2055b2a5-6a79-11e9-b6ef-0e71f8c7e888" satisfied condition "success or failure"
+Apr 29 12:20:11.798: INFO: Trying to get logs from node worker-4av2h-6665d7b5ff-9nc42 pod pod-projected-configmaps-2055b2a5-6a79-11e9-b6ef-0e71f8c7e888 container projected-configmap-volume-test: 
+STEP: delete the pod
+Apr 29 12:20:11.826: INFO: Waiting for pod pod-projected-configmaps-2055b2a5-6a79-11e9-b6ef-0e71f8c7e888 to disappear
+Apr 29 12:20:11.839: INFO: Pod pod-projected-configmaps-2055b2a5-6a79-11e9-b6ef-0e71f8c7e888 no longer exists
+[AfterEach] [sig-storage] Projected configMap
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 12:20:11.839: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "projected-7409" for this suite.
+Apr 29 12:20:17.872: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 12:20:18.110: INFO: namespace projected-7409 deletion completed in 6.25479013s
+
+• [SLOW TEST:10.562 seconds]
+[sig-storage] Projected configMap
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_configmap.go:33
+  should be consumable from pods in volume [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSSSSSSSSSSSSSSSSSSS
+------------------------------
+[sig-node] Downward API 
+  should provide default limits.cpu/memory from node allocatable [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-node] Downward API
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 12:20:18.112: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename downward-api
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in downward-api-7633
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should provide default limits.cpu/memory from node allocatable [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating a pod to test downward api env vars
+Apr 29 12:20:18.302: INFO: Waiting up to 5m0s for pod "downward-api-269cc377-6a79-11e9-b6ef-0e71f8c7e888" in namespace "downward-api-7633" to be "success or failure"
+Apr 29 12:20:18.311: INFO: Pod "downward-api-269cc377-6a79-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 9.403951ms
+Apr 29 12:20:20.320: INFO: Pod "downward-api-269cc377-6a79-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 2.018150988s
+Apr 29 12:20:22.328: INFO: Pod "downward-api-269cc377-6a79-11e9-b6ef-0e71f8c7e888": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.025823727s
+STEP: Saw pod success
+Apr 29 12:20:22.328: INFO: Pod "downward-api-269cc377-6a79-11e9-b6ef-0e71f8c7e888" satisfied condition "success or failure"
+Apr 29 12:20:22.333: INFO: Trying to get logs from node worker-4av2h-6665d7b5ff-9nc42 pod downward-api-269cc377-6a79-11e9-b6ef-0e71f8c7e888 container dapi-container: 
+STEP: delete the pod
+Apr 29 12:20:22.389: INFO: Waiting for pod downward-api-269cc377-6a79-11e9-b6ef-0e71f8c7e888 to disappear
+Apr 29 12:20:22.397: INFO: Pod downward-api-269cc377-6a79-11e9-b6ef-0e71f8c7e888 no longer exists
+[AfterEach] [sig-node] Downward API
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 12:20:22.397: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "downward-api-7633" for this suite.
+Apr 29 12:20:28.429: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 12:20:28.623: INFO: namespace downward-api-7633 deletion completed in 6.216886935s
+
+• [SLOW TEST:10.512 seconds]
+[sig-node] Downward API
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downward_api.go:38
+  should provide default limits.cpu/memory from node allocatable [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSSSSSSS
+------------------------------
+[sig-cli] Kubectl client [k8s.io] Kubectl cluster-info 
+  should check if Kubernetes master services is included in cluster-info  [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-cli] Kubectl client
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 12:20:28.625: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename kubectl
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-2806
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [sig-cli] Kubectl client
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:213
+[It] should check if Kubernetes master services is included in cluster-info  [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: validating cluster-info
+Apr 29 12:20:28.823: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 cluster-info'
+Apr 29 12:20:28.936: INFO: stderr: ""
+Apr 29 12:20:28.936: INFO: stdout: "\x1b[0;32mKubernetes master\x1b[0m is running at \x1b[0;33mhttps://172.31.0.1:443\x1b[0m\n\x1b[0;32mCoreDNS\x1b[0m is running at \x1b[0;33mhttps://172.31.0.1:443/api/v1/namespaces/kube-system/services/coredns:dns/proxy\x1b[0m\n\nTo further debug and diagnose cluster problems, use 'kubectl cluster-info dump'.\n"
+[AfterEach] [sig-cli] Kubectl client
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 12:20:28.936: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "kubectl-2806" for this suite.
+Apr 29 12:20:34.960: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 12:20:35.250: INFO: namespace kubectl-2806 deletion completed in 6.307991963s
+
+• [SLOW TEST:6.626 seconds]
+[sig-cli] Kubectl client
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23
+  [k8s.io] Kubectl cluster-info
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
+    should check if Kubernetes master services is included in cluster-info  [Conformance]
+    /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+------------------------------
+[sig-network] Services 
+  should provide secure master service  [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-network] Services
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 12:20:35.260: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename services
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in services-8755
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [sig-network] Services
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:86
+[It] should provide secure master service  [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[AfterEach] [sig-network] Services
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 12:20:35.496: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "services-8755" for this suite.
+Apr 29 12:20:41.528: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 12:20:41.734: INFO: namespace services-8755 deletion completed in 6.226791082s
+[AfterEach] [sig-network] Services
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:91
+
+• [SLOW TEST:6.475 seconds]
+[sig-network] Services
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:22
+  should provide secure master service  [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSSSSSSSSSSSSSSSSS
+------------------------------
+[k8s.io] [sig-node] Pods Extended [k8s.io] Pods Set QOS Class 
+  should be submitted and removed  [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [k8s.io] [sig-node] Pods Extended
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 12:20:41.739: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename pods
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in pods-4046
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [k8s.io] Pods Set QOS Class
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/node/pods.go:177
+[It] should be submitted and removed  [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: creating the pod
+STEP: submitting the pod to kubernetes
+STEP: verifying QOS class is set on the pod
+[AfterEach] [k8s.io] [sig-node] Pods Extended
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 12:20:41.943: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "pods-4046" for this suite.
+Apr 29 12:21:05.977: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 12:21:06.247: INFO: namespace pods-4046 deletion completed in 24.297673069s
+
+• [SLOW TEST:24.509 seconds]
+[k8s.io] [sig-node] Pods Extended
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
+  [k8s.io] Pods Set QOS Class
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
+    should be submitted and removed  [Conformance]
+    /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSS
+------------------------------
+[sig-storage] Projected downwardAPI 
+  should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-storage] Projected downwardAPI
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 12:21:06.248: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename projected
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-7874
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [sig-storage] Projected downwardAPI
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:39
+[It] should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating a pod to test downward API volume plugin
+Apr 29 12:21:06.449: INFO: Waiting up to 5m0s for pod "downwardapi-volume-434e77be-6a79-11e9-b6ef-0e71f8c7e888" in namespace "projected-7874" to be "success or failure"
+Apr 29 12:21:06.453: INFO: Pod "downwardapi-volume-434e77be-6a79-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 4.10051ms
+Apr 29 12:21:08.461: INFO: Pod "downwardapi-volume-434e77be-6a79-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 2.011738947s
+Apr 29 12:21:10.470: INFO: Pod "downwardapi-volume-434e77be-6a79-11e9-b6ef-0e71f8c7e888": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.020746085s
+STEP: Saw pod success
+Apr 29 12:21:10.470: INFO: Pod "downwardapi-volume-434e77be-6a79-11e9-b6ef-0e71f8c7e888" satisfied condition "success or failure"
+Apr 29 12:21:10.474: INFO: Trying to get logs from node worker-4av2h-6665d7b5ff-9nc42 pod downwardapi-volume-434e77be-6a79-11e9-b6ef-0e71f8c7e888 container client-container: 
+STEP: delete the pod
+Apr 29 12:21:10.503: INFO: Waiting for pod downwardapi-volume-434e77be-6a79-11e9-b6ef-0e71f8c7e888 to disappear
+Apr 29 12:21:10.508: INFO: Pod downwardapi-volume-434e77be-6a79-11e9-b6ef-0e71f8c7e888 no longer exists
+[AfterEach] [sig-storage] Projected downwardAPI
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 12:21:10.509: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "projected-7874" for this suite.
+Apr 29 12:21:16.534: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 12:21:16.710: INFO: namespace projected-7874 deletion completed in 6.195877666s
+
+• [SLOW TEST:10.462 seconds]
+[sig-storage] Projected downwardAPI
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:33
+  should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSSSSSSSSSSSSS
+------------------------------
+[sig-storage] Subpath Atomic writer volumes 
+  should support subpaths with secret pod [LinuxOnly] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-storage] Subpath
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 12:21:16.711: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename subpath
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in subpath-9039
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] Atomic writer volumes
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:38
+STEP: Setting up data
+[It] should support subpaths with secret pod [LinuxOnly] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating pod pod-subpath-test-secret-c78d
+STEP: Creating a pod to test atomic-volume-subpath
+Apr 29 12:21:16.946: INFO: Waiting up to 5m0s for pod "pod-subpath-test-secret-c78d" in namespace "subpath-9039" to be "success or failure"
+Apr 29 12:21:16.953: INFO: Pod "pod-subpath-test-secret-c78d": Phase="Pending", Reason="", readiness=false. Elapsed: 6.931692ms
+Apr 29 12:21:18.961: INFO: Pod "pod-subpath-test-secret-c78d": Phase="Pending", Reason="", readiness=false. Elapsed: 2.014836764s
+Apr 29 12:21:20.967: INFO: Pod "pod-subpath-test-secret-c78d": Phase="Running", Reason="", readiness=true. Elapsed: 4.021334306s
+Apr 29 12:21:22.974: INFO: Pod "pod-subpath-test-secret-c78d": Phase="Running", Reason="", readiness=true. Elapsed: 6.028318509s
+Apr 29 12:21:24.979: INFO: Pod "pod-subpath-test-secret-c78d": Phase="Running", Reason="", readiness=true. Elapsed: 8.033593946s
+Apr 29 12:21:26.986: INFO: Pod "pod-subpath-test-secret-c78d": Phase="Running", Reason="", readiness=true. Elapsed: 10.040407404s
+Apr 29 12:21:28.992: INFO: Pod "pod-subpath-test-secret-c78d": Phase="Running", Reason="", readiness=true. Elapsed: 12.046347232s
+Apr 29 12:21:31.000: INFO: Pod "pod-subpath-test-secret-c78d": Phase="Running", Reason="", readiness=true. Elapsed: 14.054685613s
+Apr 29 12:21:33.007: INFO: Pod "pod-subpath-test-secret-c78d": Phase="Running", Reason="", readiness=true. Elapsed: 16.061744054s
+Apr 29 12:21:35.015: INFO: Pod "pod-subpath-test-secret-c78d": Phase="Running", Reason="", readiness=true. Elapsed: 18.069418658s
+Apr 29 12:21:37.021: INFO: Pod "pod-subpath-test-secret-c78d": Phase="Running", Reason="", readiness=true. Elapsed: 20.075620885s
+Apr 29 12:21:39.028: INFO: Pod "pod-subpath-test-secret-c78d": Phase="Running", Reason="", readiness=true. Elapsed: 22.082628758s
+Apr 29 12:21:41.035: INFO: Pod "pod-subpath-test-secret-c78d": Phase="Succeeded", Reason="", readiness=false. Elapsed: 24.089306998s
+STEP: Saw pod success
+Apr 29 12:21:41.035: INFO: Pod "pod-subpath-test-secret-c78d" satisfied condition "success or failure"
+Apr 29 12:21:41.039: INFO: Trying to get logs from node worker-4av2h-6665d7b5ff-9nc42 pod pod-subpath-test-secret-c78d container test-container-subpath-secret-c78d: 
+STEP: delete the pod
+Apr 29 12:21:41.075: INFO: Waiting for pod pod-subpath-test-secret-c78d to disappear
+Apr 29 12:21:41.080: INFO: Pod pod-subpath-test-secret-c78d no longer exists
+STEP: Deleting pod pod-subpath-test-secret-c78d
+Apr 29 12:21:41.080: INFO: Deleting pod "pod-subpath-test-secret-c78d" in namespace "subpath-9039"
+[AfterEach] [sig-storage] Subpath
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 12:21:41.085: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "subpath-9039" for this suite.
+Apr 29 12:21:47.129: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 12:21:47.346: INFO: namespace subpath-9039 deletion completed in 6.239655922s
+
+• [SLOW TEST:30.635 seconds]
+[sig-storage] Subpath
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:22
+  Atomic writer volumes
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:34
+    should support subpaths with secret pod [LinuxOnly] [Conformance]
+    /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+------------------------------
+[sig-storage] ConfigMap 
+  should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-storage] ConfigMap
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 12:21:47.352: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename configmap
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in configmap-9199
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating configMap with name configmap-test-volume-map-5bcdcb1f-6a79-11e9-b6ef-0e71f8c7e888
+STEP: Creating a pod to test consume configMaps
+Apr 29 12:21:47.560: INFO: Waiting up to 5m0s for pod "pod-configmaps-5bd00125-6a79-11e9-b6ef-0e71f8c7e888" in namespace "configmap-9199" to be "success or failure"
+Apr 29 12:21:47.582: INFO: Pod "pod-configmaps-5bd00125-6a79-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 22.112551ms
+Apr 29 12:21:49.589: INFO: Pod "pod-configmaps-5bd00125-6a79-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 2.028855457s
+Apr 29 12:21:51.596: INFO: Pod "pod-configmaps-5bd00125-6a79-11e9-b6ef-0e71f8c7e888": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.035460328s
+STEP: Saw pod success
+Apr 29 12:21:51.596: INFO: Pod "pod-configmaps-5bd00125-6a79-11e9-b6ef-0e71f8c7e888" satisfied condition "success or failure"
+Apr 29 12:21:51.602: INFO: Trying to get logs from node worker-4av2h-6665d7b5ff-9nc42 pod pod-configmaps-5bd00125-6a79-11e9-b6ef-0e71f8c7e888 container configmap-volume-test: 
+STEP: delete the pod
+Apr 29 12:21:51.641: INFO: Waiting for pod pod-configmaps-5bd00125-6a79-11e9-b6ef-0e71f8c7e888 to disappear
+Apr 29 12:21:51.649: INFO: Pod pod-configmaps-5bd00125-6a79-11e9-b6ef-0e71f8c7e888 no longer exists
+[AfterEach] [sig-storage] ConfigMap
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 12:21:51.649: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "configmap-9199" for this suite.
+Apr 29 12:21:57.685: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 12:21:57.881: INFO: namespace configmap-9199 deletion completed in 6.218714694s
+
+• [SLOW TEST:10.530 seconds]
+[sig-storage] ConfigMap
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:32
+  should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+[sig-storage] EmptyDir volumes 
+  should support (root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-storage] EmptyDir volumes
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 12:21:57.884: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename emptydir
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in emptydir-188
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should support (root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating a pod to test emptydir 0666 on tmpfs
+Apr 29 12:21:58.138: INFO: Waiting up to 5m0s for pod "pod-621daea7-6a79-11e9-b6ef-0e71f8c7e888" in namespace "emptydir-188" to be "success or failure"
+Apr 29 12:21:58.144: INFO: Pod "pod-621daea7-6a79-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 6.106742ms
+Apr 29 12:22:00.153: INFO: Pod "pod-621daea7-6a79-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 2.014613829s
+Apr 29 12:22:02.161: INFO: Pod "pod-621daea7-6a79-11e9-b6ef-0e71f8c7e888": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.023184658s
+STEP: Saw pod success
+Apr 29 12:22:02.162: INFO: Pod "pod-621daea7-6a79-11e9-b6ef-0e71f8c7e888" satisfied condition "success or failure"
+Apr 29 12:22:02.167: INFO: Trying to get logs from node worker-4av2h-6665d7b5ff-9nc42 pod pod-621daea7-6a79-11e9-b6ef-0e71f8c7e888 container test-container: 
+STEP: delete the pod
+Apr 29 12:22:02.196: INFO: Waiting for pod pod-621daea7-6a79-11e9-b6ef-0e71f8c7e888 to disappear
+Apr 29 12:22:02.203: INFO: Pod pod-621daea7-6a79-11e9-b6ef-0e71f8c7e888 no longer exists
+[AfterEach] [sig-storage] EmptyDir volumes
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 12:22:02.203: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "emptydir-188" for this suite.
+Apr 29 12:22:08.240: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 12:22:08.510: INFO: namespace emptydir-188 deletion completed in 6.301379908s
+
+• [SLOW TEST:10.627 seconds]
+[sig-storage] EmptyDir volumes
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:41
+  should support (root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSS
+------------------------------
+[sig-storage] Secrets 
+  should be consumable in multiple volumes in a pod [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-storage] Secrets
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 12:22:08.511: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename secrets
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in secrets-1511
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should be consumable in multiple volumes in a pod [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating secret with name secret-test-686c1eaa-6a79-11e9-b6ef-0e71f8c7e888
+STEP: Creating a pod to test consume secrets
+Apr 29 12:22:08.726: INFO: Waiting up to 5m0s for pod "pod-secrets-686d9c84-6a79-11e9-b6ef-0e71f8c7e888" in namespace "secrets-1511" to be "success or failure"
+Apr 29 12:22:08.762: INFO: Pod "pod-secrets-686d9c84-6a79-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 35.213536ms
+Apr 29 12:22:10.769: INFO: Pod "pod-secrets-686d9c84-6a79-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 2.042177048s
+Apr 29 12:22:12.781: INFO: Pod "pod-secrets-686d9c84-6a79-11e9-b6ef-0e71f8c7e888": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.054229488s
+STEP: Saw pod success
+Apr 29 12:22:12.781: INFO: Pod "pod-secrets-686d9c84-6a79-11e9-b6ef-0e71f8c7e888" satisfied condition "success or failure"
+Apr 29 12:22:12.787: INFO: Trying to get logs from node worker-4av2h-6665d7b5ff-9nc42 pod pod-secrets-686d9c84-6a79-11e9-b6ef-0e71f8c7e888 container secret-volume-test: 
+STEP: delete the pod
+Apr 29 12:22:12.819: INFO: Waiting for pod pod-secrets-686d9c84-6a79-11e9-b6ef-0e71f8c7e888 to disappear
+Apr 29 12:22:12.824: INFO: Pod pod-secrets-686d9c84-6a79-11e9-b6ef-0e71f8c7e888 no longer exists
+[AfterEach] [sig-storage] Secrets
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 12:22:12.824: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "secrets-1511" for this suite.
+Apr 29 12:22:18.894: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 12:22:19.153: INFO: namespace secrets-1511 deletion completed in 6.318825273s
+
+• [SLOW TEST:10.643 seconds]
+[sig-storage] Secrets
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/secrets_volume.go:33
+  should be consumable in multiple volumes in a pod [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+------------------------------
+[sig-apps] ReplicaSet 
+  should adopt matching pods on creation and release no longer matching pods [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-apps] ReplicaSet
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 12:22:19.162: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename replicaset
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in replicaset-9226
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should adopt matching pods on creation and release no longer matching pods [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Given a Pod with a 'name' label pod-adoption-release is created
+STEP: When a replicaset with a matching selector is created
+STEP: Then the orphan pod is adopted
+STEP: When the matched label of one of its pods change
+Apr 29 12:22:24.430: INFO: Pod name pod-adoption-release: Found 1 pods out of 1
+STEP: Then the pod is released
+[AfterEach] [sig-apps] ReplicaSet
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 12:22:24.457: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "replicaset-9226" for this suite.
+Apr 29 12:22:48.548: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 12:22:48.790: INFO: namespace replicaset-9226 deletion completed in 24.314816747s
+
+• [SLOW TEST:29.629 seconds]
+[sig-apps] ReplicaSet
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22
+  should adopt matching pods on creation and release no longer matching pods [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SS
+------------------------------
+[sig-storage] Secrets 
+  should be consumable from pods in volume with mappings [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-storage] Secrets
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 12:22:48.798: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename secrets
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in secrets-1746
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should be consumable from pods in volume with mappings [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating secret with name secret-test-map-80771783-6a79-11e9-b6ef-0e71f8c7e888
+STEP: Creating a pod to test consume secrets
+Apr 29 12:22:49.067: INFO: Waiting up to 5m0s for pod "pod-secrets-80799b97-6a79-11e9-b6ef-0e71f8c7e888" in namespace "secrets-1746" to be "success or failure"
+Apr 29 12:22:49.076: INFO: Pod "pod-secrets-80799b97-6a79-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 9.346525ms
+Apr 29 12:22:51.086: INFO: Pod "pod-secrets-80799b97-6a79-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 2.019216907s
+Apr 29 12:22:53.093: INFO: Pod "pod-secrets-80799b97-6a79-11e9-b6ef-0e71f8c7e888": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.026559666s
+STEP: Saw pod success
+Apr 29 12:22:53.094: INFO: Pod "pod-secrets-80799b97-6a79-11e9-b6ef-0e71f8c7e888" satisfied condition "success or failure"
+Apr 29 12:22:53.104: INFO: Trying to get logs from node worker-4av2h-6665d7b5ff-9nc42 pod pod-secrets-80799b97-6a79-11e9-b6ef-0e71f8c7e888 container secret-volume-test: 
+STEP: delete the pod
+Apr 29 12:22:53.147: INFO: Waiting for pod pod-secrets-80799b97-6a79-11e9-b6ef-0e71f8c7e888 to disappear
+Apr 29 12:22:53.152: INFO: Pod pod-secrets-80799b97-6a79-11e9-b6ef-0e71f8c7e888 no longer exists
+[AfterEach] [sig-storage] Secrets
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 12:22:53.152: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "secrets-1746" for this suite.
+Apr 29 12:22:59.178: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 12:22:59.327: INFO: namespace secrets-1746 deletion completed in 6.167516145s
+
+• [SLOW TEST:10.530 seconds]
+[sig-storage] Secrets
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/secrets_volume.go:33
+  should be consumable from pods in volume with mappings [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+[sig-apps] Daemon set [Serial] 
+  should rollback without unnecessary restarts [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-apps] Daemon set [Serial]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 12:22:59.332: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename daemonsets
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in daemonsets-4782
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [sig-apps] Daemon set [Serial]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:102
+[It] should rollback without unnecessary restarts [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+Apr 29 12:22:59.543: INFO: Create a RollingUpdate DaemonSet
+Apr 29 12:22:59.562: INFO: Check that daemon pods launch on every node of the cluster
+Apr 29 12:22:59.576: INFO: DaemonSet pods can't tolerate node master-h2h3g-655686fb8c-8kgdn with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node
+Apr 29 12:22:59.587: INFO: Number of nodes with available pods: 0
+Apr 29 12:22:59.587: INFO: Node worker-3oo4n-86cbf586b7-9q98g is running more than one daemon pod
+Apr 29 12:23:00.599: INFO: DaemonSet pods can't tolerate node master-h2h3g-655686fb8c-8kgdn with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node
+Apr 29 12:23:00.606: INFO: Number of nodes with available pods: 0
+Apr 29 12:23:00.606: INFO: Node worker-3oo4n-86cbf586b7-9q98g is running more than one daemon pod
+Apr 29 12:23:01.594: INFO: DaemonSet pods can't tolerate node master-h2h3g-655686fb8c-8kgdn with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node
+Apr 29 12:23:01.601: INFO: Number of nodes with available pods: 0
+Apr 29 12:23:01.601: INFO: Node worker-3oo4n-86cbf586b7-9q98g is running more than one daemon pod
+Apr 29 12:23:02.597: INFO: DaemonSet pods can't tolerate node master-h2h3g-655686fb8c-8kgdn with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node
+Apr 29 12:23:02.603: INFO: Number of nodes with available pods: 3
+Apr 29 12:23:02.603: INFO: Number of running nodes: 3, number of available pods: 3
+Apr 29 12:23:02.604: INFO: Update the DaemonSet to trigger a rollout
+Apr 29 12:23:02.617: INFO: Updating DaemonSet daemon-set
+Apr 29 12:23:16.651: INFO: Roll back the DaemonSet before rollout is complete
+Apr 29 12:23:16.661: INFO: Updating DaemonSet daemon-set
+Apr 29 12:23:16.661: INFO: Make sure DaemonSet rollback is complete
+Apr 29 12:23:16.666: INFO: Wrong image for pod: daemon-set-w8nbj. Expected: docker.io/library/nginx:1.14-alpine, got: foo:non-existent.
+Apr 29 12:23:16.667: INFO: Pod daemon-set-w8nbj is not available
+Apr 29 12:23:16.679: INFO: DaemonSet pods can't tolerate node master-h2h3g-655686fb8c-8kgdn with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node
+Apr 29 12:23:17.685: INFO: Wrong image for pod: daemon-set-w8nbj. Expected: docker.io/library/nginx:1.14-alpine, got: foo:non-existent.
+Apr 29 12:23:17.686: INFO: Pod daemon-set-w8nbj is not available
+Apr 29 12:23:17.692: INFO: DaemonSet pods can't tolerate node master-h2h3g-655686fb8c-8kgdn with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node
+Apr 29 12:23:18.731: INFO: DaemonSet pods can't tolerate node master-h2h3g-655686fb8c-8kgdn with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node
+Apr 29 12:23:19.688: INFO: Pod daemon-set-g42rv is not available
+Apr 29 12:23:19.695: INFO: DaemonSet pods can't tolerate node master-h2h3g-655686fb8c-8kgdn with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node
+[AfterEach] [sig-apps] Daemon set [Serial]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:68
+STEP: Deleting DaemonSet "daemon-set"
+STEP: deleting DaemonSet.extensions daemon-set in namespace daemonsets-4782, will wait for the garbage collector to delete the pods
+Apr 29 12:23:19.773: INFO: Deleting DaemonSet.extensions daemon-set took: 14.791714ms
+Apr 29 12:23:20.573: INFO: Terminating DaemonSet.extensions daemon-set pods took: 800.350138ms
+Apr 29 12:24:39.880: INFO: Number of nodes with available pods: 0
+Apr 29 12:24:39.880: INFO: Number of running nodes: 0, number of available pods: 0
+Apr 29 12:24:39.885: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"selfLink":"/apis/apps/v1/namespaces/daemonsets-4782/daemonsets","resourceVersion":"16345"},"items":null}
+
+Apr 29 12:24:39.890: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"selfLink":"/api/v1/namespaces/daemonsets-4782/pods","resourceVersion":"16345"},"items":null}
+
+[AfterEach] [sig-apps] Daemon set [Serial]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 12:24:39.928: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "daemonsets-4782" for this suite.
+Apr 29 12:24:45.962: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 12:24:46.185: INFO: namespace daemonsets-4782 deletion completed in 6.250380821s
+
+• [SLOW TEST:106.853 seconds]
+[sig-apps] Daemon set [Serial]
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22
+  should rollback without unnecessary restarts [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSS
+------------------------------
+[sig-storage] ConfigMap 
+  binary data should be reflected in volume [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-storage] ConfigMap
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 12:24:46.186: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename configmap
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in configmap-1625
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] binary data should be reflected in volume [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating configMap with name configmap-test-upd-c66d134f-6a79-11e9-b6ef-0e71f8c7e888
+STEP: Creating the pod
+STEP: Waiting for pod with text data
+STEP: Waiting for pod with binary data
+[AfterEach] [sig-storage] ConfigMap
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 12:24:50.507: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "configmap-1625" for this suite.
+Apr 29 12:25:14.542: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 12:25:14.731: INFO: namespace configmap-1625 deletion completed in 24.207962269s
+
+• [SLOW TEST:28.545 seconds]
+[sig-storage] ConfigMap
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:32
+  binary data should be reflected in volume [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSSS
+------------------------------
+[k8s.io] Probing container 
+  should be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [k8s.io] Probing container
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 12:25:14.731: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename container-probe
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in container-probe-768
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [k8s.io] Probing container
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/container_probe.go:51
+[It] should be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating pod liveness-exec in namespace container-probe-768
+Apr 29 12:25:18.951: INFO: Started pod liveness-exec in namespace container-probe-768
+STEP: checking the pod's current state and verifying that restartCount is present
+Apr 29 12:25:18.963: INFO: Initial restart count of pod liveness-exec is 0
+Apr 29 12:26:05.178: INFO: Restart count of pod container-probe-768/liveness-exec is now 1 (46.214893545s elapsed)
+STEP: deleting the pod
+[AfterEach] [k8s.io] Probing container
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 12:26:05.206: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "container-probe-768" for this suite.
+Apr 29 12:26:11.245: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 12:26:11.449: INFO: namespace container-probe-768 deletion completed in 6.232629501s
+
+• [SLOW TEST:56.718 seconds]
+[k8s.io] Probing container
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
+  should be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+------------------------------
+[sig-storage] Downward API volume 
+  should update annotations on modification [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-storage] Downward API volume
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 12:26:11.450: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename downward-api
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in downward-api-9231
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [sig-storage] Downward API volume
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:39
+[It] should update annotations on modification [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating the pod
+Apr 29 12:26:16.244: INFO: Successfully updated pod "annotationupdatef93aa1f7-6a79-11e9-b6ef-0e71f8c7e888"
+[AfterEach] [sig-storage] Downward API volume
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 12:26:18.290: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "downward-api-9231" for this suite.
+Apr 29 12:26:42.320: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 12:26:42.518: INFO: namespace downward-api-9231 deletion completed in 24.219515208s
+
+• [SLOW TEST:31.068 seconds]
+[sig-storage] Downward API volume
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:34
+  should update annotations on modification [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSSSSSSSS
+------------------------------
+[sig-cli] Kubectl client [k8s.io] Kubectl api-versions 
+  should check if v1 is in available api versions  [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-cli] Kubectl client
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 12:26:42.520: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename kubectl
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-4298
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [sig-cli] Kubectl client
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:213
+[It] should check if v1 is in available api versions  [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: validating api versions
+Apr 29 12:26:42.726: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 api-versions'
+Apr 29 12:26:42.900: INFO: stderr: ""
+Apr 29 12:26:42.900: INFO: stdout: "admissionregistration.k8s.io/v1beta1\napiextensions.k8s.io/v1beta1\napiregistration.k8s.io/v1\napiregistration.k8s.io/v1beta1\napplication.giantswarm.io/v1alpha1\napps/v1\napps/v1beta1\napps/v1beta2\nauditregistration.k8s.io/v1alpha1\nauthentication.k8s.io/v1\nauthentication.k8s.io/v1beta1\nauthorization.k8s.io/v1\nauthorization.k8s.io/v1beta1\nautoscaling/v1\nautoscaling/v2beta1\nautoscaling/v2beta2\nbatch/v1\nbatch/v1beta1\nbatch/v2alpha1\ncertificates.k8s.io/v1beta1\ncoordination.k8s.io/v1\ncoordination.k8s.io/v1beta1\ncore.giantswarm.io/v1alpha1\nevents.k8s.io/v1beta1\nextensions/v1beta1\nmetrics.k8s.io/v1beta1\nnetworking.k8s.io/v1\nnetworking.k8s.io/v1beta1\nnode.k8s.io/v1alpha1\nnode.k8s.io/v1beta1\npolicy/v1beta1\nrbac.authorization.k8s.io/v1\nrbac.authorization.k8s.io/v1alpha1\nrbac.authorization.k8s.io/v1beta1\nscheduling.k8s.io/v1\nscheduling.k8s.io/v1alpha1\nscheduling.k8s.io/v1beta1\nsettings.k8s.io/v1alpha1\nstorage.k8s.io/v1\nstorage.k8s.io/v1alpha1\nstorage.k8s.io/v1beta1\nv1\n"
+[AfterEach] [sig-cli] Kubectl client
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 12:26:42.901: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "kubectl-4298" for this suite.
+Apr 29 12:26:48.939: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 12:26:49.160: INFO: namespace kubectl-4298 deletion completed in 6.252057664s
+
+• [SLOW TEST:6.641 seconds]
+[sig-cli] Kubectl client
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23
+  [k8s.io] Kubectl api-versions
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
+    should check if v1 is in available api versions  [Conformance]
+    /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+------------------------------
+[k8s.io] Kubelet when scheduling a read only busybox container 
+  should not write to root filesystem [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [k8s.io] Kubelet
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 12:26:49.173: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename kubelet-test
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubelet-test-768
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [k8s.io] Kubelet
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:37
+[It] should not write to root filesystem [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[AfterEach] [k8s.io] Kubelet
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 12:26:53.444: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "kubelet-test-768" for this suite.
+Apr 29 12:27:35.469: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 12:27:35.666: INFO: namespace kubelet-test-768 deletion completed in 42.21555258s
+
+• [SLOW TEST:46.493 seconds]
+[k8s.io] Kubelet
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
+  when scheduling a read only busybox container
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:187
+    should not write to root filesystem [LinuxOnly] [NodeConformance] [Conformance]
+    /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSS
+------------------------------
+[sig-storage] Projected downwardAPI 
+  should provide podname only [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-storage] Projected downwardAPI
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 12:27:35.666: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename projected
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-1889
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [sig-storage] Projected downwardAPI
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:39
+[It] should provide podname only [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating a pod to test downward API volume plugin
+Apr 29 12:27:35.925: INFO: Waiting up to 5m0s for pod "downwardapi-volume-2b70d117-6a7a-11e9-b6ef-0e71f8c7e888" in namespace "projected-1889" to be "success or failure"
+Apr 29 12:27:35.929: INFO: Pod "downwardapi-volume-2b70d117-6a7a-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 4.085201ms
+Apr 29 12:27:37.935: INFO: Pod "downwardapi-volume-2b70d117-6a7a-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 2.00969968s
+Apr 29 12:27:39.945: INFO: Pod "downwardapi-volume-2b70d117-6a7a-11e9-b6ef-0e71f8c7e888": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.020140925s
+STEP: Saw pod success
+Apr 29 12:27:39.945: INFO: Pod "downwardapi-volume-2b70d117-6a7a-11e9-b6ef-0e71f8c7e888" satisfied condition "success or failure"
+Apr 29 12:27:39.960: INFO: Trying to get logs from node worker-4av2h-6665d7b5ff-9nc42 pod downwardapi-volume-2b70d117-6a7a-11e9-b6ef-0e71f8c7e888 container client-container: 
+STEP: delete the pod
+Apr 29 12:27:40.003: INFO: Waiting for pod downwardapi-volume-2b70d117-6a7a-11e9-b6ef-0e71f8c7e888 to disappear
+Apr 29 12:27:40.023: INFO: Pod downwardapi-volume-2b70d117-6a7a-11e9-b6ef-0e71f8c7e888 no longer exists
+[AfterEach] [sig-storage] Projected downwardAPI
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 12:27:40.023: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "projected-1889" for this suite.
+Apr 29 12:27:46.136: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 12:27:46.410: INFO: namespace projected-1889 deletion completed in 6.301601991s
+
+• [SLOW TEST:10.744 seconds]
+[sig-storage] Projected downwardAPI
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:33
+  should provide podname only [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SS
+------------------------------
+[sig-network] DNS 
+  should provide DNS for services  [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-network] DNS
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 12:27:46.412: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename dns
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in dns-7657
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should provide DNS for services  [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating a test headless service
+STEP: Running these commands on wheezy: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search dns-test-service.dns-7657.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_udp@dns-test-service.dns-7657.svc.cluster.local;check="$$(dig +tcp +noall +answer +search dns-test-service.dns-7657.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@dns-test-service.dns-7657.svc.cluster.local;check="$$(dig +notcp +noall +answer +search _http._tcp.dns-test-service.dns-7657.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/wheezy_udp@_http._tcp.dns-test-service.dns-7657.svc.cluster.local;check="$$(dig +tcp +noall +answer +search _http._tcp.dns-test-service.dns-7657.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/wheezy_tcp@_http._tcp.dns-test-service.dns-7657.svc.cluster.local;check="$$(dig +notcp +noall +answer +search _http._tcp.test-service-2.dns-7657.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/wheezy_udp@_http._tcp.test-service-2.dns-7657.svc.cluster.local;check="$$(dig +tcp +noall +answer +search _http._tcp.test-service-2.dns-7657.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/wheezy_tcp@_http._tcp.test-service-2.dns-7657.svc.cluster.local;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-7657.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@PodARecord;check="$$(dig +notcp +noall +answer +search 226.111.31.172.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/172.31.111.226_udp@PTR;check="$$(dig +tcp +noall +answer +search 226.111.31.172.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/172.31.111.226_tcp@PTR;sleep 1; done
+
+STEP: Running these commands on jessie: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search dns-test-service.dns-7657.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_udp@dns-test-service.dns-7657.svc.cluster.local;check="$$(dig +tcp +noall +answer +search dns-test-service.dns-7657.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_tcp@dns-test-service.dns-7657.svc.cluster.local;check="$$(dig +notcp +noall +answer +search _http._tcp.dns-test-service.dns-7657.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/jessie_udp@_http._tcp.dns-test-service.dns-7657.svc.cluster.local;check="$$(dig +tcp +noall +answer +search _http._tcp.dns-test-service.dns-7657.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/jessie_tcp@_http._tcp.dns-test-service.dns-7657.svc.cluster.local;check="$$(dig +notcp +noall +answer +search _http._tcp.test-service-2.dns-7657.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/jessie_udp@_http._tcp.test-service-2.dns-7657.svc.cluster.local;check="$$(dig +tcp +noall +answer +search _http._tcp.test-service-2.dns-7657.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/jessie_tcp@_http._tcp.test-service-2.dns-7657.svc.cluster.local;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-7657.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_tcp@PodARecord;check="$$(dig +notcp +noall +answer +search 226.111.31.172.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/172.31.111.226_udp@PTR;check="$$(dig +tcp +noall +answer +search 226.111.31.172.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/172.31.111.226_tcp@PTR;sleep 1; done
+
+STEP: creating a pod to probe DNS
+STEP: submitting the pod to kubernetes
+STEP: retrieving the pod
+STEP: looking for the results for each expected name from probers
+Apr 29 12:28:08.768: INFO: Unable to read wheezy_udp@dns-test-service.dns-7657.svc.cluster.local from pod dns-7657/dns-test-31ddd420-6a7a-11e9-b6ef-0e71f8c7e888: the server could not find the requested resource (get pods dns-test-31ddd420-6a7a-11e9-b6ef-0e71f8c7e888)
+Apr 29 12:28:08.776: INFO: Unable to read wheezy_tcp@dns-test-service.dns-7657.svc.cluster.local from pod dns-7657/dns-test-31ddd420-6a7a-11e9-b6ef-0e71f8c7e888: the server could not find the requested resource (get pods dns-test-31ddd420-6a7a-11e9-b6ef-0e71f8c7e888)
+Apr 29 12:28:08.837: INFO: Unable to read jessie_udp@dns-test-service.dns-7657.svc.cluster.local from pod dns-7657/dns-test-31ddd420-6a7a-11e9-b6ef-0e71f8c7e888: the server could not find the requested resource (get pods dns-test-31ddd420-6a7a-11e9-b6ef-0e71f8c7e888)
+Apr 29 12:28:08.843: INFO: Unable to read jessie_tcp@dns-test-service.dns-7657.svc.cluster.local from pod dns-7657/dns-test-31ddd420-6a7a-11e9-b6ef-0e71f8c7e888: the server could not find the requested resource (get pods dns-test-31ddd420-6a7a-11e9-b6ef-0e71f8c7e888)
+Apr 29 12:28:08.848: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.dns-7657.svc.cluster.local from pod dns-7657/dns-test-31ddd420-6a7a-11e9-b6ef-0e71f8c7e888: the server could not find the requested resource (get pods dns-test-31ddd420-6a7a-11e9-b6ef-0e71f8c7e888)
+Apr 29 12:28:08.889: INFO: Lookups using dns-7657/dns-test-31ddd420-6a7a-11e9-b6ef-0e71f8c7e888 failed for: [wheezy_udp@dns-test-service.dns-7657.svc.cluster.local wheezy_tcp@dns-test-service.dns-7657.svc.cluster.local jessie_udp@dns-test-service.dns-7657.svc.cluster.local jessie_tcp@dns-test-service.dns-7657.svc.cluster.local jessie_udp@_http._tcp.dns-test-service.dns-7657.svc.cluster.local]
+
+Apr 29 12:28:14.076: INFO: DNS probes using dns-7657/dns-test-31ddd420-6a7a-11e9-b6ef-0e71f8c7e888 succeeded
+
+STEP: deleting the pod
+STEP: deleting the test service
+STEP: deleting the test headless service
+[AfterEach] [sig-network] DNS
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 12:28:14.304: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "dns-7657" for this suite.
+Apr 29 12:28:20.424: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 12:28:20.654: INFO: namespace dns-7657 deletion completed in 6.321400408s
+
+• [SLOW TEST:34.242 seconds]
+[sig-network] DNS
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:22
+  should provide DNS for services  [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSSSSSSS
+------------------------------
+[sig-cli] Kubectl client [k8s.io] Kubectl run default 
+  should create an rc or deployment from an image  [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-cli] Kubectl client
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 12:28:20.655: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename kubectl
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-866
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [sig-cli] Kubectl client
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:213
+[BeforeEach] [k8s.io] Kubectl run default
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1318
+[It] should create an rc or deployment from an image  [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: running the image docker.io/library/nginx:1.14-alpine
+Apr 29 12:28:20.853: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 run e2e-test-nginx-deployment --image=docker.io/library/nginx:1.14-alpine --namespace=kubectl-866'
+Apr 29 12:28:20.996: INFO: stderr: "kubectl run --generator=deployment/apps.v1 is DEPRECATED and will be removed in a future version. Use kubectl run --generator=run-pod/v1 or kubectl create instead.\n"
+Apr 29 12:28:20.996: INFO: stdout: "deployment.apps/e2e-test-nginx-deployment created\n"
+STEP: verifying the pod controlled by e2e-test-nginx-deployment gets created
+[AfterEach] [k8s.io] Kubectl run default
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1324
+Apr 29 12:28:23.015: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 delete deployment e2e-test-nginx-deployment --namespace=kubectl-866'
+Apr 29 12:28:23.146: INFO: stderr: ""
+Apr 29 12:28:23.146: INFO: stdout: "deployment.extensions \"e2e-test-nginx-deployment\" deleted\n"
+[AfterEach] [sig-cli] Kubectl client
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 12:28:23.147: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "kubectl-866" for this suite.
+Apr 29 12:30:31.185: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 12:30:31.402: INFO: namespace kubectl-866 deletion completed in 2m8.244104524s
+
+• [SLOW TEST:130.747 seconds]
+[sig-cli] Kubectl client
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23
+  [k8s.io] Kubectl run default
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
+    should create an rc or deployment from an image  [Conformance]
+    /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+------------------------------
+[sig-storage] Secrets 
+  should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-storage] Secrets
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 12:30:31.408: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename secrets
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in secrets-6776
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating secret with name secret-test-94326882-6a7a-11e9-b6ef-0e71f8c7e888
+STEP: Creating a pod to test consume secrets
+Apr 29 12:30:31.662: INFO: Waiting up to 5m0s for pod "pod-secrets-94340f4a-6a7a-11e9-b6ef-0e71f8c7e888" in namespace "secrets-6776" to be "success or failure"
+Apr 29 12:30:31.668: INFO: Pod "pod-secrets-94340f4a-6a7a-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 6.479279ms
+Apr 29 12:30:33.676: INFO: Pod "pod-secrets-94340f4a-6a7a-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 2.013997183s
+Apr 29 12:30:35.685: INFO: Pod "pod-secrets-94340f4a-6a7a-11e9-b6ef-0e71f8c7e888": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.023198928s
+STEP: Saw pod success
+Apr 29 12:30:35.685: INFO: Pod "pod-secrets-94340f4a-6a7a-11e9-b6ef-0e71f8c7e888" satisfied condition "success or failure"
+Apr 29 12:30:35.694: INFO: Trying to get logs from node worker-4av2h-6665d7b5ff-9nc42 pod pod-secrets-94340f4a-6a7a-11e9-b6ef-0e71f8c7e888 container secret-volume-test: 
+STEP: delete the pod
+Apr 29 12:30:35.730: INFO: Waiting for pod pod-secrets-94340f4a-6a7a-11e9-b6ef-0e71f8c7e888 to disappear
+Apr 29 12:30:35.734: INFO: Pod pod-secrets-94340f4a-6a7a-11e9-b6ef-0e71f8c7e888 no longer exists
+[AfterEach] [sig-storage] Secrets
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 12:30:35.734: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "secrets-6776" for this suite.
+Apr 29 12:30:41.773: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 12:30:42.123: INFO: namespace secrets-6776 deletion completed in 6.382482339s
+
+• [SLOW TEST:10.716 seconds]
+[sig-storage] Secrets
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/secrets_volume.go:33
+  should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+------------------------------
+[sig-storage] Projected configMap 
+  should be consumable from pods in volume with mappings [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-storage] Projected configMap
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 12:30:42.126: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename projected
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-9062
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should be consumable from pods in volume with mappings [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating configMap with name projected-configmap-test-volume-map-9a96f12e-6a7a-11e9-b6ef-0e71f8c7e888
+STEP: Creating a pod to test consume configMaps
+Apr 29 12:30:42.394: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-9a987060-6a7a-11e9-b6ef-0e71f8c7e888" in namespace "projected-9062" to be "success or failure"
+Apr 29 12:30:42.414: INFO: Pod "pod-projected-configmaps-9a987060-6a7a-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 20.230096ms
+Apr 29 12:30:44.423: INFO: Pod "pod-projected-configmaps-9a987060-6a7a-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 2.029114274s
+Apr 29 12:30:46.434: INFO: Pod "pod-projected-configmaps-9a987060-6a7a-11e9-b6ef-0e71f8c7e888": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.039945059s
+STEP: Saw pod success
+Apr 29 12:30:46.434: INFO: Pod "pod-projected-configmaps-9a987060-6a7a-11e9-b6ef-0e71f8c7e888" satisfied condition "success or failure"
+Apr 29 12:30:46.441: INFO: Trying to get logs from node worker-4av2h-6665d7b5ff-9nc42 pod pod-projected-configmaps-9a987060-6a7a-11e9-b6ef-0e71f8c7e888 container projected-configmap-volume-test: 
+STEP: delete the pod
+Apr 29 12:30:46.481: INFO: Waiting for pod pod-projected-configmaps-9a987060-6a7a-11e9-b6ef-0e71f8c7e888 to disappear
+Apr 29 12:30:46.487: INFO: Pod pod-projected-configmaps-9a987060-6a7a-11e9-b6ef-0e71f8c7e888 no longer exists
+[AfterEach] [sig-storage] Projected configMap
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 12:30:46.487: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "projected-9062" for this suite.
+Apr 29 12:30:52.511: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 12:30:52.712: INFO: namespace projected-9062 deletion completed in 6.218438971s
+
+• [SLOW TEST:10.586 seconds]
+[sig-storage] Projected configMap
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_configmap.go:33
+  should be consumable from pods in volume with mappings [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+------------------------------
+[sig-network] Networking Granular Checks: Pods 
+  should function for intra-pod communication: http [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-network] Networking
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 12:30:52.714: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename pod-network-test
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in pod-network-test-4192
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should function for intra-pod communication: http [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Performing setup for networking test in namespace pod-network-test-4192
+STEP: creating a selector
+STEP: Creating the service pods in kubernetes
+Apr 29 12:30:52.883: INFO: Waiting up to 10m0s for all (but 0) nodes to be schedulable
+STEP: Creating test pods
+Apr 29 12:31:13.067: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s 'http://192.168.205.96:8080/dial?request=hostName&protocol=http&host=192.168.205.93&port=8080&tries=1'] Namespace:pod-network-test-4192 PodName:host-test-container-pod ContainerName:hostexec Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
+Apr 29 12:31:13.067: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+Apr 29 12:31:13.254: INFO: Waiting for endpoints: map[]
+Apr 29 12:31:13.259: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s 'http://192.168.205.96:8080/dial?request=hostName&protocol=http&host=192.168.14.42&port=8080&tries=1'] Namespace:pod-network-test-4192 PodName:host-test-container-pod ContainerName:hostexec Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
+Apr 29 12:31:13.259: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+Apr 29 12:31:13.478: INFO: Waiting for endpoints: map[]
+Apr 29 12:31:13.484: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s 'http://192.168.205.96:8080/dial?request=hostName&protocol=http&host=192.168.148.104&port=8080&tries=1'] Namespace:pod-network-test-4192 PodName:host-test-container-pod ContainerName:hostexec Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
+Apr 29 12:31:13.484: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+Apr 29 12:31:13.647: INFO: Waiting for endpoints: map[]
+[AfterEach] [sig-network] Networking
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 12:31:13.647: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "pod-network-test-4192" for this suite.
+Apr 29 12:31:37.685: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 12:31:37.864: INFO: namespace pod-network-test-4192 deletion completed in 24.204036736s
+
+• [SLOW TEST:45.150 seconds]
+[sig-network] Networking
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:25
+  Granular Checks: Pods
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:28
+    should function for intra-pod communication: http [LinuxOnly] [NodeConformance] [Conformance]
+    /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSSSSSSSSSSSS
+------------------------------
+[sig-api-machinery] Aggregator 
+  Should be able to support the 1.10 Sample API Server using the current Aggregator [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-api-machinery] Aggregator
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 12:31:37.866: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename aggregator
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in aggregator-5069
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [sig-api-machinery] Aggregator
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/aggregator.go:69
+[It] Should be able to support the 1.10 Sample API Server using the current Aggregator [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Registering the sample API server.
+Apr 29 12:31:39.404: INFO: new replicaset for deployment "sample-apiserver-deployment" is yet to be created
+Apr 29 12:31:41.516: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63692137899, loc:(*time.Location)(0x8a060e0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63692137899, loc:(*time.Location)(0x8a060e0)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63692137899, loc:(*time.Location)(0x8a060e0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63692137899, loc:(*time.Location)(0x8a060e0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-65db6755fc\" is progressing."}}, CollisionCount:(*int32)(nil)}
+Apr 29 12:31:43.523: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63692137899, loc:(*time.Location)(0x8a060e0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63692137899, loc:(*time.Location)(0x8a060e0)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63692137899, loc:(*time.Location)(0x8a060e0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63692137899, loc:(*time.Location)(0x8a060e0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-65db6755fc\" is progressing."}}, CollisionCount:(*int32)(nil)}
+Apr 29 12:31:45.522: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63692137899, loc:(*time.Location)(0x8a060e0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63692137899, loc:(*time.Location)(0x8a060e0)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63692137899, loc:(*time.Location)(0x8a060e0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63692137899, loc:(*time.Location)(0x8a060e0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-65db6755fc\" is progressing."}}, CollisionCount:(*int32)(nil)}
+Apr 29 12:31:47.522: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63692137899, loc:(*time.Location)(0x8a060e0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63692137899, loc:(*time.Location)(0x8a060e0)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63692137899, loc:(*time.Location)(0x8a060e0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63692137899, loc:(*time.Location)(0x8a060e0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-65db6755fc\" is progressing."}}, CollisionCount:(*int32)(nil)}
+Apr 29 12:31:49.525: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63692137899, loc:(*time.Location)(0x8a060e0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63692137899, loc:(*time.Location)(0x8a060e0)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63692137899, loc:(*time.Location)(0x8a060e0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63692137899, loc:(*time.Location)(0x8a060e0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-65db6755fc\" is progressing."}}, CollisionCount:(*int32)(nil)}
+Apr 29 12:31:53.129: INFO: Waited 1.596826436s for the sample-apiserver to be ready to handle requests.
+[AfterEach] [sig-api-machinery] Aggregator
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/aggregator.go:60
+[AfterEach] [sig-api-machinery] Aggregator
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 12:31:53.966: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "aggregator-5069" for this suite.
+Apr 29 12:32:00.115: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 12:32:00.367: INFO: namespace aggregator-5069 deletion completed in 6.353737662s
+
+• [SLOW TEST:22.501 seconds]
+[sig-api-machinery] Aggregator
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:22
+  Should be able to support the 1.10 Sample API Server using the current Aggregator [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSS
+------------------------------
+[sig-auth] ServiceAccounts 
+  should mount an API token into pods  [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-auth] ServiceAccounts
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 12:32:00.368: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename svcaccounts
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in svcaccounts-5560
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should mount an API token into pods  [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: getting the auto-created API token
+STEP: reading a file in the container
+Apr 29 12:32:05.162: INFO: Running '/usr/local/bin/kubectl exec --namespace=svcaccounts-5560 pod-service-account-c9855616-6a7a-11e9-b6ef-0e71f8c7e888 -c=test -- cat /var/run/secrets/kubernetes.io/serviceaccount/token'
+STEP: reading a file in the container
+Apr 29 12:32:05.468: INFO: Running '/usr/local/bin/kubectl exec --namespace=svcaccounts-5560 pod-service-account-c9855616-6a7a-11e9-b6ef-0e71f8c7e888 -c=test -- cat /var/run/secrets/kubernetes.io/serviceaccount/ca.crt'
+STEP: reading a file in the container
+Apr 29 12:32:05.793: INFO: Running '/usr/local/bin/kubectl exec --namespace=svcaccounts-5560 pod-service-account-c9855616-6a7a-11e9-b6ef-0e71f8c7e888 -c=test -- cat /var/run/secrets/kubernetes.io/serviceaccount/namespace'
+[AfterEach] [sig-auth] ServiceAccounts
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 12:32:06.114: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "svcaccounts-5560" for this suite.
+Apr 29 12:32:12.188: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 12:32:12.407: INFO: namespace svcaccounts-5560 deletion completed in 6.266890408s
+
+• [SLOW TEST:12.039 seconds]
+[sig-auth] ServiceAccounts
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/auth/framework.go:22
+  should mount an API token into pods  [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSSSS
+------------------------------
+[sig-node] Downward API 
+  should provide pod name, namespace and IP address as env vars [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-node] Downward API
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 12:32:12.408: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename downward-api
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in downward-api-4582
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should provide pod name, namespace and IP address as env vars [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating a pod to test downward api env vars
+Apr 29 12:32:12.627: INFO: Waiting up to 5m0s for pod "downward-api-d0616419-6a7a-11e9-b6ef-0e71f8c7e888" in namespace "downward-api-4582" to be "success or failure"
+Apr 29 12:32:12.633: INFO: Pod "downward-api-d0616419-6a7a-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 5.60491ms
+Apr 29 12:32:14.648: INFO: Pod "downward-api-d0616419-6a7a-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 2.020343459s
+Apr 29 12:32:16.654: INFO: Pod "downward-api-d0616419-6a7a-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 4.026947058s
+Apr 29 12:32:18.662: INFO: Pod "downward-api-d0616419-6a7a-11e9-b6ef-0e71f8c7e888": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.035013273s
+STEP: Saw pod success
+Apr 29 12:32:18.662: INFO: Pod "downward-api-d0616419-6a7a-11e9-b6ef-0e71f8c7e888" satisfied condition "success or failure"
+Apr 29 12:32:18.667: INFO: Trying to get logs from node worker-4av2h-6665d7b5ff-9nc42 pod downward-api-d0616419-6a7a-11e9-b6ef-0e71f8c7e888 container dapi-container: 
+STEP: delete the pod
+Apr 29 12:32:18.727: INFO: Waiting for pod downward-api-d0616419-6a7a-11e9-b6ef-0e71f8c7e888 to disappear
+Apr 29 12:32:18.741: INFO: Pod downward-api-d0616419-6a7a-11e9-b6ef-0e71f8c7e888 no longer exists
+[AfterEach] [sig-node] Downward API
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 12:32:18.741: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "downward-api-4582" for this suite.
+Apr 29 12:32:24.781: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 12:32:25.047: INFO: namespace downward-api-4582 deletion completed in 6.291693132s
+
+• [SLOW TEST:12.639 seconds]
+[sig-node] Downward API
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downward_api.go:38
+  should provide pod name, namespace and IP address as env vars [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSSSSSSSSSSSSSSSSSSSSSS
+------------------------------
+[sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] 
+  should perform rolling updates and roll backs of template modifications [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-apps] StatefulSet
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 12:32:25.048: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename statefulset
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in statefulset-8657
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [sig-apps] StatefulSet
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:59
+[BeforeEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:74
+STEP: Creating service test in namespace statefulset-8657
+[It] should perform rolling updates and roll backs of template modifications [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating a new StatefulSet
+Apr 29 12:32:25.299: INFO: Found 0 stateful pods, waiting for 3
+Apr 29 12:32:35.307: INFO: Found 2 stateful pods, waiting for 3
+Apr 29 12:32:45.307: INFO: Waiting for pod ss2-0 to enter Running - Ready=true, currently Running - Ready=true
+Apr 29 12:32:45.307: INFO: Waiting for pod ss2-1 to enter Running - Ready=true, currently Running - Ready=true
+Apr 29 12:32:45.307: INFO: Waiting for pod ss2-2 to enter Running - Ready=true, currently Running - Ready=true
+Apr 29 12:32:45.319: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 exec --namespace=statefulset-8657 ss2-1 -- /bin/sh -x -c mv -v /usr/share/nginx/html/index.html /tmp/ || true'
+Apr 29 12:32:45.918: INFO: stderr: "+ mv -v /usr/share/nginx/html/index.html /tmp/\n"
+Apr 29 12:32:45.918: INFO: stdout: "'/usr/share/nginx/html/index.html' -> '/tmp/index.html'\n"
+Apr 29 12:32:45.918: INFO: stdout of mv -v /usr/share/nginx/html/index.html /tmp/ || true on ss2-1: '/usr/share/nginx/html/index.html' -> '/tmp/index.html'
+
+STEP: Updating StatefulSet template: update image from docker.io/library/nginx:1.14-alpine to docker.io/library/nginx:1.15-alpine
+Apr 29 12:32:55.989: INFO: Updating stateful set ss2
+STEP: Creating a new revision
+STEP: Updating Pods in reverse ordinal order
+Apr 29 12:33:06.027: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 exec --namespace=statefulset-8657 ss2-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
+Apr 29 12:33:06.359: INFO: stderr: "+ mv -v /tmp/index.html /usr/share/nginx/html/\n"
+Apr 29 12:33:06.359: INFO: stdout: "'/tmp/index.html' -> '/usr/share/nginx/html/index.html'\n"
+Apr 29 12:33:06.359: INFO: stdout of mv -v /tmp/index.html /usr/share/nginx/html/ || true on ss2-1: '/tmp/index.html' -> '/usr/share/nginx/html/index.html'
+
+Apr 29 12:33:16.403: INFO: Waiting for StatefulSet statefulset-8657/ss2 to complete update
+Apr 29 12:33:16.403: INFO: Waiting for Pod statefulset-8657/ss2-0 to have revision ss2-c79899b9 update revision ss2-787997d666
+Apr 29 12:33:16.403: INFO: Waiting for Pod statefulset-8657/ss2-1 to have revision ss2-c79899b9 update revision ss2-787997d666
+Apr 29 12:33:26.414: INFO: Waiting for StatefulSet statefulset-8657/ss2 to complete update
+Apr 29 12:33:26.415: INFO: Waiting for Pod statefulset-8657/ss2-0 to have revision ss2-c79899b9 update revision ss2-787997d666
+Apr 29 12:33:36.418: INFO: Waiting for StatefulSet statefulset-8657/ss2 to complete update
+STEP: Rolling back to a previous revision
+Apr 29 12:33:46.423: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 exec --namespace=statefulset-8657 ss2-1 -- /bin/sh -x -c mv -v /usr/share/nginx/html/index.html /tmp/ || true'
+Apr 29 12:33:46.728: INFO: stderr: "+ mv -v /usr/share/nginx/html/index.html /tmp/\n"
+Apr 29 12:33:46.728: INFO: stdout: "'/usr/share/nginx/html/index.html' -> '/tmp/index.html'\n"
+Apr 29 12:33:46.728: INFO: stdout of mv -v /usr/share/nginx/html/index.html /tmp/ || true on ss2-1: '/usr/share/nginx/html/index.html' -> '/tmp/index.html'
+
+Apr 29 12:33:56.772: INFO: Updating stateful set ss2
+STEP: Rolling back update in reverse ordinal order
+Apr 29 12:34:06.806: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 exec --namespace=statefulset-8657 ss2-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
+Apr 29 12:34:07.088: INFO: stderr: "+ mv -v /tmp/index.html /usr/share/nginx/html/\n"
+Apr 29 12:34:07.089: INFO: stdout: "'/tmp/index.html' -> '/usr/share/nginx/html/index.html'\n"
+Apr 29 12:34:07.089: INFO: stdout of mv -v /tmp/index.html /usr/share/nginx/html/ || true on ss2-1: '/tmp/index.html' -> '/usr/share/nginx/html/index.html'
+
+Apr 29 12:34:17.144: INFO: Waiting for StatefulSet statefulset-8657/ss2 to complete update
+Apr 29 12:34:17.145: INFO: Waiting for Pod statefulset-8657/ss2-0 to have revision ss2-787997d666 update revision ss2-c79899b9
+Apr 29 12:34:17.145: INFO: Waiting for Pod statefulset-8657/ss2-1 to have revision ss2-787997d666 update revision ss2-c79899b9
+Apr 29 12:34:27.163: INFO: Waiting for StatefulSet statefulset-8657/ss2 to complete update
+Apr 29 12:34:27.163: INFO: Waiting for Pod statefulset-8657/ss2-0 to have revision ss2-787997d666 update revision ss2-c79899b9
+[AfterEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:85
+Apr 29 12:34:37.160: INFO: Deleting all statefulset in ns statefulset-8657
+Apr 29 12:34:37.166: INFO: Scaling statefulset ss2 to 0
+Apr 29 12:35:07.198: INFO: Waiting for statefulset status.replicas updated to 0
+Apr 29 12:35:07.205: INFO: Deleting statefulset ss2
+[AfterEach] [sig-apps] StatefulSet
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 12:35:07.250: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "statefulset-8657" for this suite.
+Apr 29 12:35:15.289: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 12:35:15.473: INFO: namespace statefulset-8657 deletion completed in 8.203530069s
+
+• [SLOW TEST:170.425 seconds]
+[sig-apps] StatefulSet
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22
+  [k8s.io] Basic StatefulSet functionality [StatefulSetBasic]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
+    should perform rolling updates and roll backs of template modifications [Conformance]
+    /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+------------------------------
+[sig-storage] Projected combined 
+  should project all components that make up the projection API [Projection][NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-storage] Projected combined
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 12:35:15.474: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename projected
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-5855
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should project all components that make up the projection API [Projection][NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating configMap with name configmap-projected-all-test-volume-3d7dc7cd-6a7b-11e9-b6ef-0e71f8c7e888
+STEP: Creating secret with name secret-projected-all-test-volume-3d7dc7ac-6a7b-11e9-b6ef-0e71f8c7e888
+STEP: Creating a pod to test Check all projections for projected volume plugin
+Apr 29 12:35:15.709: INFO: Waiting up to 5m0s for pod "projected-volume-3d7dc666-6a7b-11e9-b6ef-0e71f8c7e888" in namespace "projected-5855" to be "success or failure"
+Apr 29 12:35:15.714: INFO: Pod "projected-volume-3d7dc666-6a7b-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 4.602225ms
+Apr 29 12:35:17.721: INFO: Pod "projected-volume-3d7dc666-6a7b-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 2.012043447s
+Apr 29 12:35:19.732: INFO: Pod "projected-volume-3d7dc666-6a7b-11e9-b6ef-0e71f8c7e888": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.02285357s
+STEP: Saw pod success
+Apr 29 12:35:19.733: INFO: Pod "projected-volume-3d7dc666-6a7b-11e9-b6ef-0e71f8c7e888" satisfied condition "success or failure"
+Apr 29 12:35:19.739: INFO: Trying to get logs from node worker-4av2h-6665d7b5ff-9nc42 pod projected-volume-3d7dc666-6a7b-11e9-b6ef-0e71f8c7e888 container projected-all-volume-test: 
+STEP: delete the pod
+Apr 29 12:35:19.787: INFO: Waiting for pod projected-volume-3d7dc666-6a7b-11e9-b6ef-0e71f8c7e888 to disappear
+Apr 29 12:35:19.801: INFO: Pod projected-volume-3d7dc666-6a7b-11e9-b6ef-0e71f8c7e888 no longer exists
+[AfterEach] [sig-storage] Projected combined
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 12:35:19.801: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "projected-5855" for this suite.
+Apr 29 12:35:25.832: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 12:35:26.080: INFO: namespace projected-5855 deletion completed in 6.269999723s
+
+• [SLOW TEST:10.606 seconds]
+[sig-storage] Projected combined
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_combined.go:31
+  should project all components that make up the projection API [Projection][NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSS
+------------------------------
+[sig-network] Services 
+  should serve a basic endpoint from pods  [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-network] Services
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 12:35:26.080: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename services
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in services-8146
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [sig-network] Services
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:86
+[It] should serve a basic endpoint from pods  [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: creating service endpoint-test2 in namespace services-8146
+STEP: waiting up to 3m0s for service endpoint-test2 in namespace services-8146 to expose endpoints map[]
+Apr 29 12:35:26.306: INFO: successfully validated that service endpoint-test2 in namespace services-8146 exposes endpoints map[] (10.903713ms elapsed)
+STEP: Creating pod pod1 in namespace services-8146
+STEP: waiting up to 3m0s for service endpoint-test2 in namespace services-8146 to expose endpoints map[pod1:[80]]
+Apr 29 12:35:30.450: INFO: Unexpected endpoints: found map[], expected map[pod1:[80]] (4.119292984s elapsed, will retry)
+Apr 29 12:35:33.492: INFO: successfully validated that service endpoint-test2 in namespace services-8146 exposes endpoints map[pod1:[80]] (7.160813632s elapsed)
+STEP: Creating pod pod2 in namespace services-8146
+STEP: waiting up to 3m0s for service endpoint-test2 in namespace services-8146 to expose endpoints map[pod1:[80] pod2:[80]]
+Apr 29 12:35:37.641: INFO: successfully validated that service endpoint-test2 in namespace services-8146 exposes endpoints map[pod1:[80] pod2:[80]] (4.137113729s elapsed)
+STEP: Deleting pod pod1 in namespace services-8146
+STEP: waiting up to 3m0s for service endpoint-test2 in namespace services-8146 to expose endpoints map[pod2:[80]]
+Apr 29 12:35:37.694: INFO: successfully validated that service endpoint-test2 in namespace services-8146 exposes endpoints map[pod2:[80]] (39.583151ms elapsed)
+STEP: Deleting pod pod2 in namespace services-8146
+STEP: waiting up to 3m0s for service endpoint-test2 in namespace services-8146 to expose endpoints map[]
+Apr 29 12:35:38.725: INFO: successfully validated that service endpoint-test2 in namespace services-8146 exposes endpoints map[] (1.020827861s elapsed)
+[AfterEach] [sig-network] Services
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 12:35:38.760: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "services-8146" for this suite.
+Apr 29 12:36:02.800: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 12:36:03.010: INFO: namespace services-8146 deletion completed in 24.236887891s
+[AfterEach] [sig-network] Services
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:91
+
+• [SLOW TEST:36.930 seconds]
+[sig-network] Services
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:22
+  should serve a basic endpoint from pods  [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSS
+------------------------------
+[sig-storage] Downward API volume 
+  should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-storage] Downward API volume
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 12:36:03.010: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename downward-api
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in downward-api-2569
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [sig-storage] Downward API volume
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:39
+[It] should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating a pod to test downward API volume plugin
+Apr 29 12:36:03.230: INFO: Waiting up to 5m0s for pod "downwardapi-volume-59d4bfd0-6a7b-11e9-b6ef-0e71f8c7e888" in namespace "downward-api-2569" to be "success or failure"
+Apr 29 12:36:03.244: INFO: Pod "downwardapi-volume-59d4bfd0-6a7b-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 13.578999ms
+Apr 29 12:36:05.250: INFO: Pod "downwardapi-volume-59d4bfd0-6a7b-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 2.020364555s
+Apr 29 12:36:07.257: INFO: Pod "downwardapi-volume-59d4bfd0-6a7b-11e9-b6ef-0e71f8c7e888": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.027363664s
+STEP: Saw pod success
+Apr 29 12:36:07.257: INFO: Pod "downwardapi-volume-59d4bfd0-6a7b-11e9-b6ef-0e71f8c7e888" satisfied condition "success or failure"
+Apr 29 12:36:07.261: INFO: Trying to get logs from node worker-4av2h-6665d7b5ff-9nc42 pod downwardapi-volume-59d4bfd0-6a7b-11e9-b6ef-0e71f8c7e888 container client-container: 
+STEP: delete the pod
+Apr 29 12:36:07.297: INFO: Waiting for pod downwardapi-volume-59d4bfd0-6a7b-11e9-b6ef-0e71f8c7e888 to disappear
+Apr 29 12:36:07.302: INFO: Pod downwardapi-volume-59d4bfd0-6a7b-11e9-b6ef-0e71f8c7e888 no longer exists
+[AfterEach] [sig-storage] Downward API volume
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 12:36:07.302: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "downward-api-2569" for this suite.
+Apr 29 12:36:13.342: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 12:36:13.583: INFO: namespace downward-api-2569 deletion completed in 6.271166523s
+
+• [SLOW TEST:10.574 seconds]
+[sig-storage] Downward API volume
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:34
+  should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSS
+------------------------------
+[sig-storage] ConfigMap 
+  should be consumable from pods in volume with mappings as non-root [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-storage] ConfigMap
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 12:36:13.584: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename configmap
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in configmap-453
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should be consumable from pods in volume with mappings as non-root [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating configMap with name configmap-test-volume-map-6026b5f0-6a7b-11e9-b6ef-0e71f8c7e888
+STEP: Creating a pod to test consume configMaps
+Apr 29 12:36:13.846: INFO: Waiting up to 5m0s for pod "pod-configmaps-6028642e-6a7b-11e9-b6ef-0e71f8c7e888" in namespace "configmap-453" to be "success or failure"
+Apr 29 12:36:13.856: INFO: Pod "pod-configmaps-6028642e-6a7b-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 9.80337ms
+Apr 29 12:36:15.862: INFO: Pod "pod-configmaps-6028642e-6a7b-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 2.015386348s
+Apr 29 12:36:17.867: INFO: Pod "pod-configmaps-6028642e-6a7b-11e9-b6ef-0e71f8c7e888": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.021223094s
+STEP: Saw pod success
+Apr 29 12:36:17.868: INFO: Pod "pod-configmaps-6028642e-6a7b-11e9-b6ef-0e71f8c7e888" satisfied condition "success or failure"
+Apr 29 12:36:17.873: INFO: Trying to get logs from node worker-4av2h-6665d7b5ff-9nc42 pod pod-configmaps-6028642e-6a7b-11e9-b6ef-0e71f8c7e888 container configmap-volume-test: 
+STEP: delete the pod
+Apr 29 12:36:17.912: INFO: Waiting for pod pod-configmaps-6028642e-6a7b-11e9-b6ef-0e71f8c7e888 to disappear
+Apr 29 12:36:17.917: INFO: Pod pod-configmaps-6028642e-6a7b-11e9-b6ef-0e71f8c7e888 no longer exists
+[AfterEach] [sig-storage] ConfigMap
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 12:36:17.917: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "configmap-453" for this suite.
+Apr 29 12:36:23.962: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 12:36:24.215: INFO: namespace configmap-453 deletion completed in 6.275831998s
+
+• [SLOW TEST:10.631 seconds]
+[sig-storage] ConfigMap
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:32
+  should be consumable from pods in volume with mappings as non-root [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSSSSSSSSSSSSSSSSSSS
+------------------------------
+[sig-apps] ReplicaSet 
+  should serve a basic image on each replica with a public image  [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-apps] ReplicaSet
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 12:36:24.217: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename replicaset
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in replicaset-5676
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should serve a basic image on each replica with a public image  [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+Apr 29 12:36:24.421: INFO: Creating ReplicaSet my-hostname-basic-6678a5e8-6a7b-11e9-b6ef-0e71f8c7e888
+Apr 29 12:36:24.448: INFO: Pod name my-hostname-basic-6678a5e8-6a7b-11e9-b6ef-0e71f8c7e888: Found 0 pods out of 1
+Apr 29 12:36:29.457: INFO: Pod name my-hostname-basic-6678a5e8-6a7b-11e9-b6ef-0e71f8c7e888: Found 1 pods out of 1
+Apr 29 12:36:29.457: INFO: Ensuring a pod for ReplicaSet "my-hostname-basic-6678a5e8-6a7b-11e9-b6ef-0e71f8c7e888" is running
+Apr 29 12:36:29.461: INFO: Pod "my-hostname-basic-6678a5e8-6a7b-11e9-b6ef-0e71f8c7e888-k7vfn" is running (conditions: [{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2019-04-29 12:36:24 +0000 UTC Reason: Message:} {Type:Ready Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2019-04-29 12:36:26 +0000 UTC Reason: Message:} {Type:ContainersReady Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2019-04-29 12:36:26 +0000 UTC Reason: Message:} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2019-04-29 12:36:24 +0000 UTC Reason: Message:}])
+Apr 29 12:36:29.461: INFO: Trying to dial the pod
+Apr 29 12:36:34.482: INFO: Controller my-hostname-basic-6678a5e8-6a7b-11e9-b6ef-0e71f8c7e888: Got expected result from replica 1 [my-hostname-basic-6678a5e8-6a7b-11e9-b6ef-0e71f8c7e888-k7vfn]: "my-hostname-basic-6678a5e8-6a7b-11e9-b6ef-0e71f8c7e888-k7vfn", 1 of 1 required successes so far
+[AfterEach] [sig-apps] ReplicaSet
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 12:36:34.482: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "replicaset-5676" for this suite.
+Apr 29 12:36:42.512: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 12:36:42.722: INFO: namespace replicaset-5676 deletion completed in 8.231614925s
+
+• [SLOW TEST:18.506 seconds]
+[sig-apps] ReplicaSet
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22
+  should serve a basic image on each replica with a public image  [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSSSSSSSSSSSS
+------------------------------
+[k8s.io] Container Lifecycle Hook when create a pod with lifecycle hook 
+  should execute prestop exec hook properly [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [k8s.io] Container Lifecycle Hook
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 12:36:42.728: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename container-lifecycle-hook
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in container-lifecycle-hook-4301
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] when create a pod with lifecycle hook
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:61
+STEP: create the container to handle the HTTPGet hook request.
+[It] should execute prestop exec hook properly [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: create the pod with lifecycle hook
+STEP: delete the pod with lifecycle hook
+Apr 29 12:36:51.047: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear
+Apr 29 12:36:51.051: INFO: Pod pod-with-prestop-exec-hook still exists
+Apr 29 12:36:53.051: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear
+Apr 29 12:36:53.058: INFO: Pod pod-with-prestop-exec-hook still exists
+Apr 29 12:36:55.051: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear
+Apr 29 12:36:55.057: INFO: Pod pod-with-prestop-exec-hook still exists
+Apr 29 12:36:57.051: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear
+Apr 29 12:36:57.058: INFO: Pod pod-with-prestop-exec-hook still exists
+Apr 29 12:36:59.051: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear
+Apr 29 12:36:59.057: INFO: Pod pod-with-prestop-exec-hook still exists
+Apr 29 12:37:01.056: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear
+Apr 29 12:37:01.067: INFO: Pod pod-with-prestop-exec-hook still exists
+Apr 29 12:37:03.051: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear
+Apr 29 12:37:03.060: INFO: Pod pod-with-prestop-exec-hook still exists
+Apr 29 12:37:05.051: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear
+Apr 29 12:37:05.058: INFO: Pod pod-with-prestop-exec-hook still exists
+Apr 29 12:37:07.051: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear
+Apr 29 12:37:07.062: INFO: Pod pod-with-prestop-exec-hook still exists
+Apr 29 12:37:09.051: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear
+Apr 29 12:37:09.058: INFO: Pod pod-with-prestop-exec-hook no longer exists
+STEP: check prestop hook
+[AfterEach] [k8s.io] Container Lifecycle Hook
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 12:37:09.071: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "container-lifecycle-hook-4301" for this suite.
+Apr 29 12:37:33.098: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 12:37:33.332: INFO: namespace container-lifecycle-hook-4301 deletion completed in 24.253930945s
+
+• [SLOW TEST:50.606 seconds]
+[k8s.io] Container Lifecycle Hook
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
+  when create a pod with lifecycle hook
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:40
+    should execute prestop exec hook properly [NodeConformance] [Conformance]
+    /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSSSSSSSSSSSSSSSSSSSSSSS
+------------------------------
+[sig-api-machinery] Watchers 
+  should observe add, update, and delete watch notifications on configmaps [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-api-machinery] Watchers
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 12:37:33.344: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename watch
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in watch-5821
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should observe add, update, and delete watch notifications on configmaps [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: creating a watch on configmaps with label A
+STEP: creating a watch on configmaps with label B
+STEP: creating a watch on configmaps with label A or B
+STEP: creating a configmap with label A and ensuring the correct watchers observe the notification
+Apr 29 12:37:33.593: INFO: Got : ADDED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-configmap-a,GenerateName:,Namespace:watch-5821,SelfLink:/api/v1/namespaces/watch-5821/configmaps/e2e-watch-test-configmap-a,UID:8f801a22-6a7b-11e9-92a7-deadbef3c36d,ResourceVersion:19150,Generation:0,CreationTimestamp:2019-04-29 12:37:33 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: multiple-watchers-A,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{},BinaryData:map[string][]byte{},}
+Apr 29 12:37:33.594: INFO: Got : ADDED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-configmap-a,GenerateName:,Namespace:watch-5821,SelfLink:/api/v1/namespaces/watch-5821/configmaps/e2e-watch-test-configmap-a,UID:8f801a22-6a7b-11e9-92a7-deadbef3c36d,ResourceVersion:19150,Generation:0,CreationTimestamp:2019-04-29 12:37:33 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: multiple-watchers-A,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{},BinaryData:map[string][]byte{},}
+STEP: modifying configmap A and ensuring the correct watchers observe the notification
+Apr 29 12:37:43.614: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-configmap-a,GenerateName:,Namespace:watch-5821,SelfLink:/api/v1/namespaces/watch-5821/configmaps/e2e-watch-test-configmap-a,UID:8f801a22-6a7b-11e9-92a7-deadbef3c36d,ResourceVersion:19167,Generation:0,CreationTimestamp:2019-04-29 12:37:33 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: multiple-watchers-A,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{mutation: 1,},BinaryData:map[string][]byte{},}
+Apr 29 12:37:43.614: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-configmap-a,GenerateName:,Namespace:watch-5821,SelfLink:/api/v1/namespaces/watch-5821/configmaps/e2e-watch-test-configmap-a,UID:8f801a22-6a7b-11e9-92a7-deadbef3c36d,ResourceVersion:19167,Generation:0,CreationTimestamp:2019-04-29 12:37:33 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: multiple-watchers-A,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{mutation: 1,},BinaryData:map[string][]byte{},}
+STEP: modifying configmap A again and ensuring the correct watchers observe the notification
+Apr 29 12:37:53.632: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-configmap-a,GenerateName:,Namespace:watch-5821,SelfLink:/api/v1/namespaces/watch-5821/configmaps/e2e-watch-test-configmap-a,UID:8f801a22-6a7b-11e9-92a7-deadbef3c36d,ResourceVersion:19186,Generation:0,CreationTimestamp:2019-04-29 12:37:33 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: multiple-watchers-A,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},}
+Apr 29 12:37:53.633: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-configmap-a,GenerateName:,Namespace:watch-5821,SelfLink:/api/v1/namespaces/watch-5821/configmaps/e2e-watch-test-configmap-a,UID:8f801a22-6a7b-11e9-92a7-deadbef3c36d,ResourceVersion:19186,Generation:0,CreationTimestamp:2019-04-29 12:37:33 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: multiple-watchers-A,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},}
+STEP: deleting configmap A and ensuring the correct watchers observe the notification
+Apr 29 12:38:03.645: INFO: Got : DELETED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-configmap-a,GenerateName:,Namespace:watch-5821,SelfLink:/api/v1/namespaces/watch-5821/configmaps/e2e-watch-test-configmap-a,UID:8f801a22-6a7b-11e9-92a7-deadbef3c36d,ResourceVersion:19203,Generation:0,CreationTimestamp:2019-04-29 12:37:33 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: multiple-watchers-A,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},}
+Apr 29 12:38:03.645: INFO: Got : DELETED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-configmap-a,GenerateName:,Namespace:watch-5821,SelfLink:/api/v1/namespaces/watch-5821/configmaps/e2e-watch-test-configmap-a,UID:8f801a22-6a7b-11e9-92a7-deadbef3c36d,ResourceVersion:19203,Generation:0,CreationTimestamp:2019-04-29 12:37:33 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: multiple-watchers-A,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},}
+STEP: creating a configmap with label B and ensuring the correct watchers observe the notification
+Apr 29 12:38:13.670: INFO: Got : ADDED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-configmap-b,GenerateName:,Namespace:watch-5821,SelfLink:/api/v1/namespaces/watch-5821/configmaps/e2e-watch-test-configmap-b,UID:a76203d2-6a7b-11e9-92a7-deadbef3c36d,ResourceVersion:19221,Generation:0,CreationTimestamp:2019-04-29 12:38:13 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: multiple-watchers-B,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{},BinaryData:map[string][]byte{},}
+Apr 29 12:38:13.671: INFO: Got : ADDED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-configmap-b,GenerateName:,Namespace:watch-5821,SelfLink:/api/v1/namespaces/watch-5821/configmaps/e2e-watch-test-configmap-b,UID:a76203d2-6a7b-11e9-92a7-deadbef3c36d,ResourceVersion:19221,Generation:0,CreationTimestamp:2019-04-29 12:38:13 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: multiple-watchers-B,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{},BinaryData:map[string][]byte{},}
+STEP: deleting configmap B and ensuring the correct watchers observe the notification
+Apr 29 12:38:23.683: INFO: Got : DELETED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-configmap-b,GenerateName:,Namespace:watch-5821,SelfLink:/api/v1/namespaces/watch-5821/configmaps/e2e-watch-test-configmap-b,UID:a76203d2-6a7b-11e9-92a7-deadbef3c36d,ResourceVersion:19240,Generation:0,CreationTimestamp:2019-04-29 12:38:13 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: multiple-watchers-B,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{},BinaryData:map[string][]byte{},}
+Apr 29 12:38:23.683: INFO: Got : DELETED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-configmap-b,GenerateName:,Namespace:watch-5821,SelfLink:/api/v1/namespaces/watch-5821/configmaps/e2e-watch-test-configmap-b,UID:a76203d2-6a7b-11e9-92a7-deadbef3c36d,ResourceVersion:19240,Generation:0,CreationTimestamp:2019-04-29 12:38:13 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: multiple-watchers-B,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{},BinaryData:map[string][]byte{},}
+[AfterEach] [sig-api-machinery] Watchers
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 12:38:33.686: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "watch-5821" for this suite.
+Apr 29 12:38:39.713: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 12:38:39.967: INFO: namespace watch-5821 deletion completed in 6.273069757s
+
+• [SLOW TEST:66.623 seconds]
+[sig-api-machinery] Watchers
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:22
+  should observe add, update, and delete watch notifications on configmaps [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+[sig-storage] Downward API volume 
+  should provide container's memory limit [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-storage] Downward API volume
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 12:38:39.968: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename downward-api
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in downward-api-2139
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [sig-storage] Downward API volume
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:39
+[It] should provide container's memory limit [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating a pod to test downward API volume plugin
+Apr 29 12:38:40.237: INFO: Waiting up to 5m0s for pod "downwardapi-volume-b76912e6-6a7b-11e9-b6ef-0e71f8c7e888" in namespace "downward-api-2139" to be "success or failure"
+Apr 29 12:38:40.265: INFO: Pod "downwardapi-volume-b76912e6-6a7b-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 28.115423ms
+Apr 29 12:38:42.274: INFO: Pod "downwardapi-volume-b76912e6-6a7b-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 2.036489404s
+Apr 29 12:38:44.282: INFO: Pod "downwardapi-volume-b76912e6-6a7b-11e9-b6ef-0e71f8c7e888": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.044969915s
+STEP: Saw pod success
+Apr 29 12:38:44.282: INFO: Pod "downwardapi-volume-b76912e6-6a7b-11e9-b6ef-0e71f8c7e888" satisfied condition "success or failure"
+Apr 29 12:38:44.288: INFO: Trying to get logs from node worker-4av2h-6665d7b5ff-9nc42 pod downwardapi-volume-b76912e6-6a7b-11e9-b6ef-0e71f8c7e888 container client-container: 
+STEP: delete the pod
+Apr 29 12:38:44.388: INFO: Waiting for pod downwardapi-volume-b76912e6-6a7b-11e9-b6ef-0e71f8c7e888 to disappear
+Apr 29 12:38:44.392: INFO: Pod downwardapi-volume-b76912e6-6a7b-11e9-b6ef-0e71f8c7e888 no longer exists
+[AfterEach] [sig-storage] Downward API volume
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 12:38:44.392: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "downward-api-2139" for this suite.
+Apr 29 12:38:50.424: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 12:38:50.658: INFO: namespace downward-api-2139 deletion completed in 6.255760592s
+
+• [SLOW TEST:10.691 seconds]
+[sig-storage] Downward API volume
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:34
+  should provide container's memory limit [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSSSSSSS
+------------------------------
+[k8s.io] Pods 
+  should allow activeDeadlineSeconds to be updated [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [k8s.io] Pods
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 12:38:50.663: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename pods
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in pods-6961
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [k8s.io] Pods
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:135
+[It] should allow activeDeadlineSeconds to be updated [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: creating the pod
+STEP: submitting the pod to kubernetes
+STEP: verifying the pod is in kubernetes
+STEP: updating the pod
+Apr 29 12:38:55.433: INFO: Successfully updated pod "pod-update-activedeadlineseconds-bdc28f42-6a7b-11e9-b6ef-0e71f8c7e888"
+Apr 29 12:38:55.433: INFO: Waiting up to 5m0s for pod "pod-update-activedeadlineseconds-bdc28f42-6a7b-11e9-b6ef-0e71f8c7e888" in namespace "pods-6961" to be "terminated due to deadline exceeded"
+Apr 29 12:38:55.439: INFO: Pod "pod-update-activedeadlineseconds-bdc28f42-6a7b-11e9-b6ef-0e71f8c7e888": Phase="Running", Reason="", readiness=true. Elapsed: 5.888493ms
+Apr 29 12:38:57.447: INFO: Pod "pod-update-activedeadlineseconds-bdc28f42-6a7b-11e9-b6ef-0e71f8c7e888": Phase="Failed", Reason="DeadlineExceeded", readiness=false. Elapsed: 2.01396065s
+Apr 29 12:38:57.447: INFO: Pod "pod-update-activedeadlineseconds-bdc28f42-6a7b-11e9-b6ef-0e71f8c7e888" satisfied condition "terminated due to deadline exceeded"
+[AfterEach] [k8s.io] Pods
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 12:38:57.447: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "pods-6961" for this suite.
+Apr 29 12:39:03.487: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 12:39:03.717: INFO: namespace pods-6961 deletion completed in 6.259663883s
+
+• [SLOW TEST:13.054 seconds]
+[k8s.io] Pods
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
+  should allow activeDeadlineSeconds to be updated [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSS
+------------------------------
+[sig-apps] Deployment 
+  RollingUpdateDeployment should delete old pods and create new ones [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-apps] Deployment
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 12:39:03.719: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename deployment
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in deployment-2814
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [sig-apps] Deployment
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:65
+[It] RollingUpdateDeployment should delete old pods and create new ones [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+Apr 29 12:39:03.918: INFO: Creating replica set "test-rolling-update-controller" (going to be adopted)
+Apr 29 12:39:03.955: INFO: Pod name sample-pod: Found 0 pods out of 1
+Apr 29 12:39:08.962: INFO: Pod name sample-pod: Found 1 pods out of 1
+STEP: ensuring each pod is running
+Apr 29 12:39:08.962: INFO: Creating deployment "test-rolling-update-deployment"
+Apr 29 12:39:08.973: INFO: Ensuring deployment "test-rolling-update-deployment" gets the next revision from the one the adopted replica set "test-rolling-update-controller" has
+Apr 29 12:39:08.987: INFO: new replicaset for deployment "test-rolling-update-deployment" is yet to be created
+Apr 29 12:39:10.998: INFO: Ensuring status for deployment "test-rolling-update-deployment" is the expected
+Apr 29 12:39:11.002: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:2, UpdatedReplicas:1, ReadyReplicas:1, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63692138348, loc:(*time.Location)(0x8a060e0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63692138348, loc:(*time.Location)(0x8a060e0)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63692138348, loc:(*time.Location)(0x8a060e0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63692138348, loc:(*time.Location)(0x8a060e0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rolling-update-deployment-67599b4d9\" is progressing."}}, CollisionCount:(*int32)(nil)}
+Apr 29 12:39:13.008: INFO: Ensuring deployment "test-rolling-update-deployment" has one old replica set (the one it adopted)
+[AfterEach] [sig-apps] Deployment
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:59
+Apr 29 12:39:13.024: INFO: Deployment "test-rolling-update-deployment":
+&Deployment{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-rolling-update-deployment,GenerateName:,Namespace:deployment-2814,SelfLink:/apis/apps/v1/namespaces/deployment-2814/deployments/test-rolling-update-deployment,UID:c85a4d20-6a7b-11e9-92a7-deadbef3c36d,ResourceVersion:19465,Generation:1,CreationTimestamp:2019-04-29 12:39:08 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod,},Annotations:map[string]string{deployment.kubernetes.io/revision: 3546343826724305833,},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:DeploymentSpec{Replicas:*1,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[],Containers:[{redis gcr.io/kubernetes-e2e-test-images/redis:1.0 [] []  [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,},},Strategy:DeploymentStrategy{Type:RollingUpdate,RollingUpdate:&RollingUpdateDeployment{MaxUnavailable:25%!,(MISSING)MaxSurge:25%!,(MISSING)},},MinReadySeconds:0,RevisionHistoryLimit:*10,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:1,Replicas:1,UpdatedReplicas:1,AvailableReplicas:1,UnavailableReplicas:0,Conditions:[{Available True 2019-04-29 12:39:08 +0000 UTC 2019-04-29 12:39:08 +0000 UTC MinimumReplicasAvailable Deployment has minimum availability.} {Progressing True 2019-04-29 12:39:12 +0000 UTC 2019-04-29 12:39:08 +0000 UTC NewReplicaSetAvailable ReplicaSet "test-rolling-update-deployment-67599b4d9" has successfully progressed.}],ReadyReplicas:1,CollisionCount:nil,},}
+
+Apr 29 12:39:13.028: INFO: New ReplicaSet "test-rolling-update-deployment-67599b4d9" of Deployment "test-rolling-update-deployment":
+&ReplicaSet{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-rolling-update-deployment-67599b4d9,GenerateName:,Namespace:deployment-2814,SelfLink:/apis/apps/v1/namespaces/deployment-2814/replicasets/test-rolling-update-deployment-67599b4d9,UID:c85ee068-6a7b-11e9-92a7-deadbef3c36d,ResourceVersion:19454,Generation:1,CreationTimestamp:2019-04-29 12:39:08 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod,pod-template-hash: 67599b4d9,},Annotations:map[string]string{deployment.kubernetes.io/desired-replicas: 1,deployment.kubernetes.io/max-replicas: 2,deployment.kubernetes.io/revision: 3546343826724305833,},OwnerReferences:[{apps/v1 Deployment test-rolling-update-deployment c85a4d20-6a7b-11e9-92a7-deadbef3c36d 0xc002471b10 0xc002471b11}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:ReplicaSetSpec{Replicas:*1,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod,pod-template-hash: 67599b4d9,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod,pod-template-hash: 67599b4d9,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[],Containers:[{redis gcr.io/kubernetes-e2e-test-images/redis:1.0 [] []  [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,},},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:1,FullyLabeledReplicas:1,ObservedGeneration:1,ReadyReplicas:1,AvailableReplicas:1,Conditions:[],},}
+Apr 29 12:39:13.029: INFO: All old ReplicaSets of Deployment "test-rolling-update-deployment":
+Apr 29 12:39:13.029: INFO: &ReplicaSet{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-rolling-update-controller,GenerateName:,Namespace:deployment-2814,SelfLink:/apis/apps/v1/namespaces/deployment-2814/replicasets/test-rolling-update-controller,UID:c5594f03-6a7b-11e9-92a7-deadbef3c36d,ResourceVersion:19464,Generation:2,CreationTimestamp:2019-04-29 12:39:03 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod,pod: nginx,},Annotations:map[string]string{deployment.kubernetes.io/desired-replicas: 1,deployment.kubernetes.io/max-replicas: 2,deployment.kubernetes.io/revision: 3546343826724305832,},OwnerReferences:[{apps/v1 Deployment test-rolling-update-deployment c85a4d20-6a7b-11e9-92a7-deadbef3c36d 0xc002471a47 0xc002471a48}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:ReplicaSetSpec{Replicas:*0,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod,pod: nginx,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod,pod: nginx,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] []  [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,},},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:0,FullyLabeledReplicas:0,ObservedGeneration:2,ReadyReplicas:0,AvailableReplicas:0,Conditions:[],},}
+Apr 29 12:39:13.034: INFO: Pod "test-rolling-update-deployment-67599b4d9-ptmdg" is available:
+&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-rolling-update-deployment-67599b4d9-ptmdg,GenerateName:test-rolling-update-deployment-67599b4d9-,Namespace:deployment-2814,SelfLink:/api/v1/namespaces/deployment-2814/pods/test-rolling-update-deployment-67599b4d9-ptmdg,UID:c8610e45-6a7b-11e9-92a7-deadbef3c36d,ResourceVersion:19453,Generation:0,CreationTimestamp:2019-04-29 12:39:08 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod,pod-template-hash: 67599b4d9,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet test-rolling-update-deployment-67599b4d9 c85ee068-6a7b-11e9-92a7-deadbef3c36d 0xc002d729f0 0xc002d729f1}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-n74cc {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-n74cc,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{redis gcr.io/kubernetes-e2e-test-images/redis:1.0 [] []  [] [] [] {map[] map[]} [{default-token-n74cc true /var/run/secrets/kubernetes.io/serviceaccount   }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:worker-4av2h-6665d7b5ff-9nc42,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists  NoExecute 0xc002d72a50} {node.kubernetes.io/unreachable Exists  NoExecute 0xc002d72a70}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Running,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 12:39:09 +0000 UTC  } {Ready True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 12:39:12 +0000 UTC  } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 12:39:12 +0000 UTC  } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 12:39:08 +0000 UTC  }],Message:,Reason:,HostIP:172.23.0.238,PodIP:192.168.205.116,StartTime:2019-04-29 12:39:09 +0000 UTC,ContainerStatuses:[{redis {nil ContainerStateRunning{StartedAt:2019-04-29 12:39:11 +0000 UTC,} nil} {nil nil nil} true 0 gcr.io/kubernetes-e2e-test-images/redis:1.0 docker-pullable://gcr.io/kubernetes-e2e-test-images/redis@sha256:af4748d1655c08dc54d4be5182135395db9ce87aba2d4699b26b14ae197c5830 docker://77b0940507b6ca6926597ca4f3cac3dabe3aca58518124a756c949108f03b7ef}],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},}
+[AfterEach] [sig-apps] Deployment
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 12:39:13.034: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "deployment-2814" for this suite.
+Apr 29 12:39:21.063: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 12:39:21.328: INFO: namespace deployment-2814 deletion completed in 8.28712094s
+
+• [SLOW TEST:17.608 seconds]
+[sig-apps] Deployment
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22
+  RollingUpdateDeployment should delete old pods and create new ones [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSSSSSSSSSSSSSSSSSSS
+------------------------------
+[sig-storage] ConfigMap 
+  should be consumable from pods in volume [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-storage] ConfigMap
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 12:39:21.332: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename configmap
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in configmap-5362
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should be consumable from pods in volume [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating configMap with name configmap-test-volume-d00b0091-6a7b-11e9-b6ef-0e71f8c7e888
+STEP: Creating a pod to test consume configMaps
+Apr 29 12:39:21.568: INFO: Waiting up to 5m0s for pod "pod-configmaps-d00cb8b5-6a7b-11e9-b6ef-0e71f8c7e888" in namespace "configmap-5362" to be "success or failure"
+Apr 29 12:39:21.579: INFO: Pod "pod-configmaps-d00cb8b5-6a7b-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 11.751393ms
+Apr 29 12:39:23.585: INFO: Pod "pod-configmaps-d00cb8b5-6a7b-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 2.017565626s
+Apr 29 12:39:25.594: INFO: Pod "pod-configmaps-d00cb8b5-6a7b-11e9-b6ef-0e71f8c7e888": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.026011805s
+STEP: Saw pod success
+Apr 29 12:39:25.594: INFO: Pod "pod-configmaps-d00cb8b5-6a7b-11e9-b6ef-0e71f8c7e888" satisfied condition "success or failure"
+Apr 29 12:39:25.599: INFO: Trying to get logs from node worker-4av2h-6665d7b5ff-9nc42 pod pod-configmaps-d00cb8b5-6a7b-11e9-b6ef-0e71f8c7e888 container configmap-volume-test: 
+STEP: delete the pod
+Apr 29 12:39:25.638: INFO: Waiting for pod pod-configmaps-d00cb8b5-6a7b-11e9-b6ef-0e71f8c7e888 to disappear
+Apr 29 12:39:25.645: INFO: Pod pod-configmaps-d00cb8b5-6a7b-11e9-b6ef-0e71f8c7e888 no longer exists
+[AfterEach] [sig-storage] ConfigMap
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 12:39:25.645: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "configmap-5362" for this suite.
+Apr 29 12:39:31.672: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 12:39:31.863: INFO: namespace configmap-5362 deletion completed in 6.208280009s
+
+• [SLOW TEST:10.531 seconds]
+[sig-storage] ConfigMap
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:32
+  should be consumable from pods in volume [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSSSSSSSSSSSSSSSS
+------------------------------
+[sig-storage] Projected downwardAPI 
+  should provide container's memory limit [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-storage] Projected downwardAPI
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 12:39:31.866: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename projected
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-3736
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [sig-storage] Projected downwardAPI
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:39
+[It] should provide container's memory limit [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating a pod to test downward API volume plugin
+Apr 29 12:39:32.170: INFO: Waiting up to 5m0s for pod "downwardapi-volume-d6554f45-6a7b-11e9-b6ef-0e71f8c7e888" in namespace "projected-3736" to be "success or failure"
+Apr 29 12:39:32.194: INFO: Pod "downwardapi-volume-d6554f45-6a7b-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 23.156445ms
+Apr 29 12:39:34.200: INFO: Pod "downwardapi-volume-d6554f45-6a7b-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 2.029775046s
+Apr 29 12:39:36.210: INFO: Pod "downwardapi-volume-d6554f45-6a7b-11e9-b6ef-0e71f8c7e888": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.039676344s
+STEP: Saw pod success
+Apr 29 12:39:36.211: INFO: Pod "downwardapi-volume-d6554f45-6a7b-11e9-b6ef-0e71f8c7e888" satisfied condition "success or failure"
+Apr 29 12:39:36.217: INFO: Trying to get logs from node worker-4av2h-6665d7b5ff-9nc42 pod downwardapi-volume-d6554f45-6a7b-11e9-b6ef-0e71f8c7e888 container client-container: 
+STEP: delete the pod
+Apr 29 12:39:36.248: INFO: Waiting for pod downwardapi-volume-d6554f45-6a7b-11e9-b6ef-0e71f8c7e888 to disappear
+Apr 29 12:39:36.253: INFO: Pod downwardapi-volume-d6554f45-6a7b-11e9-b6ef-0e71f8c7e888 no longer exists
+[AfterEach] [sig-storage] Projected downwardAPI
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 12:39:36.253: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "projected-3736" for this suite.
+Apr 29 12:39:42.276: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 12:39:42.470: INFO: namespace projected-3736 deletion completed in 6.211422568s
+
+• [SLOW TEST:10.604 seconds]
+[sig-storage] Projected downwardAPI
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:33
+  should provide container's memory limit [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSS
+------------------------------
+[sig-storage] Secrets 
+  should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-storage] Secrets
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 12:39:42.472: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename secrets
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in secrets-4789
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating secret with name secret-test-dca16734-6a7b-11e9-b6ef-0e71f8c7e888
+STEP: Creating a pod to test consume secrets
+Apr 29 12:39:42.681: INFO: Waiting up to 5m0s for pod "pod-secrets-dca28798-6a7b-11e9-b6ef-0e71f8c7e888" in namespace "secrets-4789" to be "success or failure"
+Apr 29 12:39:42.694: INFO: Pod "pod-secrets-dca28798-6a7b-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 12.729677ms
+Apr 29 12:39:44.701: INFO: Pod "pod-secrets-dca28798-6a7b-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 2.020110757s
+Apr 29 12:39:46.715: INFO: Pod "pod-secrets-dca28798-6a7b-11e9-b6ef-0e71f8c7e888": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.033322267s
+STEP: Saw pod success
+Apr 29 12:39:46.715: INFO: Pod "pod-secrets-dca28798-6a7b-11e9-b6ef-0e71f8c7e888" satisfied condition "success or failure"
+Apr 29 12:39:46.721: INFO: Trying to get logs from node worker-4av2h-6665d7b5ff-9nc42 pod pod-secrets-dca28798-6a7b-11e9-b6ef-0e71f8c7e888 container secret-volume-test: 
+STEP: delete the pod
+Apr 29 12:39:46.768: INFO: Waiting for pod pod-secrets-dca28798-6a7b-11e9-b6ef-0e71f8c7e888 to disappear
+Apr 29 12:39:46.773: INFO: Pod pod-secrets-dca28798-6a7b-11e9-b6ef-0e71f8c7e888 no longer exists
+[AfterEach] [sig-storage] Secrets
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 12:39:46.773: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "secrets-4789" for this suite.
+Apr 29 12:39:52.806: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 12:39:53.041: INFO: namespace secrets-4789 deletion completed in 6.261942197s
+
+• [SLOW TEST:10.570 seconds]
+[sig-storage] Secrets
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/secrets_volume.go:33
+  should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSSSSSSSSSSSSSSSSSSS
+------------------------------
+[sig-storage] EmptyDir volumes 
+  volume on default medium should have the correct mode [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-storage] EmptyDir volumes
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 12:39:53.045: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename emptydir
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in emptydir-2833
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] volume on default medium should have the correct mode [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating a pod to test emptydir volume type on node default medium
+Apr 29 12:39:53.293: INFO: Waiting up to 5m0s for pod "pod-e2f3ed16-6a7b-11e9-b6ef-0e71f8c7e888" in namespace "emptydir-2833" to be "success or failure"
+Apr 29 12:39:53.298: INFO: Pod "pod-e2f3ed16-6a7b-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 5.070637ms
+Apr 29 12:39:55.306: INFO: Pod "pod-e2f3ed16-6a7b-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 2.013401175s
+Apr 29 12:39:57.313: INFO: Pod "pod-e2f3ed16-6a7b-11e9-b6ef-0e71f8c7e888": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.019461758s
+STEP: Saw pod success
+Apr 29 12:39:57.313: INFO: Pod "pod-e2f3ed16-6a7b-11e9-b6ef-0e71f8c7e888" satisfied condition "success or failure"
+Apr 29 12:39:57.317: INFO: Trying to get logs from node worker-4av2h-6665d7b5ff-9nc42 pod pod-e2f3ed16-6a7b-11e9-b6ef-0e71f8c7e888 container test-container: 
+STEP: delete the pod
+Apr 29 12:39:57.403: INFO: Waiting for pod pod-e2f3ed16-6a7b-11e9-b6ef-0e71f8c7e888 to disappear
+Apr 29 12:39:57.411: INFO: Pod pod-e2f3ed16-6a7b-11e9-b6ef-0e71f8c7e888 no longer exists
+[AfterEach] [sig-storage] EmptyDir volumes
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 12:39:57.411: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "emptydir-2833" for this suite.
+Apr 29 12:40:03.436: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 12:40:03.665: INFO: namespace emptydir-2833 deletion completed in 6.247860586s
+
+• [SLOW TEST:10.622 seconds]
+[sig-storage] EmptyDir volumes
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:41
+  volume on default medium should have the correct mode [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SS
+------------------------------
+[sig-node] Downward API 
+  should provide host IP as an env var [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-node] Downward API
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 12:40:03.668: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename downward-api
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in downward-api-8850
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should provide host IP as an env var [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating a pod to test downward api env vars
+Apr 29 12:40:03.949: INFO: Waiting up to 5m0s for pod "downward-api-e94df221-6a7b-11e9-b6ef-0e71f8c7e888" in namespace "downward-api-8850" to be "success or failure"
+Apr 29 12:40:03.969: INFO: Pod "downward-api-e94df221-6a7b-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 19.59834ms
+Apr 29 12:40:05.978: INFO: Pod "downward-api-e94df221-6a7b-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 2.028399681s
+Apr 29 12:40:07.986: INFO: Pod "downward-api-e94df221-6a7b-11e9-b6ef-0e71f8c7e888": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.036906211s
+STEP: Saw pod success
+Apr 29 12:40:07.986: INFO: Pod "downward-api-e94df221-6a7b-11e9-b6ef-0e71f8c7e888" satisfied condition "success or failure"
+Apr 29 12:40:07.991: INFO: Trying to get logs from node worker-4av2h-6665d7b5ff-9nc42 pod downward-api-e94df221-6a7b-11e9-b6ef-0e71f8c7e888 container dapi-container: 
+STEP: delete the pod
+Apr 29 12:40:08.033: INFO: Waiting for pod downward-api-e94df221-6a7b-11e9-b6ef-0e71f8c7e888 to disappear
+Apr 29 12:40:08.040: INFO: Pod downward-api-e94df221-6a7b-11e9-b6ef-0e71f8c7e888 no longer exists
+[AfterEach] [sig-node] Downward API
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 12:40:08.041: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "downward-api-8850" for this suite.
+Apr 29 12:40:14.083: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 12:40:14.311: INFO: namespace downward-api-8850 deletion completed in 6.255369299s
+
+• [SLOW TEST:10.643 seconds]
+[sig-node] Downward API
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downward_api.go:38
+  should provide host IP as an env var [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSSSSSSSSSSSSSSSSSSSSS
+------------------------------
+[sig-storage] EmptyDir volumes 
+  should support (root,0777,default) [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-storage] EmptyDir volumes
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 12:40:14.316: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename emptydir
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in emptydir-9029
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should support (root,0777,default) [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating a pod to test emptydir 0777 on node default medium
+Apr 29 12:40:14.534: INFO: Waiting up to 5m0s for pod "pod-ef9e0d57-6a7b-11e9-b6ef-0e71f8c7e888" in namespace "emptydir-9029" to be "success or failure"
+Apr 29 12:40:14.556: INFO: Pod "pod-ef9e0d57-6a7b-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 21.207839ms
+Apr 29 12:40:16.564: INFO: Pod "pod-ef9e0d57-6a7b-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 2.029926093s
+Apr 29 12:40:18.572: INFO: Pod "pod-ef9e0d57-6a7b-11e9-b6ef-0e71f8c7e888": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.037515628s
+STEP: Saw pod success
+Apr 29 12:40:18.572: INFO: Pod "pod-ef9e0d57-6a7b-11e9-b6ef-0e71f8c7e888" satisfied condition "success or failure"
+Apr 29 12:40:18.576: INFO: Trying to get logs from node worker-4av2h-6665d7b5ff-9nc42 pod pod-ef9e0d57-6a7b-11e9-b6ef-0e71f8c7e888 container test-container: 
+STEP: delete the pod
+Apr 29 12:40:18.610: INFO: Waiting for pod pod-ef9e0d57-6a7b-11e9-b6ef-0e71f8c7e888 to disappear
+Apr 29 12:40:18.615: INFO: Pod pod-ef9e0d57-6a7b-11e9-b6ef-0e71f8c7e888 no longer exists
+[AfterEach] [sig-storage] EmptyDir volumes
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 12:40:18.615: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "emptydir-9029" for this suite.
+Apr 29 12:40:24.693: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 12:40:24.919: INFO: namespace emptydir-9029 deletion completed in 6.296192692s
+
+• [SLOW TEST:10.603 seconds]
+[sig-storage] EmptyDir volumes
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:41
+  should support (root,0777,default) [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSS
+------------------------------
+[sig-node] ConfigMap 
+  should be consumable via environment variable [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-node] ConfigMap
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 12:40:24.922: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename configmap
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in configmap-38
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should be consumable via environment variable [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating configMap configmap-38/configmap-test-f5ef6cd1-6a7b-11e9-b6ef-0e71f8c7e888
+STEP: Creating a pod to test consume configMaps
+Apr 29 12:40:25.148: INFO: Waiting up to 5m0s for pod "pod-configmaps-f5f1125e-6a7b-11e9-b6ef-0e71f8c7e888" in namespace "configmap-38" to be "success or failure"
+Apr 29 12:40:25.179: INFO: Pod "pod-configmaps-f5f1125e-6a7b-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 31.00966ms
+Apr 29 12:40:27.187: INFO: Pod "pod-configmaps-f5f1125e-6a7b-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 2.03926236s
+Apr 29 12:40:29.194: INFO: Pod "pod-configmaps-f5f1125e-6a7b-11e9-b6ef-0e71f8c7e888": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.046141261s
+STEP: Saw pod success
+Apr 29 12:40:29.194: INFO: Pod "pod-configmaps-f5f1125e-6a7b-11e9-b6ef-0e71f8c7e888" satisfied condition "success or failure"
+Apr 29 12:40:29.198: INFO: Trying to get logs from node worker-4av2h-6665d7b5ff-9nc42 pod pod-configmaps-f5f1125e-6a7b-11e9-b6ef-0e71f8c7e888 container env-test: 
+STEP: delete the pod
+Apr 29 12:40:29.235: INFO: Waiting for pod pod-configmaps-f5f1125e-6a7b-11e9-b6ef-0e71f8c7e888 to disappear
+Apr 29 12:40:29.238: INFO: Pod pod-configmaps-f5f1125e-6a7b-11e9-b6ef-0e71f8c7e888 no longer exists
+[AfterEach] [sig-node] ConfigMap
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 12:40:29.238: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "configmap-38" for this suite.
+Apr 29 12:40:35.264: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 12:40:35.453: INFO: namespace configmap-38 deletion completed in 6.20741448s
+
+• [SLOW TEST:10.532 seconds]
+[sig-node] ConfigMap
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/configmap.go:32
+  should be consumable via environment variable [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSS
+------------------------------
+[sig-storage] EmptyDir wrapper volumes 
+  should not cause race condition when used for configmaps [Serial] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-storage] EmptyDir wrapper volumes
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 12:40:35.462: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename emptydir-wrapper
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in emptydir-wrapper-7896
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should not cause race condition when used for configmaps [Serial] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating 50 configmaps
+STEP: Creating RC which spawns configmap-volume pods
+Apr 29 12:40:36.120: INFO: Pod name wrapped-volume-race-fc7ba28e-6a7b-11e9-b6ef-0e71f8c7e888: Found 0 pods out of 5
+Apr 29 12:40:41.137: INFO: Pod name wrapped-volume-race-fc7ba28e-6a7b-11e9-b6ef-0e71f8c7e888: Found 5 pods out of 5
+STEP: Ensuring each pod is running
+STEP: deleting ReplicationController wrapped-volume-race-fc7ba28e-6a7b-11e9-b6ef-0e71f8c7e888 in namespace emptydir-wrapper-7896, will wait for the garbage collector to delete the pods
+Apr 29 12:40:55.257: INFO: Deleting ReplicationController wrapped-volume-race-fc7ba28e-6a7b-11e9-b6ef-0e71f8c7e888 took: 10.810635ms
+Apr 29 12:40:55.458: INFO: Terminating ReplicationController wrapped-volume-race-fc7ba28e-6a7b-11e9-b6ef-0e71f8c7e888 pods took: 200.907931ms
+STEP: Creating RC which spawns configmap-volume pods
+Apr 29 12:41:45.893: INFO: Pod name wrapped-volume-race-26100f2b-6a7c-11e9-b6ef-0e71f8c7e888: Found 0 pods out of 5
+Apr 29 12:41:50.927: INFO: Pod name wrapped-volume-race-26100f2b-6a7c-11e9-b6ef-0e71f8c7e888: Found 5 pods out of 5
+STEP: Ensuring each pod is running
+STEP: deleting ReplicationController wrapped-volume-race-26100f2b-6a7c-11e9-b6ef-0e71f8c7e888 in namespace emptydir-wrapper-7896, will wait for the garbage collector to delete the pods
+Apr 29 12:42:05.072: INFO: Deleting ReplicationController wrapped-volume-race-26100f2b-6a7c-11e9-b6ef-0e71f8c7e888 took: 17.099245ms
+Apr 29 12:42:05.472: INFO: Terminating ReplicationController wrapped-volume-race-26100f2b-6a7c-11e9-b6ef-0e71f8c7e888 pods took: 400.46044ms
+STEP: Creating RC which spawns configmap-volume pods
+Apr 29 12:42:46.818: INFO: Pod name wrapped-volume-race-4a5eed60-6a7c-11e9-b6ef-0e71f8c7e888: Found 0 pods out of 5
+Apr 29 12:42:51.834: INFO: Pod name wrapped-volume-race-4a5eed60-6a7c-11e9-b6ef-0e71f8c7e888: Found 5 pods out of 5
+STEP: Ensuring each pod is running
+STEP: deleting ReplicationController wrapped-volume-race-4a5eed60-6a7c-11e9-b6ef-0e71f8c7e888 in namespace emptydir-wrapper-7896, will wait for the garbage collector to delete the pods
+Apr 29 12:43:08.024: INFO: Deleting ReplicationController wrapped-volume-race-4a5eed60-6a7c-11e9-b6ef-0e71f8c7e888 took: 14.677481ms
+Apr 29 12:43:08.225: INFO: Terminating ReplicationController wrapped-volume-race-4a5eed60-6a7c-11e9-b6ef-0e71f8c7e888 pods took: 200.664958ms
+STEP: Cleaning up the configMaps
+[AfterEach] [sig-storage] EmptyDir wrapper volumes
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 12:43:56.985: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "emptydir-wrapper-7896" for this suite.
+Apr 29 12:44:05.013: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 12:44:05.308: INFO: namespace emptydir-wrapper-7896 deletion completed in 8.314356865s
+
+• [SLOW TEST:209.847 seconds]
+[sig-storage] EmptyDir wrapper volumes
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:22
+  should not cause race condition when used for configmaps [Serial] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSSSSSSSSSSS
+------------------------------
+[sig-storage] Subpath Atomic writer volumes 
+  should support subpaths with downward pod [LinuxOnly] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-storage] Subpath
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 12:44:05.309: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename subpath
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in subpath-5345
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] Atomic writer volumes
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:38
+STEP: Setting up data
+[It] should support subpaths with downward pod [LinuxOnly] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating pod pod-subpath-test-downwardapi-pp9k
+STEP: Creating a pod to test atomic-volume-subpath
+Apr 29 12:44:05.519: INFO: Waiting up to 5m0s for pod "pod-subpath-test-downwardapi-pp9k" in namespace "subpath-5345" to be "success or failure"
+Apr 29 12:44:05.537: INFO: Pod "pod-subpath-test-downwardapi-pp9k": Phase="Pending", Reason="", readiness=false. Elapsed: 16.936538ms
+Apr 29 12:44:07.544: INFO: Pod "pod-subpath-test-downwardapi-pp9k": Phase="Pending", Reason="", readiness=false. Elapsed: 2.024281388s
+Apr 29 12:44:09.551: INFO: Pod "pod-subpath-test-downwardapi-pp9k": Phase="Running", Reason="", readiness=true. Elapsed: 4.031335204s
+Apr 29 12:44:11.560: INFO: Pod "pod-subpath-test-downwardapi-pp9k": Phase="Running", Reason="", readiness=true. Elapsed: 6.040532007s
+Apr 29 12:44:13.567: INFO: Pod "pod-subpath-test-downwardapi-pp9k": Phase="Running", Reason="", readiness=true. Elapsed: 8.047405387s
+Apr 29 12:44:15.574: INFO: Pod "pod-subpath-test-downwardapi-pp9k": Phase="Running", Reason="", readiness=true. Elapsed: 10.05444184s
+Apr 29 12:44:17.582: INFO: Pod "pod-subpath-test-downwardapi-pp9k": Phase="Running", Reason="", readiness=true. Elapsed: 12.062242202s
+Apr 29 12:44:19.590: INFO: Pod "pod-subpath-test-downwardapi-pp9k": Phase="Running", Reason="", readiness=true. Elapsed: 14.070210928s
+Apr 29 12:44:21.596: INFO: Pod "pod-subpath-test-downwardapi-pp9k": Phase="Running", Reason="", readiness=true. Elapsed: 16.076442705s
+Apr 29 12:44:23.604: INFO: Pod "pod-subpath-test-downwardapi-pp9k": Phase="Running", Reason="", readiness=true. Elapsed: 18.084270393s
+Apr 29 12:44:25.621: INFO: Pod "pod-subpath-test-downwardapi-pp9k": Phase="Running", Reason="", readiness=true. Elapsed: 20.10107865s
+Apr 29 12:44:27.629: INFO: Pod "pod-subpath-test-downwardapi-pp9k": Phase="Running", Reason="", readiness=true. Elapsed: 22.10872851s
+Apr 29 12:44:29.635: INFO: Pod "pod-subpath-test-downwardapi-pp9k": Phase="Succeeded", Reason="", readiness=false. Elapsed: 24.115471702s
+STEP: Saw pod success
+Apr 29 12:44:29.635: INFO: Pod "pod-subpath-test-downwardapi-pp9k" satisfied condition "success or failure"
+Apr 29 12:44:29.640: INFO: Trying to get logs from node worker-4av2h-6665d7b5ff-9nc42 pod pod-subpath-test-downwardapi-pp9k container test-container-subpath-downwardapi-pp9k: 
+STEP: delete the pod
+Apr 29 12:44:29.676: INFO: Waiting for pod pod-subpath-test-downwardapi-pp9k to disappear
+Apr 29 12:44:29.682: INFO: Pod pod-subpath-test-downwardapi-pp9k no longer exists
+STEP: Deleting pod pod-subpath-test-downwardapi-pp9k
+Apr 29 12:44:29.682: INFO: Deleting pod "pod-subpath-test-downwardapi-pp9k" in namespace "subpath-5345"
+[AfterEach] [sig-storage] Subpath
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 12:44:29.686: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "subpath-5345" for this suite.
+Apr 29 12:44:35.714: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 12:44:35.953: INFO: namespace subpath-5345 deletion completed in 6.259525067s
+
+• [SLOW TEST:30.644 seconds]
+[sig-storage] Subpath
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:22
+  Atomic writer volumes
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:34
+    should support subpaths with downward pod [LinuxOnly] [Conformance]
+    /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSSSSSSSSSSSSS
+------------------------------
+[sig-network] Service endpoints latency 
+  should not be very high  [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-network] Service endpoints latency
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 12:44:35.956: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename svc-latency
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in svc-latency-3029
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should not be very high  [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: creating replication controller svc-latency-rc in namespace svc-latency-3029
+I0429 12:44:36.169109      15 runners.go:184] Created replication controller with name: svc-latency-rc, namespace: svc-latency-3029, replica count: 1
+I0429 12:44:37.222388      15 runners.go:184] svc-latency-rc Pods: 1 out of 1 created, 0 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady 
+I0429 12:44:38.222684      15 runners.go:184] svc-latency-rc Pods: 1 out of 1 created, 0 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady 
+I0429 12:44:39.223074      15 runners.go:184] svc-latency-rc Pods: 1 out of 1 created, 1 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady 
+Apr 29 12:44:39.388: INFO: Created: latency-svc-gqnzq
+Apr 29 12:44:39.394: INFO: Got endpoints: latency-svc-gqnzq [71.115372ms]
+Apr 29 12:44:39.429: INFO: Created: latency-svc-6z8dt
+Apr 29 12:44:39.443: INFO: Created: latency-svc-dh9l7
+Apr 29 12:44:39.449: INFO: Got endpoints: latency-svc-6z8dt [54.032492ms]
+Apr 29 12:44:39.461: INFO: Created: latency-svc-gvpj4
+Apr 29 12:44:39.467: INFO: Got endpoints: latency-svc-dh9l7 [71.267878ms]
+Apr 29 12:44:39.493: INFO: Got endpoints: latency-svc-gvpj4 [97.391229ms]
+Apr 29 12:44:39.494: INFO: Created: latency-svc-64gkq
+Apr 29 12:44:39.496: INFO: Got endpoints: latency-svc-64gkq [99.709573ms]
+Apr 29 12:44:39.503: INFO: Created: latency-svc-8269l
+Apr 29 12:44:39.526: INFO: Created: latency-svc-wbgdq
+Apr 29 12:44:39.527: INFO: Got endpoints: latency-svc-8269l [131.399536ms]
+Apr 29 12:44:39.555: INFO: Got endpoints: latency-svc-wbgdq [88.083901ms]
+Apr 29 12:44:39.562: INFO: Created: latency-svc-755zs
+Apr 29 12:44:39.590: INFO: Got endpoints: latency-svc-755zs [193.754412ms]
+Apr 29 12:44:39.596: INFO: Created: latency-svc-cf6jt
+Apr 29 12:44:39.614: INFO: Got endpoints: latency-svc-cf6jt [217.038639ms]
+Apr 29 12:44:39.626: INFO: Created: latency-svc-j82v4
+Apr 29 12:44:39.632: INFO: Got endpoints: latency-svc-j82v4 [234.473087ms]
+Apr 29 12:44:39.636: INFO: Created: latency-svc-bzbpl
+Apr 29 12:44:39.646: INFO: Created: latency-svc-68jgv
+Apr 29 12:44:39.655: INFO: Got endpoints: latency-svc-bzbpl [257.040514ms]
+Apr 29 12:44:39.670: INFO: Created: latency-svc-w8p8n
+Apr 29 12:44:39.677: INFO: Got endpoints: latency-svc-68jgv [279.077455ms]
+Apr 29 12:44:39.689: INFO: Got endpoints: latency-svc-w8p8n [290.678152ms]
+Apr 29 12:44:39.704: INFO: Created: latency-svc-ddkzl
+Apr 29 12:44:39.704: INFO: Created: latency-svc-jb7z6
+Apr 29 12:44:39.712: INFO: Got endpoints: latency-svc-ddkzl [313.076241ms]
+Apr 29 12:44:39.713: INFO: Created: latency-svc-df7z9
+Apr 29 12:44:39.737: INFO: Created: latency-svc-fh9kh
+Apr 29 12:44:39.756: INFO: Got endpoints: latency-svc-fh9kh [356.951706ms]
+Apr 29 12:44:39.761: INFO: Got endpoints: latency-svc-df7z9 [361.632187ms]
+Apr 29 12:44:39.763: INFO: Got endpoints: latency-svc-jb7z6 [362.739803ms]
+Apr 29 12:44:39.743: INFO: Created: latency-svc-mgkrl
+Apr 29 12:44:39.779: INFO: Created: latency-svc-vw5m7
+Apr 29 12:44:39.789: INFO: Got endpoints: latency-svc-mgkrl [339.947449ms]
+Apr 29 12:44:39.799: INFO: Created: latency-svc-hc9hs
+Apr 29 12:44:39.803: INFO: Got endpoints: latency-svc-vw5m7 [309.260054ms]
+Apr 29 12:44:39.827: INFO: Created: latency-svc-9hpc4
+Apr 29 12:44:39.833: INFO: Created: latency-svc-lx7vt
+Apr 29 12:44:39.834: INFO: Got endpoints: latency-svc-hc9hs [337.99631ms]
+Apr 29 12:44:39.854: INFO: Got endpoints: latency-svc-lx7vt [299.589444ms]
+Apr 29 12:44:39.855: INFO: Created: latency-svc-xvg6n
+Apr 29 12:44:39.861: INFO: Got endpoints: latency-svc-9hpc4 [333.142601ms]
+Apr 29 12:44:39.882: INFO: Got endpoints: latency-svc-xvg6n [291.80464ms]
+Apr 29 12:44:39.888: INFO: Created: latency-svc-5prz8
+Apr 29 12:44:39.888: INFO: Got endpoints: latency-svc-5prz8 [274.074982ms]
+Apr 29 12:44:39.899: INFO: Created: latency-svc-lg2td
+Apr 29 12:44:39.910: INFO: Created: latency-svc-h6qbj
+Apr 29 12:44:39.919: INFO: Got endpoints: latency-svc-lg2td [286.926339ms]
+Apr 29 12:44:39.925: INFO: Got endpoints: latency-svc-h6qbj [269.772706ms]
+Apr 29 12:44:39.935: INFO: Created: latency-svc-6z5vj
+Apr 29 12:44:39.946: INFO: Created: latency-svc-jmq6l
+Apr 29 12:44:39.962: INFO: Created: latency-svc-xp494
+Apr 29 12:44:39.974: INFO: Got endpoints: latency-svc-jmq6l [285.160805ms]
+Apr 29 12:44:39.975: INFO: Got endpoints: latency-svc-6z5vj [297.450497ms]
+Apr 29 12:44:39.975: INFO: Created: latency-svc-nm8lz
+Apr 29 12:44:39.981: INFO: Created: latency-svc-jlzdb
+Apr 29 12:44:39.981: INFO: Created: latency-svc-9248z
+Apr 29 12:44:40.022: INFO: Got endpoints: latency-svc-9248z [259.951474ms]
+Apr 29 12:44:40.024: INFO: Got endpoints: latency-svc-nm8lz [312.371563ms]
+Apr 29 12:44:40.025: INFO: Got endpoints: latency-svc-xp494 [264.704428ms]
+Apr 29 12:44:40.025: INFO: Created: latency-svc-k2wn8
+Apr 29 12:44:40.027: INFO: Got endpoints: latency-svc-jlzdb [256.897164ms]
+Apr 29 12:44:40.047: INFO: Created: latency-svc-bvd6n
+Apr 29 12:44:40.053: INFO: Got endpoints: latency-svc-k2wn8 [263.464021ms]
+Apr 29 12:44:40.087: INFO: Got endpoints: latency-svc-bvd6n [284.582033ms]
+Apr 29 12:44:40.091: INFO: Created: latency-svc-qtfk7
+Apr 29 12:44:40.102: INFO: Created: latency-svc-d2b55
+Apr 29 12:44:40.123: INFO: Got endpoints: latency-svc-qtfk7 [287.685503ms]
+Apr 29 12:44:40.136: INFO: Created: latency-svc-rbf6r
+Apr 29 12:44:40.145: INFO: Got endpoints: latency-svc-rbf6r [284.054003ms]
+Apr 29 12:44:40.146: INFO: Got endpoints: latency-svc-d2b55 [290.081649ms]
+Apr 29 12:44:40.154: INFO: Created: latency-svc-82vcj
+Apr 29 12:44:40.181: INFO: Got endpoints: latency-svc-82vcj [298.040775ms]
+Apr 29 12:44:40.188: INFO: Created: latency-svc-p2jpg
+Apr 29 12:44:40.196: INFO: Created: latency-svc-5zpjx
+Apr 29 12:44:40.203: INFO: Created: latency-svc-wnxqm
+Apr 29 12:44:40.240: INFO: Created: latency-svc-bbtjs
+Apr 29 12:44:40.251: INFO: Got endpoints: latency-svc-5zpjx [362.916393ms]
+Apr 29 12:44:40.254: INFO: Got endpoints: latency-svc-wnxqm [328.362645ms]
+Apr 29 12:44:40.255: INFO: Got endpoints: latency-svc-p2jpg [335.487952ms]
+Apr 29 12:44:40.280: INFO: Got endpoints: latency-svc-bbtjs [304.155862ms]
+Apr 29 12:44:40.317: INFO: Created: latency-svc-p5pdn
+Apr 29 12:44:40.317: INFO: Created: latency-svc-v8h48
+Apr 29 12:44:40.320: INFO: Got endpoints: latency-svc-p5pdn [297.974137ms]
+Apr 29 12:44:40.320: INFO: Got endpoints: latency-svc-v8h48 [344.857009ms]
+Apr 29 12:44:40.352: INFO: Created: latency-svc-6hc6h
+Apr 29 12:44:40.413: INFO: Created: latency-svc-vfvtd
+Apr 29 12:44:40.413: INFO: Created: latency-svc-h8p8l
+Apr 29 12:44:40.414: INFO: Got endpoints: latency-svc-h8p8l [389.031071ms]
+Apr 29 12:44:40.413: INFO: Got endpoints: latency-svc-6hc6h [386.50282ms]
+Apr 29 12:44:40.416: INFO: Created: latency-svc-l9zw4
+Apr 29 12:44:40.428: INFO: Got endpoints: latency-svc-vfvtd [403.141847ms]
+Apr 29 12:44:40.443: INFO: Created: latency-svc-gkctj
+Apr 29 12:44:40.451: INFO: Got endpoints: latency-svc-l9zw4 [398.857011ms]
+Apr 29 12:44:40.463: INFO: Created: latency-svc-vmx2t
+Apr 29 12:44:40.488: INFO: Created: latency-svc-vg695
+Apr 29 12:44:40.496: INFO: Got endpoints: latency-svc-gkctj [408.631366ms]
+Apr 29 12:44:40.503: INFO: Created: latency-svc-9mc6j
+Apr 29 12:44:40.504: INFO: Got endpoints: latency-svc-vmx2t [380.935344ms]
+Apr 29 12:44:40.511: INFO: Got endpoints: latency-svc-vg695 [366.021663ms]
+Apr 29 12:44:40.533: INFO: Created: latency-svc-p69k7
+Apr 29 12:44:40.557: INFO: Got endpoints: latency-svc-9mc6j [410.493565ms]
+Apr 29 12:44:40.560: INFO: Created: latency-svc-slch5
+Apr 29 12:44:40.567: INFO: Got endpoints: latency-svc-p69k7 [384.604422ms]
+Apr 29 12:44:40.579: INFO: Created: latency-svc-6v4rt
+Apr 29 12:44:40.603: INFO: Created: latency-svc-sx5kr
+Apr 29 12:44:40.611: INFO: Got endpoints: latency-svc-slch5 [354.269807ms]
+Apr 29 12:44:40.616: INFO: Created: latency-svc-kwn7k
+Apr 29 12:44:40.640: INFO: Created: latency-svc-fhxb6
+Apr 29 12:44:40.652: INFO: Got endpoints: latency-svc-6v4rt [399.538846ms]
+Apr 29 12:44:40.665: INFO: Created: latency-svc-f5nmc
+Apr 29 12:44:40.688: INFO: Created: latency-svc-7brzr
+Apr 29 12:44:40.688: INFO: Created: latency-svc-bjxqt
+Apr 29 12:44:40.706: INFO: Created: latency-svc-bfhqc
+Apr 29 12:44:40.708: INFO: Got endpoints: latency-svc-sx5kr [453.528445ms]
+Apr 29 12:44:40.725: INFO: Created: latency-svc-phddn
+Apr 29 12:44:40.746: INFO: Got endpoints: latency-svc-kwn7k [465.123525ms]
+Apr 29 12:44:40.772: INFO: Created: latency-svc-g7mkk
+Apr 29 12:44:40.777: INFO: Created: latency-svc-zpmpm
+Apr 29 12:44:40.800: INFO: Created: latency-svc-mcsrq
+Apr 29 12:44:40.817: INFO: Created: latency-svc-zxdzz
+Apr 29 12:44:40.817: INFO: Got endpoints: latency-svc-fhxb6 [496.834131ms]
+Apr 29 12:44:40.821: INFO: Created: latency-svc-gzkgr
+Apr 29 12:44:40.835: INFO: Created: latency-svc-8qcn7
+Apr 29 12:44:40.848: INFO: Got endpoints: latency-svc-f5nmc [528.044823ms]
+Apr 29 12:44:40.850: INFO: Created: latency-svc-9lzr7
+Apr 29 12:44:40.862: INFO: Created: latency-svc-df4p6
+Apr 29 12:44:40.880: INFO: Created: latency-svc-bxh8l
+Apr 29 12:44:40.906: INFO: Got endpoints: latency-svc-7brzr [492.046465ms]
+Apr 29 12:44:40.915: INFO: Created: latency-svc-4lvm2
+Apr 29 12:44:40.941: INFO: Created: latency-svc-7rcsm
+Apr 29 12:44:40.964: INFO: Created: latency-svc-7wl94
+Apr 29 12:44:40.969: INFO: Got endpoints: latency-svc-bjxqt [553.613712ms]
+Apr 29 12:44:41.000: INFO: Got endpoints: latency-svc-bfhqc [571.504458ms]
+Apr 29 12:44:41.009: INFO: Created: latency-svc-w2rnx
+Apr 29 12:44:41.040: INFO: Created: latency-svc-n4lmq
+Apr 29 12:44:41.049: INFO: Got endpoints: latency-svc-phddn [596.44207ms]
+Apr 29 12:44:41.064: INFO: Created: latency-svc-j57qb
+Apr 29 12:44:41.091: INFO: Got endpoints: latency-svc-g7mkk [595.049536ms]
+Apr 29 12:44:41.114: INFO: Created: latency-svc-mjnnd
+Apr 29 12:44:41.155: INFO: Got endpoints: latency-svc-zpmpm [650.84656ms]
+Apr 29 12:44:41.188: INFO: Created: latency-svc-6tpxf
+Apr 29 12:44:41.210: INFO: Got endpoints: latency-svc-zxdzz [698.877243ms]
+Apr 29 12:44:41.268: INFO: Got endpoints: latency-svc-mcsrq [711.082036ms]
+Apr 29 12:44:41.286: INFO: Created: latency-svc-7vps4
+Apr 29 12:44:41.307: INFO: Got endpoints: latency-svc-gzkgr [739.933186ms]
+Apr 29 12:44:41.330: INFO: Created: latency-svc-8597z
+Apr 29 12:44:41.343: INFO: Got endpoints: latency-svc-8qcn7 [731.976871ms]
+Apr 29 12:44:41.354: INFO: Created: latency-svc-4c49s
+Apr 29 12:44:41.363: INFO: Created: latency-svc-6f2bq
+Apr 29 12:44:41.409: INFO: Got endpoints: latency-svc-9lzr7 [757.502086ms]
+Apr 29 12:44:41.462: INFO: Created: latency-svc-dvvcs
+Apr 29 12:44:41.473: INFO: Got endpoints: latency-svc-df4p6 [765.066967ms]
+Apr 29 12:44:41.514: INFO: Got endpoints: latency-svc-bxh8l [766.455249ms]
+Apr 29 12:44:41.553: INFO: Created: latency-svc-m6pt2
+Apr 29 12:44:41.553: INFO: Created: latency-svc-ff8jc
+Apr 29 12:44:41.563: INFO: Got endpoints: latency-svc-4lvm2 [745.167397ms]
+Apr 29 12:44:41.581: INFO: Created: latency-svc-fnlnb
+Apr 29 12:44:41.613: INFO: Got endpoints: latency-svc-7wl94 [764.447794ms]
+Apr 29 12:44:41.641: INFO: Got endpoints: latency-svc-7rcsm [735.43135ms]
+Apr 29 12:44:41.648: INFO: Created: latency-svc-tgrgt
+Apr 29 12:44:41.668: INFO: Created: latency-svc-nbr6q
+Apr 29 12:44:41.700: INFO: Got endpoints: latency-svc-w2rnx [731.10735ms]
+Apr 29 12:44:41.745: INFO: Got endpoints: latency-svc-n4lmq [744.619067ms]
+Apr 29 12:44:41.753: INFO: Created: latency-svc-hvvcl
+Apr 29 12:44:41.779: INFO: Created: latency-svc-slj8s
+Apr 29 12:44:41.799: INFO: Got endpoints: latency-svc-j57qb [749.404295ms]
+Apr 29 12:44:41.825: INFO: Created: latency-svc-dkpp6
+Apr 29 12:44:41.842: INFO: Got endpoints: latency-svc-mjnnd [750.964221ms]
+Apr 29 12:44:41.860: INFO: Created: latency-svc-rp47k
+Apr 29 12:44:41.889: INFO: Got endpoints: latency-svc-6tpxf [733.608469ms]
+Apr 29 12:44:41.941: INFO: Got endpoints: latency-svc-7vps4 [730.116785ms]
+Apr 29 12:44:41.953: INFO: Created: latency-svc-b6wk4
+Apr 29 12:44:41.964: INFO: Created: latency-svc-kszrh
+Apr 29 12:44:41.994: INFO: Got endpoints: latency-svc-8597z [719.85411ms]
+Apr 29 12:44:42.010: INFO: Created: latency-svc-t87p9
+Apr 29 12:44:42.061: INFO: Got endpoints: latency-svc-4c49s [753.812515ms]
+Apr 29 12:44:42.097: INFO: Got endpoints: latency-svc-6f2bq [753.243871ms]
+Apr 29 12:44:42.124: INFO: Created: latency-svc-n4b6v
+Apr 29 12:44:42.132: INFO: Created: latency-svc-2h889
+Apr 29 12:44:42.147: INFO: Got endpoints: latency-svc-dvvcs [737.068829ms]
+Apr 29 12:44:42.200: INFO: Got endpoints: latency-svc-m6pt2 [726.289165ms]
+Apr 29 12:44:42.204: INFO: Created: latency-svc-qmvfp
+Apr 29 12:44:42.234: INFO: Created: latency-svc-65s5g
+Apr 29 12:44:42.259: INFO: Got endpoints: latency-svc-ff8jc [743.979565ms]
+Apr 29 12:44:42.290: INFO: Created: latency-svc-vb9rx
+Apr 29 12:44:42.315: INFO: Got endpoints: latency-svc-fnlnb [751.827956ms]
+Apr 29 12:44:42.353: INFO: Got endpoints: latency-svc-tgrgt [738.942058ms]
+Apr 29 12:44:42.375: INFO: Created: latency-svc-lksqw
+Apr 29 12:44:42.400: INFO: Got endpoints: latency-svc-nbr6q [756.9868ms]
+Apr 29 12:44:42.412: INFO: Created: latency-svc-tgkd6
+Apr 29 12:44:42.418: INFO: Created: latency-svc-cglc2
+Apr 29 12:44:42.446: INFO: Got endpoints: latency-svc-hvvcl [744.795189ms]
+Apr 29 12:44:42.471: INFO: Created: latency-svc-5h9w4
+Apr 29 12:44:42.526: INFO: Got endpoints: latency-svc-slj8s [780.385434ms]
+Apr 29 12:44:42.558: INFO: Created: latency-svc-4z6pk
+Apr 29 12:44:42.565: INFO: Got endpoints: latency-svc-dkpp6 [765.497247ms]
+Apr 29 12:44:42.582: INFO: Created: latency-svc-9t6jd
+Apr 29 12:44:42.601: INFO: Got endpoints: latency-svc-rp47k [757.56002ms]
+Apr 29 12:44:42.624: INFO: Created: latency-svc-xpn4l
+Apr 29 12:44:42.649: INFO: Got endpoints: latency-svc-b6wk4 [759.261003ms]
+Apr 29 12:44:42.664: INFO: Created: latency-svc-bq4vk
+Apr 29 12:44:42.693: INFO: Got endpoints: latency-svc-kszrh [750.695128ms]
+Apr 29 12:44:42.704: INFO: Created: latency-svc-pskrk
+Apr 29 12:44:42.742: INFO: Got endpoints: latency-svc-t87p9 [747.156694ms]
+Apr 29 12:44:42.755: INFO: Created: latency-svc-msqrg
+Apr 29 12:44:42.791: INFO: Got endpoints: latency-svc-n4b6v [729.572733ms]
+Apr 29 12:44:42.803: INFO: Created: latency-svc-ldg9p
+Apr 29 12:44:42.847: INFO: Got endpoints: latency-svc-2h889 [748.9236ms]
+Apr 29 12:44:42.866: INFO: Created: latency-svc-d4rn6
+Apr 29 12:44:42.890: INFO: Got endpoints: latency-svc-qmvfp [740.945457ms]
+Apr 29 12:44:42.913: INFO: Created: latency-svc-dbqm5
+Apr 29 12:44:42.953: INFO: Got endpoints: latency-svc-65s5g [749.257647ms]
+Apr 29 12:44:42.994: INFO: Created: latency-svc-6m8d7
+Apr 29 12:44:43.008: INFO: Got endpoints: latency-svc-vb9rx [748.235526ms]
+Apr 29 12:44:43.028: INFO: Created: latency-svc-qwjs2
+Apr 29 12:44:43.044: INFO: Got endpoints: latency-svc-lksqw [727.311075ms]
+Apr 29 12:44:43.071: INFO: Created: latency-svc-5k86x
+Apr 29 12:44:43.107: INFO: Got endpoints: latency-svc-tgkd6 [752.791652ms]
+Apr 29 12:44:43.135: INFO: Created: latency-svc-5ckph
+Apr 29 12:44:43.142: INFO: Got endpoints: latency-svc-cglc2 [740.93063ms]
+Apr 29 12:44:43.188: INFO: Created: latency-svc-47bt9
+Apr 29 12:44:43.192: INFO: Got endpoints: latency-svc-5h9w4 [745.668882ms]
+Apr 29 12:44:43.219: INFO: Created: latency-svc-v4m7r
+Apr 29 12:44:43.250: INFO: Got endpoints: latency-svc-4z6pk [723.505261ms]
+Apr 29 12:44:43.280: INFO: Created: latency-svc-98p5m
+Apr 29 12:44:43.291: INFO: Got endpoints: latency-svc-9t6jd [725.728214ms]
+Apr 29 12:44:43.306: INFO: Created: latency-svc-ktqqk
+Apr 29 12:44:43.343: INFO: Got endpoints: latency-svc-xpn4l [741.745076ms]
+Apr 29 12:44:43.357: INFO: Created: latency-svc-jjzgz
+Apr 29 12:44:43.402: INFO: Got endpoints: latency-svc-bq4vk [753.378357ms]
+Apr 29 12:44:43.432: INFO: Created: latency-svc-fng6r
+Apr 29 12:44:43.444: INFO: Got endpoints: latency-svc-pskrk [750.694536ms]
+Apr 29 12:44:43.497: INFO: Created: latency-svc-8nglz
+Apr 29 12:44:43.517: INFO: Got endpoints: latency-svc-msqrg [774.963555ms]
+Apr 29 12:44:43.551: INFO: Got endpoints: latency-svc-ldg9p [759.487199ms]
+Apr 29 12:44:43.556: INFO: Created: latency-svc-6jsxr
+Apr 29 12:44:43.596: INFO: Created: latency-svc-c42xh
+Apr 29 12:44:43.608: INFO: Got endpoints: latency-svc-d4rn6 [761.03649ms]
+Apr 29 12:44:43.632: INFO: Created: latency-svc-fzmkz
+Apr 29 12:44:43.641: INFO: Got endpoints: latency-svc-dbqm5 [750.944789ms]
+Apr 29 12:44:43.669: INFO: Created: latency-svc-gpzp7
+Apr 29 12:44:43.696: INFO: Got endpoints: latency-svc-6m8d7 [741.592165ms]
+Apr 29 12:44:43.712: INFO: Created: latency-svc-4s6wc
+Apr 29 12:44:43.750: INFO: Got endpoints: latency-svc-qwjs2 [741.802788ms]
+Apr 29 12:44:43.787: INFO: Created: latency-svc-96pt4
+Apr 29 12:44:43.797: INFO: Got endpoints: latency-svc-5k86x [752.645939ms]
+Apr 29 12:44:43.816: INFO: Created: latency-svc-r9jnm
+Apr 29 12:44:43.858: INFO: Got endpoints: latency-svc-5ckph [750.802148ms]
+Apr 29 12:44:43.910: INFO: Got endpoints: latency-svc-47bt9 [767.18759ms]
+Apr 29 12:44:43.913: INFO: Created: latency-svc-sl7l9
+Apr 29 12:44:43.985: INFO: Got endpoints: latency-svc-v4m7r [792.876348ms]
+Apr 29 12:44:44.002: INFO: Got endpoints: latency-svc-98p5m [750.632429ms]
+Apr 29 12:44:44.005: INFO: Created: latency-svc-mt7jb
+Apr 29 12:44:44.027: INFO: Created: latency-svc-djwzb
+Apr 29 12:44:44.039: INFO: Created: latency-svc-tgznl
+Apr 29 12:44:44.048: INFO: Got endpoints: latency-svc-ktqqk [756.905571ms]
+Apr 29 12:44:44.068: INFO: Created: latency-svc-8fslw
+Apr 29 12:44:44.113: INFO: Got endpoints: latency-svc-jjzgz [769.907045ms]
+Apr 29 12:44:44.140: INFO: Created: latency-svc-ds62p
+Apr 29 12:44:44.144: INFO: Got endpoints: latency-svc-fng6r [741.901197ms]
+Apr 29 12:44:44.158: INFO: Created: latency-svc-42lq7
+Apr 29 12:44:44.198: INFO: Got endpoints: latency-svc-8nglz [751.520379ms]
+Apr 29 12:44:44.215: INFO: Created: latency-svc-qjfrz
+Apr 29 12:44:44.241: INFO: Got endpoints: latency-svc-6jsxr [723.878822ms]
+Apr 29 12:44:44.258: INFO: Created: latency-svc-dzhvx
+Apr 29 12:44:44.295: INFO: Got endpoints: latency-svc-c42xh [743.621606ms]
+Apr 29 12:44:44.326: INFO: Created: latency-svc-4zdf5
+Apr 29 12:44:44.346: INFO: Got endpoints: latency-svc-fzmkz [736.583253ms]
+Apr 29 12:44:44.366: INFO: Created: latency-svc-k9vfg
+Apr 29 12:44:44.401: INFO: Got endpoints: latency-svc-gpzp7 [758.366372ms]
+Apr 29 12:44:44.427: INFO: Created: latency-svc-cb7pb
+Apr 29 12:44:44.442: INFO: Got endpoints: latency-svc-4s6wc [744.857254ms]
+Apr 29 12:44:44.460: INFO: Created: latency-svc-n4vp5
+Apr 29 12:44:44.494: INFO: Got endpoints: latency-svc-96pt4 [741.285529ms]
+Apr 29 12:44:44.519: INFO: Created: latency-svc-zb799
+Apr 29 12:44:44.547: INFO: Got endpoints: latency-svc-r9jnm [749.690934ms]
+Apr 29 12:44:44.573: INFO: Created: latency-svc-dr5nm
+Apr 29 12:44:44.589: INFO: Got endpoints: latency-svc-sl7l9 [731.858481ms]
+Apr 29 12:44:44.609: INFO: Created: latency-svc-z9pzw
+Apr 29 12:44:44.651: INFO: Got endpoints: latency-svc-mt7jb [738.370448ms]
+Apr 29 12:44:44.669: INFO: Created: latency-svc-pmwbb
+Apr 29 12:44:44.691: INFO: Got endpoints: latency-svc-djwzb [705.512714ms]
+Apr 29 12:44:44.714: INFO: Created: latency-svc-fs755
+Apr 29 12:44:44.744: INFO: Got endpoints: latency-svc-tgznl [740.504182ms]
+Apr 29 12:44:44.757: INFO: Created: latency-svc-v6r22
+Apr 29 12:44:44.790: INFO: Got endpoints: latency-svc-8fslw [742.080698ms]
+Apr 29 12:44:44.820: INFO: Created: latency-svc-9hpf8
+Apr 29 12:44:44.845: INFO: Got endpoints: latency-svc-ds62p [730.954035ms]
+Apr 29 12:44:44.863: INFO: Created: latency-svc-766jv
+Apr 29 12:44:44.889: INFO: Got endpoints: latency-svc-42lq7 [744.901958ms]
+Apr 29 12:44:44.953: INFO: Got endpoints: latency-svc-qjfrz [753.837873ms]
+Apr 29 12:44:44.956: INFO: Created: latency-svc-bwxqh
+Apr 29 12:44:44.978: INFO: Created: latency-svc-r44jr
+Apr 29 12:44:44.993: INFO: Got endpoints: latency-svc-dzhvx [751.132003ms]
+Apr 29 12:44:45.025: INFO: Created: latency-svc-4h45m
+Apr 29 12:44:45.045: INFO: Got endpoints: latency-svc-4zdf5 [749.365787ms]
+Apr 29 12:44:45.099: INFO: Created: latency-svc-xpw6z
+Apr 29 12:44:45.102: INFO: Got endpoints: latency-svc-k9vfg [754.766852ms]
+Apr 29 12:44:45.118: INFO: Created: latency-svc-c75m7
+Apr 29 12:44:45.148: INFO: Got endpoints: latency-svc-cb7pb [740.416048ms]
+Apr 29 12:44:45.176: INFO: Created: latency-svc-sr5f9
+Apr 29 12:44:45.197: INFO: Got endpoints: latency-svc-n4vp5 [754.952088ms]
+Apr 29 12:44:45.221: INFO: Created: latency-svc-gzzrp
+Apr 29 12:44:45.243: INFO: Got endpoints: latency-svc-zb799 [747.12781ms]
+Apr 29 12:44:45.258: INFO: Created: latency-svc-ssrg8
+Apr 29 12:44:45.311: INFO: Got endpoints: latency-svc-dr5nm [763.562822ms]
+Apr 29 12:44:45.343: INFO: Created: latency-svc-rslts
+Apr 29 12:44:45.347: INFO: Got endpoints: latency-svc-z9pzw [756.583375ms]
+Apr 29 12:44:45.372: INFO: Created: latency-svc-v4p9c
+Apr 29 12:44:45.451: INFO: Got endpoints: latency-svc-pmwbb [799.157189ms]
+Apr 29 12:44:45.455: INFO: Got endpoints: latency-svc-fs755 [763.445878ms]
+Apr 29 12:44:45.475: INFO: Created: latency-svc-zqxpv
+Apr 29 12:44:45.492: INFO: Created: latency-svc-q6kcl
+Apr 29 12:44:45.499: INFO: Got endpoints: latency-svc-v6r22 [753.684762ms]
+Apr 29 12:44:45.524: INFO: Created: latency-svc-rzbr5
+Apr 29 12:44:45.543: INFO: Got endpoints: latency-svc-9hpf8 [752.0907ms]
+Apr 29 12:44:45.571: INFO: Created: latency-svc-xmx2v
+Apr 29 12:44:45.590: INFO: Got endpoints: latency-svc-766jv [744.408287ms]
+Apr 29 12:44:45.613: INFO: Created: latency-svc-jjf54
+Apr 29 12:44:45.657: INFO: Got endpoints: latency-svc-bwxqh [767.528854ms]
+Apr 29 12:44:45.737: INFO: Created: latency-svc-22dj5
+Apr 29 12:44:45.776: INFO: Got endpoints: latency-svc-4h45m [776.627024ms]
+Apr 29 12:44:45.779: INFO: Got endpoints: latency-svc-r44jr [824.796226ms]
+Apr 29 12:44:45.817: INFO: Got endpoints: latency-svc-xpw6z [769.135297ms]
+Apr 29 12:44:45.821: INFO: Created: latency-svc-zb7qb
+Apr 29 12:44:45.845: INFO: Created: latency-svc-slmhk
+Apr 29 12:44:45.859: INFO: Got endpoints: latency-svc-c75m7 [757.14812ms]
+Apr 29 12:44:45.878: INFO: Created: latency-svc-k2dwb
+Apr 29 12:44:45.955: INFO: Created: latency-svc-6k9bz
+Apr 29 12:44:45.970: INFO: Got endpoints: latency-svc-sr5f9 [818.278781ms]
+Apr 29 12:44:45.971: INFO: Got endpoints: latency-svc-gzzrp [773.409572ms]
+Apr 29 12:44:46.016: INFO: Created: latency-svc-9tgp4
+Apr 29 12:44:46.016: INFO: Created: latency-svc-d272n
+Apr 29 12:44:46.017: INFO: Got endpoints: latency-svc-ssrg8 [771.733563ms]
+Apr 29 12:44:46.030: INFO: Created: latency-svc-j4cwl
+Apr 29 12:44:46.045: INFO: Got endpoints: latency-svc-rslts [733.999661ms]
+Apr 29 12:44:46.075: INFO: Created: latency-svc-4jbr6
+Apr 29 12:44:46.102: INFO: Got endpoints: latency-svc-v4p9c [754.875415ms]
+Apr 29 12:44:46.117: INFO: Created: latency-svc-7zxdh
+Apr 29 12:44:46.169: INFO: Got endpoints: latency-svc-zqxpv [717.42808ms]
+Apr 29 12:44:46.201: INFO: Got endpoints: latency-svc-q6kcl [745.852627ms]
+Apr 29 12:44:46.229: INFO: Created: latency-svc-6g785
+Apr 29 12:44:46.257: INFO: Got endpoints: latency-svc-rzbr5 [758.010132ms]
+Apr 29 12:44:46.258: INFO: Created: latency-svc-wl67h
+Apr 29 12:44:46.305: INFO: Created: latency-svc-h4b8q
+Apr 29 12:44:46.323: INFO: Got endpoints: latency-svc-xmx2v [780.228841ms]
+Apr 29 12:44:46.346: INFO: Got endpoints: latency-svc-jjf54 [754.976777ms]
+Apr 29 12:44:46.347: INFO: Created: latency-svc-bn8nj
+Apr 29 12:44:46.363: INFO: Created: latency-svc-7x5rk
+Apr 29 12:44:46.398: INFO: Got endpoints: latency-svc-22dj5 [740.200487ms]
+Apr 29 12:44:46.417: INFO: Created: latency-svc-h59sq
+Apr 29 12:44:46.443: INFO: Got endpoints: latency-svc-zb7qb [664.515524ms]
+Apr 29 12:44:46.482: INFO: Created: latency-svc-npbpq
+Apr 29 12:44:46.491: INFO: Got endpoints: latency-svc-slmhk [709.935906ms]
+Apr 29 12:44:46.515: INFO: Created: latency-svc-k5wsq
+Apr 29 12:44:46.542: INFO: Got endpoints: latency-svc-k2dwb [721.073826ms]
+Apr 29 12:44:46.556: INFO: Created: latency-svc-h6kpm
+Apr 29 12:44:46.589: INFO: Got endpoints: latency-svc-6k9bz [729.088671ms]
+Apr 29 12:44:46.609: INFO: Created: latency-svc-ww7nk
+Apr 29 12:44:46.644: INFO: Got endpoints: latency-svc-d272n [673.889072ms]
+Apr 29 12:44:46.669: INFO: Created: latency-svc-7tp2c
+Apr 29 12:44:46.692: INFO: Got endpoints: latency-svc-9tgp4 [720.729047ms]
+Apr 29 12:44:46.708: INFO: Created: latency-svc-b4wf2
+Apr 29 12:44:46.757: INFO: Got endpoints: latency-svc-j4cwl [740.612949ms]
+Apr 29 12:44:46.781: INFO: Created: latency-svc-7csvz
+Apr 29 12:44:46.792: INFO: Got endpoints: latency-svc-4jbr6 [746.335684ms]
+Apr 29 12:44:46.806: INFO: Created: latency-svc-qx928
+Apr 29 12:44:46.842: INFO: Got endpoints: latency-svc-7zxdh [738.863261ms]
+Apr 29 12:44:46.860: INFO: Created: latency-svc-7gr99
+Apr 29 12:44:46.894: INFO: Got endpoints: latency-svc-6g785 [720.196443ms]
+Apr 29 12:44:46.922: INFO: Created: latency-svc-2wpsc
+Apr 29 12:44:46.948: INFO: Got endpoints: latency-svc-wl67h [741.124751ms]
+Apr 29 12:44:46.972: INFO: Created: latency-svc-nqc46
+Apr 29 12:44:46.989: INFO: Got endpoints: latency-svc-h4b8q [731.801919ms]
+Apr 29 12:44:47.011: INFO: Created: latency-svc-87f69
+Apr 29 12:44:47.042: INFO: Got endpoints: latency-svc-bn8nj [714.857246ms]
+Apr 29 12:44:47.060: INFO: Created: latency-svc-p26cc
+Apr 29 12:44:47.115: INFO: Got endpoints: latency-svc-7x5rk [767.842581ms]
+Apr 29 12:44:47.133: INFO: Created: latency-svc-kbsrs
+Apr 29 12:44:47.149: INFO: Got endpoints: latency-svc-h59sq [750.271477ms]
+Apr 29 12:44:47.178: INFO: Created: latency-svc-jtbnf
+Apr 29 12:44:47.197: INFO: Got endpoints: latency-svc-npbpq [752.17266ms]
+Apr 29 12:44:47.232: INFO: Created: latency-svc-6vw8j
+Apr 29 12:44:47.243: INFO: Got endpoints: latency-svc-k5wsq [751.075329ms]
+Apr 29 12:44:47.294: INFO: Got endpoints: latency-svc-h6kpm [751.647622ms]
+Apr 29 12:44:47.343: INFO: Got endpoints: latency-svc-ww7nk [752.943237ms]
+Apr 29 12:44:47.396: INFO: Got endpoints: latency-svc-7tp2c [750.353034ms]
+Apr 29 12:44:47.447: INFO: Got endpoints: latency-svc-b4wf2 [754.279525ms]
+Apr 29 12:44:47.502: INFO: Got endpoints: latency-svc-7csvz [743.721598ms]
+Apr 29 12:44:47.545: INFO: Got endpoints: latency-svc-qx928 [752.562167ms]
+Apr 29 12:44:47.597: INFO: Got endpoints: latency-svc-7gr99 [754.783351ms]
+Apr 29 12:44:47.645: INFO: Got endpoints: latency-svc-2wpsc [750.489759ms]
+Apr 29 12:44:47.702: INFO: Got endpoints: latency-svc-nqc46 [752.441762ms]
+Apr 29 12:44:47.750: INFO: Got endpoints: latency-svc-87f69 [760.0075ms]
+Apr 29 12:44:47.791: INFO: Got endpoints: latency-svc-p26cc [748.313349ms]
+Apr 29 12:44:47.846: INFO: Got endpoints: latency-svc-kbsrs [729.99598ms]
+Apr 29 12:44:47.899: INFO: Got endpoints: latency-svc-jtbnf [748.307863ms]
+Apr 29 12:44:47.958: INFO: Got endpoints: latency-svc-6vw8j [761.130743ms]
+Apr 29 12:44:47.960: INFO: Latencies: [54.032492ms 71.267878ms 88.083901ms 97.391229ms 99.709573ms 131.399536ms 193.754412ms 217.038639ms 234.473087ms 256.897164ms 257.040514ms 259.951474ms 263.464021ms 264.704428ms 269.772706ms 274.074982ms 279.077455ms 284.054003ms 284.582033ms 285.160805ms 286.926339ms 287.685503ms 290.081649ms 290.678152ms 291.80464ms 297.450497ms 297.974137ms 298.040775ms 299.589444ms 304.155862ms 309.260054ms 312.371563ms 313.076241ms 328.362645ms 333.142601ms 335.487952ms 337.99631ms 339.947449ms 344.857009ms 354.269807ms 356.951706ms 361.632187ms 362.739803ms 362.916393ms 366.021663ms 380.935344ms 384.604422ms 386.50282ms 389.031071ms 398.857011ms 399.538846ms 403.141847ms 408.631366ms 410.493565ms 453.528445ms 465.123525ms 492.046465ms 496.834131ms 528.044823ms 553.613712ms 571.504458ms 595.049536ms 596.44207ms 650.84656ms 664.515524ms 673.889072ms 698.877243ms 705.512714ms 709.935906ms 711.082036ms 714.857246ms 717.42808ms 719.85411ms 720.196443ms 720.729047ms 721.073826ms 723.505261ms 723.878822ms 725.728214ms 726.289165ms 727.311075ms 729.088671ms 729.572733ms 729.99598ms 730.116785ms 730.954035ms 731.10735ms 731.801919ms 731.858481ms 731.976871ms 733.608469ms 733.999661ms 735.43135ms 736.583253ms 737.068829ms 738.370448ms 738.863261ms 738.942058ms 739.933186ms 740.200487ms 740.416048ms 740.504182ms 740.612949ms 740.93063ms 740.945457ms 741.124751ms 741.285529ms 741.592165ms 741.745076ms 741.802788ms 741.901197ms 742.080698ms 743.621606ms 743.721598ms 743.979565ms 744.408287ms 744.619067ms 744.795189ms 744.857254ms 744.901958ms 745.167397ms 745.668882ms 745.852627ms 746.335684ms 747.12781ms 747.156694ms 748.235526ms 748.307863ms 748.313349ms 748.9236ms 749.257647ms 749.365787ms 749.404295ms 749.690934ms 750.271477ms 750.353034ms 750.489759ms 750.632429ms 750.694536ms 750.695128ms 750.802148ms 750.944789ms 750.964221ms 751.075329ms 751.132003ms 751.520379ms 751.647622ms 751.827956ms 752.0907ms 752.17266ms 752.441762ms 752.562167ms 752.645939ms 752.791652ms 752.943237ms 753.243871ms 753.378357ms 753.684762ms 753.812515ms 753.837873ms 754.279525ms 754.766852ms 754.783351ms 754.875415ms 754.952088ms 754.976777ms 756.583375ms 756.905571ms 756.9868ms 757.14812ms 757.502086ms 757.56002ms 758.010132ms 758.366372ms 759.261003ms 759.487199ms 760.0075ms 761.03649ms 761.130743ms 763.445878ms 763.562822ms 764.447794ms 765.066967ms 765.497247ms 766.455249ms 767.18759ms 767.528854ms 767.842581ms 769.135297ms 769.907045ms 771.733563ms 773.409572ms 774.963555ms 776.627024ms 780.228841ms 780.385434ms 792.876348ms 799.157189ms 818.278781ms 824.796226ms]
+Apr 29 12:44:47.960: INFO: 50 %ile: 740.416048ms
+Apr 29 12:44:47.961: INFO: 90 %ile: 763.562822ms
+Apr 29 12:44:47.962: INFO: 99 %ile: 818.278781ms
+Apr 29 12:44:47.962: INFO: Total sample count: 200
+[AfterEach] [sig-network] Service endpoints latency
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 12:44:47.963: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "svc-latency-3029" for this suite.
+Apr 29 12:45:06.006: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 12:45:06.233: INFO: namespace svc-latency-3029 deletion completed in 18.253023839s
+
+• [SLOW TEST:30.277 seconds]
+[sig-network] Service endpoints latency
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:22
+  should not be very high  [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+------------------------------
+[sig-cli] Kubectl client [k8s.io] Kubectl run deployment 
+  should create a deployment from an image  [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-cli] Kubectl client
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 12:45:06.234: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename kubectl
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-9883
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [sig-cli] Kubectl client
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:213
+[BeforeEach] [k8s.io] Kubectl run deployment
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1455
+[It] should create a deployment from an image  [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: running the image docker.io/library/nginx:1.14-alpine
+Apr 29 12:45:06.441: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 run e2e-test-nginx-deployment --image=docker.io/library/nginx:1.14-alpine --generator=deployment/v1beta1 --namespace=kubectl-9883'
+Apr 29 12:45:08.690: INFO: stderr: "kubectl run --generator=deployment/v1beta1 is DEPRECATED and will be removed in a future version. Use kubectl run --generator=run-pod/v1 or kubectl create instead.\n"
+Apr 29 12:45:08.690: INFO: stdout: "deployment.extensions/e2e-test-nginx-deployment created\n"
+STEP: verifying the deployment e2e-test-nginx-deployment was created
+STEP: verifying the pod controlled by deployment e2e-test-nginx-deployment was created
+[AfterEach] [k8s.io] Kubectl run deployment
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1460
+Apr 29 12:45:12.711: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 delete deployment e2e-test-nginx-deployment --namespace=kubectl-9883'
+Apr 29 12:45:12.840: INFO: stderr: ""
+Apr 29 12:45:12.840: INFO: stdout: "deployment.extensions \"e2e-test-nginx-deployment\" deleted\n"
+[AfterEach] [sig-cli] Kubectl client
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 12:45:12.840: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "kubectl-9883" for this suite.
+Apr 29 12:45:18.891: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 12:45:19.158: INFO: namespace kubectl-9883 deletion completed in 6.306399051s
+
+• [SLOW TEST:12.925 seconds]
+[sig-cli] Kubectl client
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23
+  [k8s.io] Kubectl run deployment
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
+    should create a deployment from an image  [Conformance]
+    /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSSSSSSSS
+------------------------------
+[k8s.io] Container Lifecycle Hook when create a pod with lifecycle hook 
+  should execute poststart exec hook properly [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [k8s.io] Container Lifecycle Hook
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 12:45:19.171: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename container-lifecycle-hook
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in container-lifecycle-hook-2711
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] when create a pod with lifecycle hook
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:61
+STEP: create the container to handle the HTTPGet hook request.
+[It] should execute poststart exec hook properly [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: create the pod with lifecycle hook
+STEP: check poststart hook
+STEP: delete the pod with lifecycle hook
+Apr 29 12:45:27.506: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear
+Apr 29 12:45:27.515: INFO: Pod pod-with-poststart-exec-hook still exists
+Apr 29 12:45:29.515: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear
+Apr 29 12:45:29.527: INFO: Pod pod-with-poststart-exec-hook still exists
+Apr 29 12:45:31.515: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear
+Apr 29 12:45:31.522: INFO: Pod pod-with-poststart-exec-hook still exists
+Apr 29 12:45:33.515: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear
+Apr 29 12:45:33.521: INFO: Pod pod-with-poststart-exec-hook still exists
+Apr 29 12:45:35.515: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear
+Apr 29 12:45:35.523: INFO: Pod pod-with-poststart-exec-hook still exists
+Apr 29 12:45:37.515: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear
+Apr 29 12:45:37.523: INFO: Pod pod-with-poststart-exec-hook still exists
+Apr 29 12:45:39.515: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear
+Apr 29 12:45:39.521: INFO: Pod pod-with-poststart-exec-hook still exists
+Apr 29 12:45:41.515: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear
+Apr 29 12:45:41.524: INFO: Pod pod-with-poststart-exec-hook still exists
+Apr 29 12:45:43.515: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear
+Apr 29 12:45:43.522: INFO: Pod pod-with-poststart-exec-hook still exists
+Apr 29 12:45:45.515: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear
+Apr 29 12:45:45.523: INFO: Pod pod-with-poststart-exec-hook still exists
+Apr 29 12:45:47.515: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear
+Apr 29 12:45:47.524: INFO: Pod pod-with-poststart-exec-hook no longer exists
+[AfterEach] [k8s.io] Container Lifecycle Hook
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 12:45:47.524: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "container-lifecycle-hook-2711" for this suite.
+Apr 29 12:46:11.580: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 12:46:11.802: INFO: namespace container-lifecycle-hook-2711 deletion completed in 24.258224153s
+
+• [SLOW TEST:52.631 seconds]
+[k8s.io] Container Lifecycle Hook
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
+  when create a pod with lifecycle hook
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:40
+    should execute poststart exec hook properly [NodeConformance] [Conformance]
+    /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+------------------------------
+[k8s.io] Kubelet when scheduling a busybox command in a pod 
+  should print the output to logs [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [k8s.io] Kubelet
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 12:46:11.811: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename kubelet-test
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubelet-test-707
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [k8s.io] Kubelet
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:37
+[It] should print the output to logs [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[AfterEach] [k8s.io] Kubelet
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 12:46:16.090: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "kubelet-test-707" for this suite.
+Apr 29 12:46:56.124: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 12:46:56.394: INFO: namespace kubelet-test-707 deletion completed in 40.296002058s
+
+• [SLOW TEST:44.583 seconds]
+[k8s.io] Kubelet
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
+  when scheduling a busybox command in a pod
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:40
+    should print the output to logs [NodeConformance] [Conformance]
+    /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+S
+------------------------------
+[sig-node] Downward API 
+  should provide container's limits.cpu/memory and requests.cpu/memory as env vars [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-node] Downward API
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 12:46:56.394: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename downward-api
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in downward-api-405
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should provide container's limits.cpu/memory and requests.cpu/memory as env vars [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating a pod to test downward api env vars
+Apr 29 12:46:56.616: INFO: Waiting up to 5m0s for pod "downward-api-df46e66f-6a7c-11e9-b6ef-0e71f8c7e888" in namespace "downward-api-405" to be "success or failure"
+Apr 29 12:46:56.754: INFO: Pod "downward-api-df46e66f-6a7c-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 138.182675ms
+Apr 29 12:46:58.761: INFO: Pod "downward-api-df46e66f-6a7c-11e9-b6ef-0e71f8c7e888": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.14543332s
+STEP: Saw pod success
+Apr 29 12:46:58.761: INFO: Pod "downward-api-df46e66f-6a7c-11e9-b6ef-0e71f8c7e888" satisfied condition "success or failure"
+Apr 29 12:46:58.765: INFO: Trying to get logs from node worker-4av2h-6665d7b5ff-9nc42 pod downward-api-df46e66f-6a7c-11e9-b6ef-0e71f8c7e888 container dapi-container: 
+STEP: delete the pod
+Apr 29 12:46:58.802: INFO: Waiting for pod downward-api-df46e66f-6a7c-11e9-b6ef-0e71f8c7e888 to disappear
+Apr 29 12:46:58.810: INFO: Pod downward-api-df46e66f-6a7c-11e9-b6ef-0e71f8c7e888 no longer exists
+[AfterEach] [sig-node] Downward API
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 12:46:58.810: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "downward-api-405" for this suite.
+Apr 29 12:47:04.838: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 12:47:05.077: INFO: namespace downward-api-405 deletion completed in 6.25751267s
+
+• [SLOW TEST:8.683 seconds]
+[sig-node] Downward API
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downward_api.go:38
+  should provide container's limits.cpu/memory and requests.cpu/memory as env vars [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSSSSSSSSSSSS
+------------------------------
+[sig-storage] Projected downwardAPI 
+  should provide container's memory request [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-storage] Projected downwardAPI
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 12:47:05.089: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename projected
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-1269
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [sig-storage] Projected downwardAPI
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:39
+[It] should provide container's memory request [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating a pod to test downward API volume plugin
+Apr 29 12:47:05.337: INFO: Waiting up to 5m0s for pod "downwardapi-volume-e47a28f1-6a7c-11e9-b6ef-0e71f8c7e888" in namespace "projected-1269" to be "success or failure"
+Apr 29 12:47:05.348: INFO: Pod "downwardapi-volume-e47a28f1-6a7c-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 10.079685ms
+Apr 29 12:47:07.358: INFO: Pod "downwardapi-volume-e47a28f1-6a7c-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 2.020321421s
+Apr 29 12:47:09.367: INFO: Pod "downwardapi-volume-e47a28f1-6a7c-11e9-b6ef-0e71f8c7e888": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.028833691s
+STEP: Saw pod success
+Apr 29 12:47:09.367: INFO: Pod "downwardapi-volume-e47a28f1-6a7c-11e9-b6ef-0e71f8c7e888" satisfied condition "success or failure"
+Apr 29 12:47:09.372: INFO: Trying to get logs from node worker-4av2h-6665d7b5ff-9nc42 pod downwardapi-volume-e47a28f1-6a7c-11e9-b6ef-0e71f8c7e888 container client-container: 
+STEP: delete the pod
+Apr 29 12:47:09.410: INFO: Waiting for pod downwardapi-volume-e47a28f1-6a7c-11e9-b6ef-0e71f8c7e888 to disappear
+Apr 29 12:47:09.415: INFO: Pod downwardapi-volume-e47a28f1-6a7c-11e9-b6ef-0e71f8c7e888 no longer exists
+[AfterEach] [sig-storage] Projected downwardAPI
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 12:47:09.415: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "projected-1269" for this suite.
+Apr 29 12:47:15.443: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 12:47:15.680: INFO: namespace projected-1269 deletion completed in 6.25824131s
+
+• [SLOW TEST:10.592 seconds]
+[sig-storage] Projected downwardAPI
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:33
+  should provide container's memory request [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSSSSSSSSSSSSSSS
+------------------------------
+[sig-storage] EmptyDir volumes 
+  should support (non-root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-storage] EmptyDir volumes
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 12:47:15.681: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename emptydir
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in emptydir-280
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should support (non-root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating a pod to test emptydir 0666 on tmpfs
+Apr 29 12:47:15.888: INFO: Waiting up to 5m0s for pod "pod-eac38209-6a7c-11e9-b6ef-0e71f8c7e888" in namespace "emptydir-280" to be "success or failure"
+Apr 29 12:47:15.904: INFO: Pod "pod-eac38209-6a7c-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 15.831042ms
+Apr 29 12:47:17.912: INFO: Pod "pod-eac38209-6a7c-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 2.023990297s
+Apr 29 12:47:19.919: INFO: Pod "pod-eac38209-6a7c-11e9-b6ef-0e71f8c7e888": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.030483898s
+STEP: Saw pod success
+Apr 29 12:47:19.919: INFO: Pod "pod-eac38209-6a7c-11e9-b6ef-0e71f8c7e888" satisfied condition "success or failure"
+Apr 29 12:47:19.925: INFO: Trying to get logs from node worker-4av2h-6665d7b5ff-9nc42 pod pod-eac38209-6a7c-11e9-b6ef-0e71f8c7e888 container test-container: 
+STEP: delete the pod
+Apr 29 12:47:19.965: INFO: Waiting for pod pod-eac38209-6a7c-11e9-b6ef-0e71f8c7e888 to disappear
+Apr 29 12:47:19.971: INFO: Pod pod-eac38209-6a7c-11e9-b6ef-0e71f8c7e888 no longer exists
+[AfterEach] [sig-storage] EmptyDir volumes
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 12:47:19.972: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "emptydir-280" for this suite.
+Apr 29 12:47:26.004: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 12:47:26.277: INFO: namespace emptydir-280 deletion completed in 6.29715001s
+
+• [SLOW TEST:10.596 seconds]
+[sig-storage] EmptyDir volumes
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:41
+  should support (non-root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSS
+------------------------------
+[sig-storage] EmptyDir volumes 
+  volume on tmpfs should have the correct mode [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-storage] EmptyDir volumes
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 12:47:26.278: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename emptydir
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in emptydir-7459
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] volume on tmpfs should have the correct mode [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating a pod to test emptydir volume type on tmpfs
+Apr 29 12:47:26.515: INFO: Waiting up to 5m0s for pod "pod-f119ad08-6a7c-11e9-b6ef-0e71f8c7e888" in namespace "emptydir-7459" to be "success or failure"
+Apr 29 12:47:26.522: INFO: Pod "pod-f119ad08-6a7c-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 7.059176ms
+Apr 29 12:47:28.529: INFO: Pod "pod-f119ad08-6a7c-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 2.013830524s
+Apr 29 12:47:30.540: INFO: Pod "pod-f119ad08-6a7c-11e9-b6ef-0e71f8c7e888": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.024894431s
+STEP: Saw pod success
+Apr 29 12:47:30.540: INFO: Pod "pod-f119ad08-6a7c-11e9-b6ef-0e71f8c7e888" satisfied condition "success or failure"
+Apr 29 12:47:30.545: INFO: Trying to get logs from node worker-4av2h-6665d7b5ff-9nc42 pod pod-f119ad08-6a7c-11e9-b6ef-0e71f8c7e888 container test-container: 
+STEP: delete the pod
+Apr 29 12:47:30.598: INFO: Waiting for pod pod-f119ad08-6a7c-11e9-b6ef-0e71f8c7e888 to disappear
+Apr 29 12:47:30.610: INFO: Pod pod-f119ad08-6a7c-11e9-b6ef-0e71f8c7e888 no longer exists
+[AfterEach] [sig-storage] EmptyDir volumes
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 12:47:30.610: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "emptydir-7459" for this suite.
+Apr 29 12:47:36.667: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 12:47:36.892: INFO: namespace emptydir-7459 deletion completed in 6.265194141s
+
+• [SLOW TEST:10.614 seconds]
+[sig-storage] EmptyDir volumes
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:41
+  volume on tmpfs should have the correct mode [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSSSSSSSSSSSS
+------------------------------
+[sig-api-machinery] Namespaces [Serial] 
+  should ensure that all services are removed when a namespace is deleted [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-api-machinery] Namespaces [Serial]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 12:47:36.896: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename namespaces
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in namespaces-8212
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should ensure that all services are removed when a namespace is deleted [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating a test namespace
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in nsdeletetest-6730
+STEP: Waiting for a default service account to be provisioned in namespace
+STEP: Creating a service in the namespace
+STEP: Deleting the namespace
+STEP: Waiting for the namespace to be removed.
+STEP: Recreating the namespace
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in nsdeletetest-6575
+STEP: Verifying there is no service in the namespace
+[AfterEach] [sig-api-machinery] Namespaces [Serial]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 12:47:43.596: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "namespaces-8212" for this suite.
+Apr 29 12:47:49.630: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 12:47:49.936: INFO: namespace namespaces-8212 deletion completed in 6.328920908s
+STEP: Destroying namespace "nsdeletetest-6730" for this suite.
+Apr 29 12:47:49.942: INFO: Namespace nsdeletetest-6730 was already deleted
+STEP: Destroying namespace "nsdeletetest-6575" for this suite.
+Apr 29 12:47:55.987: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 12:47:56.199: INFO: namespace nsdeletetest-6575 deletion completed in 6.256201264s
+
+• [SLOW TEST:19.303 seconds]
+[sig-api-machinery] Namespaces [Serial]
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:22
+  should ensure that all services are removed when a namespace is deleted [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSSS
+------------------------------
+[sig-storage] Secrets 
+  optional updates should be reflected in volume [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-storage] Secrets
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 12:47:56.199: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename secrets
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in secrets-8501
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] optional updates should be reflected in volume [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating secret with name s-test-opt-del-02f1c38c-6a7d-11e9-b6ef-0e71f8c7e888
+STEP: Creating secret with name s-test-opt-upd-02f1c404-6a7d-11e9-b6ef-0e71f8c7e888
+STEP: Creating the pod
+STEP: Deleting secret s-test-opt-del-02f1c38c-6a7d-11e9-b6ef-0e71f8c7e888
+STEP: Updating secret s-test-opt-upd-02f1c404-6a7d-11e9-b6ef-0e71f8c7e888
+STEP: Creating secret with name s-test-opt-create-02f1c43a-6a7d-11e9-b6ef-0e71f8c7e888
+STEP: waiting to observe update in volume
+[AfterEach] [sig-storage] Secrets
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 12:48:04.647: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "secrets-8501" for this suite.
+Apr 29 12:48:28.677: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 12:48:28.899: INFO: namespace secrets-8501 deletion completed in 24.246391481s
+
+• [SLOW TEST:32.700 seconds]
+[sig-storage] Secrets
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/secrets_volume.go:33
+  optional updates should be reflected in volume [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+[sig-network] Services 
+  should serve multiport endpoints from pods  [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-network] Services
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 12:48:28.900: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename services
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in services-2305
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [sig-network] Services
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:86
+[It] should serve multiport endpoints from pods  [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: creating service multi-endpoint-test in namespace services-2305
+STEP: waiting up to 3m0s for service multi-endpoint-test in namespace services-2305 to expose endpoints map[]
+Apr 29 12:48:29.189: INFO: successfully validated that service multi-endpoint-test in namespace services-2305 exposes endpoints map[] (12.793235ms elapsed)
+STEP: Creating pod pod1 in namespace services-2305
+STEP: waiting up to 3m0s for service multi-endpoint-test in namespace services-2305 to expose endpoints map[pod1:[100]]
+Apr 29 12:48:33.316: INFO: successfully validated that service multi-endpoint-test in namespace services-2305 exposes endpoints map[pod1:[100]] (4.103779395s elapsed)
+STEP: Creating pod pod2 in namespace services-2305
+STEP: waiting up to 3m0s for service multi-endpoint-test in namespace services-2305 to expose endpoints map[pod1:[100] pod2:[101]]
+Apr 29 12:48:36.426: INFO: successfully validated that service multi-endpoint-test in namespace services-2305 exposes endpoints map[pod1:[100] pod2:[101]] (3.093654747s elapsed)
+STEP: Deleting pod pod1 in namespace services-2305
+STEP: waiting up to 3m0s for service multi-endpoint-test in namespace services-2305 to expose endpoints map[pod2:[101]]
+Apr 29 12:48:36.482: INFO: successfully validated that service multi-endpoint-test in namespace services-2305 exposes endpoints map[pod2:[101]] (43.228717ms elapsed)
+STEP: Deleting pod pod2 in namespace services-2305
+STEP: waiting up to 3m0s for service multi-endpoint-test in namespace services-2305 to expose endpoints map[]
+Apr 29 12:48:36.538: INFO: successfully validated that service multi-endpoint-test in namespace services-2305 exposes endpoints map[] (20.986097ms elapsed)
+[AfterEach] [sig-network] Services
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 12:48:36.603: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "services-2305" for this suite.
+Apr 29 12:49:00.663: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 12:49:00.892: INFO: namespace services-2305 deletion completed in 24.269134918s
+[AfterEach] [sig-network] Services
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:91
+
+• [SLOW TEST:31.992 seconds]
+[sig-network] Services
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:22
+  should serve multiport endpoints from pods  [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSSSSSSSSSSSSSS
+------------------------------
+[sig-apps] Deployment 
+  RecreateDeployment should delete old pods and create new ones [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-apps] Deployment
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 12:49:00.895: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename deployment
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in deployment-3715
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [sig-apps] Deployment
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:65
+[It] RecreateDeployment should delete old pods and create new ones [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+Apr 29 12:49:01.123: INFO: Creating deployment "test-recreate-deployment"
+Apr 29 12:49:01.136: INFO: Waiting deployment "test-recreate-deployment" to be updated to revision 1
+Apr 29 12:49:01.161: INFO: new replicaset for deployment "test-recreate-deployment" is yet to be created
+Apr 29 12:49:03.182: INFO: Waiting deployment "test-recreate-deployment" to complete
+Apr 29 12:49:03.187: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63692138940, loc:(*time.Location)(0x8a060e0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63692138940, loc:(*time.Location)(0x8a060e0)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63692138940, loc:(*time.Location)(0x8a060e0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63692138940, loc:(*time.Location)(0x8a060e0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-recreate-deployment-7d57d5ff7c\" is progressing."}}, CollisionCount:(*int32)(nil)}
+Apr 29 12:49:05.195: INFO: Triggering a new rollout for deployment "test-recreate-deployment"
+Apr 29 12:49:05.209: INFO: Updating deployment test-recreate-deployment
+Apr 29 12:49:05.209: INFO: Watching deployment "test-recreate-deployment" to verify that new pods will not run with olds pods
+[AfterEach] [sig-apps] Deployment
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:59
+Apr 29 12:49:05.388: INFO: Deployment "test-recreate-deployment":
+&Deployment{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-recreate-deployment,GenerateName:,Namespace:deployment-3715,SelfLink:/apis/apps/v1/namespaces/deployment-3715/deployments/test-recreate-deployment,UID:294edfea-6a7d-11e9-92a7-deadbef3c36d,ResourceVersion:23519,Generation:2,CreationTimestamp:2019-04-29 12:49:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod-3,},Annotations:map[string]string{deployment.kubernetes.io/revision: 2,},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:DeploymentSpec{Replicas:*1,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod-3,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod-3,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] []  [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,},},Strategy:DeploymentStrategy{Type:Recreate,RollingUpdate:nil,},MinReadySeconds:0,RevisionHistoryLimit:*10,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:2,Replicas:1,UpdatedReplicas:1,AvailableReplicas:0,UnavailableReplicas:1,Conditions:[{Available False 2019-04-29 12:49:05 +0000 UTC 2019-04-29 12:49:05 +0000 UTC MinimumReplicasUnavailable Deployment does not have minimum availability.} {Progressing True 2019-04-29 12:49:05 +0000 UTC 2019-04-29 12:49:00 +0000 UTC ReplicaSetUpdated ReplicaSet "test-recreate-deployment-c9cbd8684" is progressing.}],ReadyReplicas:0,CollisionCount:nil,},}
+
+Apr 29 12:49:05.397: INFO: New ReplicaSet "test-recreate-deployment-c9cbd8684" of Deployment "test-recreate-deployment":
+&ReplicaSet{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-recreate-deployment-c9cbd8684,GenerateName:,Namespace:deployment-3715,SelfLink:/apis/apps/v1/namespaces/deployment-3715/replicasets/test-recreate-deployment-c9cbd8684,UID:2bcc7c3f-6a7d-11e9-92a7-deadbef3c36d,ResourceVersion:23518,Generation:1,CreationTimestamp:2019-04-29 12:49:04 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod-3,pod-template-hash: c9cbd8684,},Annotations:map[string]string{deployment.kubernetes.io/desired-replicas: 1,deployment.kubernetes.io/max-replicas: 1,deployment.kubernetes.io/revision: 2,},OwnerReferences:[{apps/v1 Deployment test-recreate-deployment 294edfea-6a7d-11e9-92a7-deadbef3c36d 0xc001d35cd0 0xc001d35cd1}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:ReplicaSetSpec{Replicas:*1,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod-3,pod-template-hash: c9cbd8684,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod-3,pod-template-hash: c9cbd8684,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] []  [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,},},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:1,FullyLabeledReplicas:1,ObservedGeneration:1,ReadyReplicas:0,AvailableReplicas:0,Conditions:[],},}
+Apr 29 12:49:05.397: INFO: All old ReplicaSets of Deployment "test-recreate-deployment":
+Apr 29 12:49:05.398: INFO: &ReplicaSet{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-recreate-deployment-7d57d5ff7c,GenerateName:,Namespace:deployment-3715,SelfLink:/apis/apps/v1/namespaces/deployment-3715/replicasets/test-recreate-deployment-7d57d5ff7c,UID:29527501-6a7d-11e9-92a7-deadbef3c36d,ResourceVersion:23509,Generation:2,CreationTimestamp:2019-04-29 12:49:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod-3,pod-template-hash: 7d57d5ff7c,},Annotations:map[string]string{deployment.kubernetes.io/desired-replicas: 1,deployment.kubernetes.io/max-replicas: 1,deployment.kubernetes.io/revision: 1,},OwnerReferences:[{apps/v1 Deployment test-recreate-deployment 294edfea-6a7d-11e9-92a7-deadbef3c36d 0xc001d35c07 0xc001d35c08}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:ReplicaSetSpec{Replicas:*0,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod-3,pod-template-hash: 7d57d5ff7c,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod-3,pod-template-hash: 7d57d5ff7c,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[],Containers:[{redis gcr.io/kubernetes-e2e-test-images/redis:1.0 [] []  [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,},},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:0,FullyLabeledReplicas:0,ObservedGeneration:2,ReadyReplicas:0,AvailableReplicas:0,Conditions:[],},}
+Apr 29 12:49:05.404: INFO: Pod "test-recreate-deployment-c9cbd8684-sshr2" is not available:
+&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-recreate-deployment-c9cbd8684-sshr2,GenerateName:test-recreate-deployment-c9cbd8684-,Namespace:deployment-3715,SelfLink:/api/v1/namespaces/deployment-3715/pods/test-recreate-deployment-c9cbd8684-sshr2,UID:2bce2e9e-6a7d-11e9-92a7-deadbef3c36d,ResourceVersion:23520,Generation:0,CreationTimestamp:2019-04-29 12:49:04 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod-3,pod-template-hash: c9cbd8684,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet test-recreate-deployment-c9cbd8684 2bcc7c3f-6a7d-11e9-92a7-deadbef3c36d 0xc002dd4550 0xc002dd4551}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-z8cx2 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-z8cx2,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] []  [] [] [] {map[] map[]} [{default-token-z8cx2 true /var/run/secrets/kubernetes.io/serviceaccount   }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:worker-4av2h-6665d7b5ff-9nc42,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists  NoExecute 0xc002dd45b0} {node.kubernetes.io/unreachable Exists  NoExecute 0xc002dd45d0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 12:49:05 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-04-29 12:49:05 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-04-29 12:49:05 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 12:49:05 +0000 UTC  }],Message:,Reason:,HostIP:172.23.0.238,PodIP:,StartTime:2019-04-29 12:49:05 +0000 UTC,ContainerStatuses:[{nginx {ContainerStateWaiting{Reason:ContainerCreating,Message:,} nil nil} {nil nil nil} false 0 docker.io/library/nginx:1.14-alpine  }],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},}
+[AfterEach] [sig-apps] Deployment
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 12:49:05.404: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "deployment-3715" for this suite.
+Apr 29 12:49:11.442: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 12:49:11.702: INFO: namespace deployment-3715 deletion completed in 6.28369579s
+
+• [SLOW TEST:10.807 seconds]
+[sig-apps] Deployment
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22
+  RecreateDeployment should delete old pods and create new ones [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSSSSSSSSSSSSSS
+------------------------------
+[sig-storage] Projected configMap 
+  updates should be reflected in volume [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-storage] Projected configMap
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 12:49:11.702: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename projected
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-9708
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] updates should be reflected in volume [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating projection with configMap that has name projected-configmap-test-upd-2ff38399-6a7d-11e9-b6ef-0e71f8c7e888
+STEP: Creating the pod
+STEP: Updating configmap projected-configmap-test-upd-2ff38399-6a7d-11e9-b6ef-0e71f8c7e888
+STEP: waiting to observe update in volume
+[AfterEach] [sig-storage] Projected configMap
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 12:50:24.780: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "projected-9708" for this suite.
+Apr 29 12:50:48.826: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 12:50:48.999: INFO: namespace projected-9708 deletion completed in 24.196353489s
+
+• [SLOW TEST:97.297 seconds]
+[sig-storage] Projected configMap
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_configmap.go:33
+  updates should be reflected in volume [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+[sig-storage] EmptyDir volumes 
+  should support (root,0666,default) [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-storage] EmptyDir volumes
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 12:50:48.999: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename emptydir
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in emptydir-7831
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should support (root,0666,default) [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating a pod to test emptydir 0666 on node default medium
+Apr 29 12:50:49.210: INFO: Waiting up to 5m0s for pod "pod-69ea973d-6a7d-11e9-b6ef-0e71f8c7e888" in namespace "emptydir-7831" to be "success or failure"
+Apr 29 12:50:49.221: INFO: Pod "pod-69ea973d-6a7d-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 10.300496ms
+Apr 29 12:50:51.229: INFO: Pod "pod-69ea973d-6a7d-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 2.018347505s
+Apr 29 12:50:53.236: INFO: Pod "pod-69ea973d-6a7d-11e9-b6ef-0e71f8c7e888": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.025027951s
+STEP: Saw pod success
+Apr 29 12:50:53.236: INFO: Pod "pod-69ea973d-6a7d-11e9-b6ef-0e71f8c7e888" satisfied condition "success or failure"
+Apr 29 12:50:53.241: INFO: Trying to get logs from node worker-4av2h-6665d7b5ff-9nc42 pod pod-69ea973d-6a7d-11e9-b6ef-0e71f8c7e888 container test-container: 
+STEP: delete the pod
+Apr 29 12:50:53.281: INFO: Waiting for pod pod-69ea973d-6a7d-11e9-b6ef-0e71f8c7e888 to disappear
+Apr 29 12:50:53.288: INFO: Pod pod-69ea973d-6a7d-11e9-b6ef-0e71f8c7e888 no longer exists
+[AfterEach] [sig-storage] EmptyDir volumes
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 12:50:53.288: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "emptydir-7831" for this suite.
+Apr 29 12:50:59.316: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 12:50:59.547: INFO: namespace emptydir-7831 deletion completed in 6.252481727s
+
+• [SLOW TEST:10.548 seconds]
+[sig-storage] EmptyDir volumes
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:41
+  should support (root,0666,default) [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SS
+------------------------------
+[k8s.io] [sig-node] Events 
+  should be sent by kubelets and the scheduler about pods scheduling and running  [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [k8s.io] [sig-node] Events
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 12:50:59.548: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename events
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in events-9621
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should be sent by kubelets and the scheduler about pods scheduling and running  [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: creating the pod
+STEP: submitting the pod to kubernetes
+STEP: verifying the pod is in kubernetes
+STEP: retrieving the pod
+Apr 29 12:51:03.837: INFO: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:send-events-7037c8bb-6a7d-11e9-b6ef-0e71f8c7e888,GenerateName:,Namespace:events-9621,SelfLink:/api/v1/namespaces/events-9621/pods/send-events-7037c8bb-6a7d-11e9-b6ef-0e71f8c7e888,UID:7006cc4d-6a7d-11e9-92a7-deadbef3c36d,ResourceVersion:23852,Generation:0,CreationTimestamp:2019-04-29 12:50:59 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: foo,time: 767115625,},Annotations:map[string]string{kubernetes.io/psp: cert-exporter-psp,},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-rh6dj {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-rh6dj,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{p gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1 [] []  [{ 0 80 TCP }] [] [] {map[] map[]} [{default-token-rh6dj true /var/run/secrets/kubernetes.io/serviceaccount   }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*30,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:worker-4av2h-6665d7b5ff-9nc42,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists  NoExecute 0xc002d06240} {node.kubernetes.io/unreachable Exists  NoExecute 0xc002d06260}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Running,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 12:50:59 +0000 UTC  } {Ready True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 12:51:02 +0000 UTC  } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 12:51:02 +0000 UTC  } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 12:50:59 +0000 UTC  }],Message:,Reason:,HostIP:172.23.0.238,PodIP:192.168.205.81,StartTime:2019-04-29 12:50:59 +0000 UTC,ContainerStatuses:[{p {nil ContainerStateRunning{StartedAt:2019-04-29 12:51:01 +0000 UTC,} nil} {nil nil nil} true 0 gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1 docker-pullable://gcr.io/kubernetes-e2e-test-images/serve-hostname@sha256:bab70473a6d8ef65a22625dc9a1b0f0452e811530fdbe77e4408523460177ff1 docker://4e80ed06d0957ffd22abb3c1367895b5d896d9edf9622668fb44718bbd5d18d6}],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},}
+
+STEP: checking for scheduler event about the pod
+Apr 29 12:51:05.844: INFO: Saw scheduler event for our pod.
+STEP: checking for kubelet event about the pod
+Apr 29 12:51:07.852: INFO: Saw kubelet event for our pod.
+STEP: deleting the pod
+[AfterEach] [k8s.io] [sig-node] Events
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 12:51:07.869: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "events-9621" for this suite.
+Apr 29 12:51:55.928: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 12:51:56.160: INFO: namespace events-9621 deletion completed in 48.281262597s
+
+• [SLOW TEST:56.621 seconds]
+[k8s.io] [sig-node] Events
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
+  should be sent by kubelets and the scheduler about pods scheduling and running  [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSSSSSSSSSSS
+------------------------------
+[sig-apps] ReplicationController 
+  should release no longer matching pods [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-apps] ReplicationController
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 12:51:56.170: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename replication-controller
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in replication-controller-263
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should release no longer matching pods [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Given a ReplicationController is created
+STEP: When the matched label of one of its pods change
+Apr 29 12:51:56.406: INFO: Pod name pod-release: Found 0 pods out of 1
+Apr 29 12:52:01.413: INFO: Pod name pod-release: Found 1 pods out of 1
+STEP: Then the pod is released
+[AfterEach] [sig-apps] ReplicationController
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 12:52:01.455: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "replication-controller-263" for this suite.
+Apr 29 12:52:07.517: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 12:52:07.723: INFO: namespace replication-controller-263 deletion completed in 6.252862393s
+
+• [SLOW TEST:11.553 seconds]
+[sig-apps] ReplicationController
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22
+  should release no longer matching pods [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSS
+------------------------------
+[sig-api-machinery] Garbage collector 
+  should delete pods created by rc when not orphaning [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-api-machinery] Garbage collector
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 12:52:07.724: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename gc
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in gc-6608
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should delete pods created by rc when not orphaning [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: create the rc
+STEP: delete the rc
+STEP: wait for all pods to be garbage collected
+STEP: Gathering metrics
+W0429 12:52:17.978045      15 metrics_grabber.go:79] Master node is not registered. Grabbing metrics from Scheduler, ControllerManager and ClusterAutoscaler is disabled.
+Apr 29 12:52:17.978: INFO: For apiserver_request_total:
+For apiserver_request_latencies_summary:
+For apiserver_init_events_total:
+For garbage_collector_attempt_to_delete_queue_latency:
+For garbage_collector_attempt_to_delete_work_duration:
+For garbage_collector_attempt_to_orphan_queue_latency:
+For garbage_collector_attempt_to_orphan_work_duration:
+For garbage_collector_dirty_processing_latency_microseconds:
+For garbage_collector_event_processing_latency_microseconds:
+For garbage_collector_graph_changes_queue_latency:
+For garbage_collector_graph_changes_work_duration:
+For garbage_collector_orphan_processing_latency_microseconds:
+For namespace_queue_latency:
+For namespace_queue_latency_sum:
+For namespace_queue_latency_count:
+For namespace_retries:
+For namespace_work_duration:
+For namespace_work_duration_sum:
+For namespace_work_duration_count:
+For function_duration_seconds:
+For errors_total:
+For evicted_pods_total:
+
+[AfterEach] [sig-api-machinery] Garbage collector
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 12:52:17.978: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "gc-6608" for this suite.
+Apr 29 12:52:24.012: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 12:52:24.235: INFO: namespace gc-6608 deletion completed in 6.249027244s
+
+• [SLOW TEST:16.512 seconds]
+[sig-api-machinery] Garbage collector
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:22
+  should delete pods created by rc when not orphaning [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+S
+------------------------------
+[sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] 
+  Should recreate evicted statefulset [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-apps] StatefulSet
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 12:52:24.236: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename statefulset
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in statefulset-9747
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [sig-apps] StatefulSet
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:59
+[BeforeEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:74
+STEP: Creating service test in namespace statefulset-9747
+[It] Should recreate evicted statefulset [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Looking for a node to schedule stateful set and pod
+STEP: Creating pod with conflicting port in namespace statefulset-9747
+STEP: Creating statefulset with conflicting port in namespace statefulset-9747
+STEP: Waiting until pod test-pod will start running in namespace statefulset-9747
+STEP: Waiting until stateful pod ss-0 will be recreated and deleted at least once in namespace statefulset-9747
+Apr 29 12:52:28.575: INFO: Observed stateful pod in namespace: statefulset-9747, name: ss-0, uid: a4d3c7a2-6a7d-11e9-92a7-deadbef3c36d, status phase: Pending. Waiting for statefulset controller to delete.
+Apr 29 12:52:28.929: INFO: Observed stateful pod in namespace: statefulset-9747, name: ss-0, uid: a4d3c7a2-6a7d-11e9-92a7-deadbef3c36d, status phase: Failed. Waiting for statefulset controller to delete.
+Apr 29 12:52:28.961: INFO: Observed stateful pod in namespace: statefulset-9747, name: ss-0, uid: a4d3c7a2-6a7d-11e9-92a7-deadbef3c36d, status phase: Failed. Waiting for statefulset controller to delete.
+Apr 29 12:52:28.973: INFO: Observed delete event for stateful pod ss-0 in namespace statefulset-9747
+STEP: Removing pod with conflicting port in namespace statefulset-9747
+STEP: Waiting when stateful pod ss-0 will be recreated in namespace statefulset-9747 and will be in running state
+[AfterEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:85
+Apr 29 12:52:35.099: INFO: Deleting all statefulset in ns statefulset-9747
+Apr 29 12:52:35.106: INFO: Scaling statefulset ss to 0
+Apr 29 12:52:55.138: INFO: Waiting for statefulset status.replicas updated to 0
+Apr 29 12:52:55.147: INFO: Deleting statefulset ss
+[AfterEach] [sig-apps] StatefulSet
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 12:52:55.192: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "statefulset-9747" for this suite.
+Apr 29 12:53:01.227: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 12:53:01.494: INFO: namespace statefulset-9747 deletion completed in 6.28818708s
+
+• [SLOW TEST:37.258 seconds]
+[sig-apps] StatefulSet
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22
+  [k8s.io] Basic StatefulSet functionality [StatefulSetBasic]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
+    Should recreate evicted statefulset [Conformance]
+    /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSS
+------------------------------
+[sig-apps] Deployment 
+  deployment should support rollover [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-apps] Deployment
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 12:53:01.494: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename deployment
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in deployment-7949
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [sig-apps] Deployment
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:65
+[It] deployment should support rollover [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+Apr 29 12:53:01.778: INFO: Pod name rollover-pod: Found 0 pods out of 1
+Apr 29 12:53:06.786: INFO: Pod name rollover-pod: Found 1 pods out of 1
+STEP: ensuring each pod is running
+Apr 29 12:53:06.786: INFO: Waiting for pods owned by replica set "test-rollover-controller" to become ready
+Apr 29 12:53:08.793: INFO: Creating deployment "test-rollover-deployment"
+Apr 29 12:53:08.813: INFO: Make sure deployment "test-rollover-deployment" performs scaling operations
+Apr 29 12:53:10.827: INFO: Check revision of new replica set for deployment "test-rollover-deployment"
+Apr 29 12:53:10.837: INFO: Ensure that both replica sets have 1 created replica
+Apr 29 12:53:10.846: INFO: Rollover old replica sets for deployment "test-rollover-deployment" with new image update
+Apr 29 12:53:10.880: INFO: Updating deployment test-rollover-deployment
+Apr 29 12:53:10.880: INFO: Wait deployment "test-rollover-deployment" to be observed by the deployment controller
+Apr 29 12:53:12.917: INFO: Wait for revision update of deployment "test-rollover-deployment" to 2
+Apr 29 12:53:12.934: INFO: Make sure deployment "test-rollover-deployment" is complete
+Apr 29 12:53:12.947: INFO: all replica sets need to contain the pod-template-hash label
+Apr 29 12:53:12.948: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:1, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63692139188, loc:(*time.Location)(0x8a060e0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63692139188, loc:(*time.Location)(0x8a060e0)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63692139190, loc:(*time.Location)(0x8a060e0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63692139188, loc:(*time.Location)(0x8a060e0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-766b4d6c9d\" is progressing."}}, CollisionCount:(*int32)(nil)}
+Apr 29 12:53:14.959: INFO: all replica sets need to contain the pod-template-hash label
+Apr 29 12:53:14.959: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:2, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63692139188, loc:(*time.Location)(0x8a060e0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63692139188, loc:(*time.Location)(0x8a060e0)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63692139193, loc:(*time.Location)(0x8a060e0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63692139188, loc:(*time.Location)(0x8a060e0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-766b4d6c9d\" is progressing."}}, CollisionCount:(*int32)(nil)}
+Apr 29 12:53:16.962: INFO: all replica sets need to contain the pod-template-hash label
+Apr 29 12:53:16.962: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:2, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63692139188, loc:(*time.Location)(0x8a060e0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63692139188, loc:(*time.Location)(0x8a060e0)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63692139193, loc:(*time.Location)(0x8a060e0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63692139188, loc:(*time.Location)(0x8a060e0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-766b4d6c9d\" is progressing."}}, CollisionCount:(*int32)(nil)}
+Apr 29 12:53:18.962: INFO: all replica sets need to contain the pod-template-hash label
+Apr 29 12:53:18.962: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:2, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63692139188, loc:(*time.Location)(0x8a060e0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63692139188, loc:(*time.Location)(0x8a060e0)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63692139193, loc:(*time.Location)(0x8a060e0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63692139188, loc:(*time.Location)(0x8a060e0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-766b4d6c9d\" is progressing."}}, CollisionCount:(*int32)(nil)}
+Apr 29 12:53:20.960: INFO: all replica sets need to contain the pod-template-hash label
+Apr 29 12:53:20.960: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:2, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63692139188, loc:(*time.Location)(0x8a060e0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63692139188, loc:(*time.Location)(0x8a060e0)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63692139193, loc:(*time.Location)(0x8a060e0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63692139188, loc:(*time.Location)(0x8a060e0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-766b4d6c9d\" is progressing."}}, CollisionCount:(*int32)(nil)}
+Apr 29 12:53:22.959: INFO: all replica sets need to contain the pod-template-hash label
+Apr 29 12:53:22.959: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:2, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63692139188, loc:(*time.Location)(0x8a060e0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63692139188, loc:(*time.Location)(0x8a060e0)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63692139193, loc:(*time.Location)(0x8a060e0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63692139188, loc:(*time.Location)(0x8a060e0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-766b4d6c9d\" is progressing."}}, CollisionCount:(*int32)(nil)}
+Apr 29 12:53:24.959: INFO: 
+Apr 29 12:53:24.960: INFO: Ensure that both old replica sets have no replicas
+[AfterEach] [sig-apps] Deployment
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:59
+Apr 29 12:53:24.971: INFO: Deployment "test-rollover-deployment":
+&Deployment{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-rollover-deployment,GenerateName:,Namespace:deployment-7949,SelfLink:/apis/apps/v1/namespaces/deployment-7949/deployments/test-rollover-deployment,UID:bcedd257-6a7d-11e9-92a7-deadbef3c36d,ResourceVersion:24502,Generation:2,CreationTimestamp:2019-04-29 12:53:08 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: rollover-pod,},Annotations:map[string]string{deployment.kubernetes.io/revision: 2,},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:DeploymentSpec{Replicas:*1,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: rollover-pod,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: rollover-pod,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[],Containers:[{redis gcr.io/kubernetes-e2e-test-images/redis:1.0 [] []  [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,},},Strategy:DeploymentStrategy{Type:RollingUpdate,RollingUpdate:&RollingUpdateDeployment{MaxUnavailable:0,MaxSurge:1,},},MinReadySeconds:10,RevisionHistoryLimit:*10,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:2,Replicas:1,UpdatedReplicas:1,AvailableReplicas:1,UnavailableReplicas:0,Conditions:[{Available True 2019-04-29 12:53:08 +0000 UTC 2019-04-29 12:53:08 +0000 UTC MinimumReplicasAvailable Deployment has minimum availability.} {Progressing True 2019-04-29 12:53:23 +0000 UTC 2019-04-29 12:53:08 +0000 UTC NewReplicaSetAvailable ReplicaSet "test-rollover-deployment-766b4d6c9d" has successfully progressed.}],ReadyReplicas:1,CollisionCount:nil,},}
+
+Apr 29 12:53:24.975: INFO: New ReplicaSet "test-rollover-deployment-766b4d6c9d" of Deployment "test-rollover-deployment":
+&ReplicaSet{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-rollover-deployment-766b4d6c9d,GenerateName:,Namespace:deployment-7949,SelfLink:/apis/apps/v1/namespaces/deployment-7949/replicasets/test-rollover-deployment-766b4d6c9d,UID:be2b5abf-6a7d-11e9-92a7-deadbef3c36d,ResourceVersion:24490,Generation:2,CreationTimestamp:2019-04-29 12:53:10 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: rollover-pod,pod-template-hash: 766b4d6c9d,},Annotations:map[string]string{deployment.kubernetes.io/desired-replicas: 1,deployment.kubernetes.io/max-replicas: 2,deployment.kubernetes.io/revision: 2,},OwnerReferences:[{apps/v1 Deployment test-rollover-deployment bcedd257-6a7d-11e9-92a7-deadbef3c36d 0xc002dd5d37 0xc002dd5d38}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:ReplicaSetSpec{Replicas:*1,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: rollover-pod,pod-template-hash: 766b4d6c9d,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: rollover-pod,pod-template-hash: 766b4d6c9d,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[],Containers:[{redis gcr.io/kubernetes-e2e-test-images/redis:1.0 [] []  [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,},},MinReadySeconds:10,},Status:ReplicaSetStatus{Replicas:1,FullyLabeledReplicas:1,ObservedGeneration:2,ReadyReplicas:1,AvailableReplicas:1,Conditions:[],},}
+Apr 29 12:53:24.975: INFO: All old ReplicaSets of Deployment "test-rollover-deployment":
+Apr 29 12:53:24.976: INFO: &ReplicaSet{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-rollover-controller,GenerateName:,Namespace:deployment-7949,SelfLink:/apis/apps/v1/namespaces/deployment-7949/replicasets/test-rollover-controller,UID:b8bbf117-6a7d-11e9-92a7-deadbef3c36d,ResourceVersion:24499,Generation:2,CreationTimestamp:2019-04-29 12:53:01 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: rollover-pod,pod: nginx,},Annotations:map[string]string{deployment.kubernetes.io/desired-replicas: 1,deployment.kubernetes.io/max-replicas: 2,},OwnerReferences:[{apps/v1 Deployment test-rollover-deployment bcedd257-6a7d-11e9-92a7-deadbef3c36d 0xc002dd5b87 0xc002dd5b88}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:ReplicaSetSpec{Replicas:*0,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: rollover-pod,pod: nginx,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: rollover-pod,pod: nginx,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] []  [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,},},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:0,FullyLabeledReplicas:0,ObservedGeneration:2,ReadyReplicas:0,AvailableReplicas:0,Conditions:[],},}
+Apr 29 12:53:24.977: INFO: &ReplicaSet{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-rollover-deployment-6455657675,GenerateName:,Namespace:deployment-7949,SelfLink:/apis/apps/v1/namespaces/deployment-7949/replicasets/test-rollover-deployment-6455657675,UID:bcf5d99c-6a7d-11e9-92a7-deadbef3c36d,ResourceVersion:24451,Generation:2,CreationTimestamp:2019-04-29 12:53:08 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: rollover-pod,pod-template-hash: 6455657675,},Annotations:map[string]string{deployment.kubernetes.io/desired-replicas: 1,deployment.kubernetes.io/max-replicas: 2,deployment.kubernetes.io/revision: 1,},OwnerReferences:[{apps/v1 Deployment test-rollover-deployment bcedd257-6a7d-11e9-92a7-deadbef3c36d 0xc002dd5c57 0xc002dd5c58}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:ReplicaSetSpec{Replicas:*0,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: rollover-pod,pod-template-hash: 6455657675,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: rollover-pod,pod-template-hash: 6455657675,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[],Containers:[{redis-slave gcr.io/google_samples/gb-redisslave:nonexistent [] []  [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,},},MinReadySeconds:10,},Status:ReplicaSetStatus{Replicas:0,FullyLabeledReplicas:0,ObservedGeneration:2,ReadyReplicas:0,AvailableReplicas:0,Conditions:[],},}
+Apr 29 12:53:24.981: INFO: Pod "test-rollover-deployment-766b4d6c9d-lbkcd" is available:
+&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-rollover-deployment-766b4d6c9d-lbkcd,GenerateName:test-rollover-deployment-766b4d6c9d-,Namespace:deployment-7949,SelfLink:/api/v1/namespaces/deployment-7949/pods/test-rollover-deployment-766b4d6c9d-lbkcd,UID:be3b4826-6a7d-11e9-92a7-deadbef3c36d,ResourceVersion:24470,Generation:0,CreationTimestamp:2019-04-29 12:53:10 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: rollover-pod,pod-template-hash: 766b4d6c9d,},Annotations:map[string]string{kubernetes.io/psp: e2e-test-privileged-psp,},OwnerReferences:[{apps/v1 ReplicaSet test-rollover-deployment-766b4d6c9d be2b5abf-6a7d-11e9-92a7-deadbef3c36d 0xc002e62fa7 0xc002e62fa8}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-2w8fx {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-2w8fx,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{redis gcr.io/kubernetes-e2e-test-images/redis:1.0 [] []  [] [] [] {map[] map[]} [{default-token-2w8fx true /var/run/secrets/kubernetes.io/serviceaccount   }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:worker-4av2h-6665d7b5ff-9nc42,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists  NoExecute 0xc002e63070} {node.kubernetes.io/unreachable Exists  NoExecute 0xc002e630e0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Running,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 12:53:11 +0000 UTC  } {Ready True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 12:53:13 +0000 UTC  } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 12:53:13 +0000 UTC  } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 12:53:10 +0000 UTC  }],Message:,Reason:,HostIP:172.23.0.238,PodIP:192.168.205.88,StartTime:2019-04-29 12:53:11 +0000 UTC,ContainerStatuses:[{redis {nil ContainerStateRunning{StartedAt:2019-04-29 12:53:13 +0000 UTC,} nil} {nil nil nil} true 0 gcr.io/kubernetes-e2e-test-images/redis:1.0 docker-pullable://gcr.io/kubernetes-e2e-test-images/redis@sha256:af4748d1655c08dc54d4be5182135395db9ce87aba2d4699b26b14ae197c5830 docker://5ca1161744390df4dc1da0defe621254b8adfd19f81e8871d7beb36f64b6b1a1}],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},}
+[AfterEach] [sig-apps] Deployment
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 12:53:24.981: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "deployment-7949" for this suite.
+Apr 29 12:53:33.005: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 12:53:33.213: INFO: namespace deployment-7949 deletion completed in 8.22588234s
+
+• [SLOW TEST:31.720 seconds]
+[sig-apps] Deployment
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22
+  deployment should support rollover [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSSSSSSSSSSSS
+------------------------------
+[sig-network] DNS 
+  should provide DNS for the cluster  [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-network] DNS
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 12:53:33.215: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename dns
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in dns-1954
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should provide DNS for the cluster  [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Running these commands on wheezy: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search kubernetes.default.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_udp@kubernetes.default.svc.cluster.local;check="$$(dig +tcp +noall +answer +search kubernetes.default.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@kubernetes.default.svc.cluster.local;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-1954.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@PodARecord;sleep 1; done
+
+STEP: Running these commands on jessie: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search kubernetes.default.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_udp@kubernetes.default.svc.cluster.local;check="$$(dig +tcp +noall +answer +search kubernetes.default.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_tcp@kubernetes.default.svc.cluster.local;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-1954.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_tcp@PodARecord;sleep 1; done
+
+STEP: creating a pod to probe DNS
+STEP: submitting the pod to kubernetes
+STEP: retrieving the pod
+STEP: looking for the results for each expected name from probers
+Apr 29 12:53:49.511: INFO: DNS probes using dns-1954/dns-test-cbc9f5d1-6a7d-11e9-b6ef-0e71f8c7e888 succeeded
+
+STEP: deleting the pod
+[AfterEach] [sig-network] DNS
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 12:53:49.569: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "dns-1954" for this suite.
+Apr 29 12:53:55.597: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 12:53:55.854: INFO: namespace dns-1954 deletion completed in 6.275799482s
+
+• [SLOW TEST:22.640 seconds]
+[sig-network] DNS
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:22
+  should provide DNS for the cluster  [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSS
+------------------------------
+[k8s.io] Probing container 
+  should *not* be restarted with a /healthz http liveness probe [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [k8s.io] Probing container
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 12:53:55.858: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename container-probe
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in container-probe-8638
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [k8s.io] Probing container
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/container_probe.go:51
+[It] should *not* be restarted with a /healthz http liveness probe [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating pod liveness-http in namespace container-probe-8638
+Apr 29 12:54:00.114: INFO: Started pod liveness-http in namespace container-probe-8638
+STEP: checking the pod's current state and verifying that restartCount is present
+Apr 29 12:54:00.123: INFO: Initial restart count of pod liveness-http is 0
+STEP: deleting the pod
+[AfterEach] [k8s.io] Probing container
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 12:58:01.197: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "container-probe-8638" for this suite.
+Apr 29 12:58:07.238: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 12:58:07.519: INFO: namespace container-probe-8638 deletion completed in 6.309047031s
+
+• [SLOW TEST:251.662 seconds]
+[k8s.io] Probing container
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
+  should *not* be restarted with a /healthz http liveness probe [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+------------------------------
+[sig-apps] Daemon set [Serial] 
+  should run and stop simple daemon [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-apps] Daemon set [Serial]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 12:58:07.531: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename daemonsets
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in daemonsets-9228
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [sig-apps] Daemon set [Serial]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:102
+[It] should run and stop simple daemon [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating simple DaemonSet "daemon-set"
+STEP: Check that daemon pods launch on every node of the cluster.
+Apr 29 12:58:07.806: INFO: DaemonSet pods can't tolerate node master-h2h3g-655686fb8c-8kgdn with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node
+Apr 29 12:58:07.811: INFO: Number of nodes with available pods: 0
+Apr 29 12:58:07.811: INFO: Node worker-3oo4n-86cbf586b7-9q98g is running more than one daemon pod
+Apr 29 12:58:08.822: INFO: DaemonSet pods can't tolerate node master-h2h3g-655686fb8c-8kgdn with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node
+Apr 29 12:58:08.831: INFO: Number of nodes with available pods: 0
+Apr 29 12:58:08.832: INFO: Node worker-3oo4n-86cbf586b7-9q98g is running more than one daemon pod
+Apr 29 12:58:09.829: INFO: DaemonSet pods can't tolerate node master-h2h3g-655686fb8c-8kgdn with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node
+Apr 29 12:58:09.837: INFO: Number of nodes with available pods: 0
+Apr 29 12:58:09.837: INFO: Node worker-3oo4n-86cbf586b7-9q98g is running more than one daemon pod
+Apr 29 12:58:10.821: INFO: DaemonSet pods can't tolerate node master-h2h3g-655686fb8c-8kgdn with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node
+Apr 29 12:58:10.828: INFO: Number of nodes with available pods: 3
+Apr 29 12:58:10.828: INFO: Number of running nodes: 3, number of available pods: 3
+STEP: Stop a daemon pod, check that the daemon pod is revived.
+Apr 29 12:58:10.866: INFO: DaemonSet pods can't tolerate node master-h2h3g-655686fb8c-8kgdn with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node
+Apr 29 12:58:10.876: INFO: Number of nodes with available pods: 2
+Apr 29 12:58:10.877: INFO: Node worker-4av2h-6665d7b5ff-9nc42 is running more than one daemon pod
+Apr 29 12:58:11.888: INFO: DaemonSet pods can't tolerate node master-h2h3g-655686fb8c-8kgdn with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node
+Apr 29 12:58:11.894: INFO: Number of nodes with available pods: 2
+Apr 29 12:58:11.894: INFO: Node worker-4av2h-6665d7b5ff-9nc42 is running more than one daemon pod
+Apr 29 12:58:12.886: INFO: DaemonSet pods can't tolerate node master-h2h3g-655686fb8c-8kgdn with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node
+Apr 29 12:58:12.893: INFO: Number of nodes with available pods: 2
+Apr 29 12:58:12.893: INFO: Node worker-4av2h-6665d7b5ff-9nc42 is running more than one daemon pod
+Apr 29 12:58:13.887: INFO: DaemonSet pods can't tolerate node master-h2h3g-655686fb8c-8kgdn with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node
+Apr 29 12:58:13.894: INFO: Number of nodes with available pods: 2
+Apr 29 12:58:13.894: INFO: Node worker-4av2h-6665d7b5ff-9nc42 is running more than one daemon pod
+Apr 29 12:58:14.888: INFO: DaemonSet pods can't tolerate node master-h2h3g-655686fb8c-8kgdn with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node
+Apr 29 12:58:14.894: INFO: Number of nodes with available pods: 2
+Apr 29 12:58:14.894: INFO: Node worker-4av2h-6665d7b5ff-9nc42 is running more than one daemon pod
+Apr 29 12:58:15.888: INFO: DaemonSet pods can't tolerate node master-h2h3g-655686fb8c-8kgdn with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node
+Apr 29 12:58:15.898: INFO: Number of nodes with available pods: 2
+Apr 29 12:58:15.898: INFO: Node worker-4av2h-6665d7b5ff-9nc42 is running more than one daemon pod
+Apr 29 12:58:16.884: INFO: DaemonSet pods can't tolerate node master-h2h3g-655686fb8c-8kgdn with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node
+Apr 29 12:58:16.891: INFO: Number of nodes with available pods: 2
+Apr 29 12:58:16.891: INFO: Node worker-4av2h-6665d7b5ff-9nc42 is running more than one daemon pod
+Apr 29 12:58:17.896: INFO: DaemonSet pods can't tolerate node master-h2h3g-655686fb8c-8kgdn with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node
+Apr 29 12:58:17.904: INFO: Number of nodes with available pods: 2
+Apr 29 12:58:17.904: INFO: Node worker-4av2h-6665d7b5ff-9nc42 is running more than one daemon pod
+Apr 29 12:58:18.896: INFO: DaemonSet pods can't tolerate node master-h2h3g-655686fb8c-8kgdn with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node
+Apr 29 12:58:18.911: INFO: Number of nodes with available pods: 2
+Apr 29 12:58:18.911: INFO: Node worker-4av2h-6665d7b5ff-9nc42 is running more than one daemon pod
+Apr 29 12:58:19.888: INFO: DaemonSet pods can't tolerate node master-h2h3g-655686fb8c-8kgdn with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node
+Apr 29 12:58:19.894: INFO: Number of nodes with available pods: 2
+Apr 29 12:58:19.894: INFO: Node worker-4av2h-6665d7b5ff-9nc42 is running more than one daemon pod
+Apr 29 12:58:20.888: INFO: DaemonSet pods can't tolerate node master-h2h3g-655686fb8c-8kgdn with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node
+Apr 29 12:58:20.895: INFO: Number of nodes with available pods: 2
+Apr 29 12:58:20.895: INFO: Node worker-4av2h-6665d7b5ff-9nc42 is running more than one daemon pod
+Apr 29 12:58:21.883: INFO: DaemonSet pods can't tolerate node master-h2h3g-655686fb8c-8kgdn with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node
+Apr 29 12:58:21.894: INFO: Number of nodes with available pods: 2
+Apr 29 12:58:21.895: INFO: Node worker-4av2h-6665d7b5ff-9nc42 is running more than one daemon pod
+Apr 29 12:58:22.885: INFO: DaemonSet pods can't tolerate node master-h2h3g-655686fb8c-8kgdn with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node
+Apr 29 12:58:22.892: INFO: Number of nodes with available pods: 2
+Apr 29 12:58:22.892: INFO: Node worker-4av2h-6665d7b5ff-9nc42 is running more than one daemon pod
+Apr 29 12:58:23.884: INFO: DaemonSet pods can't tolerate node master-h2h3g-655686fb8c-8kgdn with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node
+Apr 29 12:58:23.889: INFO: Number of nodes with available pods: 2
+Apr 29 12:58:23.890: INFO: Node worker-4av2h-6665d7b5ff-9nc42 is running more than one daemon pod
+Apr 29 12:58:24.888: INFO: DaemonSet pods can't tolerate node master-h2h3g-655686fb8c-8kgdn with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node
+Apr 29 12:58:24.896: INFO: Number of nodes with available pods: 2
+Apr 29 12:58:24.896: INFO: Node worker-4av2h-6665d7b5ff-9nc42 is running more than one daemon pod
+Apr 29 12:58:25.887: INFO: DaemonSet pods can't tolerate node master-h2h3g-655686fb8c-8kgdn with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node
+Apr 29 12:58:25.893: INFO: Number of nodes with available pods: 2
+Apr 29 12:58:25.893: INFO: Node worker-4av2h-6665d7b5ff-9nc42 is running more than one daemon pod
+Apr 29 12:58:26.892: INFO: DaemonSet pods can't tolerate node master-h2h3g-655686fb8c-8kgdn with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node
+Apr 29 12:58:26.902: INFO: Number of nodes with available pods: 2
+Apr 29 12:58:26.902: INFO: Node worker-4av2h-6665d7b5ff-9nc42 is running more than one daemon pod
+Apr 29 12:58:27.886: INFO: DaemonSet pods can't tolerate node master-h2h3g-655686fb8c-8kgdn with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node
+Apr 29 12:58:27.892: INFO: Number of nodes with available pods: 3
+Apr 29 12:58:27.892: INFO: Number of running nodes: 3, number of available pods: 3
+[AfterEach] [sig-apps] Daemon set [Serial]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:68
+STEP: Deleting DaemonSet "daemon-set"
+STEP: deleting DaemonSet.extensions daemon-set in namespace daemonsets-9228, will wait for the garbage collector to delete the pods
+Apr 29 12:58:27.978: INFO: Deleting DaemonSet.extensions daemon-set took: 12.115261ms
+Apr 29 12:58:28.079: INFO: Terminating DaemonSet.extensions daemon-set pods took: 101.014488ms
+Apr 29 12:58:35.886: INFO: Number of nodes with available pods: 0
+Apr 29 12:58:35.886: INFO: Number of running nodes: 0, number of available pods: 0
+Apr 29 12:58:35.891: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"selfLink":"/apis/apps/v1/namespaces/daemonsets-9228/daemonsets","resourceVersion":"25268"},"items":null}
+
+Apr 29 12:58:35.902: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"selfLink":"/api/v1/namespaces/daemonsets-9228/pods","resourceVersion":"25268"},"items":null}
+
+[AfterEach] [sig-apps] Daemon set [Serial]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 12:58:35.946: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "daemonsets-9228" for this suite.
+Apr 29 12:58:41.978: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 12:58:42.176: INFO: namespace daemonsets-9228 deletion completed in 6.223839197s
+
+• [SLOW TEST:34.646 seconds]
+[sig-apps] Daemon set [Serial]
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22
+  should run and stop simple daemon [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSS
+------------------------------
+[sig-api-machinery] Garbage collector 
+  should orphan RS created by deployment when deleteOptions.PropagationPolicy is Orphan [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-api-machinery] Garbage collector
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 12:58:42.181: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename gc
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in gc-5715
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should orphan RS created by deployment when deleteOptions.PropagationPolicy is Orphan [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: create the deployment
+STEP: Wait for the Deployment to create new ReplicaSet
+STEP: delete the deployment
+STEP: wait for 30 seconds to see if the garbage collector mistakenly deletes the rs
+STEP: Gathering metrics
+W0429 12:59:12.963430      15 metrics_grabber.go:79] Master node is not registered. Grabbing metrics from Scheduler, ControllerManager and ClusterAutoscaler is disabled.
+Apr 29 12:59:12.963: INFO: For apiserver_request_total:
+For apiserver_request_latencies_summary:
+For apiserver_init_events_total:
+For garbage_collector_attempt_to_delete_queue_latency:
+For garbage_collector_attempt_to_delete_work_duration:
+For garbage_collector_attempt_to_orphan_queue_latency:
+For garbage_collector_attempt_to_orphan_work_duration:
+For garbage_collector_dirty_processing_latency_microseconds:
+For garbage_collector_event_processing_latency_microseconds:
+For garbage_collector_graph_changes_queue_latency:
+For garbage_collector_graph_changes_work_duration:
+For garbage_collector_orphan_processing_latency_microseconds:
+For namespace_queue_latency:
+For namespace_queue_latency_sum:
+For namespace_queue_latency_count:
+For namespace_retries:
+For namespace_work_duration:
+For namespace_work_duration_sum:
+For namespace_work_duration_count:
+For function_duration_seconds:
+For errors_total:
+For evicted_pods_total:
+
+[AfterEach] [sig-api-machinery] Garbage collector
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 12:59:12.963: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "gc-5715" for this suite.
+Apr 29 12:59:18.989: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 12:59:19.251: INFO: namespace gc-5715 deletion completed in 6.281554133s
+
+• [SLOW TEST:37.070 seconds]
+[sig-api-machinery] Garbage collector
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:22
+  should orphan RS created by deployment when deleteOptions.PropagationPolicy is Orphan [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSS
+------------------------------
+[k8s.io] Container Runtime blackbox test when starting a container that exits 
+  should run with the expected status [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [k8s.io] Container Runtime
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 12:59:19.257: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename container-runtime
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in container-runtime-5699
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should run with the expected status [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Container 'terminate-cmd-rpa': should get the expected 'RestartCount'
+STEP: Container 'terminate-cmd-rpa': should get the expected 'Phase'
+STEP: Container 'terminate-cmd-rpa': should get the expected 'Ready' condition
+STEP: Container 'terminate-cmd-rpa': should get the expected 'State'
+STEP: Container 'terminate-cmd-rpa': should be possible to delete [NodeConformance]
+STEP: Container 'terminate-cmd-rpof': should get the expected 'RestartCount'
+STEP: Container 'terminate-cmd-rpof': should get the expected 'Phase'
+STEP: Container 'terminate-cmd-rpof': should get the expected 'Ready' condition
+STEP: Container 'terminate-cmd-rpof': should get the expected 'State'
+STEP: Container 'terminate-cmd-rpof': should be possible to delete [NodeConformance]
+STEP: Container 'terminate-cmd-rpn': should get the expected 'RestartCount'
+STEP: Container 'terminate-cmd-rpn': should get the expected 'Phase'
+STEP: Container 'terminate-cmd-rpn': should get the expected 'Ready' condition
+STEP: Container 'terminate-cmd-rpn': should get the expected 'State'
+STEP: Container 'terminate-cmd-rpn': should be possible to delete [NodeConformance]
+[AfterEach] [k8s.io] Container Runtime
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 12:59:48.962: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "container-runtime-5699" for this suite.
+Apr 29 12:59:54.994: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 12:59:55.176: INFO: namespace container-runtime-5699 deletion completed in 6.207742413s
+
+• [SLOW TEST:35.920 seconds]
+[k8s.io] Container Runtime
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
+  blackbox test
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/runtime.go:37
+    when starting a container that exits
+    /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/runtime.go:38
+      should run with the expected status [NodeConformance] [Conformance]
+      /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSSSSSS
+------------------------------
+[k8s.io] Docker Containers 
+  should be able to override the image's default arguments (docker cmd) [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [k8s.io] Docker Containers
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 12:59:55.179: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename containers
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in containers-412
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should be able to override the image's default arguments (docker cmd) [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating a pod to test override arguments
+Apr 29 12:59:55.397: INFO: Waiting up to 5m0s for pod "client-containers-af782d43-6a7e-11e9-b6ef-0e71f8c7e888" in namespace "containers-412" to be "success or failure"
+Apr 29 12:59:55.402: INFO: Pod "client-containers-af782d43-6a7e-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 5.541111ms
+Apr 29 12:59:57.417: INFO: Pod "client-containers-af782d43-6a7e-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 2.019612679s
+Apr 29 12:59:59.422: INFO: Pod "client-containers-af782d43-6a7e-11e9-b6ef-0e71f8c7e888": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.025135908s
+STEP: Saw pod success
+Apr 29 12:59:59.422: INFO: Pod "client-containers-af782d43-6a7e-11e9-b6ef-0e71f8c7e888" satisfied condition "success or failure"
+Apr 29 12:59:59.426: INFO: Trying to get logs from node worker-4av2h-6665d7b5ff-9nc42 pod client-containers-af782d43-6a7e-11e9-b6ef-0e71f8c7e888 container test-container: 
+STEP: delete the pod
+Apr 29 12:59:59.463: INFO: Waiting for pod client-containers-af782d43-6a7e-11e9-b6ef-0e71f8c7e888 to disappear
+Apr 29 12:59:59.467: INFO: Pod client-containers-af782d43-6a7e-11e9-b6ef-0e71f8c7e888 no longer exists
+[AfterEach] [k8s.io] Docker Containers
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 12:59:59.467: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "containers-412" for this suite.
+Apr 29 13:00:05.487: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 13:00:05.693: INFO: namespace containers-412 deletion completed in 6.219102426s
+
+• [SLOW TEST:10.514 seconds]
+[k8s.io] Docker Containers
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
+  should be able to override the image's default arguments (docker cmd) [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSSSSS
+------------------------------
+[sig-storage] ConfigMap 
+  should be consumable from pods in volume with mappings [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-storage] ConfigMap
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 13:00:05.699: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename configmap
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in configmap-2607
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should be consumable from pods in volume with mappings [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating configMap with name configmap-test-volume-map-b5bf4f1e-6a7e-11e9-b6ef-0e71f8c7e888
+STEP: Creating a pod to test consume configMaps
+Apr 29 13:00:05.939: INFO: Waiting up to 5m0s for pod "pod-configmaps-b5c09cd2-6a7e-11e9-b6ef-0e71f8c7e888" in namespace "configmap-2607" to be "success or failure"
+Apr 29 13:00:05.949: INFO: Pod "pod-configmaps-b5c09cd2-6a7e-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 9.258069ms
+Apr 29 13:00:07.956: INFO: Pod "pod-configmaps-b5c09cd2-6a7e-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 2.016850788s
+Apr 29 13:00:09.963: INFO: Pod "pod-configmaps-b5c09cd2-6a7e-11e9-b6ef-0e71f8c7e888": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.02400119s
+STEP: Saw pod success
+Apr 29 13:00:09.964: INFO: Pod "pod-configmaps-b5c09cd2-6a7e-11e9-b6ef-0e71f8c7e888" satisfied condition "success or failure"
+Apr 29 13:00:09.969: INFO: Trying to get logs from node worker-4av2h-6665d7b5ff-9nc42 pod pod-configmaps-b5c09cd2-6a7e-11e9-b6ef-0e71f8c7e888 container configmap-volume-test: 
+STEP: delete the pod
+Apr 29 13:00:10.039: INFO: Waiting for pod pod-configmaps-b5c09cd2-6a7e-11e9-b6ef-0e71f8c7e888 to disappear
+Apr 29 13:00:10.078: INFO: Pod pod-configmaps-b5c09cd2-6a7e-11e9-b6ef-0e71f8c7e888 no longer exists
+[AfterEach] [sig-storage] ConfigMap
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 13:00:10.078: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "configmap-2607" for this suite.
+Apr 29 13:00:16.131: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 13:00:16.500: INFO: namespace configmap-2607 deletion completed in 6.39986502s
+
+• [SLOW TEST:10.802 seconds]
+[sig-storage] ConfigMap
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:32
+  should be consumable from pods in volume with mappings [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSSSSSSSSSSSSSSSSSS
+------------------------------
+[sig-api-machinery] Secrets 
+  should be consumable from pods in env vars [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-api-machinery] Secrets
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 13:00:16.501: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename secrets
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in secrets-4532
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should be consumable from pods in env vars [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating secret with name secret-test-bc2c2d2f-6a7e-11e9-b6ef-0e71f8c7e888
+STEP: Creating a pod to test consume secrets
+Apr 29 13:00:16.710: INFO: Waiting up to 5m0s for pod "pod-secrets-bc2d4339-6a7e-11e9-b6ef-0e71f8c7e888" in namespace "secrets-4532" to be "success or failure"
+Apr 29 13:00:16.717: INFO: Pod "pod-secrets-bc2d4339-6a7e-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 6.370819ms
+Apr 29 13:00:18.724: INFO: Pod "pod-secrets-bc2d4339-6a7e-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 2.01335036s
+Apr 29 13:00:20.730: INFO: Pod "pod-secrets-bc2d4339-6a7e-11e9-b6ef-0e71f8c7e888": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.019982686s
+STEP: Saw pod success
+Apr 29 13:00:20.731: INFO: Pod "pod-secrets-bc2d4339-6a7e-11e9-b6ef-0e71f8c7e888" satisfied condition "success or failure"
+Apr 29 13:00:20.737: INFO: Trying to get logs from node worker-4av2h-6665d7b5ff-9nc42 pod pod-secrets-bc2d4339-6a7e-11e9-b6ef-0e71f8c7e888 container secret-env-test: 
+STEP: delete the pod
+Apr 29 13:00:20.806: INFO: Waiting for pod pod-secrets-bc2d4339-6a7e-11e9-b6ef-0e71f8c7e888 to disappear
+Apr 29 13:00:20.822: INFO: Pod pod-secrets-bc2d4339-6a7e-11e9-b6ef-0e71f8c7e888 no longer exists
+[AfterEach] [sig-api-machinery] Secrets
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 13:00:20.822: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "secrets-4532" for this suite.
+Apr 29 13:00:26.861: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 13:00:27.065: INFO: namespace secrets-4532 deletion completed in 6.233745131s
+
+• [SLOW TEST:10.564 seconds]
+[sig-api-machinery] Secrets
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/secrets.go:32
+  should be consumable from pods in env vars [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+[sig-api-machinery] CustomResourceDefinition resources Simple CustomResourceDefinition 
+  creating/deleting custom resource definition objects works  [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-api-machinery] CustomResourceDefinition resources
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 13:00:27.065: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename custom-resource-definition
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in custom-resource-definition-2429
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] creating/deleting custom resource definition objects works  [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+Apr 29 13:00:27.288: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+[AfterEach] [sig-api-machinery] CustomResourceDefinition resources
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 13:00:28.470: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "custom-resource-definition-2429" for this suite.
+Apr 29 13:00:34.500: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 13:00:34.740: INFO: namespace custom-resource-definition-2429 deletion completed in 6.258019375s
+
+• [SLOW TEST:7.675 seconds]
+[sig-api-machinery] CustomResourceDefinition resources
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:22
+  Simple CustomResourceDefinition
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/custom_resource_definition.go:35
+    creating/deleting custom resource definition objects works  [Conformance]
+    /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+------------------------------
+[sig-storage] Downward API volume 
+  should set mode on item file [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-storage] Downward API volume
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 13:00:34.743: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename downward-api
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in downward-api-3123
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [sig-storage] Downward API volume
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:39
+[It] should set mode on item file [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating a pod to test downward API volume plugin
+Apr 29 13:00:34.949: INFO: Waiting up to 5m0s for pod "downwardapi-volume-c70a09a2-6a7e-11e9-b6ef-0e71f8c7e888" in namespace "downward-api-3123" to be "success or failure"
+Apr 29 13:00:34.954: INFO: Pod "downwardapi-volume-c70a09a2-6a7e-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 5.044374ms
+Apr 29 13:00:36.962: INFO: Pod "downwardapi-volume-c70a09a2-6a7e-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 2.012963746s
+Apr 29 13:00:38.972: INFO: Pod "downwardapi-volume-c70a09a2-6a7e-11e9-b6ef-0e71f8c7e888": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.023135989s
+STEP: Saw pod success
+Apr 29 13:00:38.972: INFO: Pod "downwardapi-volume-c70a09a2-6a7e-11e9-b6ef-0e71f8c7e888" satisfied condition "success or failure"
+Apr 29 13:00:38.977: INFO: Trying to get logs from node worker-4av2h-6665d7b5ff-9nc42 pod downwardapi-volume-c70a09a2-6a7e-11e9-b6ef-0e71f8c7e888 container client-container: 
+STEP: delete the pod
+Apr 29 13:00:39.016: INFO: Waiting for pod downwardapi-volume-c70a09a2-6a7e-11e9-b6ef-0e71f8c7e888 to disappear
+Apr 29 13:00:39.021: INFO: Pod downwardapi-volume-c70a09a2-6a7e-11e9-b6ef-0e71f8c7e888 no longer exists
+[AfterEach] [sig-storage] Downward API volume
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 13:00:39.022: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "downward-api-3123" for this suite.
+Apr 29 13:00:45.051: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 13:00:45.253: INFO: namespace downward-api-3123 deletion completed in 6.223392121s
+
+• [SLOW TEST:10.510 seconds]
+[sig-storage] Downward API volume
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:34
+  should set mode on item file [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSSSSSS
+------------------------------
+[sig-storage] EmptyDir volumes 
+  should support (root,0644,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-storage] EmptyDir volumes
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 13:00:45.254: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename emptydir
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in emptydir-4877
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should support (root,0644,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating a pod to test emptydir 0644 on tmpfs
+Apr 29 13:00:45.442: INFO: Waiting up to 5m0s for pod "pod-cd4cf9f9-6a7e-11e9-b6ef-0e71f8c7e888" in namespace "emptydir-4877" to be "success or failure"
+Apr 29 13:00:45.450: INFO: Pod "pod-cd4cf9f9-6a7e-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 8.159184ms
+Apr 29 13:00:47.457: INFO: Pod "pod-cd4cf9f9-6a7e-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 2.015210795s
+Apr 29 13:00:49.467: INFO: Pod "pod-cd4cf9f9-6a7e-11e9-b6ef-0e71f8c7e888": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.024505778s
+STEP: Saw pod success
+Apr 29 13:00:49.467: INFO: Pod "pod-cd4cf9f9-6a7e-11e9-b6ef-0e71f8c7e888" satisfied condition "success or failure"
+Apr 29 13:00:49.472: INFO: Trying to get logs from node worker-4av2h-6665d7b5ff-9nc42 pod pod-cd4cf9f9-6a7e-11e9-b6ef-0e71f8c7e888 container test-container: 
+STEP: delete the pod
+Apr 29 13:00:49.508: INFO: Waiting for pod pod-cd4cf9f9-6a7e-11e9-b6ef-0e71f8c7e888 to disappear
+Apr 29 13:00:49.516: INFO: Pod pod-cd4cf9f9-6a7e-11e9-b6ef-0e71f8c7e888 no longer exists
+[AfterEach] [sig-storage] EmptyDir volumes
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 13:00:49.516: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "emptydir-4877" for this suite.
+Apr 29 13:00:55.556: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 13:00:55.744: INFO: namespace emptydir-4877 deletion completed in 6.220449674s
+
+• [SLOW TEST:10.489 seconds]
+[sig-storage] EmptyDir volumes
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:41
+  should support (root,0644,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+S
+------------------------------
+[sig-storage] Downward API volume 
+  should provide container's cpu limit [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-storage] Downward API volume
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 13:00:55.744: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename downward-api
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in downward-api-1493
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [sig-storage] Downward API volume
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:39
+[It] should provide container's cpu limit [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating a pod to test downward API volume plugin
+Apr 29 13:00:55.950: INFO: Waiting up to 5m0s for pod "downwardapi-volume-d39029e7-6a7e-11e9-b6ef-0e71f8c7e888" in namespace "downward-api-1493" to be "success or failure"
+Apr 29 13:00:55.959: INFO: Pod "downwardapi-volume-d39029e7-6a7e-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 8.819882ms
+Apr 29 13:00:57.966: INFO: Pod "downwardapi-volume-d39029e7-6a7e-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 2.015423296s
+Apr 29 13:00:59.973: INFO: Pod "downwardapi-volume-d39029e7-6a7e-11e9-b6ef-0e71f8c7e888": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.023122215s
+STEP: Saw pod success
+Apr 29 13:00:59.973: INFO: Pod "downwardapi-volume-d39029e7-6a7e-11e9-b6ef-0e71f8c7e888" satisfied condition "success or failure"
+Apr 29 13:00:59.979: INFO: Trying to get logs from node worker-4av2h-6665d7b5ff-9nc42 pod downwardapi-volume-d39029e7-6a7e-11e9-b6ef-0e71f8c7e888 container client-container: 
+STEP: delete the pod
+Apr 29 13:01:00.026: INFO: Waiting for pod downwardapi-volume-d39029e7-6a7e-11e9-b6ef-0e71f8c7e888 to disappear
+Apr 29 13:01:00.039: INFO: Pod downwardapi-volume-d39029e7-6a7e-11e9-b6ef-0e71f8c7e888 no longer exists
+[AfterEach] [sig-storage] Downward API volume
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 13:01:00.040: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "downward-api-1493" for this suite.
+Apr 29 13:01:06.119: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 13:01:06.441: INFO: namespace downward-api-1493 deletion completed in 6.382215686s
+
+• [SLOW TEST:10.698 seconds]
+[sig-storage] Downward API volume
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:34
+  should provide container's cpu limit [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSSSSSSSSS
+------------------------------
+[sig-storage] Projected secret 
+  should be consumable from pods in volume with mappings [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-storage] Projected secret
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 13:01:06.442: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename projected
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-8053
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should be consumable from pods in volume with mappings [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating projection with secret that has name projected-secret-test-map-d9f3f080-6a7e-11e9-b6ef-0e71f8c7e888
+STEP: Creating a pod to test consume secrets
+Apr 29 13:01:06.691: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-d9f5afa4-6a7e-11e9-b6ef-0e71f8c7e888" in namespace "projected-8053" to be "success or failure"
+Apr 29 13:01:06.703: INFO: Pod "pod-projected-secrets-d9f5afa4-6a7e-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 11.672507ms
+Apr 29 13:01:08.709: INFO: Pod "pod-projected-secrets-d9f5afa4-6a7e-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 2.017550694s
+Apr 29 13:01:10.715: INFO: Pod "pod-projected-secrets-d9f5afa4-6a7e-11e9-b6ef-0e71f8c7e888": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.023977604s
+STEP: Saw pod success
+Apr 29 13:01:10.716: INFO: Pod "pod-projected-secrets-d9f5afa4-6a7e-11e9-b6ef-0e71f8c7e888" satisfied condition "success or failure"
+Apr 29 13:01:10.722: INFO: Trying to get logs from node worker-4av2h-6665d7b5ff-9nc42 pod pod-projected-secrets-d9f5afa4-6a7e-11e9-b6ef-0e71f8c7e888 container projected-secret-volume-test: 
+STEP: delete the pod
+Apr 29 13:01:10.758: INFO: Waiting for pod pod-projected-secrets-d9f5afa4-6a7e-11e9-b6ef-0e71f8c7e888 to disappear
+Apr 29 13:01:10.775: INFO: Pod pod-projected-secrets-d9f5afa4-6a7e-11e9-b6ef-0e71f8c7e888 no longer exists
+[AfterEach] [sig-storage] Projected secret
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 13:01:10.775: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "projected-8053" for this suite.
+Apr 29 13:01:16.805: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 13:01:17.065: INFO: namespace projected-8053 deletion completed in 6.279124668s
+
+• [SLOW TEST:10.622 seconds]
+[sig-storage] Projected secret
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_secret.go:33
+  should be consumable from pods in volume with mappings [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSSSSSSSSSSSSSS
+------------------------------
+[sig-cli] Kubectl client [k8s.io] Kubectl run rc 
+  should create an rc from an image  [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-cli] Kubectl client
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 13:01:17.075: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename kubectl
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-4785
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [sig-cli] Kubectl client
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:213
+[BeforeEach] [k8s.io] Kubectl run rc
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1354
+[It] should create an rc from an image  [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: running the image docker.io/library/nginx:1.14-alpine
+Apr 29 13:01:17.294: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 run e2e-test-nginx-rc --image=docker.io/library/nginx:1.14-alpine --generator=run/v1 --namespace=kubectl-4785'
+Apr 29 13:01:19.339: INFO: stderr: "kubectl run --generator=run/v1 is DEPRECATED and will be removed in a future version. Use kubectl run --generator=run-pod/v1 or kubectl create instead.\n"
+Apr 29 13:01:19.339: INFO: stdout: "replicationcontroller/e2e-test-nginx-rc created\n"
+STEP: verifying the rc e2e-test-nginx-rc was created
+STEP: verifying the pod controlled by rc e2e-test-nginx-rc was created
+STEP: confirm that you can get logs from an rc
+Apr 29 13:01:19.369: INFO: Waiting up to 5m0s for 1 pods to be running and ready: [e2e-test-nginx-rc-nzjzv]
+Apr 29 13:01:19.369: INFO: Waiting up to 5m0s for pod "e2e-test-nginx-rc-nzjzv" in namespace "kubectl-4785" to be "running and ready"
+Apr 29 13:01:19.382: INFO: Pod "e2e-test-nginx-rc-nzjzv": Phase="Pending", Reason="", readiness=false. Elapsed: 13.488705ms
+Apr 29 13:01:21.391: INFO: Pod "e2e-test-nginx-rc-nzjzv": Phase="Pending", Reason="", readiness=false. Elapsed: 2.022014393s
+Apr 29 13:01:23.397: INFO: Pod "e2e-test-nginx-rc-nzjzv": Phase="Running", Reason="", readiness=true. Elapsed: 4.028165315s
+Apr 29 13:01:23.397: INFO: Pod "e2e-test-nginx-rc-nzjzv" satisfied condition "running and ready"
+Apr 29 13:01:23.397: INFO: Wanted all 1 pods to be running and ready. Result: true. Pods: [e2e-test-nginx-rc-nzjzv]
+Apr 29 13:01:23.397: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 logs rc/e2e-test-nginx-rc --namespace=kubectl-4785'
+Apr 29 13:01:23.544: INFO: stderr: ""
+Apr 29 13:01:23.545: INFO: stdout: ""
+[AfterEach] [k8s.io] Kubectl run rc
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1359
+Apr 29 13:01:23.545: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 delete rc e2e-test-nginx-rc --namespace=kubectl-4785'
+Apr 29 13:01:23.667: INFO: stderr: ""
+Apr 29 13:01:23.667: INFO: stdout: "replicationcontroller \"e2e-test-nginx-rc\" deleted\n"
+[AfterEach] [sig-cli] Kubectl client
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 13:01:23.668: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "kubectl-4785" for this suite.
+Apr 29 13:01:47.702: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 13:01:47.946: INFO: namespace kubectl-4785 deletion completed in 24.266921598s
+
+• [SLOW TEST:30.871 seconds]
+[sig-cli] Kubectl client
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23
+  [k8s.io] Kubectl run rc
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
+    should create an rc from an image  [Conformance]
+    /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SS
+------------------------------
+[sig-cli] Kubectl client [k8s.io] Kubectl patch 
+  should add annotations for pods in rc  [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-cli] Kubectl client
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 13:01:47.946: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename kubectl
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-7802
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [sig-cli] Kubectl client
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:213
+[It] should add annotations for pods in rc  [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: creating Redis RC
+Apr 29 13:01:48.162: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 create -f - --namespace=kubectl-7802'
+Apr 29 13:01:52.047: INFO: stderr: ""
+Apr 29 13:01:52.047: INFO: stdout: "replicationcontroller/redis-master created\n"
+STEP: Waiting for Redis master to start.
+Apr 29 13:01:53.055: INFO: Selector matched 1 pods for map[app:redis]
+Apr 29 13:01:53.055: INFO: Found 0 / 1
+Apr 29 13:01:54.054: INFO: Selector matched 1 pods for map[app:redis]
+Apr 29 13:01:54.054: INFO: Found 0 / 1
+Apr 29 13:01:55.053: INFO: Selector matched 1 pods for map[app:redis]
+Apr 29 13:01:55.053: INFO: Found 0 / 1
+Apr 29 13:01:56.064: INFO: Selector matched 1 pods for map[app:redis]
+Apr 29 13:01:56.064: INFO: Found 0 / 1
+Apr 29 13:01:57.055: INFO: Selector matched 1 pods for map[app:redis]
+Apr 29 13:01:57.055: INFO: Found 1 / 1
+Apr 29 13:01:57.055: INFO: WaitFor completed with timeout 5m0s.  Pods found = 1 out of 1
+STEP: patching all pods
+Apr 29 13:01:57.060: INFO: Selector matched 1 pods for map[app:redis]
+Apr 29 13:01:57.060: INFO: ForEach: Found 1 pods from the filter.  Now looping through them.
+Apr 29 13:01:57.060: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 patch pod redis-master-tclcs --namespace=kubectl-7802 -p {"metadata":{"annotations":{"x":"y"}}}'
+Apr 29 13:01:57.194: INFO: stderr: ""
+Apr 29 13:01:57.194: INFO: stdout: "pod/redis-master-tclcs patched\n"
+STEP: checking annotations
+Apr 29 13:01:57.199: INFO: Selector matched 1 pods for map[app:redis]
+Apr 29 13:01:57.199: INFO: ForEach: Found 1 pods from the filter.  Now looping through them.
+[AfterEach] [sig-cli] Kubectl client
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 13:01:57.199: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "kubectl-7802" for this suite.
+Apr 29 13:02:21.229: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 13:02:21.459: INFO: namespace kubectl-7802 deletion completed in 24.253168609s
+
+• [SLOW TEST:33.513 seconds]
+[sig-cli] Kubectl client
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23
+  [k8s.io] Kubectl patch
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
+    should add annotations for pods in rc  [Conformance]
+    /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+[sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] 
+  Scaling should happen in predictable order and halt if any stateful pod is unhealthy [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-apps] StatefulSet
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 13:02:21.461: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename statefulset
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in statefulset-7999
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [sig-apps] StatefulSet
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:59
+[BeforeEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:74
+STEP: Creating service test in namespace statefulset-7999
+[It] Scaling should happen in predictable order and halt if any stateful pod is unhealthy [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Initializing watcher for selector baz=blah,foo=bar
+STEP: Creating stateful set ss in namespace statefulset-7999
+STEP: Waiting until all stateful set ss replicas will be running in namespace statefulset-7999
+Apr 29 13:02:21.708: INFO: Found 0 stateful pods, waiting for 1
+Apr 29 13:02:31.720: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true
+STEP: Confirming that stateful set scale up will halt with unhealthy stateful pod
+Apr 29 13:02:31.725: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 exec --namespace=statefulset-7999 ss-0 -- /bin/sh -x -c mv -v /usr/share/nginx/html/index.html /tmp/ || true'
+Apr 29 13:02:32.111: INFO: stderr: "+ mv -v /usr/share/nginx/html/index.html /tmp/\n"
+Apr 29 13:02:32.111: INFO: stdout: "'/usr/share/nginx/html/index.html' -> '/tmp/index.html'\n"
+Apr 29 13:02:32.112: INFO: stdout of mv -v /usr/share/nginx/html/index.html /tmp/ || true on ss-0: '/usr/share/nginx/html/index.html' -> '/tmp/index.html'
+
+Apr 29 13:02:32.118: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=true
+Apr 29 13:02:42.128: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=false
+Apr 29 13:02:42.128: INFO: Waiting for statefulset status.replicas updated to 0
+Apr 29 13:02:42.158: INFO: Verifying statefulset ss doesn't scale past 1 for another 9.999998945s
+Apr 29 13:02:43.164: INFO: Verifying statefulset ss doesn't scale past 1 for another 8.990114572s
+Apr 29 13:02:44.170: INFO: Verifying statefulset ss doesn't scale past 1 for another 7.983494075s
+Apr 29 13:02:45.175: INFO: Verifying statefulset ss doesn't scale past 1 for another 6.978128036s
+Apr 29 13:02:46.182: INFO: Verifying statefulset ss doesn't scale past 1 for another 5.972404875s
+Apr 29 13:02:47.190: INFO: Verifying statefulset ss doesn't scale past 1 for another 4.963937791s
+Apr 29 13:02:48.196: INFO: Verifying statefulset ss doesn't scale past 1 for another 3.958111061s
+Apr 29 13:02:49.204: INFO: Verifying statefulset ss doesn't scale past 1 for another 2.951079503s
+Apr 29 13:02:50.211: INFO: Verifying statefulset ss doesn't scale past 1 for another 1.943350025s
+Apr 29 13:02:51.216: INFO: Verifying statefulset ss doesn't scale past 1 for another 937.117538ms
+STEP: Scaling up stateful set ss to 3 replicas and waiting until all of them will be running in namespace statefulset-7999
+Apr 29 13:02:52.224: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 exec --namespace=statefulset-7999 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
+Apr 29 13:02:52.531: INFO: stderr: "+ mv -v /tmp/index.html /usr/share/nginx/html/\n"
+Apr 29 13:02:52.531: INFO: stdout: "'/tmp/index.html' -> '/usr/share/nginx/html/index.html'\n"
+Apr 29 13:02:52.531: INFO: stdout of mv -v /tmp/index.html /usr/share/nginx/html/ || true on ss-0: '/tmp/index.html' -> '/usr/share/nginx/html/index.html'
+
+Apr 29 13:02:52.537: INFO: Found 1 stateful pods, waiting for 3
+Apr 29 13:03:02.544: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true
+Apr 29 13:03:02.544: INFO: Waiting for pod ss-1 to enter Running - Ready=true, currently Running - Ready=true
+Apr 29 13:03:02.544: INFO: Waiting for pod ss-2 to enter Running - Ready=true, currently Running - Ready=true
+STEP: Verifying that stateful set ss was scaled up in order
+STEP: Scale down will halt with unhealthy stateful pod
+Apr 29 13:03:02.552: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 exec --namespace=statefulset-7999 ss-0 -- /bin/sh -x -c mv -v /usr/share/nginx/html/index.html /tmp/ || true'
+Apr 29 13:03:02.848: INFO: stderr: "+ mv -v /usr/share/nginx/html/index.html /tmp/\n"
+Apr 29 13:03:02.848: INFO: stdout: "'/usr/share/nginx/html/index.html' -> '/tmp/index.html'\n"
+Apr 29 13:03:02.848: INFO: stdout of mv -v /usr/share/nginx/html/index.html /tmp/ || true on ss-0: '/usr/share/nginx/html/index.html' -> '/tmp/index.html'
+
+Apr 29 13:03:02.848: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 exec --namespace=statefulset-7999 ss-1 -- /bin/sh -x -c mv -v /usr/share/nginx/html/index.html /tmp/ || true'
+Apr 29 13:03:03.164: INFO: stderr: "+ mv -v /usr/share/nginx/html/index.html /tmp/\n"
+Apr 29 13:03:03.164: INFO: stdout: "'/usr/share/nginx/html/index.html' -> '/tmp/index.html'\n"
+Apr 29 13:03:03.164: INFO: stdout of mv -v /usr/share/nginx/html/index.html /tmp/ || true on ss-1: '/usr/share/nginx/html/index.html' -> '/tmp/index.html'
+
+Apr 29 13:03:03.164: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 exec --namespace=statefulset-7999 ss-2 -- /bin/sh -x -c mv -v /usr/share/nginx/html/index.html /tmp/ || true'
+Apr 29 13:03:03.505: INFO: stderr: "+ mv -v /usr/share/nginx/html/index.html /tmp/\n"
+Apr 29 13:03:03.506: INFO: stdout: "'/usr/share/nginx/html/index.html' -> '/tmp/index.html'\n"
+Apr 29 13:03:03.506: INFO: stdout of mv -v /usr/share/nginx/html/index.html /tmp/ || true on ss-2: '/usr/share/nginx/html/index.html' -> '/tmp/index.html'
+
+Apr 29 13:03:03.506: INFO: Waiting for statefulset status.replicas updated to 0
+Apr 29 13:03:03.515: INFO: Waiting for stateful set status.readyReplicas to become 0, currently 2
+Apr 29 13:03:13.525: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=false
+Apr 29 13:03:13.525: INFO: Waiting for pod ss-1 to enter Running - Ready=false, currently Running - Ready=false
+Apr 29 13:03:13.525: INFO: Waiting for pod ss-2 to enter Running - Ready=false, currently Running - Ready=false
+Apr 29 13:03:13.539: INFO: Verifying statefulset ss doesn't scale past 3 for another 9.999998848s
+Apr 29 13:03:14.546: INFO: Verifying statefulset ss doesn't scale past 3 for another 8.994141454s
+Apr 29 13:03:15.554: INFO: Verifying statefulset ss doesn't scale past 3 for another 7.986918252s
+Apr 29 13:03:16.560: INFO: Verifying statefulset ss doesn't scale past 3 for another 6.979255225s
+Apr 29 13:03:17.567: INFO: Verifying statefulset ss doesn't scale past 3 for another 5.9729222s
+Apr 29 13:03:18.575: INFO: Verifying statefulset ss doesn't scale past 3 for another 4.965922645s
+Apr 29 13:03:19.582: INFO: Verifying statefulset ss doesn't scale past 3 for another 3.957546121s
+Apr 29 13:03:20.592: INFO: Verifying statefulset ss doesn't scale past 3 for another 2.950942818s
+Apr 29 13:03:21.600: INFO: Verifying statefulset ss doesn't scale past 3 for another 1.941081318s
+Apr 29 13:03:22.607: INFO: Verifying statefulset ss doesn't scale past 3 for another 933.043283ms
+STEP: Scaling down stateful set ss to 0 replicas and waiting until none of pods will run in namespacestatefulset-7999
+Apr 29 13:03:23.617: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 exec --namespace=statefulset-7999 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
+Apr 29 13:03:23.924: INFO: stderr: "+ mv -v /tmp/index.html /usr/share/nginx/html/\n"
+Apr 29 13:03:23.924: INFO: stdout: "'/tmp/index.html' -> '/usr/share/nginx/html/index.html'\n"
+Apr 29 13:03:23.924: INFO: stdout of mv -v /tmp/index.html /usr/share/nginx/html/ || true on ss-0: '/tmp/index.html' -> '/usr/share/nginx/html/index.html'
+
+Apr 29 13:03:23.924: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 exec --namespace=statefulset-7999 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
+Apr 29 13:03:24.283: INFO: stderr: "+ mv -v /tmp/index.html /usr/share/nginx/html/\n"
+Apr 29 13:03:24.283: INFO: stdout: "'/tmp/index.html' -> '/usr/share/nginx/html/index.html'\n"
+Apr 29 13:03:24.283: INFO: stdout of mv -v /tmp/index.html /usr/share/nginx/html/ || true on ss-1: '/tmp/index.html' -> '/usr/share/nginx/html/index.html'
+
+Apr 29 13:03:24.283: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 exec --namespace=statefulset-7999 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
+Apr 29 13:03:24.583: INFO: stderr: "+ mv -v /tmp/index.html /usr/share/nginx/html/\n"
+Apr 29 13:03:24.583: INFO: stdout: "'/tmp/index.html' -> '/usr/share/nginx/html/index.html'\n"
+Apr 29 13:03:24.583: INFO: stdout of mv -v /tmp/index.html /usr/share/nginx/html/ || true on ss-2: '/tmp/index.html' -> '/usr/share/nginx/html/index.html'
+
+Apr 29 13:03:24.583: INFO: Scaling statefulset ss to 0
+STEP: Verifying that stateful set ss was scaled down in reverse order
+[AfterEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:85
+Apr 29 13:03:44.618: INFO: Deleting all statefulset in ns statefulset-7999
+Apr 29 13:03:44.623: INFO: Scaling statefulset ss to 0
+Apr 29 13:03:44.638: INFO: Waiting for statefulset status.replicas updated to 0
+Apr 29 13:03:44.642: INFO: Deleting statefulset ss
+[AfterEach] [sig-apps] StatefulSet
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 13:03:44.674: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "statefulset-7999" for this suite.
+Apr 29 13:03:50.703: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 13:03:50.891: INFO: namespace statefulset-7999 deletion completed in 6.209519316s
+
+• [SLOW TEST:89.431 seconds]
+[sig-apps] StatefulSet
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22
+  [k8s.io] Basic StatefulSet functionality [StatefulSetBasic]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
+    Scaling should happen in predictable order and halt if any stateful pod is unhealthy [Conformance]
+    /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSSSSSSSSSSSSSSSSSSSSS
+------------------------------
+[k8s.io] Kubelet when scheduling a busybox Pod with hostAliases 
+  should write entries to /etc/hosts [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [k8s.io] Kubelet
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 13:03:50.892: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename kubelet-test
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubelet-test-783
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [k8s.io] Kubelet
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:37
+[It] should write entries to /etc/hosts [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[AfterEach] [k8s.io] Kubelet
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 13:03:57.219: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "kubelet-test-783" for this suite.
+Apr 29 13:04:37.245: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 13:04:37.431: INFO: namespace kubelet-test-783 deletion completed in 40.206275349s
+
+• [SLOW TEST:46.539 seconds]
+[k8s.io] Kubelet
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
+  when scheduling a busybox Pod with hostAliases
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:136
+    should write entries to /etc/hosts [LinuxOnly] [NodeConformance] [Conformance]
+    /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSS
+------------------------------
+[sig-storage] Projected configMap 
+  should be consumable from pods in volume with mappings as non-root [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-storage] Projected configMap
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 13:04:37.439: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename projected
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-1806
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should be consumable from pods in volume with mappings as non-root [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating configMap with name projected-configmap-test-volume-map-57b385e5-6a7f-11e9-b6ef-0e71f8c7e888
+STEP: Creating a pod to test consume configMaps
+Apr 29 13:04:37.668: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-57b64d0c-6a7f-11e9-b6ef-0e71f8c7e888" in namespace "projected-1806" to be "success or failure"
+Apr 29 13:04:37.688: INFO: Pod "pod-projected-configmaps-57b64d0c-6a7f-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 20.386796ms
+Apr 29 13:04:39.695: INFO: Pod "pod-projected-configmaps-57b64d0c-6a7f-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 2.027610557s
+Apr 29 13:04:41.703: INFO: Pod "pod-projected-configmaps-57b64d0c-6a7f-11e9-b6ef-0e71f8c7e888": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.035564975s
+STEP: Saw pod success
+Apr 29 13:04:41.703: INFO: Pod "pod-projected-configmaps-57b64d0c-6a7f-11e9-b6ef-0e71f8c7e888" satisfied condition "success or failure"
+Apr 29 13:04:41.707: INFO: Trying to get logs from node worker-4av2h-6665d7b5ff-9nc42 pod pod-projected-configmaps-57b64d0c-6a7f-11e9-b6ef-0e71f8c7e888 container projected-configmap-volume-test: 
+STEP: delete the pod
+Apr 29 13:04:41.743: INFO: Waiting for pod pod-projected-configmaps-57b64d0c-6a7f-11e9-b6ef-0e71f8c7e888 to disappear
+Apr 29 13:04:41.750: INFO: Pod pod-projected-configmaps-57b64d0c-6a7f-11e9-b6ef-0e71f8c7e888 no longer exists
+[AfterEach] [sig-storage] Projected configMap
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 13:04:41.750: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "projected-1806" for this suite.
+Apr 29 13:04:47.779: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 13:04:48.059: INFO: namespace projected-1806 deletion completed in 6.301304617s
+
+• [SLOW TEST:10.621 seconds]
+[sig-storage] Projected configMap
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_configmap.go:33
+  should be consumable from pods in volume with mappings as non-root [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSSSSSSSSSSSSSSSSSS
+------------------------------
+[sig-cli] Kubectl client [k8s.io] Kubectl logs 
+  should be able to retrieve and filter logs  [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-cli] Kubectl client
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 13:04:48.060: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename kubectl
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-7303
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [sig-cli] Kubectl client
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:213
+[BeforeEach] [k8s.io] Kubectl logs
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1190
+STEP: creating an rc
+Apr 29 13:04:48.333: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 create -f - --namespace=kubectl-7303'
+Apr 29 13:04:49.313: INFO: stderr: ""
+Apr 29 13:04:49.313: INFO: stdout: "replicationcontroller/redis-master created\n"
+[It] should be able to retrieve and filter logs  [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Waiting for Redis master to start.
+Apr 29 13:04:50.334: INFO: Selector matched 1 pods for map[app:redis]
+Apr 29 13:04:50.334: INFO: Found 0 / 1
+Apr 29 13:04:51.321: INFO: Selector matched 1 pods for map[app:redis]
+Apr 29 13:04:51.321: INFO: Found 0 / 1
+Apr 29 13:04:52.320: INFO: Selector matched 1 pods for map[app:redis]
+Apr 29 13:04:52.320: INFO: Found 0 / 1
+Apr 29 13:04:53.320: INFO: Selector matched 1 pods for map[app:redis]
+Apr 29 13:04:53.321: INFO: Found 0 / 1
+Apr 29 13:04:54.319: INFO: Selector matched 1 pods for map[app:redis]
+Apr 29 13:04:54.319: INFO: Found 0 / 1
+Apr 29 13:04:55.320: INFO: Selector matched 1 pods for map[app:redis]
+Apr 29 13:04:55.320: INFO: Found 0 / 1
+Apr 29 13:04:56.320: INFO: Selector matched 1 pods for map[app:redis]
+Apr 29 13:04:56.320: INFO: Found 0 / 1
+Apr 29 13:04:57.319: INFO: Selector matched 1 pods for map[app:redis]
+Apr 29 13:04:57.320: INFO: Found 0 / 1
+Apr 29 13:04:58.320: INFO: Selector matched 1 pods for map[app:redis]
+Apr 29 13:04:58.320: INFO: Found 0 / 1
+Apr 29 13:04:59.321: INFO: Selector matched 1 pods for map[app:redis]
+Apr 29 13:04:59.321: INFO: Found 0 / 1
+Apr 29 13:05:00.322: INFO: Selector matched 1 pods for map[app:redis]
+Apr 29 13:05:00.322: INFO: Found 0 / 1
+Apr 29 13:05:01.320: INFO: Selector matched 1 pods for map[app:redis]
+Apr 29 13:05:01.320: INFO: Found 0 / 1
+Apr 29 13:05:02.318: INFO: Selector matched 1 pods for map[app:redis]
+Apr 29 13:05:02.319: INFO: Found 1 / 1
+Apr 29 13:05:02.319: INFO: WaitFor completed with timeout 5m0s.  Pods found = 1 out of 1
+Apr 29 13:05:02.323: INFO: Selector matched 1 pods for map[app:redis]
+Apr 29 13:05:02.323: INFO: ForEach: Found 1 pods from the filter.  Now looping through them.
+STEP: checking for a matching strings
+Apr 29 13:05:02.323: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 logs redis-master-rfn5x redis-master --namespace=kubectl-7303'
+Apr 29 13:05:02.450: INFO: stderr: ""
+Apr 29 13:05:02.450: INFO: stdout: "                _._                                                  \n           _.-``__ ''-._                                             \n      _.-``    `.  `_.  ''-._           Redis 3.2.12 (35a5711f/0) 64 bit\n  .-`` .-```.  ```\\/    _.,_ ''-._                                   \n (    '      ,       .-`  | `,    )     Running in standalone mode\n |`-._`-...-` __...-.``-._|'` _.-'|     Port: 6379\n |    `-._   `._    /     _.-'    |     PID: 1\n  `-._    `-._  `-./  _.-'    _.-'                                   \n |`-._`-._    `-.__.-'    _.-'_.-'|                                  \n |    `-._`-._        _.-'_.-'    |           http://redis.io        \n  `-._    `-._`-.__.-'_.-'    _.-'                                   \n |`-._`-._    `-.__.-'    _.-'_.-'|                                  \n |    `-._`-._        _.-'_.-'    |                                  \n  `-._    `-._`-.__.-'_.-'    _.-'                                   \n      `-._    `-.__.-'    _.-'                                       \n          `-._        _.-'                                           \n              `-.__.-'                                               \n\n1:M 29 Apr 13:05:00.512 # WARNING: The TCP backlog setting of 511 cannot be enforced because /proc/sys/net/core/somaxconn is set to the lower value of 128.\n1:M 29 Apr 13:05:00.512 # Server started, Redis version 3.2.12\n1:M 29 Apr 13:05:00.512 # WARNING you have Transparent Huge Pages (THP) support enabled in your kernel. This will create latency and memory usage issues with Redis. To fix this issue run the command 'echo never > /sys/kernel/mm/transparent_hugepage/enabled' as root, and add it to your /etc/rc.local in order to retain the setting after a reboot. Redis must be restarted after THP is disabled.\n1:M 29 Apr 13:05:00.512 * The server is now ready to accept connections on port 6379\n"
+STEP: limiting log lines
+Apr 29 13:05:02.450: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 log redis-master-rfn5x redis-master --namespace=kubectl-7303 --tail=1'
+Apr 29 13:05:02.570: INFO: stderr: ""
+Apr 29 13:05:02.570: INFO: stdout: "1:M 29 Apr 13:05:00.512 * The server is now ready to accept connections on port 6379\n"
+STEP: limiting log bytes
+Apr 29 13:05:02.571: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 log redis-master-rfn5x redis-master --namespace=kubectl-7303 --limit-bytes=1'
+Apr 29 13:05:02.697: INFO: stderr: ""
+Apr 29 13:05:02.697: INFO: stdout: " "
+STEP: exposing timestamps
+Apr 29 13:05:02.697: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 log redis-master-rfn5x redis-master --namespace=kubectl-7303 --tail=1 --timestamps'
+Apr 29 13:05:02.837: INFO: stderr: ""
+Apr 29 13:05:02.837: INFO: stdout: "2019-04-29T13:05:00.512911169Z 1:M 29 Apr 13:05:00.512 * The server is now ready to accept connections on port 6379\n"
+STEP: restricting to a time range
+Apr 29 13:05:05.338: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 log redis-master-rfn5x redis-master --namespace=kubectl-7303 --since=1s'
+Apr 29 13:05:05.480: INFO: stderr: ""
+Apr 29 13:05:05.480: INFO: stdout: ""
+Apr 29 13:05:05.480: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 log redis-master-rfn5x redis-master --namespace=kubectl-7303 --since=24h'
+Apr 29 13:05:05.624: INFO: stderr: ""
+Apr 29 13:05:05.624: INFO: stdout: "                _._                                                  \n           _.-``__ ''-._                                             \n      _.-``    `.  `_.  ''-._           Redis 3.2.12 (35a5711f/0) 64 bit\n  .-`` .-```.  ```\\/    _.,_ ''-._                                   \n (    '      ,       .-`  | `,    )     Running in standalone mode\n |`-._`-...-` __...-.``-._|'` _.-'|     Port: 6379\n |    `-._   `._    /     _.-'    |     PID: 1\n  `-._    `-._  `-./  _.-'    _.-'                                   \n |`-._`-._    `-.__.-'    _.-'_.-'|                                  \n |    `-._`-._        _.-'_.-'    |           http://redis.io        \n  `-._    `-._`-.__.-'_.-'    _.-'                                   \n |`-._`-._    `-.__.-'    _.-'_.-'|                                  \n |    `-._`-._        _.-'_.-'    |                                  \n  `-._    `-._`-.__.-'_.-'    _.-'                                   \n      `-._    `-.__.-'    _.-'                                       \n          `-._        _.-'                                           \n              `-.__.-'                                               \n\n1:M 29 Apr 13:05:00.512 # WARNING: The TCP backlog setting of 511 cannot be enforced because /proc/sys/net/core/somaxconn is set to the lower value of 128.\n1:M 29 Apr 13:05:00.512 # Server started, Redis version 3.2.12\n1:M 29 Apr 13:05:00.512 # WARNING you have Transparent Huge Pages (THP) support enabled in your kernel. This will create latency and memory usage issues with Redis. To fix this issue run the command 'echo never > /sys/kernel/mm/transparent_hugepage/enabled' as root, and add it to your /etc/rc.local in order to retain the setting after a reboot. Redis must be restarted after THP is disabled.\n1:M 29 Apr 13:05:00.512 * The server is now ready to accept connections on port 6379\n"
+[AfterEach] [k8s.io] Kubectl logs
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1196
+STEP: using delete to clean up resources
+Apr 29 13:05:05.624: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 delete --grace-period=0 --force -f - --namespace=kubectl-7303'
+Apr 29 13:05:05.775: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n"
+Apr 29 13:05:05.775: INFO: stdout: "replicationcontroller \"redis-master\" force deleted\n"
+Apr 29 13:05:05.775: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 get rc,svc -l name=nginx --no-headers --namespace=kubectl-7303'
+Apr 29 13:05:05.956: INFO: stderr: "No resources found.\n"
+Apr 29 13:05:05.956: INFO: stdout: ""
+Apr 29 13:05:05.956: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 get pods -l name=nginx --namespace=kubectl-7303 -o go-template={{ range .items }}{{ if not .metadata.deletionTimestamp }}{{ .metadata.name }}{{ "\n" }}{{ end }}{{ end }}'
+Apr 29 13:05:06.109: INFO: stderr: ""
+Apr 29 13:05:06.109: INFO: stdout: ""
+[AfterEach] [sig-cli] Kubectl client
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 13:05:06.110: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "kubectl-7303" for this suite.
+Apr 29 13:05:12.139: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 13:05:12.320: INFO: namespace kubectl-7303 deletion completed in 6.203855357s
+
+• [SLOW TEST:24.260 seconds]
+[sig-cli] Kubectl client
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23
+  [k8s.io] Kubectl logs
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
+    should be able to retrieve and filter logs  [Conformance]
+    /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSS
+------------------------------
+[sig-storage] Downward API volume 
+  should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-storage] Downward API volume
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 13:05:12.326: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename downward-api
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in downward-api-3608
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [sig-storage] Downward API volume
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:39
+[It] should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating a pod to test downward API volume plugin
+Apr 29 13:05:12.519: INFO: Waiting up to 5m0s for pod "downwardapi-volume-6c7db0ae-6a7f-11e9-b6ef-0e71f8c7e888" in namespace "downward-api-3608" to be "success or failure"
+Apr 29 13:05:12.523: INFO: Pod "downwardapi-volume-6c7db0ae-6a7f-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 4.556892ms
+Apr 29 13:05:14.531: INFO: Pod "downwardapi-volume-6c7db0ae-6a7f-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 2.011772881s
+Apr 29 13:05:16.557: INFO: Pod "downwardapi-volume-6c7db0ae-6a7f-11e9-b6ef-0e71f8c7e888": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.038080052s
+STEP: Saw pod success
+Apr 29 13:05:16.557: INFO: Pod "downwardapi-volume-6c7db0ae-6a7f-11e9-b6ef-0e71f8c7e888" satisfied condition "success or failure"
+Apr 29 13:05:16.569: INFO: Trying to get logs from node worker-4av2h-6665d7b5ff-9nc42 pod downwardapi-volume-6c7db0ae-6a7f-11e9-b6ef-0e71f8c7e888 container client-container: 
+STEP: delete the pod
+Apr 29 13:05:16.605: INFO: Waiting for pod downwardapi-volume-6c7db0ae-6a7f-11e9-b6ef-0e71f8c7e888 to disappear
+Apr 29 13:05:16.612: INFO: Pod downwardapi-volume-6c7db0ae-6a7f-11e9-b6ef-0e71f8c7e888 no longer exists
+[AfterEach] [sig-storage] Downward API volume
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 13:05:16.612: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "downward-api-3608" for this suite.
+Apr 29 13:05:22.636: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 13:05:22.864: INFO: namespace downward-api-3608 deletion completed in 6.245351692s
+
+• [SLOW TEST:10.539 seconds]
+[sig-storage] Downward API volume
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:34
+  should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSSSSSSSSSSSSSSSSSSS
+------------------------------
+[sig-storage] Projected downwardAPI 
+  should provide container's cpu limit [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-storage] Projected downwardAPI
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 13:05:22.865: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename projected
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-3546
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [sig-storage] Projected downwardAPI
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:39
+[It] should provide container's cpu limit [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating a pod to test downward API volume plugin
+Apr 29 13:05:23.129: INFO: Waiting up to 5m0s for pod "downwardapi-volume-72cc0f6a-6a7f-11e9-b6ef-0e71f8c7e888" in namespace "projected-3546" to be "success or failure"
+Apr 29 13:05:23.141: INFO: Pod "downwardapi-volume-72cc0f6a-6a7f-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 11.650098ms
+Apr 29 13:05:25.163: INFO: Pod "downwardapi-volume-72cc0f6a-6a7f-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 2.032890094s
+Apr 29 13:05:27.171: INFO: Pod "downwardapi-volume-72cc0f6a-6a7f-11e9-b6ef-0e71f8c7e888": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.041769934s
+STEP: Saw pod success
+Apr 29 13:05:27.172: INFO: Pod "downwardapi-volume-72cc0f6a-6a7f-11e9-b6ef-0e71f8c7e888" satisfied condition "success or failure"
+Apr 29 13:05:27.177: INFO: Trying to get logs from node worker-4av2h-6665d7b5ff-9nc42 pod downwardapi-volume-72cc0f6a-6a7f-11e9-b6ef-0e71f8c7e888 container client-container: 
+STEP: delete the pod
+Apr 29 13:05:27.242: INFO: Waiting for pod downwardapi-volume-72cc0f6a-6a7f-11e9-b6ef-0e71f8c7e888 to disappear
+Apr 29 13:05:27.250: INFO: Pod downwardapi-volume-72cc0f6a-6a7f-11e9-b6ef-0e71f8c7e888 no longer exists
+[AfterEach] [sig-storage] Projected downwardAPI
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 13:05:27.250: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "projected-3546" for this suite.
+Apr 29 13:05:33.285: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 13:05:33.482: INFO: namespace projected-3546 deletion completed in 6.224071168s
+
+• [SLOW TEST:10.617 seconds]
+[sig-storage] Projected downwardAPI
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:33
+  should provide container's cpu limit [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SS
+------------------------------
+[sig-storage] Downward API volume 
+  should provide container's memory request [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-storage] Downward API volume
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 13:05:33.488: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename downward-api
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in downward-api-8543
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [sig-storage] Downward API volume
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:39
+[It] should provide container's memory request [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating a pod to test downward API volume plugin
+Apr 29 13:05:33.708: INFO: Waiting up to 5m0s for pod "downwardapi-volume-791e5379-6a7f-11e9-b6ef-0e71f8c7e888" in namespace "downward-api-8543" to be "success or failure"
+Apr 29 13:05:33.718: INFO: Pod "downwardapi-volume-791e5379-6a7f-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 9.487691ms
+Apr 29 13:05:35.725: INFO: Pod "downwardapi-volume-791e5379-6a7f-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 2.017441092s
+Apr 29 13:05:37.734: INFO: Pod "downwardapi-volume-791e5379-6a7f-11e9-b6ef-0e71f8c7e888": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.025908965s
+STEP: Saw pod success
+Apr 29 13:05:37.734: INFO: Pod "downwardapi-volume-791e5379-6a7f-11e9-b6ef-0e71f8c7e888" satisfied condition "success or failure"
+Apr 29 13:05:37.740: INFO: Trying to get logs from node worker-4av2h-6665d7b5ff-9nc42 pod downwardapi-volume-791e5379-6a7f-11e9-b6ef-0e71f8c7e888 container client-container: 
+STEP: delete the pod
+Apr 29 13:05:37.773: INFO: Waiting for pod downwardapi-volume-791e5379-6a7f-11e9-b6ef-0e71f8c7e888 to disappear
+Apr 29 13:05:37.779: INFO: Pod downwardapi-volume-791e5379-6a7f-11e9-b6ef-0e71f8c7e888 no longer exists
+[AfterEach] [sig-storage] Downward API volume
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 13:05:37.779: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "downward-api-8543" for this suite.
+Apr 29 13:05:43.816: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 13:05:44.040: INFO: namespace downward-api-8543 deletion completed in 6.254130669s
+
+• [SLOW TEST:10.553 seconds]
+[sig-storage] Downward API volume
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:34
+  should provide container's memory request [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSSSSSSSSSSSSS
+------------------------------
+[k8s.io] Pods 
+  should get a host IP [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [k8s.io] Pods
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 13:05:44.040: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename pods
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in pods-5548
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [k8s.io] Pods
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:135
+[It] should get a host IP [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: creating pod
+Apr 29 13:05:50.286: INFO: Pod pod-hostip-7f6692b9-6a7f-11e9-b6ef-0e71f8c7e888 has hostIP: 172.23.0.238
+[AfterEach] [k8s.io] Pods
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 13:05:50.286: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "pods-5548" for this suite.
+Apr 29 13:06:14.347: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 13:06:14.523: INFO: namespace pods-5548 deletion completed in 24.211201898s
+
+• [SLOW TEST:30.483 seconds]
+[k8s.io] Pods
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
+  should get a host IP [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSS
+------------------------------
+[k8s.io] Container Lifecycle Hook when create a pod with lifecycle hook 
+  should execute poststart http hook properly [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [k8s.io] Container Lifecycle Hook
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 13:06:14.526: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename container-lifecycle-hook
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in container-lifecycle-hook-8923
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] when create a pod with lifecycle hook
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:61
+STEP: create the container to handle the HTTPGet hook request.
+[It] should execute poststart http hook properly [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: create the pod with lifecycle hook
+STEP: check poststart hook
+STEP: delete the pod with lifecycle hook
+Apr 29 13:06:22.815: INFO: Waiting for pod pod-with-poststart-http-hook to disappear
+Apr 29 13:06:22.821: INFO: Pod pod-with-poststart-http-hook still exists
+Apr 29 13:06:24.822: INFO: Waiting for pod pod-with-poststart-http-hook to disappear
+Apr 29 13:06:24.829: INFO: Pod pod-with-poststart-http-hook still exists
+Apr 29 13:06:26.822: INFO: Waiting for pod pod-with-poststart-http-hook to disappear
+Apr 29 13:06:26.829: INFO: Pod pod-with-poststart-http-hook still exists
+Apr 29 13:06:28.822: INFO: Waiting for pod pod-with-poststart-http-hook to disappear
+Apr 29 13:06:28.831: INFO: Pod pod-with-poststart-http-hook still exists
+Apr 29 13:06:30.822: INFO: Waiting for pod pod-with-poststart-http-hook to disappear
+Apr 29 13:06:30.830: INFO: Pod pod-with-poststart-http-hook still exists
+Apr 29 13:06:32.822: INFO: Waiting for pod pod-with-poststart-http-hook to disappear
+Apr 29 13:06:32.829: INFO: Pod pod-with-poststart-http-hook still exists
+Apr 29 13:06:34.822: INFO: Waiting for pod pod-with-poststart-http-hook to disappear
+Apr 29 13:06:34.829: INFO: Pod pod-with-poststart-http-hook no longer exists
+[AfterEach] [k8s.io] Container Lifecycle Hook
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 13:06:34.830: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "container-lifecycle-hook-8923" for this suite.
+Apr 29 13:06:58.865: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 13:06:59.051: INFO: namespace container-lifecycle-hook-8923 deletion completed in 24.211864892s
+
+• [SLOW TEST:44.525 seconds]
+[k8s.io] Container Lifecycle Hook
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
+  when create a pod with lifecycle hook
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:40
+    should execute poststart http hook properly [NodeConformance] [Conformance]
+    /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSSSSSSSSSSSSSSS
+------------------------------
+[sig-network] Proxy version v1 
+  should proxy logs on node with explicit kubelet port using proxy subresource  [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] version v1
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 13:06:59.054: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename proxy
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in proxy-8859
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should proxy logs on node with explicit kubelet port using proxy subresource  [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+Apr 29 13:06:59.258: INFO: (0) /api/v1/nodes/worker-3oo4n-86cbf586b7-9q98g:10250/proxy/logs/: 
+btmp
+containers/
+faillog... (200; 5.871108ms)
+Apr 29 13:06:59.263: INFO: (1) /api/v1/nodes/worker-3oo4n-86cbf586b7-9q98g:10250/proxy/logs/: 
+btmp
+containers/
+faillog... (200; 4.828415ms)
+Apr 29 13:06:59.268: INFO: (2) /api/v1/nodes/worker-3oo4n-86cbf586b7-9q98g:10250/proxy/logs/: 
+btmp
+containers/
+faillog... (200; 5.214916ms)
+Apr 29 13:06:59.277: INFO: (3) /api/v1/nodes/worker-3oo4n-86cbf586b7-9q98g:10250/proxy/logs/: 
+btmp
+containers/
+faillog... (200; 8.659416ms)
+Apr 29 13:06:59.284: INFO: (4) /api/v1/nodes/worker-3oo4n-86cbf586b7-9q98g:10250/proxy/logs/: 
+btmp
+containers/
+faillog... (200; 6.441247ms)
+Apr 29 13:06:59.290: INFO: (5) /api/v1/nodes/worker-3oo4n-86cbf586b7-9q98g:10250/proxy/logs/: 
+btmp
+containers/
+faillog... (200; 5.974255ms)
+Apr 29 13:06:59.295: INFO: (6) /api/v1/nodes/worker-3oo4n-86cbf586b7-9q98g:10250/proxy/logs/: 
+btmp
+containers/
+faillog... (200; 5.367642ms)
+Apr 29 13:06:59.303: INFO: (7) /api/v1/nodes/worker-3oo4n-86cbf586b7-9q98g:10250/proxy/logs/: 
+btmp
+containers/
+faillog... (200; 7.208307ms)
+Apr 29 13:06:59.309: INFO: (8) /api/v1/nodes/worker-3oo4n-86cbf586b7-9q98g:10250/proxy/logs/: 
+btmp
+containers/
+faillog... (200; 6.365762ms)
+Apr 29 13:06:59.317: INFO: (9) /api/v1/nodes/worker-3oo4n-86cbf586b7-9q98g:10250/proxy/logs/: 
+btmp
+containers/
+faillog... (200; 7.334905ms)
+Apr 29 13:06:59.324: INFO: (10) /api/v1/nodes/worker-3oo4n-86cbf586b7-9q98g:10250/proxy/logs/: 
+btmp
+containers/
+faillog... (200; 6.976124ms)
+Apr 29 13:06:59.333: INFO: (11) /api/v1/nodes/worker-3oo4n-86cbf586b7-9q98g:10250/proxy/logs/: 
+btmp
+containers/
+faillog... (200; 9.275261ms)
+Apr 29 13:06:59.341: INFO: (12) /api/v1/nodes/worker-3oo4n-86cbf586b7-9q98g:10250/proxy/logs/: 
+btmp
+containers/
+faillog... (200; 8.016306ms)
+Apr 29 13:06:59.347: INFO: (13) /api/v1/nodes/worker-3oo4n-86cbf586b7-9q98g:10250/proxy/logs/: 
+btmp
+containers/
+faillog... (200; 5.793297ms)
+Apr 29 13:06:59.352: INFO: (14) /api/v1/nodes/worker-3oo4n-86cbf586b7-9q98g:10250/proxy/logs/: 
+btmp
+containers/
+faillog... (200; 4.868304ms)
+Apr 29 13:06:59.358: INFO: (15) /api/v1/nodes/worker-3oo4n-86cbf586b7-9q98g:10250/proxy/logs/: 
+btmp
+containers/
+faillog... (200; 5.925233ms)
+Apr 29 13:06:59.364: INFO: (16) /api/v1/nodes/worker-3oo4n-86cbf586b7-9q98g:10250/proxy/logs/: 
+btmp
+containers/
+faillog... (200; 5.671472ms)
+Apr 29 13:06:59.371: INFO: (17) /api/v1/nodes/worker-3oo4n-86cbf586b7-9q98g:10250/proxy/logs/: 
+btmp
+containers/
+faillog... (200; 7.605323ms)
+Apr 29 13:06:59.379: INFO: (18) /api/v1/nodes/worker-3oo4n-86cbf586b7-9q98g:10250/proxy/logs/: 
+btmp
+containers/
+faillog... (200; 7.421335ms)
+Apr 29 13:06:59.384: INFO: (19) /api/v1/nodes/worker-3oo4n-86cbf586b7-9q98g:10250/proxy/logs/: 
+btmp
+containers/
+faillog... (200; 4.678549ms)
+[AfterEach] version v1
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 13:06:59.384: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "proxy-8859" for this suite.
+Apr 29 13:07:05.408: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 13:07:05.643: INFO: namespace proxy-8859 deletion completed in 6.254754253s
+
+• [SLOW TEST:6.590 seconds]
+[sig-network] Proxy
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:22
+  version v1
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/proxy.go:56
+    should proxy logs on node with explicit kubelet port using proxy subresource  [Conformance]
+    /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+------------------------------
+[sig-api-machinery] Watchers 
+  should be able to restart watching from the last resource version observed by the previous watch [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-api-machinery] Watchers
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 13:07:05.659: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename watch
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in watch-9629
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should be able to restart watching from the last resource version observed by the previous watch [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: creating a watch on configmaps
+STEP: creating a new configmap
+STEP: modifying the configmap once
+STEP: closing the watch once it receives two notifications
+Apr 29 13:07:05.884: INFO: Got : ADDED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-watch-closed,GenerateName:,Namespace:watch-9629,SelfLink:/api/v1/namespaces/watch-9629/configmaps/e2e-watch-test-watch-closed,UID:afd9c213-6a7f-11e9-92a7-deadbef3c36d,ResourceVersion:27292,Generation:0,CreationTimestamp:2019-04-29 13:07:05 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: watch-closed-and-restarted,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{},BinaryData:map[string][]byte{},}
+Apr 29 13:07:05.884: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-watch-closed,GenerateName:,Namespace:watch-9629,SelfLink:/api/v1/namespaces/watch-9629/configmaps/e2e-watch-test-watch-closed,UID:afd9c213-6a7f-11e9-92a7-deadbef3c36d,ResourceVersion:27293,Generation:0,CreationTimestamp:2019-04-29 13:07:05 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: watch-closed-and-restarted,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{mutation: 1,},BinaryData:map[string][]byte{},}
+STEP: modifying the configmap a second time, while the watch is closed
+STEP: creating a new watch on configmaps from the last resource version observed by the first watch
+STEP: deleting the configmap
+STEP: Expecting to observe notifications for all changes to the configmap since the first watch closed
+Apr 29 13:07:05.930: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-watch-closed,GenerateName:,Namespace:watch-9629,SelfLink:/api/v1/namespaces/watch-9629/configmaps/e2e-watch-test-watch-closed,UID:afd9c213-6a7f-11e9-92a7-deadbef3c36d,ResourceVersion:27294,Generation:0,CreationTimestamp:2019-04-29 13:07:05 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: watch-closed-and-restarted,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},}
+Apr 29 13:07:05.930: INFO: Got : DELETED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-watch-closed,GenerateName:,Namespace:watch-9629,SelfLink:/api/v1/namespaces/watch-9629/configmaps/e2e-watch-test-watch-closed,UID:afd9c213-6a7f-11e9-92a7-deadbef3c36d,ResourceVersion:27295,Generation:0,CreationTimestamp:2019-04-29 13:07:05 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: watch-closed-and-restarted,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},}
+[AfterEach] [sig-api-machinery] Watchers
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 13:07:05.930: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "watch-9629" for this suite.
+Apr 29 13:07:11.964: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 13:07:12.182: INFO: namespace watch-9629 deletion completed in 6.244171994s
+
+• [SLOW TEST:6.524 seconds]
+[sig-api-machinery] Watchers
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:22
+  should be able to restart watching from the last resource version observed by the previous watch [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSS
+------------------------------
+[k8s.io] Probing container 
+  with readiness probe that fails should never be ready and never restart [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [k8s.io] Probing container
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 13:07:12.184: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename container-probe
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in container-probe-8419
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [k8s.io] Probing container
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/container_probe.go:51
+[It] with readiness probe that fails should never be ready and never restart [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[AfterEach] [k8s.io] Probing container
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 13:08:12.398: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "container-probe-8419" for this suite.
+Apr 29 13:08:36.432: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 13:08:36.666: INFO: namespace container-probe-8419 deletion completed in 24.260007221s
+
+• [SLOW TEST:84.482 seconds]
+[k8s.io] Probing container
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
+  with readiness probe that fails should never be ready and never restart [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSSSSSSSSSSSSSSS
+------------------------------
+[sig-cli] Kubectl client [k8s.io] Kubectl run pod 
+  should create a pod from an image when restart is Never  [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-cli] Kubectl client
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 13:08:36.667: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename kubectl
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-1749
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [sig-cli] Kubectl client
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:213
+[BeforeEach] [k8s.io] Kubectl run pod
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1583
+[It] should create a pod from an image when restart is Never  [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: running the image docker.io/library/nginx:1.14-alpine
+Apr 29 13:08:36.856: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 run e2e-test-nginx-pod --restart=Never --generator=run-pod/v1 --image=docker.io/library/nginx:1.14-alpine --namespace=kubectl-1749'
+Apr 29 13:08:36.984: INFO: stderr: ""
+Apr 29 13:08:36.984: INFO: stdout: "pod/e2e-test-nginx-pod created\n"
+STEP: verifying the pod e2e-test-nginx-pod was created
+[AfterEach] [k8s.io] Kubectl run pod
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1588
+Apr 29 13:08:36.994: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 delete pods e2e-test-nginx-pod --namespace=kubectl-1749'
+Apr 29 13:08:44.424: INFO: stderr: ""
+Apr 29 13:08:44.424: INFO: stdout: "pod \"e2e-test-nginx-pod\" deleted\n"
+[AfterEach] [sig-cli] Kubectl client
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 13:08:44.424: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "kubectl-1749" for this suite.
+Apr 29 13:08:50.455: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 13:08:50.661: INFO: namespace kubectl-1749 deletion completed in 6.229986655s
+
+• [SLOW TEST:13.994 seconds]
+[sig-cli] Kubectl client
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23
+  [k8s.io] Kubectl run pod
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
+    should create a pod from an image when restart is Never  [Conformance]
+    /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSS
+------------------------------
+[k8s.io] Docker Containers 
+  should be able to override the image's default command (docker entrypoint) [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [k8s.io] Docker Containers
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 13:08:50.661: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename containers
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in containers-8643
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should be able to override the image's default command (docker entrypoint) [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating a pod to test override command
+Apr 29 13:08:50.874: INFO: Waiting up to 5m0s for pod "client-containers-eea2d61a-6a7f-11e9-b6ef-0e71f8c7e888" in namespace "containers-8643" to be "success or failure"
+Apr 29 13:08:50.884: INFO: Pod "client-containers-eea2d61a-6a7f-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 9.620911ms
+Apr 29 13:08:52.892: INFO: Pod "client-containers-eea2d61a-6a7f-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 2.017061919s
+Apr 29 13:08:54.900: INFO: Pod "client-containers-eea2d61a-6a7f-11e9-b6ef-0e71f8c7e888": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.024908177s
+STEP: Saw pod success
+Apr 29 13:08:54.900: INFO: Pod "client-containers-eea2d61a-6a7f-11e9-b6ef-0e71f8c7e888" satisfied condition "success or failure"
+Apr 29 13:08:54.905: INFO: Trying to get logs from node worker-4av2h-6665d7b5ff-9nc42 pod client-containers-eea2d61a-6a7f-11e9-b6ef-0e71f8c7e888 container test-container: 
+STEP: delete the pod
+Apr 29 13:08:54.937: INFO: Waiting for pod client-containers-eea2d61a-6a7f-11e9-b6ef-0e71f8c7e888 to disappear
+Apr 29 13:08:54.943: INFO: Pod client-containers-eea2d61a-6a7f-11e9-b6ef-0e71f8c7e888 no longer exists
+[AfterEach] [k8s.io] Docker Containers
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 13:08:54.944: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "containers-8643" for this suite.
+Apr 29 13:09:00.975: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 13:09:01.253: INFO: namespace containers-8643 deletion completed in 6.299385099s
+
+• [SLOW TEST:10.592 seconds]
+[k8s.io] Docker Containers
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
+  should be able to override the image's default command (docker entrypoint) [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSS
+------------------------------
+[sig-api-machinery] Garbage collector 
+  should not be blocked by dependency circle [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-api-machinery] Garbage collector
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 13:09:01.253: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename gc
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in gc-967
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should not be blocked by dependency circle [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+Apr 29 13:09:01.533: INFO: pod1.ObjectMeta.OwnerReferences=[]v1.OwnerReference{v1.OwnerReference{APIVersion:"v1", Kind:"Pod", Name:"pod3", UID:"f4c739e0-6a7f-11e9-92a7-deadbef3c36d", Controller:(*bool)(0xc00268af5e), BlockOwnerDeletion:(*bool)(0xc00268af5f)}}
+Apr 29 13:09:01.548: INFO: pod2.ObjectMeta.OwnerReferences=[]v1.OwnerReference{v1.OwnerReference{APIVersion:"v1", Kind:"Pod", Name:"pod1", UID:"f4c3828d-6a7f-11e9-92a7-deadbef3c36d", Controller:(*bool)(0xc000c4de5e), BlockOwnerDeletion:(*bool)(0xc000c4de5f)}}
+Apr 29 13:09:01.564: INFO: pod3.ObjectMeta.OwnerReferences=[]v1.OwnerReference{v1.OwnerReference{APIVersion:"v1", Kind:"Pod", Name:"pod2", UID:"f4c5b1d0-6a7f-11e9-92a7-deadbef3c36d", Controller:(*bool)(0xc00268b11e), BlockOwnerDeletion:(*bool)(0xc00268b11f)}}
+[AfterEach] [sig-api-machinery] Garbage collector
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 13:09:06.578: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "gc-967" for this suite.
+Apr 29 13:09:12.611: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 13:09:12.810: INFO: namespace gc-967 deletion completed in 6.222759057s
+
+• [SLOW TEST:11.557 seconds]
+[sig-api-machinery] Garbage collector
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:22
+  should not be blocked by dependency circle [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSSSSSSSSSSSSSSSSS
+------------------------------
+[sig-storage] Subpath Atomic writer volumes 
+  should support subpaths with configmap pod with mountPath of existing file [LinuxOnly] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-storage] Subpath
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 13:09:12.811: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename subpath
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in subpath-6204
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] Atomic writer volumes
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:38
+STEP: Setting up data
+[It] should support subpaths with configmap pod with mountPath of existing file [LinuxOnly] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating pod pod-subpath-test-configmap-crxb
+STEP: Creating a pod to test atomic-volume-subpath
+Apr 29 13:09:13.022: INFO: Waiting up to 5m0s for pod "pod-subpath-test-configmap-crxb" in namespace "subpath-6204" to be "success or failure"
+Apr 29 13:09:13.028: INFO: Pod "pod-subpath-test-configmap-crxb": Phase="Pending", Reason="", readiness=false. Elapsed: 5.829328ms
+Apr 29 13:09:15.035: INFO: Pod "pod-subpath-test-configmap-crxb": Phase="Pending", Reason="", readiness=false. Elapsed: 2.013278214s
+Apr 29 13:09:17.042: INFO: Pod "pod-subpath-test-configmap-crxb": Phase="Running", Reason="", readiness=true. Elapsed: 4.019418225s
+Apr 29 13:09:19.048: INFO: Pod "pod-subpath-test-configmap-crxb": Phase="Running", Reason="", readiness=true. Elapsed: 6.025472175s
+Apr 29 13:09:21.056: INFO: Pod "pod-subpath-test-configmap-crxb": Phase="Running", Reason="", readiness=true. Elapsed: 8.034108499s
+Apr 29 13:09:23.063: INFO: Pod "pod-subpath-test-configmap-crxb": Phase="Running", Reason="", readiness=true. Elapsed: 10.040901676s
+Apr 29 13:09:25.070: INFO: Pod "pod-subpath-test-configmap-crxb": Phase="Running", Reason="", readiness=true. Elapsed: 12.047402965s
+Apr 29 13:09:27.076: INFO: Pod "pod-subpath-test-configmap-crxb": Phase="Running", Reason="", readiness=true. Elapsed: 14.053926036s
+Apr 29 13:09:29.083: INFO: Pod "pod-subpath-test-configmap-crxb": Phase="Running", Reason="", readiness=true. Elapsed: 16.060994573s
+Apr 29 13:09:31.094: INFO: Pod "pod-subpath-test-configmap-crxb": Phase="Running", Reason="", readiness=true. Elapsed: 18.072115655s
+Apr 29 13:09:33.103: INFO: Pod "pod-subpath-test-configmap-crxb": Phase="Running", Reason="", readiness=true. Elapsed: 20.080852725s
+Apr 29 13:09:35.113: INFO: Pod "pod-subpath-test-configmap-crxb": Phase="Running", Reason="", readiness=true. Elapsed: 22.09087081s
+Apr 29 13:09:37.120: INFO: Pod "pod-subpath-test-configmap-crxb": Phase="Succeeded", Reason="", readiness=false. Elapsed: 24.098158531s
+STEP: Saw pod success
+Apr 29 13:09:37.121: INFO: Pod "pod-subpath-test-configmap-crxb" satisfied condition "success or failure"
+Apr 29 13:09:37.126: INFO: Trying to get logs from node worker-4av2h-6665d7b5ff-9nc42 pod pod-subpath-test-configmap-crxb container test-container-subpath-configmap-crxb: 
+STEP: delete the pod
+Apr 29 13:09:37.177: INFO: Waiting for pod pod-subpath-test-configmap-crxb to disappear
+Apr 29 13:09:37.182: INFO: Pod pod-subpath-test-configmap-crxb no longer exists
+STEP: Deleting pod pod-subpath-test-configmap-crxb
+Apr 29 13:09:37.182: INFO: Deleting pod "pod-subpath-test-configmap-crxb" in namespace "subpath-6204"
+[AfterEach] [sig-storage] Subpath
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 13:09:37.187: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "subpath-6204" for this suite.
+Apr 29 13:09:43.223: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 13:09:43.382: INFO: namespace subpath-6204 deletion completed in 6.18593432s
+
+• [SLOW TEST:30.571 seconds]
+[sig-storage] Subpath
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:22
+  Atomic writer volumes
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:34
+    should support subpaths with configmap pod with mountPath of existing file [LinuxOnly] [Conformance]
+    /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSSSSSSSS
+------------------------------
+[sig-scheduling] SchedulerPredicates [Serial] 
+  validates resource limits of pods that are allowed to run  [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-scheduling] SchedulerPredicates [Serial]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 13:09:43.383: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename sched-pred
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in sched-pred-2584
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [sig-scheduling] SchedulerPredicates [Serial]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:79
+Apr 29 13:09:43.587: INFO: Waiting up to 1m0s for all (but 0) nodes to be ready
+Apr 29 13:09:43.598: INFO: Waiting for terminating namespaces to be deleted...
+Apr 29 13:09:43.602: INFO: 
+Logging pods the kubelet thinks is on node worker-3oo4n-86cbf586b7-9q98g before test
+Apr 29 13:09:43.630: INFO: calico-node-fmqgw from kube-system started at 2019-04-29 11:13:35 +0000 UTC (1 container statuses recorded)
+Apr 29 13:09:43.630: INFO: 	Container calico-node ready: true, restart count 0
+Apr 29 13:09:43.630: INFO: kube-proxy-dw2qf from kube-system started at 2019-04-29 11:13:55 +0000 UTC (1 container statuses recorded)
+Apr 29 13:09:43.631: INFO: 	Container kube-proxy ready: true, restart count 0
+Apr 29 13:09:43.631: INFO: node-exporter-j2kbr from kube-system started at 2019-04-29 11:21:42 +0000 UTC (1 container statuses recorded)
+Apr 29 13:09:43.631: INFO: 	Container node-exporter ready: true, restart count 0
+Apr 29 13:09:43.631: INFO: nginx-ingress-controller-6cffb97fb5-5fbv2 from kube-system started at 2019-04-29 11:23:55 +0000 UTC (1 container statuses recorded)
+Apr 29 13:09:43.632: INFO: 	Container nginx-ingress-controller ready: true, restart count 0
+Apr 29 13:09:43.632: INFO: net-exporter-wbqq5 from kube-system started at 2019-04-29 11:21:44 +0000 UTC (1 container statuses recorded)
+Apr 29 13:09:43.632: INFO: 	Container net-exporter ready: true, restart count 0
+Apr 29 13:09:43.632: INFO: sonobuoy-systemd-logs-daemon-set-8191105fdead4396-6mc85 from heptio-sonobuoy started at 2019-04-29 11:30:45 +0000 UTC (2 container statuses recorded)
+Apr 29 13:09:43.633: INFO: 	Container sonobuoy-worker ready: true, restart count 1
+Apr 29 13:09:43.634: INFO: 	Container systemd-logs ready: true, restart count 1
+Apr 29 13:09:43.634: INFO: kube-state-metrics-5fdb649879-lqkfk from kube-system started at 2019-04-29 11:17:01 +0000 UTC (2 container statuses recorded)
+Apr 29 13:09:43.634: INFO: 	Container addon-resizer ready: true, restart count 0
+Apr 29 13:09:43.634: INFO: 	Container kube-state-metrics ready: true, restart count 0
+Apr 29 13:09:43.635: INFO: coredns-58f7d854b4-xlcnp from kube-system started at 2019-04-29 11:16:23 +0000 UTC (1 container statuses recorded)
+Apr 29 13:09:43.635: INFO: 	Container coredns ready: true, restart count 0
+Apr 29 13:09:43.635: INFO: cert-exporter-kgwcd from kube-system started at 2019-04-29 11:16:36 +0000 UTC (1 container statuses recorded)
+Apr 29 13:09:43.635: INFO: 	Container cert-exporter ready: true, restart count 0
+Apr 29 13:09:43.636: INFO: 
+Logging pods the kubelet thinks is on node worker-4av2h-6665d7b5ff-9nc42 before test
+Apr 29 13:09:43.650: INFO: coredns-58f7d854b4-c4lkh from kube-system started at 2019-04-29 11:16:23 +0000 UTC (1 container statuses recorded)
+Apr 29 13:09:43.650: INFO: 	Container coredns ready: true, restart count 0
+Apr 29 13:09:43.650: INFO: sonobuoy-systemd-logs-daemon-set-8191105fdead4396-gcldz from heptio-sonobuoy started at 2019-04-29 11:30:45 +0000 UTC (2 container statuses recorded)
+Apr 29 13:09:43.650: INFO: 	Container sonobuoy-worker ready: true, restart count 1
+Apr 29 13:09:43.650: INFO: 	Container systemd-logs ready: true, restart count 1
+Apr 29 13:09:43.650: INFO: cert-exporter-pn7vb from kube-system started at 2019-04-29 11:16:35 +0000 UTC (1 container statuses recorded)
+Apr 29 13:09:43.650: INFO: 	Container cert-exporter ready: true, restart count 0
+Apr 29 13:09:43.650: INFO: node-exporter-pcjs5 from kube-system started at 2019-04-29 11:21:43 +0000 UTC (1 container statuses recorded)
+Apr 29 13:09:43.650: INFO: 	Container node-exporter ready: true, restart count 0
+Apr 29 13:09:43.650: INFO: nginx-ingress-controller-6cffb97fb5-mfcvr from kube-system started at 2019-04-29 11:23:04 +0000 UTC (1 container statuses recorded)
+Apr 29 13:09:43.650: INFO: 	Container nginx-ingress-controller ready: true, restart count 0
+Apr 29 13:09:43.650: INFO: tiller-deploy-54494c4fb6-k7mrg from giantswarm started at 2019-04-29 11:15:23 +0000 UTC (1 container statuses recorded)
+Apr 29 13:09:43.650: INFO: 	Container tiller ready: true, restart count 0
+Apr 29 13:09:43.650: INFO: calico-node-k64gm from kube-system started at 2019-04-29 11:13:35 +0000 UTC (1 container statuses recorded)
+Apr 29 13:09:43.650: INFO: 	Container calico-node ready: true, restart count 0
+Apr 29 13:09:43.650: INFO: kube-proxy-mf5pp from kube-system started at 2019-04-29 11:13:55 +0000 UTC (1 container statuses recorded)
+Apr 29 13:09:43.650: INFO: 	Container kube-proxy ready: true, restart count 0
+Apr 29 13:09:43.650: INFO: net-exporter-vhll2 from kube-system started at 2019-04-29 11:21:44 +0000 UTC (1 container statuses recorded)
+Apr 29 13:09:43.650: INFO: 	Container net-exporter ready: true, restart count 0
+Apr 29 13:09:43.650: INFO: default-http-backend-77b6545877-zd4s2 from kube-system started at 2019-04-29 11:21:34 +0000 UTC (1 container statuses recorded)
+Apr 29 13:09:43.650: INFO: 	Container default-http-backend ready: true, restart count 0
+Apr 29 13:09:43.650: INFO: 
+Logging pods the kubelet thinks is on node worker-p95fb-54c845dd45-crdqs before test
+Apr 29 13:09:43.666: INFO: metrics-server-b94b95fb4-ncg25 from kube-system started at 2019-04-29 11:16:39 +0000 UTC (1 container statuses recorded)
+Apr 29 13:09:43.667: INFO: 	Container metrics-server ready: true, restart count 0
+Apr 29 13:09:43.667: INFO: sonobuoy-systemd-logs-daemon-set-8191105fdead4396-w5l9x from heptio-sonobuoy started at 2019-04-29 11:30:45 +0000 UTC (2 container statuses recorded)
+Apr 29 13:09:43.667: INFO: 	Container sonobuoy-worker ready: true, restart count 1
+Apr 29 13:09:43.667: INFO: 	Container systemd-logs ready: true, restart count 1
+Apr 29 13:09:43.667: INFO: calico-node-hwth9 from kube-system started at 2019-04-29 11:13:35 +0000 UTC (1 container statuses recorded)
+Apr 29 13:09:43.668: INFO: 	Container calico-node ready: true, restart count 0
+Apr 29 13:09:43.668: INFO: cert-exporter-jkxfh from kube-system started at 2019-04-29 11:16:35 +0000 UTC (1 container statuses recorded)
+Apr 29 13:09:43.668: INFO: 	Container cert-exporter ready: true, restart count 0
+Apr 29 13:09:43.668: INFO: chart-operator-58979ccd8d-wvbzg from giantswarm started at 2019-04-29 11:16:10 +0000 UTC (1 container statuses recorded)
+Apr 29 13:09:43.669: INFO: 	Container chart-operator ready: true, restart count 0
+Apr 29 13:09:43.669: INFO: node-exporter-98xpc from kube-system started at 2019-04-29 11:21:42 +0000 UTC (1 container statuses recorded)
+Apr 29 13:09:43.669: INFO: 	Container node-exporter ready: true, restart count 0
+Apr 29 13:09:43.669: INFO: kube-proxy-tppgw from kube-system started at 2019-04-29 11:13:55 +0000 UTC (1 container statuses recorded)
+Apr 29 13:09:43.670: INFO: 	Container kube-proxy ready: true, restart count 0
+Apr 29 13:09:43.670: INFO: net-exporter-xcjtp from kube-system started at 2019-04-29 11:21:44 +0000 UTC (1 container statuses recorded)
+Apr 29 13:09:43.670: INFO: 	Container net-exporter ready: true, restart count 0
+Apr 29 13:09:43.670: INFO: default-http-backend-77b6545877-jwz64 from kube-system started at 2019-04-29 11:22:25 +0000 UTC (1 container statuses recorded)
+Apr 29 13:09:43.670: INFO: 	Container default-http-backend ready: true, restart count 0
+Apr 29 13:09:43.671: INFO: sonobuoy from heptio-sonobuoy started at 2019-04-29 11:30:35 +0000 UTC (1 container statuses recorded)
+Apr 29 13:09:43.671: INFO: 	Container kube-sonobuoy ready: true, restart count 0
+Apr 29 13:09:43.672: INFO: sonobuoy-e2e-job-9cf815dd5d0c4d66 from heptio-sonobuoy started at 2019-04-29 11:30:45 +0000 UTC (2 container statuses recorded)
+Apr 29 13:09:43.672: INFO: 	Container e2e ready: true, restart count 0
+Apr 29 13:09:43.672: INFO: 	Container sonobuoy-worker ready: true, restart count 0
+Apr 29 13:09:43.673: INFO: nginx-ingress-controller-6cffb97fb5-vqksl from kube-system started at 2019-04-29 11:21:47 +0000 UTC (1 container statuses recorded)
+Apr 29 13:09:43.673: INFO: 	Container nginx-ingress-controller ready: true, restart count 0
+[It] validates resource limits of pods that are allowed to run  [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: verifying the node has the label node worker-3oo4n-86cbf586b7-9q98g
+STEP: verifying the node has the label node worker-4av2h-6665d7b5ff-9nc42
+STEP: verifying the node has the label node worker-p95fb-54c845dd45-crdqs
+Apr 29 13:09:43.837: INFO: Pod chart-operator-58979ccd8d-wvbzg requesting resource cpu=250m on Node worker-p95fb-54c845dd45-crdqs
+Apr 29 13:09:43.837: INFO: Pod tiller-deploy-54494c4fb6-k7mrg requesting resource cpu=0m on Node worker-4av2h-6665d7b5ff-9nc42
+Apr 29 13:09:43.837: INFO: Pod sonobuoy requesting resource cpu=0m on Node worker-p95fb-54c845dd45-crdqs
+Apr 29 13:09:43.837: INFO: Pod sonobuoy-e2e-job-9cf815dd5d0c4d66 requesting resource cpu=0m on Node worker-p95fb-54c845dd45-crdqs
+Apr 29 13:09:43.837: INFO: Pod sonobuoy-systemd-logs-daemon-set-8191105fdead4396-6mc85 requesting resource cpu=0m on Node worker-3oo4n-86cbf586b7-9q98g
+Apr 29 13:09:43.837: INFO: Pod sonobuoy-systemd-logs-daemon-set-8191105fdead4396-gcldz requesting resource cpu=0m on Node worker-4av2h-6665d7b5ff-9nc42
+Apr 29 13:09:43.837: INFO: Pod sonobuoy-systemd-logs-daemon-set-8191105fdead4396-w5l9x requesting resource cpu=0m on Node worker-p95fb-54c845dd45-crdqs
+Apr 29 13:09:43.838: INFO: Pod calico-node-fmqgw requesting resource cpu=250m on Node worker-3oo4n-86cbf586b7-9q98g
+Apr 29 13:09:43.838: INFO: Pod calico-node-hwth9 requesting resource cpu=250m on Node worker-p95fb-54c845dd45-crdqs
+Apr 29 13:09:43.838: INFO: Pod calico-node-k64gm requesting resource cpu=250m on Node worker-4av2h-6665d7b5ff-9nc42
+Apr 29 13:09:43.838: INFO: Pod cert-exporter-jkxfh requesting resource cpu=50m on Node worker-p95fb-54c845dd45-crdqs
+Apr 29 13:09:43.838: INFO: Pod cert-exporter-kgwcd requesting resource cpu=50m on Node worker-3oo4n-86cbf586b7-9q98g
+Apr 29 13:09:43.838: INFO: Pod cert-exporter-pn7vb requesting resource cpu=50m on Node worker-4av2h-6665d7b5ff-9nc42
+Apr 29 13:09:43.838: INFO: Pod coredns-58f7d854b4-c4lkh requesting resource cpu=250m on Node worker-4av2h-6665d7b5ff-9nc42
+Apr 29 13:09:43.838: INFO: Pod coredns-58f7d854b4-xlcnp requesting resource cpu=250m on Node worker-3oo4n-86cbf586b7-9q98g
+Apr 29 13:09:43.838: INFO: Pod default-http-backend-77b6545877-jwz64 requesting resource cpu=10m on Node worker-p95fb-54c845dd45-crdqs
+Apr 29 13:09:43.838: INFO: Pod default-http-backend-77b6545877-zd4s2 requesting resource cpu=10m on Node worker-4av2h-6665d7b5ff-9nc42
+Apr 29 13:09:43.838: INFO: Pod kube-proxy-dw2qf requesting resource cpu=75m on Node worker-3oo4n-86cbf586b7-9q98g
+Apr 29 13:09:43.838: INFO: Pod kube-proxy-mf5pp requesting resource cpu=75m on Node worker-4av2h-6665d7b5ff-9nc42
+Apr 29 13:09:43.838: INFO: Pod kube-proxy-tppgw requesting resource cpu=75m on Node worker-p95fb-54c845dd45-crdqs
+Apr 29 13:09:43.838: INFO: Pod kube-state-metrics-5fdb649879-lqkfk requesting resource cpu=354m on Node worker-3oo4n-86cbf586b7-9q98g
+Apr 29 13:09:43.838: INFO: Pod metrics-server-b94b95fb4-ncg25 requesting resource cpu=0m on Node worker-p95fb-54c845dd45-crdqs
+Apr 29 13:09:43.838: INFO: Pod net-exporter-vhll2 requesting resource cpu=50m on Node worker-4av2h-6665d7b5ff-9nc42
+Apr 29 13:09:43.838: INFO: Pod net-exporter-wbqq5 requesting resource cpu=50m on Node worker-3oo4n-86cbf586b7-9q98g
+Apr 29 13:09:43.838: INFO: Pod net-exporter-xcjtp requesting resource cpu=50m on Node worker-p95fb-54c845dd45-crdqs
+Apr 29 13:09:43.838: INFO: Pod nginx-ingress-controller-6cffb97fb5-5fbv2 requesting resource cpu=500m on Node worker-3oo4n-86cbf586b7-9q98g
+Apr 29 13:09:43.838: INFO: Pod nginx-ingress-controller-6cffb97fb5-mfcvr requesting resource cpu=500m on Node worker-4av2h-6665d7b5ff-9nc42
+Apr 29 13:09:43.838: INFO: Pod nginx-ingress-controller-6cffb97fb5-vqksl requesting resource cpu=500m on Node worker-p95fb-54c845dd45-crdqs
+Apr 29 13:09:43.839: INFO: Pod node-exporter-98xpc requesting resource cpu=200m on Node worker-p95fb-54c845dd45-crdqs
+Apr 29 13:09:43.839: INFO: Pod node-exporter-j2kbr requesting resource cpu=200m on Node worker-3oo4n-86cbf586b7-9q98g
+Apr 29 13:09:43.839: INFO: Pod node-exporter-pcjs5 requesting resource cpu=200m on Node worker-4av2h-6665d7b5ff-9nc42
+STEP: Starting Pods to consume most of the cluster CPU.
+STEP: Creating another pod that requires unavailable amount of CPU.
+STEP: Considering event: 
+Type = [Normal], Name = [filler-pod-0e379d67-6a80-11e9-b6ef-0e71f8c7e888.1599f3f5dc4f91ae], Reason = [Scheduled], Message = [Successfully assigned sched-pred-2584/filler-pod-0e379d67-6a80-11e9-b6ef-0e71f8c7e888 to worker-4av2h-6665d7b5ff-9nc42]
+STEP: Considering event: 
+Type = [Normal], Name = [filler-pod-0e379d67-6a80-11e9-b6ef-0e71f8c7e888.1599f3f7457581c8], Reason = [Pulled], Message = [Container image "k8s.gcr.io/pause:3.1" already present on machine]
+STEP: Considering event: 
+Type = [Normal], Name = [filler-pod-0e379d67-6a80-11e9-b6ef-0e71f8c7e888.1599f3f74fae86cc], Reason = [Created], Message = [Created container filler-pod-0e379d67-6a80-11e9-b6ef-0e71f8c7e888]
+STEP: Considering event: 
+Type = [Normal], Name = [filler-pod-0e379d67-6a80-11e9-b6ef-0e71f8c7e888.1599f3f75cda9d53], Reason = [Started], Message = [Started container filler-pod-0e379d67-6a80-11e9-b6ef-0e71f8c7e888]
+STEP: Considering event: 
+Type = [Normal], Name = [filler-pod-0e3aa7b4-6a80-11e9-b6ef-0e71f8c7e888.1599f3f5dd9aa681], Reason = [Scheduled], Message = [Successfully assigned sched-pred-2584/filler-pod-0e3aa7b4-6a80-11e9-b6ef-0e71f8c7e888 to worker-p95fb-54c845dd45-crdqs]
+STEP: Considering event: 
+Type = [Normal], Name = [filler-pod-0e3aa7b4-6a80-11e9-b6ef-0e71f8c7e888.1599f3f74287d8ac], Reason = [Pulled], Message = [Container image "k8s.gcr.io/pause:3.1" already present on machine]
+STEP: Considering event: 
+Type = [Normal], Name = [filler-pod-0e3aa7b4-6a80-11e9-b6ef-0e71f8c7e888.1599f3f750574c95], Reason = [Created], Message = [Created container filler-pod-0e3aa7b4-6a80-11e9-b6ef-0e71f8c7e888]
+STEP: Considering event: 
+Type = [Normal], Name = [filler-pod-0e3aa7b4-6a80-11e9-b6ef-0e71f8c7e888.1599f3f75d815654], Reason = [Started], Message = [Started container filler-pod-0e3aa7b4-6a80-11e9-b6ef-0e71f8c7e888]
+STEP: Considering event: 
+Type = [Normal], Name = [filler-pod-0e3bff7d-6a80-11e9-b6ef-0e71f8c7e888.1599f3f5df102d2b], Reason = [Scheduled], Message = [Successfully assigned sched-pred-2584/filler-pod-0e3bff7d-6a80-11e9-b6ef-0e71f8c7e888 to worker-3oo4n-86cbf586b7-9q98g]
+STEP: Considering event: 
+Type = [Normal], Name = [filler-pod-0e3bff7d-6a80-11e9-b6ef-0e71f8c7e888.1599f3f727740321], Reason = [Pulled], Message = [Container image "k8s.gcr.io/pause:3.1" already present on machine]
+STEP: Considering event: 
+Type = [Normal], Name = [filler-pod-0e3bff7d-6a80-11e9-b6ef-0e71f8c7e888.1599f3f73385f14f], Reason = [Created], Message = [Created container filler-pod-0e3bff7d-6a80-11e9-b6ef-0e71f8c7e888]
+STEP: Considering event: 
+Type = [Normal], Name = [filler-pod-0e3bff7d-6a80-11e9-b6ef-0e71f8c7e888.1599f3f7486d3e4b], Reason = [Started], Message = [Started container filler-pod-0e3bff7d-6a80-11e9-b6ef-0e71f8c7e888]
+STEP: Considering event: 
+Type = [Warning], Name = [additional-pod.1599f3f7bdc76a28], Reason = [FailedScheduling], Message = [0/4 nodes are available: 4 Insufficient cpu.]
+STEP: removing the label node off the node worker-p95fb-54c845dd45-crdqs
+STEP: verifying the node doesn't have the label node
+STEP: removing the label node off the node worker-3oo4n-86cbf586b7-9q98g
+STEP: verifying the node doesn't have the label node
+STEP: removing the label node off the node worker-4av2h-6665d7b5ff-9nc42
+STEP: verifying the node doesn't have the label node
+[AfterEach] [sig-scheduling] SchedulerPredicates [Serial]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 13:09:53.064: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "sched-pred-2584" for this suite.
+Apr 29 13:10:01.091: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 13:10:01.327: INFO: namespace sched-pred-2584 deletion completed in 8.254676326s
+[AfterEach] [sig-scheduling] SchedulerPredicates [Serial]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:70
+
+• [SLOW TEST:17.945 seconds]
+[sig-scheduling] SchedulerPredicates [Serial]
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/framework.go:22
+  validates resource limits of pods that are allowed to run  [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSSSSSSSSSSSSSSSS
+------------------------------
+[sig-network] DNS 
+  should provide /etc/hosts entries for the cluster [LinuxOnly] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-network] DNS
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 13:10:01.333: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename dns
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in dns-1220
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should provide /etc/hosts entries for the cluster [LinuxOnly] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Running these commands on wheezy: for i in `seq 1 600`; do test -n "$$(getent hosts dns-querier-1.dns-test-service.dns-1220.svc.cluster.local)" && echo OK > /results/wheezy_hosts@dns-querier-1.dns-test-service.dns-1220.svc.cluster.local;test -n "$$(getent hosts dns-querier-1)" && echo OK > /results/wheezy_hosts@dns-querier-1;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-1220.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@PodARecord;sleep 1; done
+
+STEP: Running these commands on jessie: for i in `seq 1 600`; do test -n "$$(getent hosts dns-querier-1.dns-test-service.dns-1220.svc.cluster.local)" && echo OK > /results/jessie_hosts@dns-querier-1.dns-test-service.dns-1220.svc.cluster.local;test -n "$$(getent hosts dns-querier-1)" && echo OK > /results/jessie_hosts@dns-querier-1;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-1220.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_tcp@PodARecord;sleep 1; done
+
+STEP: creating a pod to probe /etc/hosts
+STEP: submitting the pod to kubernetes
+STEP: retrieving the pod
+STEP: looking for the results for each expected name from probers
+Apr 29 13:10:05.658: INFO: DNS probes using dns-1220/dns-test-18c6b493-6a80-11e9-b6ef-0e71f8c7e888 succeeded
+
+STEP: deleting the pod
+[AfterEach] [sig-network] DNS
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 13:10:05.691: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "dns-1220" for this suite.
+Apr 29 13:10:11.726: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 13:10:11.939: INFO: namespace dns-1220 deletion completed in 6.237932901s
+
+• [SLOW TEST:10.607 seconds]
+[sig-network] DNS
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:22
+  should provide /etc/hosts entries for the cluster [LinuxOnly] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+------------------------------
+[sig-storage] Projected configMap 
+  should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-storage] Projected configMap
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 13:10:11.941: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename projected
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-8774
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating configMap with name projected-configmap-test-volume-map-1f164e9f-6a80-11e9-b6ef-0e71f8c7e888
+STEP: Creating a pod to test consume configMaps
+Apr 29 13:10:12.162: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-1f174e1b-6a80-11e9-b6ef-0e71f8c7e888" in namespace "projected-8774" to be "success or failure"
+Apr 29 13:10:12.172: INFO: Pod "pod-projected-configmaps-1f174e1b-6a80-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 9.846841ms
+Apr 29 13:10:14.179: INFO: Pod "pod-projected-configmaps-1f174e1b-6a80-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 2.016699687s
+Apr 29 13:10:16.191: INFO: Pod "pod-projected-configmaps-1f174e1b-6a80-11e9-b6ef-0e71f8c7e888": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.028900662s
+STEP: Saw pod success
+Apr 29 13:10:16.191: INFO: Pod "pod-projected-configmaps-1f174e1b-6a80-11e9-b6ef-0e71f8c7e888" satisfied condition "success or failure"
+Apr 29 13:10:16.204: INFO: Trying to get logs from node worker-4av2h-6665d7b5ff-9nc42 pod pod-projected-configmaps-1f174e1b-6a80-11e9-b6ef-0e71f8c7e888 container projected-configmap-volume-test: 
+STEP: delete the pod
+Apr 29 13:10:16.241: INFO: Waiting for pod pod-projected-configmaps-1f174e1b-6a80-11e9-b6ef-0e71f8c7e888 to disappear
+Apr 29 13:10:16.248: INFO: Pod pod-projected-configmaps-1f174e1b-6a80-11e9-b6ef-0e71f8c7e888 no longer exists
+[AfterEach] [sig-storage] Projected configMap
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 13:10:16.248: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "projected-8774" for this suite.
+Apr 29 13:10:22.505: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 13:10:22.692: INFO: namespace projected-8774 deletion completed in 6.436897019s
+
+• [SLOW TEST:10.751 seconds]
+[sig-storage] Projected configMap
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_configmap.go:33
+  should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSSSSSSSSSSSSS
+------------------------------
+[sig-storage] Secrets 
+  should be able to mount in a volume regardless of a different secret existing with same name in different namespace [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-storage] Secrets
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 13:10:22.695: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename secrets
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in secrets-9999
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should be able to mount in a volume regardless of a different secret existing with same name in different namespace [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in secret-namespace-6720
+STEP: Creating secret with name secret-test-257ea471-6a80-11e9-b6ef-0e71f8c7e888
+STEP: Creating a pod to test consume secrets
+Apr 29 13:10:23.085: INFO: Waiting up to 5m0s for pod "pod-secrets-2599ccf8-6a80-11e9-b6ef-0e71f8c7e888" in namespace "secrets-9999" to be "success or failure"
+Apr 29 13:10:23.090: INFO: Pod "pod-secrets-2599ccf8-6a80-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 4.880266ms
+Apr 29 13:10:25.103: INFO: Pod "pod-secrets-2599ccf8-6a80-11e9-b6ef-0e71f8c7e888": Phase="Pending", Reason="", readiness=false. Elapsed: 2.017258785s
+Apr 29 13:10:27.117: INFO: Pod "pod-secrets-2599ccf8-6a80-11e9-b6ef-0e71f8c7e888": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.032168118s
+STEP: Saw pod success
+Apr 29 13:10:27.118: INFO: Pod "pod-secrets-2599ccf8-6a80-11e9-b6ef-0e71f8c7e888" satisfied condition "success or failure"
+Apr 29 13:10:27.124: INFO: Trying to get logs from node worker-4av2h-6665d7b5ff-9nc42 pod pod-secrets-2599ccf8-6a80-11e9-b6ef-0e71f8c7e888 container secret-volume-test: 
+STEP: delete the pod
+Apr 29 13:10:27.156: INFO: Waiting for pod pod-secrets-2599ccf8-6a80-11e9-b6ef-0e71f8c7e888 to disappear
+Apr 29 13:10:27.160: INFO: Pod pod-secrets-2599ccf8-6a80-11e9-b6ef-0e71f8c7e888 no longer exists
+[AfterEach] [sig-storage] Secrets
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 13:10:27.162: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "secrets-9999" for this suite.
+Apr 29 13:10:33.203: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 13:10:33.444: INFO: namespace secrets-9999 deletion completed in 6.271996034s
+STEP: Destroying namespace "secret-namespace-6720" for this suite.
+Apr 29 13:10:39.467: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 13:10:39.682: INFO: namespace secret-namespace-6720 deletion completed in 6.237323485s
+
+• [SLOW TEST:16.987 seconds]
+[sig-storage] Secrets
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/secrets_volume.go:33
+  should be able to mount in a volume regardless of a different secret existing with same name in different namespace [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+------------------------------
+[sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] 
+  Burst scaling should run to completion even with unhealthy pods [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-apps] StatefulSet
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 13:10:39.692: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename statefulset
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in statefulset-7031
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [sig-apps] StatefulSet
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:59
+[BeforeEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:74
+STEP: Creating service test in namespace statefulset-7031
+[It] Burst scaling should run to completion even with unhealthy pods [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating stateful set ss in namespace statefulset-7031
+STEP: Waiting until all stateful set ss replicas will be running in namespace statefulset-7031
+Apr 29 13:10:39.928: INFO: Found 0 stateful pods, waiting for 1
+Apr 29 13:10:49.947: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true
+STEP: Confirming that stateful set scale up will not halt with unhealthy stateful pod
+Apr 29 13:10:49.955: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 exec --namespace=statefulset-7031 ss-0 -- /bin/sh -x -c mv -v /usr/share/nginx/html/index.html /tmp/ || true'
+Apr 29 13:10:50.324: INFO: stderr: "+ mv -v /usr/share/nginx/html/index.html /tmp/\n"
+Apr 29 13:10:50.324: INFO: stdout: "'/usr/share/nginx/html/index.html' -> '/tmp/index.html'\n"
+Apr 29 13:10:50.324: INFO: stdout of mv -v /usr/share/nginx/html/index.html /tmp/ || true on ss-0: '/usr/share/nginx/html/index.html' -> '/tmp/index.html'
+
+Apr 29 13:10:50.329: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=true
+Apr 29 13:11:00.336: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=false
+Apr 29 13:11:00.336: INFO: Waiting for statefulset status.replicas updated to 0
+Apr 29 13:11:00.381: INFO: POD   NODE                           PHASE    GRACE  CONDITIONS
+Apr 29 13:11:00.381: INFO: ss-0  worker-4av2h-6665d7b5ff-9nc42  Running         [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 13:10:39 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-04-29 13:10:50 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-04-29 13:10:50 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 13:10:39 +0000 UTC  }]
+Apr 29 13:11:00.381: INFO: ss-1                                 Pending         []
+Apr 29 13:11:00.381: INFO: 
+Apr 29 13:11:00.381: INFO: StatefulSet ss has not reached scale 3, at 2
+Apr 29 13:11:01.392: INFO: Verifying statefulset ss doesn't scale past 3 for another 8.977986293s
+Apr 29 13:11:02.400: INFO: Verifying statefulset ss doesn't scale past 3 for another 7.967705769s
+Apr 29 13:11:03.417: INFO: Verifying statefulset ss doesn't scale past 3 for another 6.958874994s
+Apr 29 13:11:04.426: INFO: Verifying statefulset ss doesn't scale past 3 for another 5.943041412s
+Apr 29 13:11:05.433: INFO: Verifying statefulset ss doesn't scale past 3 for another 4.933618618s
+Apr 29 13:11:06.441: INFO: Verifying statefulset ss doesn't scale past 3 for another 3.926801855s
+Apr 29 13:11:07.448: INFO: Verifying statefulset ss doesn't scale past 3 for another 2.918364267s
+Apr 29 13:11:08.458: INFO: Verifying statefulset ss doesn't scale past 3 for another 1.911474057s
+Apr 29 13:11:09.467: INFO: Verifying statefulset ss doesn't scale past 3 for another 901.843921ms
+STEP: Scaling up stateful set ss to 3 replicas and waiting until all of them will be running in namespace statefulset-7031
+Apr 29 13:11:10.476: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 exec --namespace=statefulset-7031 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
+Apr 29 13:11:10.806: INFO: stderr: "+ mv -v /tmp/index.html /usr/share/nginx/html/\n"
+Apr 29 13:11:10.806: INFO: stdout: "'/tmp/index.html' -> '/usr/share/nginx/html/index.html'\n"
+Apr 29 13:11:10.806: INFO: stdout of mv -v /tmp/index.html /usr/share/nginx/html/ || true on ss-0: '/tmp/index.html' -> '/usr/share/nginx/html/index.html'
+
+Apr 29 13:11:10.806: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 exec --namespace=statefulset-7031 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
+Apr 29 13:11:11.090: INFO: stderr: "+ mv -v /tmp/index.html /usr/share/nginx/html/\nmv: can't rename '/tmp/index.html': No such file or directory\n+ true\n"
+Apr 29 13:11:11.090: INFO: stdout: "'/tmp/index.html' -> '/usr/share/nginx/html/index.html'\n"
+Apr 29 13:11:11.090: INFO: stdout of mv -v /tmp/index.html /usr/share/nginx/html/ || true on ss-1: '/tmp/index.html' -> '/usr/share/nginx/html/index.html'
+
+Apr 29 13:11:11.090: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 exec --namespace=statefulset-7031 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
+Apr 29 13:11:11.432: INFO: stderr: "+ mv -v /tmp/index.html /usr/share/nginx/html/\nmv: can't rename '/tmp/index.html': No such file or directory\n+ true\n"
+Apr 29 13:11:11.432: INFO: stdout: "'/tmp/index.html' -> '/usr/share/nginx/html/index.html'\n"
+Apr 29 13:11:11.432: INFO: stdout of mv -v /tmp/index.html /usr/share/nginx/html/ || true on ss-2: '/tmp/index.html' -> '/usr/share/nginx/html/index.html'
+
+Apr 29 13:11:11.438: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true
+Apr 29 13:11:11.438: INFO: Waiting for pod ss-1 to enter Running - Ready=true, currently Running - Ready=true
+Apr 29 13:11:11.438: INFO: Waiting for pod ss-2 to enter Running - Ready=true, currently Running - Ready=true
+STEP: Scale down will not halt with unhealthy stateful pod
+Apr 29 13:11:11.458: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 exec --namespace=statefulset-7031 ss-0 -- /bin/sh -x -c mv -v /usr/share/nginx/html/index.html /tmp/ || true'
+Apr 29 13:11:11.742: INFO: stderr: "+ mv -v /usr/share/nginx/html/index.html /tmp/\n"
+Apr 29 13:11:11.742: INFO: stdout: "'/usr/share/nginx/html/index.html' -> '/tmp/index.html'\n"
+Apr 29 13:11:11.742: INFO: stdout of mv -v /usr/share/nginx/html/index.html /tmp/ || true on ss-0: '/usr/share/nginx/html/index.html' -> '/tmp/index.html'
+
+Apr 29 13:11:11.742: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 exec --namespace=statefulset-7031 ss-1 -- /bin/sh -x -c mv -v /usr/share/nginx/html/index.html /tmp/ || true'
+Apr 29 13:11:12.065: INFO: stderr: "+ mv -v /usr/share/nginx/html/index.html /tmp/\n"
+Apr 29 13:11:12.065: INFO: stdout: "'/usr/share/nginx/html/index.html' -> '/tmp/index.html'\n"
+Apr 29 13:11:12.065: INFO: stdout of mv -v /usr/share/nginx/html/index.html /tmp/ || true on ss-1: '/usr/share/nginx/html/index.html' -> '/tmp/index.html'
+
+Apr 29 13:11:12.065: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 exec --namespace=statefulset-7031 ss-2 -- /bin/sh -x -c mv -v /usr/share/nginx/html/index.html /tmp/ || true'
+Apr 29 13:11:12.368: INFO: stderr: "+ mv -v /usr/share/nginx/html/index.html /tmp/\n"
+Apr 29 13:11:12.368: INFO: stdout: "'/usr/share/nginx/html/index.html' -> '/tmp/index.html'\n"
+Apr 29 13:11:12.368: INFO: stdout of mv -v /usr/share/nginx/html/index.html /tmp/ || true on ss-2: '/usr/share/nginx/html/index.html' -> '/tmp/index.html'
+
+Apr 29 13:11:12.368: INFO: Waiting for statefulset status.replicas updated to 0
+Apr 29 13:11:12.375: INFO: Waiting for stateful set status.readyReplicas to become 0, currently 1
+Apr 29 13:11:22.393: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=false
+Apr 29 13:11:22.393: INFO: Waiting for pod ss-1 to enter Running - Ready=false, currently Running - Ready=false
+Apr 29 13:11:22.393: INFO: Waiting for pod ss-2 to enter Running - Ready=false, currently Running - Ready=false
+Apr 29 13:11:22.422: INFO: POD   NODE                           PHASE    GRACE  CONDITIONS
+Apr 29 13:11:22.422: INFO: ss-0  worker-4av2h-6665d7b5ff-9nc42  Running         [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 13:10:39 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-04-29 13:11:11 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-04-29 13:11:11 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 13:10:39 +0000 UTC  }]
+Apr 29 13:11:22.422: INFO: ss-1  worker-3oo4n-86cbf586b7-9q98g  Running         [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 13:11:00 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-04-29 13:11:12 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-04-29 13:11:12 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 13:11:00 +0000 UTC  }]
+Apr 29 13:11:22.422: INFO: ss-2  worker-p95fb-54c845dd45-crdqs  Running         [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 13:11:00 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-04-29 13:11:12 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-04-29 13:11:12 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 13:11:00 +0000 UTC  }]
+Apr 29 13:11:22.422: INFO: 
+Apr 29 13:11:22.422: INFO: StatefulSet ss has not reached scale 0, at 3
+Apr 29 13:11:23.428: INFO: POD   NODE                           PHASE    GRACE  CONDITIONS
+Apr 29 13:11:23.428: INFO: ss-0  worker-4av2h-6665d7b5ff-9nc42  Running  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 13:10:39 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-04-29 13:11:11 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-04-29 13:11:11 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 13:10:39 +0000 UTC  }]
+Apr 29 13:11:23.429: INFO: ss-1  worker-3oo4n-86cbf586b7-9q98g  Running  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 13:11:00 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-04-29 13:11:12 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-04-29 13:11:12 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 13:11:00 +0000 UTC  }]
+Apr 29 13:11:23.429: INFO: ss-2  worker-p95fb-54c845dd45-crdqs  Running  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 13:11:00 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-04-29 13:11:12 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-04-29 13:11:12 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 13:11:00 +0000 UTC  }]
+Apr 29 13:11:23.429: INFO: 
+Apr 29 13:11:23.429: INFO: StatefulSet ss has not reached scale 0, at 3
+Apr 29 13:11:24.437: INFO: POD   NODE                           PHASE    GRACE  CONDITIONS
+Apr 29 13:11:24.437: INFO: ss-0  worker-4av2h-6665d7b5ff-9nc42  Running  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 13:10:39 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-04-29 13:11:11 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-04-29 13:11:11 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 13:10:39 +0000 UTC  }]
+Apr 29 13:11:24.437: INFO: ss-1  worker-3oo4n-86cbf586b7-9q98g  Running  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 13:11:00 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-04-29 13:11:12 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-04-29 13:11:12 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 13:11:00 +0000 UTC  }]
+Apr 29 13:11:24.437: INFO: ss-2  worker-p95fb-54c845dd45-crdqs  Running  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 13:11:00 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-04-29 13:11:12 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-04-29 13:11:12 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 13:11:00 +0000 UTC  }]
+Apr 29 13:11:24.437: INFO: 
+Apr 29 13:11:24.437: INFO: StatefulSet ss has not reached scale 0, at 3
+Apr 29 13:11:25.444: INFO: POD   NODE                           PHASE    GRACE  CONDITIONS
+Apr 29 13:11:25.445: INFO: ss-0  worker-4av2h-6665d7b5ff-9nc42  Running  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 13:10:39 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-04-29 13:11:11 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-04-29 13:11:11 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 13:10:39 +0000 UTC  }]
+Apr 29 13:11:25.445: INFO: ss-1  worker-3oo4n-86cbf586b7-9q98g  Pending  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 13:11:00 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-04-29 13:11:12 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-04-29 13:11:12 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 13:11:00 +0000 UTC  }]
+Apr 29 13:11:25.445: INFO: ss-2  worker-p95fb-54c845dd45-crdqs  Running  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 13:11:00 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-04-29 13:11:12 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-04-29 13:11:12 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 13:11:00 +0000 UTC  }]
+Apr 29 13:11:25.445: INFO: 
+Apr 29 13:11:25.445: INFO: StatefulSet ss has not reached scale 0, at 3
+Apr 29 13:11:26.451: INFO: POD   NODE                           PHASE    GRACE  CONDITIONS
+Apr 29 13:11:26.451: INFO: ss-0  worker-4av2h-6665d7b5ff-9nc42  Pending  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 13:10:39 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-04-29 13:11:11 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-04-29 13:11:11 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 13:10:39 +0000 UTC  }]
+Apr 29 13:11:26.452: INFO: ss-2  worker-p95fb-54c845dd45-crdqs  Pending  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 13:11:00 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-04-29 13:11:12 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-04-29 13:11:12 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 13:11:00 +0000 UTC  }]
+Apr 29 13:11:26.452: INFO: 
+Apr 29 13:11:26.452: INFO: StatefulSet ss has not reached scale 0, at 2
+Apr 29 13:11:27.458: INFO: POD   NODE                           PHASE    GRACE  CONDITIONS
+Apr 29 13:11:27.458: INFO: ss-0  worker-4av2h-6665d7b5ff-9nc42  Pending  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 13:10:39 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-04-29 13:11:11 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-04-29 13:11:11 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 13:10:39 +0000 UTC  }]
+Apr 29 13:11:27.458: INFO: ss-2  worker-p95fb-54c845dd45-crdqs  Pending  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 13:11:00 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-04-29 13:11:12 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-04-29 13:11:12 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 13:11:00 +0000 UTC  }]
+Apr 29 13:11:27.458: INFO: 
+Apr 29 13:11:27.458: INFO: StatefulSet ss has not reached scale 0, at 2
+Apr 29 13:11:28.465: INFO: POD   NODE                           PHASE    GRACE  CONDITIONS
+Apr 29 13:11:28.466: INFO: ss-0  worker-4av2h-6665d7b5ff-9nc42  Pending  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 13:10:39 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-04-29 13:11:11 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-04-29 13:11:11 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 13:10:39 +0000 UTC  }]
+Apr 29 13:11:28.466: INFO: ss-2  worker-p95fb-54c845dd45-crdqs  Pending  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 13:11:00 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-04-29 13:11:12 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-04-29 13:11:12 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 13:11:00 +0000 UTC  }]
+Apr 29 13:11:28.466: INFO: 
+Apr 29 13:11:28.466: INFO: StatefulSet ss has not reached scale 0, at 2
+Apr 29 13:11:29.472: INFO: POD   NODE                           PHASE    GRACE  CONDITIONS
+Apr 29 13:11:29.472: INFO: ss-0  worker-4av2h-6665d7b5ff-9nc42  Pending  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 13:10:39 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-04-29 13:11:11 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-04-29 13:11:11 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 13:10:39 +0000 UTC  }]
+Apr 29 13:11:29.472: INFO: ss-2  worker-p95fb-54c845dd45-crdqs  Pending  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 13:11:00 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-04-29 13:11:12 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-04-29 13:11:12 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 13:11:00 +0000 UTC  }]
+Apr 29 13:11:29.472: INFO: 
+Apr 29 13:11:29.472: INFO: StatefulSet ss has not reached scale 0, at 2
+Apr 29 13:11:30.479: INFO: POD   NODE                           PHASE    GRACE  CONDITIONS
+Apr 29 13:11:30.480: INFO: ss-0  worker-4av2h-6665d7b5ff-9nc42  Pending  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 13:10:39 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-04-29 13:11:11 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-04-29 13:11:11 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 13:10:39 +0000 UTC  }]
+Apr 29 13:11:30.481: INFO: ss-2  worker-p95fb-54c845dd45-crdqs  Pending  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 13:11:00 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-04-29 13:11:12 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-04-29 13:11:12 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 13:11:00 +0000 UTC  }]
+Apr 29 13:11:30.482: INFO: 
+Apr 29 13:11:30.482: INFO: StatefulSet ss has not reached scale 0, at 2
+Apr 29 13:11:31.495: INFO: POD   NODE                           PHASE    GRACE  CONDITIONS
+Apr 29 13:11:31.495: INFO: ss-0  worker-4av2h-6665d7b5ff-9nc42  Pending  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 13:10:39 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-04-29 13:11:11 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-04-29 13:11:11 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 13:10:39 +0000 UTC  }]
+Apr 29 13:11:31.495: INFO: ss-2  worker-p95fb-54c845dd45-crdqs  Pending  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 13:11:00 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2019-04-29 13:11:12 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2019-04-29 13:11:12 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2019-04-29 13:11:00 +0000 UTC  }]
+Apr 29 13:11:31.496: INFO: 
+Apr 29 13:11:31.496: INFO: StatefulSet ss has not reached scale 0, at 2
+STEP: Scaling down stateful set ss to 0 replicas and waiting until none of pods will run in namespacestatefulset-7031
+Apr 29 13:11:32.502: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 exec --namespace=statefulset-7031 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
+Apr 29 13:11:32.671: INFO: rc: 1
+Apr 29 13:11:32.672: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/tmp/kubeconfig-185508174 exec --namespace=statefulset-7031 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] []    error: unable to upgrade connection: container not found ("nginx")
+ []  0xc001675dd0 exit status 1   true [0xc0027df0c8 0xc0027df0e0 0xc0027df0f8] [0xc0027df0c8 0xc0027df0e0 0xc0027df0f8] [0xc0027df0d8 0xc0027df0f0] [0x9bf9f0 0x9bf9f0] 0xc001d86780 }:
+Command stdout:
+
+stderr:
+error: unable to upgrade connection: container not found ("nginx")
+
+error:
+exit status 1
+
+Apr 29 13:11:42.672: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 exec --namespace=statefulset-7031 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
+Apr 29 13:11:42.771: INFO: rc: 1
+Apr 29 13:11:42.771: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/tmp/kubeconfig-185508174 exec --namespace=statefulset-7031 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] []    Error from server (NotFound): pods "ss-0" not found
+ []  0xc001264660 exit status 1   true [0xc001501660 0xc001501678 0xc001501690] [0xc001501660 0xc001501678 0xc001501690] [0xc001501670 0xc001501688] [0x9bf9f0 0x9bf9f0] 0xc001a945a0 }:
+Command stdout:
+
+stderr:
+Error from server (NotFound): pods "ss-0" not found
+
+error:
+exit status 1
+
+Apr 29 13:11:52.772: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 exec --namespace=statefulset-7031 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
+Apr 29 13:11:52.864: INFO: rc: 1
+Apr 29 13:11:52.864: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/tmp/kubeconfig-185508174 exec --namespace=statefulset-7031 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] []    Error from server (NotFound): pods "ss-0" not found
+ []  0xc001cd01b0 exit status 1   true [0xc0027df100 0xc0027df118 0xc0027df130] [0xc0027df100 0xc0027df118 0xc0027df130] [0xc0027df110 0xc0027df128] [0x9bf9f0 0x9bf9f0] 0xc001bbe600 }:
+Command stdout:
+
+stderr:
+Error from server (NotFound): pods "ss-0" not found
+
+error:
+exit status 1
+
+Apr 29 13:12:02.864: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 exec --namespace=statefulset-7031 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
+Apr 29 13:12:02.958: INFO: rc: 1
+Apr 29 13:12:02.958: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/tmp/kubeconfig-185508174 exec --namespace=statefulset-7031 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] []    Error from server (NotFound): pods "ss-0" not found
+ []  0xc0012649f0 exit status 1   true [0xc001501698 0xc0015016b0 0xc0015016c8] [0xc001501698 0xc0015016b0 0xc0015016c8] [0xc0015016a8 0xc0015016c0] [0x9bf9f0 0x9bf9f0] 0xc001e8bc20 }:
+Command stdout:
+
+stderr:
+Error from server (NotFound): pods "ss-0" not found
+
+error:
+exit status 1
+
+Apr 29 13:12:12.959: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 exec --namespace=statefulset-7031 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
+Apr 29 13:12:13.057: INFO: rc: 1
+Apr 29 13:12:13.057: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/tmp/kubeconfig-185508174 exec --namespace=statefulset-7031 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] []    Error from server (NotFound): pods "ss-0" not found
+ []  0xc001cd0510 exit status 1   true [0xc0027df138 0xc0027df150 0xc0027df168] [0xc0027df138 0xc0027df150 0xc0027df168] [0xc0027df148 0xc0027df160] [0x9bf9f0 0x9bf9f0] 0xc0018a2b40 }:
+Command stdout:
+
+stderr:
+Error from server (NotFound): pods "ss-0" not found
+
+error:
+exit status 1
+
+Apr 29 13:12:23.057: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 exec --namespace=statefulset-7031 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
+Apr 29 13:12:23.146: INFO: rc: 1
+Apr 29 13:12:23.147: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/tmp/kubeconfig-185508174 exec --namespace=statefulset-7031 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] []    Error from server (NotFound): pods "ss-0" not found
+ []  0xc001264d50 exit status 1   true [0xc0015016d0 0xc0015016e8 0xc001501700] [0xc0015016d0 0xc0015016e8 0xc001501700] [0xc0015016e0 0xc0015016f8] [0x9bf9f0 0x9bf9f0] 0xc0017b8b40 }:
+Command stdout:
+
+stderr:
+Error from server (NotFound): pods "ss-0" not found
+
+error:
+exit status 1
+
+Apr 29 13:12:33.147: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 exec --namespace=statefulset-7031 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
+Apr 29 13:12:33.250: INFO: rc: 1
+Apr 29 13:12:33.250: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/tmp/kubeconfig-185508174 exec --namespace=statefulset-7031 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] []    Error from server (NotFound): pods "ss-0" not found
+ []  0xc001674210 exit status 1   true [0xc000010690 0xc0000107b0 0xc000010a90] [0xc000010690 0xc0000107b0 0xc000010a90] [0xc000010780 0xc0000109e0] [0x9bf9f0 0x9bf9f0] 0xc001c514a0 }:
+Command stdout:
+
+stderr:
+Error from server (NotFound): pods "ss-0" not found
+
+error:
+exit status 1
+
+Apr 29 13:12:43.251: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 exec --namespace=statefulset-7031 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
+Apr 29 13:12:43.326: INFO: rc: 1
+Apr 29 13:12:43.326: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/tmp/kubeconfig-185508174 exec --namespace=statefulset-7031 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] []    Error from server (NotFound): pods "ss-0" not found
+ []  0xc001674570 exit status 1   true [0xc000010ab0 0xc000010d80 0xc000010e78] [0xc000010ab0 0xc000010d80 0xc000010e78] [0xc000010c10 0xc000010e50] [0x9bf9f0 0x9bf9f0] 0xc001bbed80 }:
+Command stdout:
+
+stderr:
+Error from server (NotFound): pods "ss-0" not found
+
+error:
+exit status 1
+
+Apr 29 13:12:53.327: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 exec --namespace=statefulset-7031 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
+Apr 29 13:12:53.420: INFO: rc: 1
+Apr 29 13:12:53.420: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/tmp/kubeconfig-185508174 exec --namespace=statefulset-7031 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] []    Error from server (NotFound): pods "ss-0" not found
+ []  0xc001674900 exit status 1   true [0xc000010e88 0xc000010f10 0xc000010f60] [0xc000010e88 0xc000010f10 0xc000010f60] [0xc000010ee0 0xc000010f50] [0x9bf9f0 0x9bf9f0] 0xc001a95380 }:
+Command stdout:
+
+stderr:
+Error from server (NotFound): pods "ss-0" not found
+
+error:
+exit status 1
+
+Apr 29 13:13:03.420: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 exec --namespace=statefulset-7031 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
+Apr 29 13:13:03.516: INFO: rc: 1
+Apr 29 13:13:03.516: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/tmp/kubeconfig-185508174 exec --namespace=statefulset-7031 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] []    Error from server (NotFound): pods "ss-0" not found
+ []  0xc001674c60 exit status 1   true [0xc000010f80 0xc000010fe8 0xc0000115c8] [0xc000010f80 0xc000010fe8 0xc0000115c8] [0xc000010fd8 0xc000011538] [0x9bf9f0 0x9bf9f0] 0xc001d87800 }:
+Command stdout:
+
+stderr:
+Error from server (NotFound): pods "ss-0" not found
+
+error:
+exit status 1
+
+Apr 29 13:13:13.516: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 exec --namespace=statefulset-7031 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
+Apr 29 13:13:13.611: INFO: rc: 1
+Apr 29 13:13:13.611: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/tmp/kubeconfig-185508174 exec --namespace=statefulset-7031 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] []    Error from server (NotFound): pods "ss-0" not found
+ []  0xc001675140 exit status 1   true [0xc0000115f0 0xc0000116a8 0xc000011778] [0xc0000115f0 0xc0000116a8 0xc000011778] [0xc000011650 0xc000011718] [0x9bf9f0 0x9bf9f0] 0xc002061320 }:
+Command stdout:
+
+stderr:
+Error from server (NotFound): pods "ss-0" not found
+
+error:
+exit status 1
+
+Apr 29 13:13:23.611: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 exec --namespace=statefulset-7031 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
+Apr 29 13:13:23.708: INFO: rc: 1
+Apr 29 13:13:23.708: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/tmp/kubeconfig-185508174 exec --namespace=statefulset-7031 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] []    Error from server (NotFound): pods "ss-0" not found
+ []  0xc002958330 exit status 1   true [0xc0027de000 0xc0027de018 0xc0027de058] [0xc0027de000 0xc0027de018 0xc0027de058] [0xc0027de010 0xc0027de038] [0x9bf9f0 0x9bf9f0] 0xc0023a3560 }:
+Command stdout:
+
+stderr:
+Error from server (NotFound): pods "ss-0" not found
+
+error:
+exit status 1
+
+Apr 29 13:13:33.709: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 exec --namespace=statefulset-7031 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
+Apr 29 13:13:33.837: INFO: rc: 1
+Apr 29 13:13:33.837: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/tmp/kubeconfig-185508174 exec --namespace=statefulset-7031 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] []    Error from server (NotFound): pods "ss-0" not found
+ []  0xc002958690 exit status 1   true [0xc0027de060 0xc0027de088 0xc0027de0a0] [0xc0027de060 0xc0027de088 0xc0027de0a0] [0xc0027de080 0xc0027de098] [0x9bf9f0 0x9bf9f0] 0xc002508300 }:
+Command stdout:
+
+stderr:
+Error from server (NotFound): pods "ss-0" not found
+
+error:
+exit status 1
+
+Apr 29 13:13:43.838: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 exec --namespace=statefulset-7031 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
+Apr 29 13:13:43.946: INFO: rc: 1
+Apr 29 13:13:43.946: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/tmp/kubeconfig-185508174 exec --namespace=statefulset-7031 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] []    Error from server (NotFound): pods "ss-0" not found
+ []  0xc0016754a0 exit status 1   true [0xc000011798 0xc000011838 0xc0000118e8] [0xc000011798 0xc000011838 0xc0000118e8] [0xc0000117d0 0xc0000118b8] [0x9bf9f0 0x9bf9f0] 0xc001e6cae0 }:
+Command stdout:
+
+stderr:
+Error from server (NotFound): pods "ss-0" not found
+
+error:
+exit status 1
+
+Apr 29 13:13:53.947: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 exec --namespace=statefulset-7031 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
+Apr 29 13:13:54.042: INFO: rc: 1
+Apr 29 13:13:54.043: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/tmp/kubeconfig-185508174 exec --namespace=statefulset-7031 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] []    Error from server (NotFound): pods "ss-0" not found
+ []  0xc0029589f0 exit status 1   true [0xc0027de0a8 0xc0027de0c0 0xc0027de0d8] [0xc0027de0a8 0xc0027de0c0 0xc0027de0d8] [0xc0027de0b8 0xc0027de0d0] [0x9bf9f0 0x9bf9f0] 0xc002509320 }:
+Command stdout:
+
+stderr:
+Error from server (NotFound): pods "ss-0" not found
+
+error:
+exit status 1
+
+Apr 29 13:14:04.043: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 exec --namespace=statefulset-7031 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
+Apr 29 13:14:04.160: INFO: rc: 1
+Apr 29 13:14:04.160: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/tmp/kubeconfig-185508174 exec --namespace=statefulset-7031 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] []    Error from server (NotFound): pods "ss-0" not found
+ []  0xc002958d80 exit status 1   true [0xc0027de0e0 0xc0027de0f8 0xc0027de110] [0xc0027de0e0 0xc0027de0f8 0xc0027de110] [0xc0027de0f0 0xc0027de108] [0x9bf9f0 0x9bf9f0] 0xc003002060 }:
+Command stdout:
+
+stderr:
+Error from server (NotFound): pods "ss-0" not found
+
+error:
+exit status 1
+
+Apr 29 13:14:14.161: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 exec --namespace=statefulset-7031 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
+Apr 29 13:14:14.274: INFO: rc: 1
+Apr 29 13:14:14.274: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/tmp/kubeconfig-185508174 exec --namespace=statefulset-7031 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] []    Error from server (NotFound): pods "ss-0" not found
+ []  0xc002959200 exit status 1   true [0xc0027de118 0xc0027de130 0xc0027de148] [0xc0027de118 0xc0027de130 0xc0027de148] [0xc0027de128 0xc0027de140] [0x9bf9f0 0x9bf9f0] 0xc003002540 }:
+Command stdout:
+
+stderr:
+Error from server (NotFound): pods "ss-0" not found
+
+error:
+exit status 1
+
+Apr 29 13:14:24.275: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 exec --namespace=statefulset-7031 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
+Apr 29 13:14:24.373: INFO: rc: 1
+Apr 29 13:14:24.373: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/tmp/kubeconfig-185508174 exec --namespace=statefulset-7031 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] []    Error from server (NotFound): pods "ss-0" not found
+ []  0xc001675bc0 exit status 1   true [0xc000011900 0xc000011990 0xc0000119f0] [0xc000011900 0xc000011990 0xc0000119f0] [0xc000011978 0xc0000119c8] [0x9bf9f0 0x9bf9f0] 0xc001e6d6e0 }:
+Command stdout:
+
+stderr:
+Error from server (NotFound): pods "ss-0" not found
+
+error:
+exit status 1
+
+Apr 29 13:14:34.374: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 exec --namespace=statefulset-7031 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
+Apr 29 13:14:34.462: INFO: rc: 1
+Apr 29 13:14:34.462: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/tmp/kubeconfig-185508174 exec --namespace=statefulset-7031 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] []    Error from server (NotFound): pods "ss-0" not found
+ []  0xc001674240 exit status 1   true [0xc000010690 0xc0000107b0 0xc000010a90] [0xc000010690 0xc0000107b0 0xc000010a90] [0xc000010780 0xc0000109e0] [0x9bf9f0 0x9bf9f0] 0xc002508cc0 }:
+Command stdout:
+
+stderr:
+Error from server (NotFound): pods "ss-0" not found
+
+error:
+exit status 1
+
+Apr 29 13:14:44.462: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 exec --namespace=statefulset-7031 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
+Apr 29 13:14:44.574: INFO: rc: 1
+Apr 29 13:14:44.574: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/tmp/kubeconfig-185508174 exec --namespace=statefulset-7031 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] []    Error from server (NotFound): pods "ss-0" not found
+ []  0xc002958390 exit status 1   true [0xc0027de000 0xc0027de018 0xc0027de058] [0xc0027de000 0xc0027de018 0xc0027de058] [0xc0027de010 0xc0027de038] [0x9bf9f0 0x9bf9f0] 0xc0023a3560 }:
+Command stdout:
+
+stderr:
+Error from server (NotFound): pods "ss-0" not found
+
+error:
+exit status 1
+
+Apr 29 13:14:54.574: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 exec --namespace=statefulset-7031 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
+Apr 29 13:14:54.685: INFO: rc: 1
+Apr 29 13:14:54.686: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/tmp/kubeconfig-185508174 exec --namespace=statefulset-7031 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] []    Error from server (NotFound): pods "ss-0" not found
+ []  0xc0016745a0 exit status 1   true [0xc000010ab0 0xc000010d80 0xc000010e78] [0xc000010ab0 0xc000010d80 0xc000010e78] [0xc000010c10 0xc000010e50] [0x9bf9f0 0x9bf9f0] 0xc0025099e0 }:
+Command stdout:
+
+stderr:
+Error from server (NotFound): pods "ss-0" not found
+
+error:
+exit status 1
+
+Apr 29 13:15:04.686: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 exec --namespace=statefulset-7031 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
+Apr 29 13:15:04.780: INFO: rc: 1
+Apr 29 13:15:04.780: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/tmp/kubeconfig-185508174 exec --namespace=statefulset-7031 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] []    Error from server (NotFound): pods "ss-0" not found
+ []  0xc002958720 exit status 1   true [0xc0027de060 0xc0027de088 0xc0027de0a0] [0xc0027de060 0xc0027de088 0xc0027de0a0] [0xc0027de080 0xc0027de098] [0x9bf9f0 0x9bf9f0] 0xc002060660 }:
+Command stdout:
+
+stderr:
+Error from server (NotFound): pods "ss-0" not found
+
+error:
+exit status 1
+
+Apr 29 13:15:14.781: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 exec --namespace=statefulset-7031 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
+Apr 29 13:15:14.877: INFO: rc: 1
+Apr 29 13:15:14.877: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/tmp/kubeconfig-185508174 exec --namespace=statefulset-7031 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] []    Error from server (NotFound): pods "ss-0" not found
+ []  0xc002958ab0 exit status 1   true [0xc0027de0a8 0xc0027de0c0 0xc0027de0d8] [0xc0027de0a8 0xc0027de0c0 0xc0027de0d8] [0xc0027de0b8 0xc0027de0d0] [0x9bf9f0 0x9bf9f0] 0xc001d86780 }:
+Command stdout:
+
+stderr:
+Error from server (NotFound): pods "ss-0" not found
+
+error:
+exit status 1
+
+Apr 29 13:15:24.878: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 exec --namespace=statefulset-7031 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
+Apr 29 13:15:24.980: INFO: rc: 1
+Apr 29 13:15:24.981: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/tmp/kubeconfig-185508174 exec --namespace=statefulset-7031 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] []    Error from server (NotFound): pods "ss-0" not found
+ []  0xc002958e70 exit status 1   true [0xc0027de0e0 0xc0027de0f8 0xc0027de110] [0xc0027de0e0 0xc0027de0f8 0xc0027de110] [0xc0027de0f0 0xc0027de108] [0x9bf9f0 0x9bf9f0] 0xc001a945a0 }:
+Command stdout:
+
+stderr:
+Error from server (NotFound): pods "ss-0" not found
+
+error:
+exit status 1
+
+Apr 29 13:15:34.982: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 exec --namespace=statefulset-7031 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
+Apr 29 13:15:35.087: INFO: rc: 1
+Apr 29 13:15:35.087: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/tmp/kubeconfig-185508174 exec --namespace=statefulset-7031 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] []    Error from server (NotFound): pods "ss-0" not found
+ []  0xc001674960 exit status 1   true [0xc000010e88 0xc000010f10 0xc000010f60] [0xc000010e88 0xc000010f10 0xc000010f60] [0xc000010ee0 0xc000010f50] [0x9bf9f0 0x9bf9f0] 0xc001bbfd40 }:
+Command stdout:
+
+stderr:
+Error from server (NotFound): pods "ss-0" not found
+
+error:
+exit status 1
+
+Apr 29 13:15:45.087: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 exec --namespace=statefulset-7031 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
+Apr 29 13:15:45.180: INFO: rc: 1
+Apr 29 13:15:45.180: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/tmp/kubeconfig-185508174 exec --namespace=statefulset-7031 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] []    Error from server (NotFound): pods "ss-0" not found
+ []  0xc002959320 exit status 1   true [0xc0027de118 0xc0027de130 0xc0027de148] [0xc0027de118 0xc0027de130 0xc0027de148] [0xc0027de128 0xc0027de140] [0x9bf9f0 0x9bf9f0] 0xc001f4d680 }:
+Command stdout:
+
+stderr:
+Error from server (NotFound): pods "ss-0" not found
+
+error:
+exit status 1
+
+Apr 29 13:15:55.180: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 exec --namespace=statefulset-7031 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
+Apr 29 13:15:55.265: INFO: rc: 1
+Apr 29 13:15:55.266: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/tmp/kubeconfig-185508174 exec --namespace=statefulset-7031 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] []    Error from server (NotFound): pods "ss-0" not found
+ []  0xc002959800 exit status 1   true [0xc0027de150 0xc0027de168 0xc0027de188] [0xc0027de150 0xc0027de168 0xc0027de188] [0xc0027de160 0xc0027de178] [0x9bf9f0 0x9bf9f0] 0xc001e6c120 }:
+Command stdout:
+
+stderr:
+Error from server (NotFound): pods "ss-0" not found
+
+error:
+exit status 1
+
+Apr 29 13:16:05.266: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 exec --namespace=statefulset-7031 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
+Apr 29 13:16:05.367: INFO: rc: 1
+Apr 29 13:16:05.368: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/tmp/kubeconfig-185508174 exec --namespace=statefulset-7031 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] []    Error from server (NotFound): pods "ss-0" not found
+ []  0xc001674e10 exit status 1   true [0xc000010f80 0xc000010fe8 0xc0000115c8] [0xc000010f80 0xc000010fe8 0xc0000115c8] [0xc000010fd8 0xc000011538] [0x9bf9f0 0x9bf9f0] 0xc0030022a0 }:
+Command stdout:
+
+stderr:
+Error from server (NotFound): pods "ss-0" not found
+
+error:
+exit status 1
+
+Apr 29 13:16:15.368: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 exec --namespace=statefulset-7031 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
+Apr 29 13:16:15.468: INFO: rc: 1
+Apr 29 13:16:15.468: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/tmp/kubeconfig-185508174 exec --namespace=statefulset-7031 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] []    Error from server (NotFound): pods "ss-0" not found
+ []  0xc002959b60 exit status 1   true [0xc0027de1a0 0xc0027de1c0 0xc0027de1d8] [0xc0027de1a0 0xc0027de1c0 0xc0027de1d8] [0xc0027de1b8 0xc0027de1d0] [0x9bf9f0 0x9bf9f0] 0xc001e6d260 }:
+Command stdout:
+
+stderr:
+Error from server (NotFound): pods "ss-0" not found
+
+error:
+exit status 1
+
+Apr 29 13:16:25.468: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 exec --namespace=statefulset-7031 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
+Apr 29 13:16:25.569: INFO: rc: 1
+Apr 29 13:16:25.569: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/tmp/kubeconfig-185508174 exec --namespace=statefulset-7031 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] []    Error from server (NotFound): pods "ss-0" not found
+ []  0xc001675230 exit status 1   true [0xc0000115f0 0xc0000116a8 0xc000011778] [0xc0000115f0 0xc0000116a8 0xc000011778] [0xc000011650 0xc000011718] [0x9bf9f0 0x9bf9f0] 0xc0030027e0 }:
+Command stdout:
+
+stderr:
+Error from server (NotFound): pods "ss-0" not found
+
+error:
+exit status 1
+
+Apr 29 13:16:35.569: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-185508174 exec --namespace=statefulset-7031 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
+Apr 29 13:16:35.665: INFO: rc: 1
+Apr 29 13:16:35.665: INFO: stdout of mv -v /tmp/index.html /usr/share/nginx/html/ || true on ss-0: 
+Apr 29 13:16:35.665: INFO: Scaling statefulset ss to 0
+Apr 29 13:16:35.679: INFO: Waiting for statefulset status.replicas updated to 0
+[AfterEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:85
+Apr 29 13:16:35.683: INFO: Deleting all statefulset in ns statefulset-7031
+Apr 29 13:16:35.686: INFO: Scaling statefulset ss to 0
+Apr 29 13:16:35.699: INFO: Waiting for statefulset status.replicas updated to 0
+Apr 29 13:16:35.706: INFO: Deleting statefulset ss
+[AfterEach] [sig-apps] StatefulSet
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 13:16:35.726: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "statefulset-7031" for this suite.
+Apr 29 13:16:43.764: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 13:16:44.044: INFO: namespace statefulset-7031 deletion completed in 8.306889372s
+
+• [SLOW TEST:364.352 seconds]
+[sig-apps] StatefulSet
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22
+  [k8s.io] Basic StatefulSet functionality [StatefulSetBasic]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
+    Burst scaling should run to completion even with unhealthy pods [Conformance]
+    /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSSSSSS
+------------------------------
+[sig-network] Networking Granular Checks: Pods 
+  should function for intra-pod communication: udp [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-network] Networking
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 13:16:44.047: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename pod-network-test
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in pod-network-test-9878
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should function for intra-pod communication: udp [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Performing setup for networking test in namespace pod-network-test-9878
+STEP: creating a selector
+STEP: Creating the service pods in kubernetes
+Apr 29 13:16:44.239: INFO: Waiting up to 10m0s for all (but 0) nodes to be schedulable
+STEP: Creating test pods
+Apr 29 13:17:04.452: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s 'http://192.168.205.124:8080/dial?request=hostName&protocol=udp&host=192.168.14.10&port=8081&tries=1'] Namespace:pod-network-test-9878 PodName:host-test-container-pod ContainerName:hostexec Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
+Apr 29 13:17:04.452: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+Apr 29 13:17:04.620: INFO: Waiting for endpoints: map[]
+Apr 29 13:17:04.629: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s 'http://192.168.205.124:8080/dial?request=hostName&protocol=udp&host=192.168.205.123&port=8081&tries=1'] Namespace:pod-network-test-9878 PodName:host-test-container-pod ContainerName:hostexec Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
+Apr 29 13:17:04.629: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+Apr 29 13:17:04.804: INFO: Waiting for endpoints: map[]
+Apr 29 13:17:04.814: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s 'http://192.168.205.124:8080/dial?request=hostName&protocol=udp&host=192.168.148.114&port=8081&tries=1'] Namespace:pod-network-test-9878 PodName:host-test-container-pod ContainerName:hostexec Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
+Apr 29 13:17:04.814: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+Apr 29 13:17:04.982: INFO: Waiting for endpoints: map[]
+[AfterEach] [sig-network] Networking
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 13:17:04.982: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "pod-network-test-9878" for this suite.
+Apr 29 13:17:29.017: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 13:17:29.224: INFO: namespace pod-network-test-9878 deletion completed in 24.229629207s
+
+• [SLOW TEST:45.177 seconds]
+[sig-network] Networking
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:25
+  Granular Checks: Pods
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:28
+    should function for intra-pod communication: udp [LinuxOnly] [NodeConformance] [Conformance]
+    /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+S
+------------------------------
+[k8s.io] [sig-node] PreStop 
+  should call prestop when killing a pod  [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [k8s.io] [sig-node] PreStop
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 13:17:29.224: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename prestop
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in prestop-7231
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [k8s.io] [sig-node] PreStop
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/node/pre_stop.go:167
+[It] should call prestop when killing a pod  [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating server pod server in namespace prestop-7231
+STEP: Waiting for pods to come up.
+STEP: Creating tester pod tester in namespace prestop-7231
+STEP: Deleting pre-stop pod
+Apr 29 13:17:44.474: INFO: Saw: {
+	"Hostname": "server",
+	"Sent": null,
+	"Received": {
+		"prestop": 1
+	},
+	"Errors": null,
+	"Log": [
+		"default/nettest has 0 endpoints ([]), which is less than 8 as expected. Waiting for all endpoints to come up.",
+		"default/nettest has 0 endpoints ([]), which is less than 8 as expected. Waiting for all endpoints to come up.",
+		"default/nettest has 0 endpoints ([]), which is less than 8 as expected. Waiting for all endpoints to come up."
+	],
+	"StillContactingPeers": true
+}
+STEP: Deleting the server pod
+[AfterEach] [k8s.io] [sig-node] PreStop
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 13:17:44.500: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "prestop-7231" for this suite.
+Apr 29 13:18:26.582: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 13:18:26.774: INFO: namespace prestop-7231 deletion completed in 42.25267319s
+
+• [SLOW TEST:57.550 seconds]
+[k8s.io] [sig-node] PreStop
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:687
+  should call prestop when killing a pod  [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSSSSSSSSSSS
+------------------------------
+[sig-apps] ReplicationController 
+  should serve a basic image on each replica with a public image  [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+[BeforeEach] [sig-apps] ReplicationController
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:149
+STEP: Creating a kubernetes client
+Apr 29 13:18:26.774: INFO: >>> kubeConfig: /tmp/kubeconfig-185508174
+STEP: Building a namespace api object, basename replication-controller
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in replication-controller-5107
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should serve a basic image on each replica with a public image  [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+STEP: Creating replication controller my-hostname-basic-4605e0a6-6a81-11e9-b6ef-0e71f8c7e888
+Apr 29 13:18:26.982: INFO: Pod name my-hostname-basic-4605e0a6-6a81-11e9-b6ef-0e71f8c7e888: Found 0 pods out of 1
+Apr 29 13:18:31.992: INFO: Pod name my-hostname-basic-4605e0a6-6a81-11e9-b6ef-0e71f8c7e888: Found 1 pods out of 1
+Apr 29 13:18:31.992: INFO: Ensuring all pods for ReplicationController "my-hostname-basic-4605e0a6-6a81-11e9-b6ef-0e71f8c7e888" are running
+Apr 29 13:18:31.999: INFO: Pod "my-hostname-basic-4605e0a6-6a81-11e9-b6ef-0e71f8c7e888-x2tkx" is running (conditions: [{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2019-04-29 13:18:27 +0000 UTC Reason: Message:} {Type:Ready Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2019-04-29 13:18:29 +0000 UTC Reason: Message:} {Type:ContainersReady Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2019-04-29 13:18:29 +0000 UTC Reason: Message:} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2019-04-29 13:18:26 +0000 UTC Reason: Message:}])
+Apr 29 13:18:32.000: INFO: Trying to dial the pod
+Apr 29 13:18:37.022: INFO: Controller my-hostname-basic-4605e0a6-6a81-11e9-b6ef-0e71f8c7e888: Got expected result from replica 1 [my-hostname-basic-4605e0a6-6a81-11e9-b6ef-0e71f8c7e888-x2tkx]: "my-hostname-basic-4605e0a6-6a81-11e9-b6ef-0e71f8c7e888-x2tkx", 1 of 1 required successes so far
+[AfterEach] [sig-apps] ReplicationController
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150
+Apr 29 13:18:37.022: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "replication-controller-5107" for this suite.
+Apr 29 13:18:43.051: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
+Apr 29 13:18:43.271: INFO: namespace replication-controller-5107 deletion completed in 6.240910989s
+
+• [SLOW TEST:16.497 seconds]
+[sig-apps] ReplicationController
+/workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22
+  should serve a basic image on each replica with a public image  [Conformance]
+  /workspace/anago-v1.14.1-beta.0.44+b7394102d6ef77/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692
+------------------------------
+SSSSSSSSSSSApr 29 13:18:43.276: INFO: Running AfterSuite actions on all nodes
+Apr 29 13:18:43.277: INFO: Running AfterSuite actions on node 1
+Apr 29 13:18:43.278: INFO: Skipping dumping logs from cluster
+
+Ran 204 of 3584 Specs in 6445.836 seconds
+SUCCESS! -- 204 Passed | 0 Failed | 0 Pending | 3380 Skipped PASS
+
+Ginkgo ran 1 suite in 1h47m27.844138548s
+Test Suite Passed
diff --git a/v1.14/giantswarm-kvm/junit_01.xml b/v1.14/giantswarm-kvm/junit_01.xml
new file mode 100644
index 0000000000..ed48c7c8c9
--- /dev/null
+++ b/v1.14/giantswarm-kvm/junit_01.xml
@@ -0,0 +1,10347 @@
+
+  
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+  
\ No newline at end of file